feat: 不再默认提供max_tokens

This commit is contained in:
Rock Chin
2023-03-16 13:37:48 +08:00
parent ca4eeda6f0
commit 087d097204

View File

@@ -159,7 +159,6 @@ prompt_submit_length = 1024
completion_api_params = {
"model": "gpt-3.5-turbo",
"temperature": 0.9, # 数值越低得到的回答越理性,取值范围[0, 1]
"max_tokens": 1024, # 每次获取OpenAI接口响应的文字量上限, 不高于4096
"top_p": 1, # 生成的文本的文本与要求的符合度, 取值范围[0, 1]
"frequency_penalty": 0.2,
"presence_penalty": 1.0,