fix: #3186 enable max_tokens in chat payload

This commit is contained in:
Yidadaa
2023-11-09 03:19:13 +08:00
parent 33b97082fa
commit d0a1d910d4
3 changed files with 5 additions and 4 deletions

View File

@@ -70,6 +70,7 @@ export class ChatGPTApi implements LLMApi {
presence_penalty: modelConfig.presence_penalty,
frequency_penalty: modelConfig.frequency_penalty,
top_p: modelConfig.top_p,
max_tokens: Math.max(modelConfig.max_tokens, 1024),
};
console.log("[Request] openai payload: ", requestPayload);