From 42eff644b4a169c8b812f6a2b6b493f9f9a655eb Mon Sep 17 00:00:00 2001 From: Sam Date: Fri, 8 Aug 2025 15:28:54 +0800 Subject: [PATCH] use max_completion_tokens --- app/client/platforms/openai.ts | 5 +++++ 1 file changed, 5 insertions(+) diff --git a/app/client/platforms/openai.ts b/app/client/platforms/openai.ts index 51ae71ea6..715637692 100644 --- a/app/client/platforms/openai.ts +++ b/app/client/platforms/openai.ts @@ -200,6 +200,7 @@ export class ChatGPTApi implements LLMApi { options.config.model.startsWith("o1") || options.config.model.startsWith("o3") || options.config.model.startsWith("o4-mini"); + const isGpt5 = options.config.model.startsWith("gpt-5"); if (isDalle3) { const prompt = getMessageTextContent( options.messages.slice(-1)?.pop() as any, @@ -251,6 +252,10 @@ export class ChatGPTApi implements LLMApi { requestPayload["max_completion_tokens"] = modelConfig.max_tokens; } + if (isGpt5) { + requestPayload["max_completion_tokens"] = modelConfig.max_tokens; + } + // add max_tokens to vision model if (visionModel && !isO1OrO3) { requestPayload["max_tokens"] = Math.max(modelConfig.max_tokens, 4000);