fix: remove parameter stream for o3 from the request
This commit is contained in:
parent
566406f62d
commit
9117a10b47
|
@ -196,6 +196,7 @@ export class ChatGPTApi implements LLMApi {
|
|||
|
||||
const isDalle3 = _isDalle3(options.config.model);
|
||||
const isO1 = options.config.model.startsWith("o1");
|
||||
const isO1Exact = options.config.model === "o1";
|
||||
const isO3 = options.config.model.startsWith("o3");
|
||||
if (isDalle3) {
|
||||
const prompt = getMessageTextContent(
|
||||
|
@ -225,7 +226,7 @@ export class ChatGPTApi implements LLMApi {
|
|||
// O1 not support image, tools (plugin in ChatGPTNextWeb) and system, stream, logprobs, temperature, top_p, n, presence_penalty, frequency_penalty yet.
|
||||
requestPayload = {
|
||||
messages,
|
||||
stream: options.config.stream,
|
||||
stream: isO1Exact ? false : options.config.stream,
|
||||
model: modelConfig.model,
|
||||
temperature: !(isO1 || isO3)? modelConfig.temperature : 1,
|
||||
presence_penalty: !(isO1 || isO3) ? modelConfig.presence_penalty : 0,
|
||||
|
|
Loading…
Reference in New Issue