189 lines
5.9 KiB
TypeScript
189 lines
5.9 KiB
TypeScript
import type { NextRequest } from 'next/server';
|
|
import { NextResponse } from 'next/server';
|
|
import { getServerSideConfig } from '../config/server';
|
|
import { OPENAI_BASE_URL, ServiceProvider } from '../constant';
|
|
import { cloudflareAIGatewayUrl } from '../utils/cloudflare';
|
|
import { getModelProvider, isModelAvailableInServer } from '../utils/model';
|
|
|
|
const serverConfig = getServerSideConfig();
|
|
|
|
export async function requestOpenai(req: NextRequest) {
|
|
const controller = new AbortController();
|
|
|
|
const isAzure = req.nextUrl.pathname.includes('azure/deployments');
|
|
|
|
let authValue;
|
|
let authHeaderName = '';
|
|
if (isAzure) {
|
|
authValue
|
|
= req.headers
|
|
.get('Authorization')
|
|
?.trim()
|
|
.replaceAll('Bearer ', '')
|
|
.trim() ?? '';
|
|
|
|
authHeaderName = 'api-key';
|
|
} else {
|
|
authValue = req.headers.get('Authorization') ?? '';
|
|
authHeaderName = 'Authorization';
|
|
}
|
|
|
|
let path = `${req.nextUrl.pathname}`.replaceAll('/api/openai/', '');
|
|
|
|
let baseUrl
|
|
= (isAzure ? serverConfig.azureUrl : serverConfig.baseUrl) || OPENAI_BASE_URL;
|
|
|
|
if (!baseUrl.startsWith('http')) {
|
|
baseUrl = `https://${baseUrl}`;
|
|
}
|
|
|
|
if (baseUrl.endsWith('/')) {
|
|
baseUrl = baseUrl.slice(0, -1);
|
|
}
|
|
|
|
console.log('[Proxy] ', path);
|
|
console.log('[Base Url]', baseUrl);
|
|
|
|
const timeoutId = setTimeout(
|
|
() => {
|
|
controller.abort();
|
|
},
|
|
10 * 60 * 1000,
|
|
);
|
|
|
|
if (isAzure) {
|
|
const azureApiVersion
|
|
= req?.nextUrl?.searchParams?.get('api-version')
|
|
|| serverConfig.azureApiVersion;
|
|
baseUrl = baseUrl.split('/deployments').shift() as string;
|
|
path = `${req.nextUrl.pathname.replaceAll(
|
|
'/api/azure/',
|
|
'',
|
|
)}?api-version=${azureApiVersion}`;
|
|
|
|
// Forward compatibility:
|
|
// if display_name(deployment_name) not set, and '{deploy-id}' in AZURE_URL
|
|
// then using default '{deploy-id}'
|
|
if (serverConfig.customModels && serverConfig.azureUrl) {
|
|
const modelName = path.split('/')[1];
|
|
let realDeployName = '';
|
|
serverConfig.customModels
|
|
.split(',')
|
|
.filter(v => !!v && !v.startsWith('-') && v.includes(modelName))
|
|
.forEach((m) => {
|
|
const [fullName, displayName] = m.split('=');
|
|
const [_, providerName] = getModelProvider(fullName);
|
|
if (providerName === 'azure' && !displayName) {
|
|
const [_, deployId] = (serverConfig?.azureUrl ?? '').split(
|
|
'deployments/',
|
|
);
|
|
if (deployId) {
|
|
realDeployName = deployId;
|
|
}
|
|
}
|
|
});
|
|
if (realDeployName) {
|
|
console.log('[Replace with DeployId', realDeployName);
|
|
path = path.replaceAll(modelName, realDeployName);
|
|
}
|
|
}
|
|
}
|
|
|
|
const fetchUrl = cloudflareAIGatewayUrl(`${baseUrl}/${path}`);
|
|
console.log('fetchUrl', fetchUrl);
|
|
const fetchOptions: RequestInit = {
|
|
headers: {
|
|
'Content-Type': 'application/json',
|
|
'Cache-Control': 'no-store',
|
|
[authHeaderName]: authValue,
|
|
...(serverConfig.openaiOrgId && {
|
|
'OpenAI-Organization': serverConfig.openaiOrgId,
|
|
}),
|
|
},
|
|
method: req.method,
|
|
body: req.body,
|
|
// to fix #2485: https://stackoverflow.com/questions/55920957/cloudflare-worker-typeerror-one-time-use-body
|
|
redirect: 'manual',
|
|
// @ts-ignore
|
|
duplex: 'half',
|
|
signal: controller.signal,
|
|
};
|
|
|
|
// #1815 try to refuse gpt4 request
|
|
if (serverConfig.customModels && req.body) {
|
|
try {
|
|
const clonedBody = await req.text();
|
|
fetchOptions.body = clonedBody;
|
|
|
|
const jsonBody = JSON.parse(clonedBody) as { model?: string };
|
|
|
|
// not undefined and is false
|
|
if (
|
|
isModelAvailableInServer(
|
|
serverConfig.customModels,
|
|
jsonBody?.model as string,
|
|
ServiceProvider.OpenAI as string,
|
|
)
|
|
|| isModelAvailableInServer(
|
|
serverConfig.customModels,
|
|
jsonBody?.model as string,
|
|
ServiceProvider.Azure as string,
|
|
)
|
|
) {
|
|
return NextResponse.json(
|
|
{
|
|
error: true,
|
|
message: `you are not allowed to use ${jsonBody?.model} model`,
|
|
},
|
|
{
|
|
status: 403,
|
|
},
|
|
);
|
|
}
|
|
} catch (e) {
|
|
console.error('[OpenAI] gpt4 filter', e);
|
|
}
|
|
}
|
|
|
|
try {
|
|
const res = await fetch(fetchUrl, fetchOptions);
|
|
|
|
// Extract the OpenAI-Organization header from the response
|
|
const openaiOrganizationHeader = res.headers.get('OpenAI-Organization');
|
|
|
|
// Check if serverConfig.openaiOrgId is defined and not an empty string
|
|
if (serverConfig.openaiOrgId && serverConfig.openaiOrgId.trim() !== '') {
|
|
// If openaiOrganizationHeader is present, log it; otherwise, log that the header is not present
|
|
console.log('[Org ID]', openaiOrganizationHeader);
|
|
} else {
|
|
console.log('[Org ID] is not set up.');
|
|
}
|
|
|
|
// to prevent browser prompt for credentials
|
|
const newHeaders = new Headers(res.headers);
|
|
newHeaders.delete('www-authenticate');
|
|
// to disable nginx buffering
|
|
newHeaders.set('X-Accel-Buffering', 'no');
|
|
|
|
// Conditionally delete the OpenAI-Organization header from the response if [Org ID] is undefined or empty (not setup in ENV)
|
|
// Also, this is to prevent the header from being sent to the client
|
|
if (!serverConfig.openaiOrgId || serverConfig.openaiOrgId.trim() === '') {
|
|
newHeaders.delete('OpenAI-Organization');
|
|
}
|
|
|
|
// The latest version of the OpenAI API forced the content-encoding to be "br" in json response
|
|
// So if the streaming is disabled, we need to remove the content-encoding header
|
|
// Because Vercel uses gzip to compress the response, if we don't remove the content-encoding header
|
|
// The browser will try to decode the response with brotli and fail
|
|
newHeaders.delete('content-encoding');
|
|
|
|
return new Response(res.body, {
|
|
status: res.status,
|
|
statusText: res.statusText,
|
|
headers: newHeaders,
|
|
});
|
|
} finally {
|
|
clearTimeout(timeoutId);
|
|
}
|
|
}
|