import { NextRequest, NextResponse } from "next/server"; import { getServerSideConfig } from "../config/server"; import { DEFAULT_MODELS, OPENAI_BASE_URL, GEMINI_BASE_URL } from "../constant"; import { collectModelTable } from "../utils/model"; import { makeAzurePath } from "../azure"; const serverConfig = getServerSideConfig(); export async function requestOpenai( req: NextRequest, cloneBody: any, isAzure: boolean, current_model: string, ) { const controller = new AbortController(); var authValue, authHeaderName = ""; if (isAzure) { authValue = req.headers .get("Authorization") ?.trim() .replaceAll("Bearer ", "") .trim() ?? ""; authHeaderName = "api-key"; } else { authValue = req.headers.get("Authorization") ?? ""; authHeaderName = "Authorization"; } // const authValue = req.headers.get("Authorization") ?? ""; // const authHeaderName = isAzure ? "api-key" : "Authorization"; let path = `${req.nextUrl.pathname}${req.nextUrl.search}`.replaceAll( "/api/openai/", "", ); let baseUrl = isAzure ? serverConfig.azureUrl : serverConfig.baseUrl || OPENAI_BASE_URL; if (!baseUrl.startsWith("http")) { baseUrl = `https://${baseUrl}`; } if (baseUrl.endsWith("/")) { baseUrl = baseUrl.slice(0, -1); } // console.log("[Proxy] ", path); // console.log("[Base Url]", baseUrl); // // this fix [Org ID] undefined in server side if not using custom point // if (serverConfig.openaiOrgId !== undefined) { // console.log("[Org ID]", serverConfig.openaiOrgId); // } const timeoutId = setTimeout( () => { controller.abort(); }, 10 * 60 * 1000, ); const fetchUrl = `${baseUrl}/${path}`; const fetchOptions: RequestInit = { headers: { "Content-Type": "application/json", "Cache-Control": "no-store", [authHeaderName]: authValue, ...(serverConfig.openaiOrgId && { "OpenAI-Organization": serverConfig.openaiOrgId, }), }, method: req.method, body: cloneBody, // to fix #2485: https://stackoverflow.com/questions/55920957/cloudflare-worker-typeerror-one-time-use-body redirect: "manual", // @ts-ignore duplex: "half", signal: controller.signal, }; // #1815 try to refuse some model request if (current_model) { try { const modelTable = collectModelTable( DEFAULT_MODELS, serverConfig.customModels, ); // not undefined and is false if (!modelTable[current_model ?? ""].available) { return NextResponse.json( { error: true, message: `you are not allowed to use ${current_model} model`, }, { status: 403, }, ); } } catch (e) { console.error("[OpenAI] gpt model filter", e); } } try { const res = await fetch(fetchUrl, fetchOptions); // to prevent browser prompt for credentials const newHeaders = new Headers(res.headers); newHeaders.delete("www-authenticate"); // to disable nginx buffering newHeaders.set("X-Accel-Buffering", "no"); // The latest version of the OpenAI API forced the content-encoding to be "br" in json response // So if the streaming is disabled, we need to remove the content-encoding header // Because Vercel uses gzip to compress the response, if we don't remove the content-encoding header // The browser will try to decode the response with brotli and fail newHeaders.delete("content-encoding"); return new Response(res.body, { status: res.status, statusText: res.statusText, headers: newHeaders, }); } finally { clearTimeout(timeoutId); } }