mirror of
https://github.com/ChatGPTNextWeb/ChatGPT-Next-Web.git
synced 2025-11-16 14:03:43 +08:00
merge
This commit is contained in:
@@ -2,7 +2,7 @@ import { getClientConfig } from "../config/client";
|
||||
import {
|
||||
ACCESS_CODE_PREFIX,
|
||||
Azure,
|
||||
AZURE_MODELS,
|
||||
// AZURE_MODELS,
|
||||
ModelProvider,
|
||||
ServiceProvider,
|
||||
} from "../constant";
|
||||
@@ -176,7 +176,7 @@ export class ClientApi {
|
||||
}
|
||||
}
|
||||
|
||||
export function getHeaders(isAzure?: boolean) {
|
||||
export function getHeaders() {
|
||||
const accessStore = useAccessStore.getState();
|
||||
const chatStore = useChatStore.getState();
|
||||
const headers: Record<string, string> = {
|
||||
@@ -189,7 +189,7 @@ export function getHeaders(isAzure?: boolean) {
|
||||
function getConfig() {
|
||||
const modelConfig = chatStore.currentSession().mask.modelConfig;
|
||||
const isGoogle = modelConfig.providerName == ServiceProvider.Google;
|
||||
// const isAzure = modelConfig.providerName === ServiceProvider.Azure;
|
||||
const isAzure = modelConfig.providerName === ServiceProvider.Azure;
|
||||
const isAnthropic = modelConfig.providerName === ServiceProvider.Anthropic;
|
||||
const isBaidu = modelConfig.providerName == ServiceProvider.Baidu;
|
||||
const isByteDance = modelConfig.providerName === ServiceProvider.ByteDance;
|
||||
@@ -233,7 +233,7 @@ export function getHeaders(isAzure?: boolean) {
|
||||
}
|
||||
const {
|
||||
isGoogle,
|
||||
// isAzure,
|
||||
isAzure,
|
||||
isAnthropic,
|
||||
isBaidu,
|
||||
apiKey,
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
// azure and openai, using same models. so using same LLMApi.
|
||||
import {
|
||||
ApiPath,
|
||||
AZURE_MODELS,
|
||||
// AZURE_MODELS,
|
||||
DEFAULT_API_HOST,
|
||||
DEFAULT_MODELS,
|
||||
OpenaiPath,
|
||||
@@ -60,12 +60,12 @@ export interface RequestPayload {
|
||||
export class ChatGPTApi implements LLMApi {
|
||||
private disableListModels = true;
|
||||
|
||||
path(path: string, isAzure?: boolean, azureModel?: string): string {
|
||||
path(path: string): string {
|
||||
const accessStore = useAccessStore.getState();
|
||||
|
||||
let baseUrl = "";
|
||||
|
||||
// const isAzure = path.includes("deployments");
|
||||
const isAzure = path.includes("deployments");
|
||||
if (accessStore.useCustomConfig) {
|
||||
if (isAzure && !accessStore.isValidAzure()) {
|
||||
throw Error(
|
||||
@@ -93,10 +93,6 @@ export class ChatGPTApi implements LLMApi {
|
||||
baseUrl = "https://" + baseUrl;
|
||||
}
|
||||
|
||||
if (isAzure) {
|
||||
path = makeAzurePath(path, accessStore.azureApiVersion, azureModel);
|
||||
}
|
||||
|
||||
console.log("[Proxy Endpoint] ", baseUrl, path);
|
||||
|
||||
return [baseUrl, path].join("/");
|
||||
@@ -121,7 +117,6 @@ export class ChatGPTApi implements LLMApi {
|
||||
providerName: options.config.providerName,
|
||||
},
|
||||
};
|
||||
const is_azure = AZURE_MODELS.includes(modelConfig.model);
|
||||
const requestPayload: RequestPayload = {
|
||||
messages,
|
||||
stream: options.config.stream,
|
||||
@@ -172,8 +167,6 @@ export class ChatGPTApi implements LLMApi {
|
||||
(model?.displayName ?? model?.name) as string,
|
||||
useCustomConfig ? useAccessStore.getState().azureApiVersion : "",
|
||||
),
|
||||
is_azure,
|
||||
modelConfig.model,
|
||||
);
|
||||
} else {
|
||||
chatPath = this.path(OpenaiPath.ChatPath);
|
||||
@@ -182,7 +175,7 @@ export class ChatGPTApi implements LLMApi {
|
||||
method: "POST",
|
||||
body: JSON.stringify(requestPayload),
|
||||
signal: controller.signal,
|
||||
headers: getHeaders(is_azure),
|
||||
headers: getHeaders(),
|
||||
};
|
||||
// make a fetch request
|
||||
const requestTimeoutId = setTimeout(
|
||||
|
||||
Reference in New Issue
Block a user