mirror of
https://github.com/ChatGPTNextWeb/ChatGPT-Next-Web.git
synced 2025-10-15 22:43:43 +08:00
Compare commits
11 Commits
Leizhenpen
...
afce63b6e3
Author | SHA1 | Date | |
---|---|---|---|
|
afce63b6e3 | ||
|
705dffc664 | ||
|
02f7e6de98 | ||
|
3809375694 | ||
|
1b0de25986 | ||
|
865c45dd29 | ||
|
1f5d8e6d9c | ||
|
2d7229d2b8 | ||
|
ecc65fa775 | ||
|
da9963d4ee | ||
|
24ff78e1c5 |
@@ -22,7 +22,7 @@ English / [简体中文](./README_CN.md)
|
|||||||
[![MacOS][MacOS-image]][download-url]
|
[![MacOS][MacOS-image]][download-url]
|
||||||
[![Linux][Linux-image]][download-url]
|
[![Linux][Linux-image]][download-url]
|
||||||
|
|
||||||
[NextChatAI](https://nextchat.club?utm_source=readme) / [Web App Demo](https://app.nextchat.dev) / [Desktop App](https://github.com/Yidadaa/ChatGPT-Next-Web/releases) / [Discord](https://discord.gg/YCkeafCafC) / [Enterprise Edition](#enterprise-edition) / [Twitter](https://twitter.com/NextChatDev)
|
[NextChatAI](https://nextchat.club?utm_source=readme) / [iOS APP](https://apps.apple.com/us/app/nextchat-ai/id6743085599) / [Web App Demo](https://app.nextchat.dev) / [Desktop App](https://github.com/Yidadaa/ChatGPT-Next-Web/releases) / [Enterprise Edition](#enterprise-edition)
|
||||||
|
|
||||||
|
|
||||||
[saas-url]: https://nextchat.club?utm_source=readme
|
[saas-url]: https://nextchat.club?utm_source=readme
|
||||||
@@ -41,7 +41,9 @@ English / [简体中文](./README_CN.md)
|
|||||||
</div>
|
</div>
|
||||||
|
|
||||||
## 🥳 Cheer for NextChat iOS Version Online!
|
## 🥳 Cheer for NextChat iOS Version Online!
|
||||||
> [ 👉 Click Here Install Now](https://apps.apple.com/us/app/nextchat-ai/id6743085599)
|
> [👉 Click Here to Install Now](https://apps.apple.com/us/app/nextchat-ai/id6743085599)
|
||||||
|
|
||||||
|
> [❤️ Source Code Coming Soon](https://github.com/ChatGPTNextWeb/NextChat-iOS)
|
||||||
|
|
||||||

|

|
||||||
|
|
||||||
|
@@ -1,6 +1,6 @@
|
|||||||
import { NextRequest, NextResponse } from "next/server";
|
import { NextRequest, NextResponse } from "next/server";
|
||||||
import { getServerSideConfig } from "../config/server";
|
import { getServerSideConfig } from "../config/server";
|
||||||
import { OPENAI_BASE_URL, ServiceProvider } from "../constant";
|
import { OPENAI_BASE_URL, ServiceProvider, OpenaiPath } from "../constant";
|
||||||
import { cloudflareAIGatewayUrl } from "../utils/cloudflare";
|
import { cloudflareAIGatewayUrl } from "../utils/cloudflare";
|
||||||
import { getModelProvider, isModelNotavailableInServer } from "../utils/model";
|
import { getModelProvider, isModelNotavailableInServer } from "../utils/model";
|
||||||
|
|
||||||
@@ -26,8 +26,8 @@ export async function requestOpenai(req: NextRequest) {
|
|||||||
authValue = req.headers.get("Authorization") ?? "";
|
authValue = req.headers.get("Authorization") ?? "";
|
||||||
authHeaderName = "Authorization";
|
authHeaderName = "Authorization";
|
||||||
}
|
}
|
||||||
|
|
||||||
let path = `${req.nextUrl.pathname}`.replaceAll("/api/openai/", "");
|
let path = `${req.nextUrl.pathname}`.replaceAll("/api/openai/", "");
|
||||||
|
let isChatRequest = path.includes(OpenaiPath.ChatPath);
|
||||||
|
|
||||||
let baseUrl =
|
let baseUrl =
|
||||||
(isAzure ? serverConfig.azureUrl : serverConfig.baseUrl) || OPENAI_BASE_URL;
|
(isAzure ? serverConfig.azureUrl : serverConfig.baseUrl) || OPENAI_BASE_URL;
|
||||||
@@ -117,14 +117,14 @@ export async function requestOpenai(req: NextRequest) {
|
|||||||
const jsonBody = JSON.parse(clonedBody) as { model?: string };
|
const jsonBody = JSON.parse(clonedBody) as { model?: string };
|
||||||
|
|
||||||
// not undefined and is false
|
// not undefined and is false
|
||||||
if (
|
if ( isChatRequest &&
|
||||||
isModelNotavailableInServer(
|
isModelNotavailableInServer(
|
||||||
serverConfig.customModels,
|
serverConfig.customModels,
|
||||||
jsonBody?.model as string,
|
jsonBody?.model as string,
|
||||||
[
|
[
|
||||||
ServiceProvider.OpenAI,
|
ServiceProvider.OpenAI,
|
||||||
ServiceProvider.Azure,
|
ServiceProvider.Azure,
|
||||||
jsonBody?.model as string, // support provider-unspecified model
|
"custom" as string, // support provider-unspecified model
|
||||||
],
|
],
|
||||||
)
|
)
|
||||||
) {
|
) {
|
||||||
|
@@ -198,7 +198,8 @@ export class ChatGPTApi implements LLMApi {
|
|||||||
const isDalle3 = _isDalle3(options.config.model);
|
const isDalle3 = _isDalle3(options.config.model);
|
||||||
const isO1OrO3 =
|
const isO1OrO3 =
|
||||||
options.config.model.startsWith("o1") ||
|
options.config.model.startsWith("o1") ||
|
||||||
options.config.model.startsWith("o3");
|
options.config.model.startsWith("o3") ||
|
||||||
|
options.config.model.startsWith("o4-mini");
|
||||||
if (isDalle3) {
|
if (isDalle3) {
|
||||||
const prompt = getMessageTextContent(
|
const prompt = getMessageTextContent(
|
||||||
options.messages.slice(-1)?.pop() as any,
|
options.messages.slice(-1)?.pop() as any,
|
||||||
@@ -243,7 +244,7 @@ export class ChatGPTApi implements LLMApi {
|
|||||||
}
|
}
|
||||||
|
|
||||||
// add max_tokens to vision model
|
// add max_tokens to vision model
|
||||||
if (visionModel) {
|
if (visionModel && !isO1OrO3) {
|
||||||
requestPayload["max_tokens"] = Math.max(modelConfig.max_tokens, 4000);
|
requestPayload["max_tokens"] = Math.max(modelConfig.max_tokens, 4000);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -478,6 +478,8 @@ export const VISION_MODEL_REGEXES = [
|
|||||||
/^dall-e-3$/, // Matches exactly "dall-e-3"
|
/^dall-e-3$/, // Matches exactly "dall-e-3"
|
||||||
/glm-4v/,
|
/glm-4v/,
|
||||||
/vl/i,
|
/vl/i,
|
||||||
|
/o3/,
|
||||||
|
/o4-mini/,
|
||||||
];
|
];
|
||||||
|
|
||||||
export const EXCLUDE_VISION_MODEL_REGEXES = [/claude-3-5-haiku-20241022/];
|
export const EXCLUDE_VISION_MODEL_REGEXES = [/claude-3-5-haiku-20241022/];
|
||||||
@@ -516,6 +518,8 @@ const openaiModels = [
|
|||||||
"o1-mini",
|
"o1-mini",
|
||||||
"o1-preview",
|
"o1-preview",
|
||||||
"o3-mini",
|
"o3-mini",
|
||||||
|
"o3",
|
||||||
|
"o4-mini",
|
||||||
];
|
];
|
||||||
|
|
||||||
const googleModels = [
|
const googleModels = [
|
||||||
@@ -542,6 +546,7 @@ const googleModels = [
|
|||||||
"gemini-2.0-flash-thinking-exp-01-21",
|
"gemini-2.0-flash-thinking-exp-01-21",
|
||||||
"gemini-2.0-pro-exp",
|
"gemini-2.0-pro-exp",
|
||||||
"gemini-2.0-pro-exp-02-05",
|
"gemini-2.0-pro-exp-02-05",
|
||||||
|
"gemini-2.5-pro-preview-06-05",
|
||||||
];
|
];
|
||||||
|
|
||||||
const anthropicModels = [
|
const anthropicModels = [
|
||||||
|
@@ -66,8 +66,8 @@ export function collectModelTable(
|
|||||||
|
|
||||||
// default models
|
// default models
|
||||||
models.forEach((m) => {
|
models.forEach((m) => {
|
||||||
// using <modelName>@<providerId> as fullName
|
// using <modelName>@<providerType> as fullName
|
||||||
modelTable[`${m.name}@${m?.provider?.id}`] = {
|
modelTable[`${m.name}@${m?.provider?.providerType}`] = {
|
||||||
...m,
|
...m,
|
||||||
displayName: m.name, // 'provider' is copied over if it exists
|
displayName: m.name, // 'provider' is copied over if it exists
|
||||||
};
|
};
|
||||||
@@ -121,12 +121,14 @@ export function collectModelTable(
|
|||||||
if (displayName && provider.providerName == "ByteDance") {
|
if (displayName && provider.providerName == "ByteDance") {
|
||||||
[customModelName, displayName] = [displayName, customModelName];
|
[customModelName, displayName] = [displayName, customModelName];
|
||||||
}
|
}
|
||||||
modelTable[`${customModelName}@${provider?.id}`] = {
|
modelTable[`${customModelName}@${provider?.providerType}`] = {
|
||||||
name: customModelName,
|
name: customModelName,
|
||||||
displayName: displayName || customModelName,
|
displayName: displayName || customModelName,
|
||||||
available,
|
available,
|
||||||
provider, // Use optional chaining
|
provider, // Use optional chaining
|
||||||
sorted: CustomSeq.next(`${customModelName}@${provider?.id}`),
|
sorted: CustomSeq.next(
|
||||||
|
`${customModelName}@${provider?.providerType}`,
|
||||||
|
),
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@@ -53,7 +53,7 @@ describe("isModelNotavailableInServer", () => {
|
|||||||
expect(result).toBe(true);
|
expect(result).toBe(true);
|
||||||
});
|
});
|
||||||
|
|
||||||
// FIXME: 这个测试用例有问题,需要修复
|
// FIXME: 这个测试用例有问题,需要修复 ???
|
||||||
// test("support passing multiple providers, model available on one of the providers will return false", () => {
|
// test("support passing multiple providers, model available on one of the providers will return false", () => {
|
||||||
// const customModels = "-all,gpt-4@google";
|
// const customModels = "-all,gpt-4@google";
|
||||||
// const modelName = "gpt-4";
|
// const modelName = "gpt-4";
|
||||||
@@ -69,7 +69,19 @@ describe("isModelNotavailableInServer", () => {
|
|||||||
test("test custom model without setting provider", () => {
|
test("test custom model without setting provider", () => {
|
||||||
const customModels = "-all,mistral-large";
|
const customModels = "-all,mistral-large";
|
||||||
const modelName = "mistral-large";
|
const modelName = "mistral-large";
|
||||||
const providerNames = modelName;
|
const providerNames = "custom";
|
||||||
|
const result = isModelNotavailableInServer(
|
||||||
|
customModels,
|
||||||
|
modelName,
|
||||||
|
providerNames,
|
||||||
|
);
|
||||||
|
expect(result).toBe(false);
|
||||||
|
});
|
||||||
|
|
||||||
|
test("test custom model with non-standard provider", () => {
|
||||||
|
const customModels = "-all,deepseek-chat@DeepSeek";
|
||||||
|
const modelName = "deepseek-chat";
|
||||||
|
const providerNames = "custom";
|
||||||
const result = isModelNotavailableInServer(
|
const result = isModelNotavailableInServer(
|
||||||
customModels,
|
customModels,
|
||||||
modelName,
|
modelName,
|
||||||
|
Reference in New Issue
Block a user