Compare commits

...

7 Commits

Author SHA1 Message Date
sanmao
58e1b9e39e Merge e2429d444b into 11b37c15bd 2025-04-17 12:51:31 +08:00
RiverRay
11b37c15bd Merge pull request #6450 from stephen-zeng/main
Some checks failed
Run Tests / test (push) Has been cancelled
Add gpt-4.1 family & gpt-4.5-preview support
2025-04-17 08:29:19 +08:00
QwQwQ
1d0038f17d add gpt-4.5-preview support 2025-04-16 22:10:47 +08:00
QwQwQ
619fa519c0 add gpt-4.1 family support 2025-04-16 22:02:35 +08:00
RiverRay
48469bd8ca Merge pull request #6392 from ChatGPTNextWeb/Leizhenpeng-patch-6
Some checks failed
Run Tests / test (push) Has been cancelled
Update README.md
2025-03-20 17:52:02 +08:00
suruiqiang
e2429d444b support bytedance api start with "bot-" (with internet ability) 2025-02-14 16:14:41 +08:00
suruiqiang
c15dbf5296 in chat list, show model display name if it exists 2025-02-13 18:00:32 +08:00
6 changed files with 57 additions and 5 deletions

View File

@@ -117,7 +117,7 @@ export class DoubaoApi implements LLMApi {
options.onController?.(controller);
try {
const chatPath = this.path(ByteDance.ChatPath);
const chatPath = this.path(ByteDance.ChatPath(modelConfig.model));
const chatPayload = {
method: "POST",
body: JSON.stringify(requestPayload),

View File

@@ -1868,7 +1868,7 @@ function _Chat() {
</div>
{!isUser && (
<div className={styles["chat-model-name"]}>
{message.model}
{message.modelDisplayName ?? message.model}
</div>
)}

View File

@@ -82,7 +82,11 @@ export function Avatar(props: { model?: ModelType; avatar?: string }) {
LlmIcon = BotIconGrok;
} else if (modelName.startsWith("hunyuan")) {
LlmIcon = BotIconHunyuan;
} else if (modelName.startsWith("doubao") || modelName.startsWith("ep-")) {
} else if (
modelName.startsWith("doubao") ||
modelName.startsWith("ep-") ||
modelName.startsWith("bot-")
) {
LlmIcon = BotIconDoubao;
} else if (
modelName.includes("glm") ||

View File

@@ -216,7 +216,13 @@ export const Baidu = {
export const ByteDance = {
ExampleEndpoint: "https://ark.cn-beijing.volces.com/api/",
ChatPath: "api/v3/chat/completions",
ChatPath: (modelName: string) => {
if (modelName.startsWith("bot-")) {
return "api/v3/bots/chat/completions";
} else {
return "api/v3/chat/completions";
}
},
};
export const Alibaba = {
@@ -417,6 +423,14 @@ export const KnowledgeCutOffDate: Record<string, string> = {
"gpt-4-turbo": "2023-12",
"gpt-4-turbo-2024-04-09": "2023-12",
"gpt-4-turbo-preview": "2023-12",
"gpt-4.1": "2024-06",
"gpt-4.1-2025-04-14": "2024-06",
"gpt-4.1-mini": "2024-06",
"gpt-4.1-mini-2025-04-14": "2024-06",
"gpt-4.1-nano": "2024-06",
"gpt-4.1-nano-2025-04-14": "2024-06",
"gpt-4.5-preview": "2023-10",
"gpt-4.5-preview-2025-02-27": "2023-10",
"gpt-4o": "2023-10",
"gpt-4o-2024-05-13": "2023-10",
"gpt-4o-2024-08-06": "2023-10",
@@ -458,6 +472,7 @@ export const DEFAULT_TTS_VOICES = [
export const VISION_MODEL_REGEXES = [
/vision/,
/gpt-4o/,
/gpt-4\.1/,
/claude-3/,
/gemini-1\.5/,
/gemini-exp/,
@@ -485,6 +500,14 @@ const openaiModels = [
"gpt-4-32k-0613",
"gpt-4-turbo",
"gpt-4-turbo-preview",
"gpt-4.1",
"gpt-4.1-2025-04-14",
"gpt-4.1-mini",
"gpt-4.1-mini-2025-04-14",
"gpt-4.1-nano",
"gpt-4.1-nano-2025-04-14",
"gpt-4.5-preview",
"gpt-4.5-preview-2025-02-27",
"gpt-4o",
"gpt-4o-2024-05-13",
"gpt-4o-2024-08-06",

View File

@@ -60,6 +60,7 @@ export type ChatMessage = RequestMessage & {
isError?: boolean;
id: string;
model?: ModelType;
modelDisplayName?: string;
tools?: ChatMessageTool[];
audio_url?: string;
isMcpResponse?: boolean;
@@ -151,6 +152,24 @@ function getSummarizeModel(
return [currentModel, providerName];
}
function getModelDisplayName(
model: ModelType,
providerName: ServiceProvider,
): string | undefined {
const configStore = useAppConfig.getState();
const accessStore = useAccessStore.getState();
const allModel = collectModelsWithDefaultModel(
configStore.models,
[configStore.customModels, accessStore.customModels].join(","),
accessStore.defaultModel,
);
const matchedModel = allModel.find(
(m) => m.name === model && m.provider?.providerName === providerName,
);
return matchedModel ? matchedModel.displayName : undefined;
}
function countMessages(msgs: ChatMessage[]) {
return msgs.reduce(
(pre, cur) => pre + estimateTokenLength(getMessageTextContent(cur)),
@@ -437,6 +456,10 @@ export const useChatStore = createPersistStore(
role: "assistant",
streaming: true,
model: modelConfig.model,
modelDisplayName: getModelDisplayName(
modelConfig.model,
modelConfig.providerName,
),
});
// get recent messages

View File

@@ -304,7 +304,9 @@ export function getTimeoutMSByModel(model: string) {
model.startsWith("o1") ||
model.startsWith("o3") ||
model.includes("deepseek-r") ||
model.includes("-thinking")
model.includes("-thinking") ||
model.startsWith("ep-") ||
model.startsWith("bot-")
)
return REQUEST_TIMEOUT_MS_FOR_THINKING;
return REQUEST_TIMEOUT_MS;