Compare commits

...

2 Commits

Author SHA1 Message Date
Davidlasky
ac7b720b5b send longer chat history 2025-04-15 13:37:26 -05:00
Davidlasky
d02f9b0dd4 added gpt-4.1 and removed gpt-4.5-preview 2025-04-15 12:52:30 -05:00
4 changed files with 10 additions and 23 deletions

View File

@@ -408,7 +408,7 @@ You are an AI assistant with access to system tools. Your role is to help users
`;
export const SUMMARIZE_MODEL = "gpt-4o-mini";
export const SUMMARIZE_MODEL = "gpt-4.1-mini";
export const GEMINI_SUMMARIZE_MODEL = "gemini-2.0-flash";
export const DEEPSEEK_SUMMARIZE_MODEL = "deepseek-chat";
@@ -420,6 +420,8 @@ export const KnowledgeCutOffDate: Record<string, string> = {
"gemini-2.0-flash": "2024-08",
"claude-3-7-sonnet-latest": "2024-10",
"claude-3-5-haiku-latest": "2024-10",
"gpt-4.1": "2024-06",
"gpt-4.1-mini": "2024-06",
"deepseek-chat": "2024-07",
"deepseek-coder": "2024-07",
};
@@ -440,7 +442,7 @@ export const DEFAULT_TTS_VOICES = [
export const VISION_MODEL_REGEXES = [
/vision/,
/gpt-4o/,
/gpt-4\.1/,
/claude-3/,
/gemini-1\.5/,
/gemini-exp/,
@@ -458,24 +460,9 @@ export const VISION_MODEL_REGEXES = [
export const EXCLUDE_VISION_MODEL_REGEXES = [/claude-3-5-haiku-20241022/];
const openaiModels = [
// As of July 2024, gpt-4o-mini should be used in place of gpt-3.5-turbo,
// as it is cheaper, more capable, multimodal, and just as fast. gpt-3.5-turbo is still available for use in the API.
"gpt-3.5-turbo",
"gpt-4o",
"gpt-4o-mini",
"dall-e-3",
"o1-mini",
"o1-preview",
"o1",
"o3-mini",
"gpt-4.5-preview",
];
const openaiModels = ["dall-e-3", "o1", "o3-mini", "gpt-4.1", "gpt-4.1-mini"];
const googleModels = [
"gemini-1.5-pro",
"gemini-1.5-flash",
"gemini-1.5-flash-8b",
"gemini-2.0-flash",
"gemini-2.0-flash-lite",
"gemini-2.5-pro-exp-03-25",

View File

@@ -66,14 +66,14 @@ export const DEFAULT_CONFIG = {
modelConfig: {
model: "gpt-4o-mini" as ModelType,
providerName: "OpenAI" as ServiceProvider,
temperature: 0.5,
temperature: 0.2,
top_p: 1,
max_tokens: 4000,
presence_penalty: 0,
frequency_penalty: 0,
sendMemory: true,
historyMessageCount: 4,
compressMessageLengthThreshold: 1000,
historyMessageCount: 20,
compressMessageLengthThreshold: 5000,
compressModel: "",
compressProviderName: "",
enableInjectSystemPrompts: true,

View File

@@ -3,7 +3,7 @@ import { isModelNotavailableInServer } from "../app/utils/model";
describe("isModelNotavailableInServer", () => {
test("test model will return false, which means the model is available", () => {
const customModels = "";
const modelName = "gpt-4o";
const modelName = "gpt-4.1";
const providerNames = "OpenAI";
const result = isModelNotavailableInServer(
customModels,

View File

@@ -15,7 +15,7 @@ describe("isVisionModel", () => {
test("should identify vision models using regex patterns", () => {
const visionModels = [
"gpt-4o",
"gpt-4.1",
"claude-3-opus",
"gemini-1.5-pro",
"gemini-2.0",