This commit is contained in:
GH Action - Upstream Sync 2024-02-28 00:41:24 +00:00
commit 6bd4c050d0

View File

@ -110,6 +110,16 @@ export class ChatGPTApi implements LLMApi {
// Please do not ask me why not send max_tokens, no reason, this param is just shit, I dont want to explain anymore.
};
// add max_tokens to vision model
if (visionModel) {
Object.defineProperty(requestPayload, "max_tokens", {
enumerable: true,
configurable: true,
writable: true,
value: Math.max(modelConfig.max_tokens, 4096),
});
}
console.log("[Request] openai payload: ", requestPayload);
const shouldStream = !!options.config.stream;