From 1ea322f3716d70f848691dba74a401a7eca659f4 Mon Sep 17 00:00:00 2001 From: "Laisky.Cai" Date: Fri, 24 Jan 2025 13:41:00 +0000 Subject: [PATCH] fix: update OpenAI adaptor to set default temperature and clarify model limitations --- relay/adaptor/openai/adaptor.go | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/relay/adaptor/openai/adaptor.go b/relay/adaptor/openai/adaptor.go index 5612105c..ea001bee 100644 --- a/relay/adaptor/openai/adaptor.go +++ b/relay/adaptor/openai/adaptor.go @@ -85,9 +85,13 @@ func (a *Adaptor) ConvertRequest(c *gin.Context, relayMode int, request *model.G request.StreamOptions.IncludeUsage = true } - // o1/o1-mini/o1-preview do not support system prompt and max_tokens + // o1/o1-mini/o1-preview do not support system prompt/max_tokens/temperature if strings.HasPrefix(request.Model, "o1") { + temperature := float64(1) + request.Temperature = &temperature // Only the default (1) value is supported + request.MaxTokens = 0 + request.Messages = func(raw []model.Message) (filtered []model.Message) { for i := range raw { if raw[i].Role != "system" {