From a97bdebd0af78573d9abcb9f44b814db54b1dfaf Mon Sep 17 00:00:00 2001 From: CaIon <1808837298@qq.com> Date: Fri, 17 Nov 2023 18:24:37 +0800 Subject: [PATCH 1/3] support gpt-4-1106-vision-preview --- controller/channel-test.go | 7 ++++++- controller/relay-aiproxy.go | 4 ++-- controller/relay-ali.go | 11 ++++++----- controller/relay-baidu.go | 7 ++++--- controller/relay-claude.go | 3 ++- controller/relay-openai.go | 2 +- controller/relay-palm.go | 5 +++-- controller/relay-tencent.go | 7 ++++--- controller/relay-text.go | 6 +++++- controller/relay-utils.go | 31 +++++++++++++++++++++++++------ controller/relay-xunfei.go | 7 ++++--- controller/relay-zhipu.go | 7 ++++--- controller/relay.go | 18 +++++++++++++++--- 13 files changed, 81 insertions(+), 34 deletions(-) diff --git a/controller/channel-test.go b/controller/channel-test.go index 18f1e9b..c0ac9a6 100644 --- a/controller/channel-test.go +++ b/controller/channel-test.go @@ -86,9 +86,10 @@ func buildTestRequest() *ChatRequest { Model: "", // this will be set later MaxTokens: 1, } + content, _ := json.Marshal("hi") testMessage := Message{ Role: "user", - Content: "hi", + Content: content, } testRequest.Messages = append(testRequest.Messages, testMessage) return testRequest @@ -186,6 +187,10 @@ func testAllChannels(notify bool) error { err = errors.New(fmt.Sprintf("响应时间 %.2fs 超过阈值 %.2fs", float64(milliseconds)/1000.0, float64(disableThreshold)/1000.0)) ban = true } + if openaiErr != nil { + err = errors.New(fmt.Sprintf("type %s, code %v, message %s", openaiErr.Type, openaiErr.Code, openaiErr.Message)) + ban = true + } // parse *int to bool if channel.AutoBan != nil && *channel.AutoBan == 0 { ban = false diff --git a/controller/relay-aiproxy.go b/controller/relay-aiproxy.go index d0159ce..7dbf679 100644 --- a/controller/relay-aiproxy.go +++ b/controller/relay-aiproxy.go @@ -48,7 +48,7 @@ type AIProxyLibraryStreamResponse struct { func requestOpenAI2AIProxyLibrary(request GeneralOpenAIRequest) *AIProxyLibraryRequest { query := "" if len(request.Messages) != 0 { - query = request.Messages[len(request.Messages)-1].Content + query = string(request.Messages[len(request.Messages)-1].Content) } return &AIProxyLibraryRequest{ Model: request.Model, @@ -69,7 +69,7 @@ func aiProxyDocuments2Markdown(documents []AIProxyLibraryDocument) string { } func responseAIProxyLibrary2OpenAI(response *AIProxyLibraryResponse) *OpenAITextResponse { - content := response.Answer + aiProxyDocuments2Markdown(response.Documents) + content, _ := json.Marshal(response.Answer + aiProxyDocuments2Markdown(response.Documents)) choice := OpenAITextResponseChoice{ Index: 0, Message: Message{ diff --git a/controller/relay-ali.go b/controller/relay-ali.go index 50dc743..6a79d2b 100644 --- a/controller/relay-ali.go +++ b/controller/relay-ali.go @@ -88,18 +88,18 @@ func requestOpenAI2Ali(request GeneralOpenAIRequest) *AliChatRequest { message := request.Messages[i] if message.Role == "system" { messages = append(messages, AliMessage{ - User: message.Content, + User: string(message.Content), Bot: "Okay", }) continue } else { if i == len(request.Messages)-1 { - prompt = message.Content + prompt = string(message.Content) break } messages = append(messages, AliMessage{ - User: message.Content, - Bot: request.Messages[i+1].Content, + User: string(message.Content), + Bot: string(request.Messages[i+1].Content), }) i++ } @@ -184,11 +184,12 @@ func embeddingResponseAli2OpenAI(response *AliEmbeddingResponse) *OpenAIEmbeddin } func responseAli2OpenAI(response *AliChatResponse) *OpenAITextResponse { + content, _ := json.Marshal(response.Output.Text) choice := OpenAITextResponseChoice{ Index: 0, Message: Message{ Role: "assistant", - Content: response.Output.Text, + Content: content, }, FinishReason: response.Output.FinishReason, } diff --git a/controller/relay-baidu.go b/controller/relay-baidu.go index ed08ac0..05bbad0 100644 --- a/controller/relay-baidu.go +++ b/controller/relay-baidu.go @@ -89,7 +89,7 @@ func requestOpenAI2Baidu(request GeneralOpenAIRequest) *BaiduChatRequest { if message.Role == "system" { messages = append(messages, BaiduMessage{ Role: "user", - Content: message.Content, + Content: string(message.Content), }) messages = append(messages, BaiduMessage{ Role: "assistant", @@ -98,7 +98,7 @@ func requestOpenAI2Baidu(request GeneralOpenAIRequest) *BaiduChatRequest { } else { messages = append(messages, BaiduMessage{ Role: message.Role, - Content: message.Content, + Content: string(message.Content), }) } } @@ -109,11 +109,12 @@ func requestOpenAI2Baidu(request GeneralOpenAIRequest) *BaiduChatRequest { } func responseBaidu2OpenAI(response *BaiduChatResponse) *OpenAITextResponse { + content, _ := json.Marshal(response.Result) choice := OpenAITextResponseChoice{ Index: 0, Message: Message{ Role: "assistant", - Content: response.Result, + Content: content, }, FinishReason: "stop", } diff --git a/controller/relay-claude.go b/controller/relay-claude.go index 1f4a3e7..e131263 100644 --- a/controller/relay-claude.go +++ b/controller/relay-claude.go @@ -93,11 +93,12 @@ func streamResponseClaude2OpenAI(claudeResponse *ClaudeResponse) *ChatCompletion } func responseClaude2OpenAI(claudeResponse *ClaudeResponse) *OpenAITextResponse { + content, _ := json.Marshal(strings.TrimPrefix(claudeResponse.Completion, " ")) choice := OpenAITextResponseChoice{ Index: 0, Message: Message{ Role: "assistant", - Content: strings.TrimPrefix(claudeResponse.Completion, " "), + Content: content, Name: nil, }, FinishReason: stopReasonClaude2OpenAI(claudeResponse.StopReason), diff --git a/controller/relay-openai.go b/controller/relay-openai.go index 6bdfbc0..9b08f85 100644 --- a/controller/relay-openai.go +++ b/controller/relay-openai.go @@ -132,7 +132,7 @@ func openaiHandler(c *gin.Context, resp *http.Response, consumeQuota bool, promp if textResponse.Usage.TotalTokens == 0 { completionTokens := 0 for _, choice := range textResponse.Choices { - completionTokens += countTokenText(choice.Message.Content, model) + completionTokens += countTokenText(string(choice.Message.Content), model) } textResponse.Usage = Usage{ PromptTokens: promptTokens, diff --git a/controller/relay-palm.go b/controller/relay-palm.go index a705b31..a7b0c1f 100644 --- a/controller/relay-palm.go +++ b/controller/relay-palm.go @@ -59,7 +59,7 @@ func requestOpenAI2PaLM(textRequest GeneralOpenAIRequest) *PaLMChatRequest { } for _, message := range textRequest.Messages { palmMessage := PaLMChatMessage{ - Content: message.Content, + Content: string(message.Content), } if message.Role == "user" { palmMessage.Author = "0" @@ -76,11 +76,12 @@ func responsePaLM2OpenAI(response *PaLMChatResponse) *OpenAITextResponse { Choices: make([]OpenAITextResponseChoice, 0, len(response.Candidates)), } for i, candidate := range response.Candidates { + content, _ := json.Marshal(candidate.Content) choice := OpenAITextResponseChoice{ Index: i, Message: Message{ Role: "assistant", - Content: candidate.Content, + Content: content, }, FinishReason: "stop", } diff --git a/controller/relay-tencent.go b/controller/relay-tencent.go index 024468b..c96e6d4 100644 --- a/controller/relay-tencent.go +++ b/controller/relay-tencent.go @@ -84,7 +84,7 @@ func requestOpenAI2Tencent(request GeneralOpenAIRequest) *TencentChatRequest { if message.Role == "system" { messages = append(messages, TencentMessage{ Role: "user", - Content: message.Content, + Content: string(message.Content), }) messages = append(messages, TencentMessage{ Role: "assistant", @@ -93,7 +93,7 @@ func requestOpenAI2Tencent(request GeneralOpenAIRequest) *TencentChatRequest { continue } messages = append(messages, TencentMessage{ - Content: message.Content, + Content: string(message.Content), Role: message.Role, }) } @@ -119,11 +119,12 @@ func responseTencent2OpenAI(response *TencentChatResponse) *OpenAITextResponse { Usage: response.Usage, } if len(response.Choices) > 0 { + content, _ := json.Marshal(response.Choices[0].Messages.Content) choice := OpenAITextResponseChoice{ Index: 0, Message: Message{ Role: "assistant", - Content: response.Choices[0].Messages.Content, + Content: content, }, FinishReason: response.Choices[0].FinishReason, } diff --git a/controller/relay-text.go b/controller/relay-text.go index 2729650..a009267 100644 --- a/controller/relay-text.go +++ b/controller/relay-text.go @@ -199,9 +199,13 @@ func relayTextHelper(c *gin.Context, relayMode int) *OpenAIErrorWithStatusCode { } var promptTokens int var completionTokens int + var err error switch relayMode { case RelayModeChatCompletions: - promptTokens = countTokenMessages(textRequest.Messages, textRequest.Model) + promptTokens, err = countTokenMessages(textRequest.Messages, textRequest.Model) + if err != nil { + return errorWrapper(err, "count_token_messages_failed", http.StatusInternalServerError) + } case RelayModeCompletions: promptTokens = countTokenInput(textRequest.Prompt, textRequest.Model) case RelayModeModerations: diff --git a/controller/relay-utils.go b/controller/relay-utils.go index 40aa547..177d853 100644 --- a/controller/relay-utils.go +++ b/controller/relay-utils.go @@ -63,7 +63,8 @@ func getTokenNum(tokenEncoder *tiktoken.Tiktoken, text string) int { return len(tokenEncoder.Encode(text, nil, nil)) } -func countTokenMessages(messages []Message, model string) int { +func countTokenMessages(messages []Message, model string) (int, error) { + //recover when panic tokenEncoder := getTokenEncoder(model) // Reference: // https://github.com/openai/openai-cookbook/blob/main/examples/How_to_count_tokens_with_tiktoken.ipynb @@ -82,15 +83,33 @@ func countTokenMessages(messages []Message, model string) int { tokenNum := 0 for _, message := range messages { tokenNum += tokensPerMessage - tokenNum += getTokenNum(tokenEncoder, message.Content) tokenNum += getTokenNum(tokenEncoder, message.Role) - if message.Name != nil { - tokenNum += tokensPerName - tokenNum += getTokenNum(tokenEncoder, *message.Name) + var arrayContent []MediaMessage + if err := json.Unmarshal(message.Content, &arrayContent); err != nil { + + var stringContent string + if err := json.Unmarshal(message.Content, &stringContent); err != nil { + return 0, err + } else { + tokenNum += getTokenNum(tokenEncoder, stringContent) + if message.Name != nil { + tokenNum += tokensPerName + tokenNum += getTokenNum(tokenEncoder, *message.Name) + } + } + } else { + for _, m := range arrayContent { + if m.Type == "image_url" { + //TODO: getImageToken + tokenNum += 1000 + } else { + tokenNum += getTokenNum(tokenEncoder, m.Text) + } + } } } tokenNum += 3 // Every reply is primed with <|start|>assistant<|message|> - return tokenNum + return tokenNum, nil } func countTokenInput(input any, model string) int { diff --git a/controller/relay-xunfei.go b/controller/relay-xunfei.go index 91fb604..33383d8 100644 --- a/controller/relay-xunfei.go +++ b/controller/relay-xunfei.go @@ -81,7 +81,7 @@ func requestOpenAI2Xunfei(request GeneralOpenAIRequest, xunfeiAppId string, doma if message.Role == "system" { messages = append(messages, XunfeiMessage{ Role: "user", - Content: message.Content, + Content: string(message.Content), }) messages = append(messages, XunfeiMessage{ Role: "assistant", @@ -90,7 +90,7 @@ func requestOpenAI2Xunfei(request GeneralOpenAIRequest, xunfeiAppId string, doma } else { messages = append(messages, XunfeiMessage{ Role: message.Role, - Content: message.Content, + Content: string(message.Content), }) } } @@ -112,11 +112,12 @@ func responseXunfei2OpenAI(response *XunfeiChatResponse) *OpenAITextResponse { }, } } + content, _ := json.Marshal(response.Payload.Choices.Text[0].Content) choice := OpenAITextResponseChoice{ Index: 0, Message: Message{ Role: "assistant", - Content: response.Payload.Choices.Text[0].Content, + Content: content, }, FinishReason: stopFinishReason, } diff --git a/controller/relay-zhipu.go b/controller/relay-zhipu.go index 7a4a582..5ad4151 100644 --- a/controller/relay-zhipu.go +++ b/controller/relay-zhipu.go @@ -114,7 +114,7 @@ func requestOpenAI2Zhipu(request GeneralOpenAIRequest) *ZhipuRequest { if message.Role == "system" { messages = append(messages, ZhipuMessage{ Role: "system", - Content: message.Content, + Content: string(message.Content), }) messages = append(messages, ZhipuMessage{ Role: "user", @@ -123,7 +123,7 @@ func requestOpenAI2Zhipu(request GeneralOpenAIRequest) *ZhipuRequest { } else { messages = append(messages, ZhipuMessage{ Role: message.Role, - Content: message.Content, + Content: string(message.Content), }) } } @@ -144,11 +144,12 @@ func responseZhipu2OpenAI(response *ZhipuResponse) *OpenAITextResponse { Usage: response.Data.Usage, } for i, choice := range response.Data.Choices { + content, _ := json.Marshal(strings.Trim(choice.Content, "\"")) openaiChoice := OpenAITextResponseChoice{ Index: i, Message: Message{ Role: choice.Role, - Content: strings.Trim(choice.Content, "\""), + Content: content, }, FinishReason: "", } diff --git a/controller/relay.go b/controller/relay.go index 21cbfba..9e910fa 100644 --- a/controller/relay.go +++ b/controller/relay.go @@ -1,6 +1,7 @@ package controller import ( + "encoding/json" "fmt" "log" "net/http" @@ -12,9 +13,20 @@ import ( ) type Message struct { - Role string `json:"role"` - Content string `json:"content"` - Name *string `json:"name,omitempty"` + Role string `json:"role"` + Content json.RawMessage `json:"content"` + Name *string `json:"name,omitempty"` +} + +type MediaMessage struct { + Type string `json:"type"` + Text string `json:"text"` + ImageUrl MessageImageUrl `json:"image_url,omitempty"` +} + +type MessageImageUrl struct { + Url string `json:"url"` + Detail string `json:"detail"` } const ( From 2d1ca2d9be66f960fc2cb1fdbc35896f9a7218d1 Mon Sep 17 00:00:00 2001 From: CaIon <1808837298@qq.com> Date: Fri, 17 Nov 2023 20:32:11 +0800 Subject: [PATCH 2/3] fix image token calculate --- controller/relay-utils.go | 74 +++++++++++++++++++++++++++++++++++++-- go.mod | 1 + go.sum | 2 ++ 3 files changed, 75 insertions(+), 2 deletions(-) diff --git a/controller/relay-utils.go b/controller/relay-utils.go index 177d853..1873cab 100644 --- a/controller/relay-utils.go +++ b/controller/relay-utils.go @@ -2,10 +2,18 @@ package controller import ( "encoding/json" + "errors" "fmt" + "github.com/chai2010/webp" "github.com/gin-gonic/gin" "github.com/pkoukk/tiktoken-go" + "image" + _ "image/gif" + _ "image/jpeg" + _ "image/png" "io" + "log" + "math" "net/http" "one-api/common" "strconv" @@ -63,6 +71,64 @@ func getTokenNum(tokenEncoder *tiktoken.Tiktoken, text string) int { return len(tokenEncoder.Encode(text, nil, nil)) } +func getImageToken(imageUrl MessageImageUrl) (int, error) { + if imageUrl.Detail == "low" { + return 85, nil + } + + response, err := http.Get(imageUrl.Url) + if err != nil { + fmt.Println("Error: Failed to get the URL") + return 0, err + } + + defer response.Body.Close() + + // 限制读取的字节数,防止下载整个图片 + limitReader := io.LimitReader(response.Body, 8192) + + // 读取图片的头部信息来获取图片尺寸 + config, _, err := image.DecodeConfig(limitReader) + if err != nil { + common.SysLog(fmt.Sprintf("fail to decode image config(gif, jpg, png): %s", err.Error())) + config, err = webp.DecodeConfig(limitReader) + if err != nil { + common.SysLog(fmt.Sprintf("fail to decode image config(webp): %s", err.Error())) + } + } + if config.Width == 0 || config.Height == 0 { + return 0, errors.New(fmt.Sprintf("fail to decode image config: %s", err.Error())) + } + if config.Width < 512 && config.Height < 512 { + if imageUrl.Detail == "auto" || imageUrl.Detail == "" { + return 85, nil + } + } + + shortSide := config.Width + otherSide := config.Height + log.Printf("width: %d, height: %d", config.Width, config.Height) + // 缩放倍数 + scale := 1.0 + if config.Height < shortSide { + shortSide = config.Height + otherSide = config.Width + } + + // 将最小变的尺寸缩小到768以下,如果大于768,则缩放到768 + if shortSide > 768 { + scale = float64(shortSide) / 768 + shortSide = 768 + } + // 将另一边按照相同的比例缩小,向上取整 + otherSide = int(math.Ceil(float64(otherSide) / scale)) + log.Printf("shortSide: %d, otherSide: %d, scale: %f", shortSide, otherSide, scale) + // 计算图片的token数量(边的长度除以512,向上取整) + tiles := (shortSide + 511) / 512 * ((otherSide + 511) / 512) + log.Printf("tiles: %d", tiles) + return tiles*170 + 85, nil +} + func countTokenMessages(messages []Message, model string) (int, error) { //recover when panic tokenEncoder := getTokenEncoder(model) @@ -100,8 +166,12 @@ func countTokenMessages(messages []Message, model string) (int, error) { } else { for _, m := range arrayContent { if m.Type == "image_url" { - //TODO: getImageToken - tokenNum += 1000 + imageTokenNum, err := getImageToken(m.ImageUrl) + if err != nil { + return 0, err + } + tokenNum += imageTokenNum + log.Printf("image token num: %d", imageTokenNum) } else { tokenNum += getTokenNum(tokenEncoder, m.Text) } diff --git a/go.mod b/go.mod index a82121b..3a75341 100644 --- a/go.mod +++ b/go.mod @@ -4,6 +4,7 @@ module one-api go 1.18 require ( + github.com/chai2010/webp v1.1.1 github.com/gin-contrib/cors v1.4.0 github.com/gin-contrib/gzip v0.0.6 github.com/gin-contrib/sessions v0.0.5 diff --git a/go.sum b/go.sum index 2d64620..6e7f963 100644 --- a/go.sum +++ b/go.sum @@ -3,6 +3,8 @@ github.com/bytedance/sonic v1.9.1 h1:6iJ6NqdoxCDr6mbY8h18oSO+cShGSMRGCEo7F2h0x8s github.com/bytedance/sonic v1.9.1/go.mod h1:i736AoUSYt75HyZLoJW9ERYxcy6eaN6h4BZXU064P/U= github.com/cespare/xxhash/v2 v2.1.2 h1:YRXhKfTDauu4ajMg1TPgFO5jnlC2HCbmLXMcTG5cbYE= github.com/cespare/xxhash/v2 v2.1.2/go.mod h1:VGX0DQ3Q6kWi7AoAeZDth3/j3BFtOZR5XLFGgcrjCOs= +github.com/chai2010/webp v1.1.1 h1:jTRmEccAJ4MGrhFOrPMpNGIJ/eybIgwKpcACsrTEapk= +github.com/chai2010/webp v1.1.1/go.mod h1:0XVwvZWdjjdxpUEIf7b9g9VkHFnInUSYujwqTLEuldU= github.com/chenzhuoyu/base64x v0.0.0-20211019084208-fb5309c8db06/go.mod h1:DH46F32mSOjUmXrMHnKwZdA8wcEefY7UVqBKYGjpdQY= github.com/chenzhuoyu/base64x v0.0.0-20221115062448-fe3a3abad311 h1:qSGYFH7+jGhDF8vLC+iwCD4WpbV1EBDSzWkJODFLams= github.com/chenzhuoyu/base64x v0.0.0-20221115062448-fe3a3abad311/go.mod h1:b583jCggY9gE99b6G5LEC39OIiVsWj+R97kbl5odCEk= From 3f085b612694f943a380f0ab6d508e5b8f07d932 Mon Sep 17 00:00:00 2001 From: CaIon <1808837298@qq.com> Date: Fri, 17 Nov 2023 20:47:51 +0800 Subject: [PATCH 3/3] update README.md --- README.md | 1 + 1 file changed, 1 insertion(+) diff --git a/README.md b/README.md index 89dfba3..edc34d9 100644 --- a/README.md +++ b/README.md @@ -28,6 +28,7 @@ + 配合项目[neko-api-key-tool](https://github.com/Calcium-Ion/neko-api-key-tool)可实现用key查询使用情况,方便二次分销 5. 渠道显示已使用额度,支持指定组织访问 6. 分页支持选择每页显示数量 +7. 支持gpt-4-1106-vision-preview,dall-e-3,tts-1 ## 界面截图 ![image](https://github.com/Calcium-Ion/new-api/assets/61247483/3ca0b282-00ff-4c96-bf9d-e29ef615c605)