mirror of
https://github.com/songquanpeng/one-api.git
synced 2025-11-17 21:53:42 +08:00
🔖 chore: Rename relay/util to relay/relay_util package and add utils package
This commit is contained in:
@@ -5,6 +5,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
)
|
||||
@@ -92,7 +93,7 @@ func (p *AliProvider) convertToChatOpenai(response *AliChatResponse, request *ty
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
ID: response.RequestId,
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: request.Model,
|
||||
Choices: response.Output.ToChatCompletionChoices(),
|
||||
Usage: &types.Usage{
|
||||
@@ -223,7 +224,7 @@ func (h *aliStreamHandler) convertToOpenaiStream(aliResponse *AliChatResponse, d
|
||||
streamResponse := types.ChatCompletionStreamResponse{
|
||||
ID: aliResponse.RequestId,
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
Choices: []types.ChatCompletionStreamChoice{choice},
|
||||
}
|
||||
|
||||
@@ -8,6 +8,7 @@ import (
|
||||
"one-api/common"
|
||||
"one-api/common/image"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/providers/base"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
@@ -172,7 +173,7 @@ func ConvertToChatOpenai(provider base.ProviderInterface, response *ClaudeRespon
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
ID: response.Id,
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Choices: []types.ChatCompletionChoice{choice},
|
||||
Model: request.Model,
|
||||
Usage: &types.Usage{
|
||||
@@ -264,9 +265,9 @@ func (h *ClaudeStreamHandler) convertToOpenaiStream(claudeResponse *ClaudeStream
|
||||
choice.FinishReason = &finishReason
|
||||
}
|
||||
chatCompletion := types.ChatCompletionStreamResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
Choices: []types.ChatCompletionStreamChoice{choice},
|
||||
}
|
||||
|
||||
@@ -7,6 +7,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
)
|
||||
@@ -85,9 +86,9 @@ func (p *CloudflareAIProvider) convertToChatOpenai(response *ChatRespone, reques
|
||||
}
|
||||
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: request.Model,
|
||||
Choices: []types.ChatCompletionChoice{{
|
||||
Index: 0,
|
||||
@@ -155,9 +156,9 @@ func (h *CloudflareAIStreamHandler) handlerStream(rawLine *[]byte, dataChan chan
|
||||
|
||||
func (h *CloudflareAIStreamHandler) convertToOpenaiStream(chatResponse *ChatResult, dataChan chan string, isStop bool) {
|
||||
streamResponse := types.ChatCompletionStreamResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
}
|
||||
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/storage"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"time"
|
||||
)
|
||||
@@ -46,7 +47,7 @@ func (p *CloudflareAIProvider) CreateImageGenerations(request *types.ImageReques
|
||||
|
||||
url := ""
|
||||
if request.ResponseFormat == "" || request.ResponseFormat == "url" {
|
||||
url = storage.Upload(body, common.GetUUID()+".png")
|
||||
url = storage.Upload(body, utils.GetUUID()+".png")
|
||||
}
|
||||
|
||||
openaiResponse := &types.ImageResponse{
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/providers/base"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
@@ -138,7 +139,7 @@ func ConvertToChatOpenai(provider base.ProviderInterface, response *CohereRespon
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
ID: response.GenerationID,
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Choices: []types.ChatCompletionChoice{choice},
|
||||
Model: request.Model,
|
||||
Usage: &types.Usage{},
|
||||
@@ -190,9 +191,9 @@ func (h *CohereStreamHandler) convertToOpenaiStream(cohereResponse *CohereStream
|
||||
}
|
||||
|
||||
chatCompletion := types.ChatCompletionStreamResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
Choices: []types.ChatCompletionStreamChoice{choice},
|
||||
}
|
||||
|
||||
@@ -7,6 +7,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
)
|
||||
@@ -89,9 +90,9 @@ func (p *CozeProvider) convertToChatOpenai(response *CozeResponse, request *type
|
||||
}
|
||||
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: request.Model,
|
||||
Choices: []types.ChatCompletionChoice{{
|
||||
Index: 0,
|
||||
@@ -168,9 +169,9 @@ func (h *CozeStreamHandler) handlerStream(rawLine *[]byte, dataChan chan string,
|
||||
|
||||
func (h *CozeStreamHandler) convertToOpenaiStream(chatResponse *CozeStreamResponse, dataChan chan string) {
|
||||
streamResponse := types.ChatCompletionStreamResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
}
|
||||
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
)
|
||||
@@ -145,9 +146,9 @@ func (p *GeminiProvider) convertToChatOpenai(response *GeminiChatResponse, reque
|
||||
}
|
||||
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: request.Model,
|
||||
Choices: make([]types.ChatCompletionChoice, 0, len(response.Candidates)),
|
||||
}
|
||||
@@ -191,9 +192,9 @@ func (h *geminiStreamHandler) handlerStream(rawLine *[]byte, dataChan chan strin
|
||||
|
||||
func (h *geminiStreamHandler) convertToOpenaiStream(geminiResponse *GeminiChatResponse, dataChan chan string) {
|
||||
streamResponse := types.ChatCompletionStreamResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
// Choices: choices,
|
||||
}
|
||||
|
||||
@@ -5,6 +5,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/image"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
)
|
||||
|
||||
@@ -120,7 +121,7 @@ func (g *GeminiFunctionCall) ToOpenAITool() *types.ChatCompletionToolCalls {
|
||||
args, _ := json.Marshal(g.Args)
|
||||
|
||||
return &types.ChatCompletionToolCalls{
|
||||
Id: "call_" + common.GetRandomString(24),
|
||||
Id: "call_" + utils.GetRandomString(24),
|
||||
Type: types.ChatMessageRoleFunction,
|
||||
Index: 0,
|
||||
Function: &types.ChatCompletionToolCallsFunction{
|
||||
|
||||
@@ -7,6 +7,7 @@ import (
|
||||
"one-api/common"
|
||||
"one-api/common/image"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
)
|
||||
@@ -100,9 +101,9 @@ func (p *OllamaProvider) convertToChatOpenai(response *ChatResponse, request *ty
|
||||
}
|
||||
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: request.Model,
|
||||
Choices: []types.ChatCompletionChoice{choices},
|
||||
Usage: &types.Usage{
|
||||
@@ -195,9 +196,9 @@ func (h *ollamaStreamHandler) handlerStream(rawLine *[]byte, dataChan chan strin
|
||||
}
|
||||
|
||||
chatCompletion := types.ChatCompletionStreamResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
Choices: []types.ChatCompletionStreamChoice{choice},
|
||||
}
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
)
|
||||
@@ -177,11 +178,11 @@ func (h *palmStreamHandler) convertToOpenaiStream(palmChatResponse *PaLMChatResp
|
||||
choice.FinishReason = types.FinishReasonStop
|
||||
|
||||
streamResponse := types.ChatCompletionStreamResponse{
|
||||
ID: fmt.Sprintf("chatcmpl-%s", common.GetUUID()),
|
||||
ID: fmt.Sprintf("chatcmpl-%s", utils.GetUUID()),
|
||||
Object: "chat.completion.chunk",
|
||||
Model: h.Request.Model,
|
||||
Choices: []types.ChatCompletionStreamChoice{choice},
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
}
|
||||
|
||||
responseBody, _ := json.Marshal(streamResponse)
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/storage"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"time"
|
||||
)
|
||||
@@ -71,7 +72,7 @@ func (p *StabilityAIProvider) CreateImageGenerations(request *types.ImageRequest
|
||||
if request.ResponseFormat == "" || request.ResponseFormat == "url" {
|
||||
body, err := base64.StdEncoding.DecodeString(stabilityAIResponse.Image)
|
||||
if err == nil {
|
||||
imgUrl = storage.Upload(body, common.GetUUID()+".png")
|
||||
imgUrl = storage.Upload(body, utils.GetUUID()+".png")
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -6,6 +6,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
)
|
||||
@@ -101,7 +102,7 @@ func (p *TencentProvider) convertToChatOpenai(response *TencentChatResponse, req
|
||||
|
||||
openaiResponse = &types.ChatCompletionResponse{
|
||||
Object: "chat.completion",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Usage: response.Usage,
|
||||
Model: request.Model,
|
||||
}
|
||||
@@ -137,9 +138,9 @@ func convertFromChatOpenai(request *types.ChatCompletionRequest) *TencentChatReq
|
||||
stream = 1
|
||||
}
|
||||
return &TencentChatRequest{
|
||||
Timestamp: common.GetTimestamp(),
|
||||
Expired: common.GetTimestamp() + 24*60*60,
|
||||
QueryID: common.GetUUID(),
|
||||
Timestamp: utils.GetTimestamp(),
|
||||
Expired: utils.GetTimestamp() + 24*60*60,
|
||||
QueryID: utils.GetUUID(),
|
||||
Temperature: request.Temperature,
|
||||
TopP: request.TopP,
|
||||
Stream: stream,
|
||||
@@ -178,7 +179,7 @@ func (h *tencentStreamHandler) handlerStream(rawLine *[]byte, dataChan chan stri
|
||||
func (h *tencentStreamHandler) convertToOpenaiStream(tencentChatResponse *TencentChatResponse, dataChan chan string) {
|
||||
streamResponse := types.ChatCompletionStreamResponse{
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
}
|
||||
if len(tencentChatResponse.Choices) > 0 {
|
||||
|
||||
@@ -8,6 +8,7 @@ import (
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/common/requester"
|
||||
"one-api/common/utils"
|
||||
"one-api/types"
|
||||
"strings"
|
||||
|
||||
@@ -194,7 +195,7 @@ func (h *xunfeiHandler) convertToChatOpenai(stream requester.StreamReaderInterfa
|
||||
ID: xunfeiResponse.Header.Sid,
|
||||
Object: "chat.completion",
|
||||
Model: h.Request.Model,
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Choices: []types.ChatCompletionChoice{choice},
|
||||
Usage: &xunfeiResponse.Payload.Usage.Text,
|
||||
}
|
||||
@@ -310,7 +311,7 @@ func (h *xunfeiHandler) convertToOpenaiStream(xunfeiChatResponse *XunfeiChatResp
|
||||
chatCompletion := types.ChatCompletionStreamResponse{
|
||||
ID: xunfeiChatResponse.Header.Sid,
|
||||
Object: "chat.completion.chunk",
|
||||
Created: common.GetTimestamp(),
|
||||
Created: utils.GetTimestamp(),
|
||||
Model: h.Request.Model,
|
||||
}
|
||||
|
||||
|
||||
Reference in New Issue
Block a user