mirror of
https://github.com/songquanpeng/one-api.git
synced 2025-09-29 22:56:39 +08:00
145 lines
3.7 KiB
Go
145 lines
3.7 KiB
Go
package openai
|
|
|
|
import (
|
|
"encoding/json"
|
|
"io"
|
|
"net/http"
|
|
"one-api/common"
|
|
"one-api/common/requester"
|
|
"one-api/types"
|
|
"strings"
|
|
)
|
|
|
|
type OpenAIStreamHandler struct {
|
|
Usage *types.Usage
|
|
ModelName string
|
|
isAzure bool
|
|
}
|
|
|
|
func (p *OpenAIProvider) CreateChatCompletion(request *types.ChatCompletionRequest) (openaiResponse *types.ChatCompletionResponse, errWithCode *types.OpenAIErrorWithStatusCode) {
|
|
req, errWithCode := p.GetRequestTextBody(common.RelayModeChatCompletions, request.Model, request)
|
|
if errWithCode != nil {
|
|
return nil, errWithCode
|
|
}
|
|
defer req.Body.Close()
|
|
|
|
response := &OpenAIProviderChatResponse{}
|
|
// 发送请求
|
|
_, errWithCode = p.Requester.SendRequest(req, response, false)
|
|
if errWithCode != nil {
|
|
return nil, errWithCode
|
|
}
|
|
|
|
// 检测是否错误
|
|
openaiErr := ErrorHandle(&response.OpenAIErrorResponse)
|
|
if openaiErr != nil {
|
|
errWithCode = &types.OpenAIErrorWithStatusCode{
|
|
OpenAIError: *openaiErr,
|
|
StatusCode: http.StatusBadRequest,
|
|
}
|
|
return nil, errWithCode
|
|
}
|
|
|
|
if response.Usage == nil {
|
|
response.Usage = &types.Usage{
|
|
PromptTokens: p.Usage.PromptTokens,
|
|
CompletionTokens: 0,
|
|
TotalTokens: 0,
|
|
}
|
|
// 那么需要计算
|
|
response.Usage.CompletionTokens = common.CountTokenText(response.GetContent(), request.Model)
|
|
response.Usage.TotalTokens = response.Usage.PromptTokens + response.Usage.CompletionTokens
|
|
}
|
|
|
|
*p.Usage = *response.Usage
|
|
|
|
return &response.ChatCompletionResponse, nil
|
|
}
|
|
|
|
func (p *OpenAIProvider) CreateChatCompletionStream(request *types.ChatCompletionRequest) (requester.StreamReaderInterface[string], *types.OpenAIErrorWithStatusCode) {
|
|
streamOptions := request.StreamOptions
|
|
// 如果支持流式返回Usage 则需要更改配置:
|
|
if p.SupportStreamOptions {
|
|
request.StreamOptions = &types.StreamOptions{
|
|
IncludeUsage: true,
|
|
}
|
|
} else {
|
|
// 避免误传导致报错
|
|
request.StreamOptions = nil
|
|
}
|
|
req, errWithCode := p.GetRequestTextBody(common.RelayModeChatCompletions, request.Model, request)
|
|
if errWithCode != nil {
|
|
return nil, errWithCode
|
|
}
|
|
defer req.Body.Close()
|
|
|
|
// 恢复原来的配置
|
|
request.StreamOptions = streamOptions
|
|
|
|
// 发送请求
|
|
resp, errWithCode := p.Requester.SendRequestRaw(req)
|
|
if errWithCode != nil {
|
|
return nil, errWithCode
|
|
}
|
|
|
|
chatHandler := OpenAIStreamHandler{
|
|
Usage: p.Usage,
|
|
ModelName: request.Model,
|
|
isAzure: p.IsAzure,
|
|
}
|
|
|
|
return requester.RequestStream[string](p.Requester, resp, chatHandler.HandlerChatStream)
|
|
}
|
|
|
|
func (h *OpenAIStreamHandler) HandlerChatStream(rawLine *[]byte, dataChan chan string, errChan chan error) {
|
|
// 如果rawLine 前缀不为data:,则直接返回
|
|
if !strings.HasPrefix(string(*rawLine), "data: ") {
|
|
*rawLine = nil
|
|
return
|
|
}
|
|
|
|
// 去除前缀
|
|
*rawLine = (*rawLine)[6:]
|
|
|
|
// 如果等于 DONE 则结束
|
|
if string(*rawLine) == "[DONE]" {
|
|
errChan <- io.EOF
|
|
*rawLine = requester.StreamClosed
|
|
return
|
|
}
|
|
|
|
var openaiResponse OpenAIProviderChatStreamResponse
|
|
err := json.Unmarshal(*rawLine, &openaiResponse)
|
|
if err != nil {
|
|
errChan <- common.ErrorToOpenAIError(err)
|
|
return
|
|
}
|
|
|
|
error := ErrorHandle(&openaiResponse.OpenAIErrorResponse)
|
|
if error != nil {
|
|
errChan <- error
|
|
return
|
|
}
|
|
|
|
if len(openaiResponse.Choices) == 0 {
|
|
if openaiResponse.Usage != nil {
|
|
*h.Usage = *openaiResponse.Usage
|
|
}
|
|
*rawLine = nil
|
|
return
|
|
}
|
|
|
|
dataChan <- string(*rawLine)
|
|
|
|
if len(openaiResponse.Choices) > 0 && openaiResponse.Choices[0].Usage != nil {
|
|
*h.Usage = *openaiResponse.Choices[0].Usage
|
|
} else {
|
|
if h.Usage.TotalTokens == 0 {
|
|
h.Usage.TotalTokens = h.Usage.PromptTokens
|
|
}
|
|
countTokenText := common.CountTokenText(openaiResponse.getResponseText(), h.ModelName)
|
|
h.Usage.CompletionTokens += countTokenText
|
|
h.Usage.TotalTokens += countTokenText
|
|
}
|
|
}
|