mirror of
https://github.com/songquanpeng/one-api.git
synced 2025-09-29 22:56:39 +08:00
135 lines
4.5 KiB
Go
135 lines
4.5 KiB
Go
package controller
|
|
|
|
import (
|
|
"bytes"
|
|
"encoding/json"
|
|
"fmt"
|
|
"io"
|
|
"net/http"
|
|
|
|
"github.com/gin-gonic/gin"
|
|
"github.com/songquanpeng/one-api/common/ctxkey"
|
|
"github.com/songquanpeng/one-api/common/logger"
|
|
"github.com/songquanpeng/one-api/relay"
|
|
"github.com/songquanpeng/one-api/relay/adaptor/openai"
|
|
"github.com/songquanpeng/one-api/relay/apitype"
|
|
"github.com/songquanpeng/one-api/relay/billing"
|
|
"github.com/songquanpeng/one-api/relay/channeltype"
|
|
"github.com/songquanpeng/one-api/relay/meta"
|
|
"github.com/songquanpeng/one-api/relay/model"
|
|
)
|
|
|
|
func (rl *defaultRelay) RelayTextHelper(c *gin.Context) *model.ErrorWithStatusCode {
|
|
meta := meta.GetByContext(c)
|
|
// get & validate textRequest
|
|
textRequest, err := getAndValidateTextRequest(c, meta.Mode)
|
|
if err != nil {
|
|
logger.Errorf(c, "getAndValidateTextRequest failed: %s", err.Error())
|
|
return openai.ErrorWrapper(err, "invalid_text_request", http.StatusBadRequest)
|
|
}
|
|
meta.IsStream = textRequest.Stream
|
|
|
|
// map model name
|
|
var isModelMapped bool
|
|
meta.OriginModelName = textRequest.Model
|
|
textRequest.Model, isModelMapped = getMappedModelName(textRequest.Model, meta.ModelMapping)
|
|
meta.ActualModelName = textRequest.Model
|
|
// get model ratio & group ratio
|
|
var (
|
|
preConsumedQuota int64
|
|
modelRatio float64
|
|
groupRatio float64
|
|
ratio float64
|
|
)
|
|
if rl.Bookkeeper != nil {
|
|
modelRatio = rl.ModelRatio(textRequest.Model)
|
|
groupRatio = rl.GroupRation(meta.Group)
|
|
ratio = modelRatio * groupRatio
|
|
// pre-consume quota
|
|
meta.PromptTokens = getPromptTokens(textRequest, meta.Mode)
|
|
preConsumeQuota := getPreConsumedQuota(textRequest, meta.PromptTokens, ratio)
|
|
consumedQuota, bizErr := rl.PreConsumeQuota(c, preConsumeQuota, meta.UserId, meta.TokenId)
|
|
if bizErr != nil {
|
|
logger.Warnf(c, "preConsumeQuota failed: %+v", *bizErr)
|
|
return bizErr
|
|
}
|
|
preConsumedQuota = consumedQuota
|
|
}
|
|
adaptor := relay.GetAdaptor(meta.APIType)
|
|
if adaptor == nil {
|
|
return openai.ErrorWrapper(fmt.Errorf("invalid api type: %d", meta.APIType), "invalid_api_type", http.StatusBadRequest)
|
|
}
|
|
adaptor.Init(meta)
|
|
|
|
// get request body
|
|
var requestBody io.Reader
|
|
if meta.APIType == apitype.OpenAI {
|
|
// no need to convert request for openai
|
|
shouldResetRequestBody := isModelMapped || meta.ChannelType == channeltype.Baichuan // frequency_penalty 0 is not acceptable for baichuan
|
|
if shouldResetRequestBody {
|
|
jsonStr, err := json.Marshal(textRequest)
|
|
if err != nil {
|
|
return openai.ErrorWrapper(err, "json_marshal_failed", http.StatusInternalServerError)
|
|
}
|
|
requestBody = bytes.NewBuffer(jsonStr)
|
|
} else {
|
|
requestBody = c.Request.Body
|
|
}
|
|
} else {
|
|
convertedRequest, err := adaptor.ConvertRequest(c, meta.Mode, textRequest)
|
|
if err != nil {
|
|
return openai.ErrorWrapper(err, "convert_request_failed", http.StatusInternalServerError)
|
|
}
|
|
jsonData, err := json.Marshal(convertedRequest)
|
|
if err != nil {
|
|
return openai.ErrorWrapper(err, "json_marshal_failed", http.StatusInternalServerError)
|
|
}
|
|
logger.Debugf(c, "converted request: \n%s", string(jsonData))
|
|
requestBody = bytes.NewBuffer(jsonData)
|
|
}
|
|
|
|
// do request
|
|
resp, err := adaptor.DoRequest(c, meta, requestBody)
|
|
if err != nil {
|
|
logger.Errorf(c, "DoRequest failed: %s", err.Error())
|
|
return openai.ErrorWrapper(err, "do_request_failed", http.StatusInternalServerError)
|
|
}
|
|
refund := func() {
|
|
if rl.Bookkeeper != nil && preConsumedQuota > 0 {
|
|
rl.RefundQuota(c, preConsumedQuota, meta.TokenId)
|
|
}
|
|
}
|
|
if isErrorHappened(meta, resp) {
|
|
refund()
|
|
return RelayErrorHandler(resp)
|
|
}
|
|
|
|
// do response
|
|
usage, respErr := adaptor.DoResponse(c, resp, meta)
|
|
if respErr != nil {
|
|
logger.Errorf(c, "respErr is not nil: %+v", respErr)
|
|
refund()
|
|
return respErr
|
|
}
|
|
// post-consume quota
|
|
if rl.Bookkeeper != nil {
|
|
// go postConsumeQuota(c, usage, meta, textRequest, ratio, preConsumedQuota, modelRatio, groupRatio)
|
|
completionRatio := rl.ModelCompletionRatio(textRequest.Model)
|
|
logContent := fmt.Sprintf("模型倍率 %.2f,分组倍率 %.2f,补全倍率 %.2f", modelRatio, groupRatio, completionRatio)
|
|
consumeLog := &billing.ConsumeLog{
|
|
UserId: meta.UserId,
|
|
ChannelId: meta.ChannelId,
|
|
ModelName: textRequest.Model,
|
|
TokenName: c.GetString(ctxkey.TokenName),
|
|
TokenId: meta.TokenId,
|
|
Quota: usage.Quota(completionRatio, ratio),
|
|
Content: logContent,
|
|
PromptTokens: usage.PromptTokens,
|
|
CompletionTokens: usage.CompletionTokens,
|
|
PreConsumedQuota: preConsumedQuota,
|
|
}
|
|
rl.Bookkeeper.Consume(c, consumeLog)
|
|
}
|
|
return nil
|
|
}
|