mirror of
https://github.com/songquanpeng/one-api.git
synced 2025-11-15 12:43:41 +08:00
🎨 调整供应商目录结构,合并文本输出函数
This commit is contained in:
@@ -70,7 +70,7 @@ func testChannel(channel *model.Channel, request types.ChatCompletionRequest) (e
|
||||
}
|
||||
|
||||
promptTokens := common.CountTokenMessages(request.Messages, request.Model)
|
||||
_, openAIErrorWithStatusCode := chatProvider.ChatCompleteResponse(&request, isModelMapped, promptTokens)
|
||||
_, openAIErrorWithStatusCode := chatProvider.ChatAction(&request, isModelMapped, promptTokens)
|
||||
if openAIErrorWithStatusCode != nil {
|
||||
return nil, &openAIErrorWithStatusCode.OpenAIError
|
||||
}
|
||||
|
||||
@@ -2,6 +2,7 @@ package controller
|
||||
|
||||
import (
|
||||
"fmt"
|
||||
"one-api/types"
|
||||
|
||||
"github.com/gin-gonic/gin"
|
||||
)
|
||||
@@ -541,7 +542,7 @@ func RetrieveModel(c *gin.Context) {
|
||||
if model, ok := openAIModelsMap[modelId]; ok {
|
||||
c.JSON(200, model)
|
||||
} else {
|
||||
openAIError := OpenAIError{
|
||||
openAIError := types.OpenAIError{
|
||||
Message: fmt.Sprintf("The model '%s' does not exist", modelId),
|
||||
Type: "invalid_request_error",
|
||||
Param: "model",
|
||||
|
||||
@@ -1,127 +0,0 @@
|
||||
package controller
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/model"
|
||||
"one-api/providers"
|
||||
"one-api/types"
|
||||
|
||||
"github.com/gin-gonic/gin"
|
||||
)
|
||||
|
||||
func relayChatHelper(c *gin.Context) *types.OpenAIErrorWithStatusCode {
|
||||
|
||||
// 获取请求参数
|
||||
channelType := c.GetInt("channel")
|
||||
channelId := c.GetInt("channel_id")
|
||||
tokenId := c.GetInt("token_id")
|
||||
userId := c.GetInt("id")
|
||||
// consumeQuota := c.GetBool("consume_quota")
|
||||
group := c.GetString("group")
|
||||
|
||||
// 获取 Provider
|
||||
chatProvider := GetChatProvider(channelType, c)
|
||||
if chatProvider == nil {
|
||||
return types.ErrorWrapper(errors.New("API not implemented"), "api_not_implemented", http.StatusNotImplemented)
|
||||
}
|
||||
|
||||
// 获取请求体
|
||||
var chatRequest types.ChatCompletionRequest
|
||||
err := common.UnmarshalBodyReusable(c, &chatRequest)
|
||||
if err != nil {
|
||||
return types.ErrorWrapper(err, "bind_request_body_failed", http.StatusBadRequest)
|
||||
}
|
||||
|
||||
// 检查模型映射
|
||||
isModelMapped := false
|
||||
modelMap, err := parseModelMapping(c.GetString("model_mapping"))
|
||||
if err != nil {
|
||||
return types.ErrorWrapper(err, "unmarshal_model_mapping_failed", http.StatusInternalServerError)
|
||||
}
|
||||
if modelMap != nil && modelMap[chatRequest.Model] != "" {
|
||||
chatRequest.Model = modelMap[chatRequest.Model]
|
||||
isModelMapped = true
|
||||
}
|
||||
|
||||
// 开始计算Tokens
|
||||
var promptTokens int
|
||||
promptTokens = common.CountTokenMessages(chatRequest.Messages, chatRequest.Model)
|
||||
|
||||
// 计算预付费配额
|
||||
quotaInfo := &QuotaInfo{
|
||||
modelName: chatRequest.Model,
|
||||
promptTokens: promptTokens,
|
||||
userId: userId,
|
||||
channelId: channelId,
|
||||
tokenId: tokenId,
|
||||
}
|
||||
quotaInfo.initQuotaInfo(group)
|
||||
quota_err := quotaInfo.preQuotaConsumption()
|
||||
if quota_err != nil {
|
||||
return quota_err
|
||||
}
|
||||
|
||||
usage, openAIErrorWithStatusCode := chatProvider.ChatCompleteResponse(&chatRequest, isModelMapped, promptTokens)
|
||||
|
||||
if openAIErrorWithStatusCode != nil {
|
||||
if quotaInfo.preConsumedQuota != 0 {
|
||||
go func(ctx context.Context) {
|
||||
// return pre-consumed quota
|
||||
err := model.PostConsumeTokenQuota(tokenId, -quotaInfo.preConsumedQuota)
|
||||
if err != nil {
|
||||
common.LogError(ctx, "error return pre-consumed quota: "+err.Error())
|
||||
}
|
||||
}(c.Request.Context())
|
||||
}
|
||||
return openAIErrorWithStatusCode
|
||||
}
|
||||
|
||||
tokenName := c.GetString("token_name")
|
||||
defer func(ctx context.Context) {
|
||||
go func() {
|
||||
err = quotaInfo.completedQuotaConsumption(usage, tokenName, ctx)
|
||||
if err != nil {
|
||||
common.LogError(ctx, err.Error())
|
||||
}
|
||||
}()
|
||||
}(c.Request.Context())
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func GetChatProvider(channelType int, c *gin.Context) providers.ChatProviderAction {
|
||||
switch channelType {
|
||||
case common.ChannelTypeOpenAI:
|
||||
return providers.CreateOpenAIProvider(c, "")
|
||||
case common.ChannelTypeAzure:
|
||||
return providers.CreateAzureProvider(c)
|
||||
case common.ChannelTypeAli:
|
||||
return providers.CreateAliAIProvider(c)
|
||||
case common.ChannelTypeTencent:
|
||||
return providers.CreateTencentProvider(c)
|
||||
case common.ChannelTypeBaidu:
|
||||
return providers.CreateBaiduProvider(c)
|
||||
case common.ChannelTypeAnthropic:
|
||||
return providers.CreateClaudeProvider(c)
|
||||
case common.ChannelTypePaLM:
|
||||
return providers.CreatePalmProvider(c)
|
||||
case common.ChannelTypeZhipu:
|
||||
return providers.CreateZhipuProvider(c)
|
||||
case common.ChannelTypeXunfei:
|
||||
return providers.CreateXunfeiProvider(c)
|
||||
}
|
||||
|
||||
baseURL := common.ChannelBaseURLs[channelType]
|
||||
if c.GetString("base_url") != "" {
|
||||
baseURL = c.GetString("base_url")
|
||||
}
|
||||
|
||||
if baseURL != "" {
|
||||
return providers.CreateOpenAIProvider(c, baseURL)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
@@ -1,113 +0,0 @@
|
||||
package controller
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/model"
|
||||
"one-api/providers"
|
||||
"one-api/types"
|
||||
|
||||
"github.com/gin-gonic/gin"
|
||||
)
|
||||
|
||||
func relayCompletionHelper(c *gin.Context) *types.OpenAIErrorWithStatusCode {
|
||||
|
||||
// 获取请求参数
|
||||
channelType := c.GetInt("channel")
|
||||
channelId := c.GetInt("channel_id")
|
||||
tokenId := c.GetInt("token_id")
|
||||
userId := c.GetInt("id")
|
||||
// consumeQuota := c.GetBool("consume_quota")
|
||||
group := c.GetString("group")
|
||||
|
||||
// 获取 Provider
|
||||
completionProvider := GetCompletionProvider(channelType, c)
|
||||
if completionProvider == nil {
|
||||
return types.ErrorWrapper(errors.New("API not implemented"), "api_not_implemented", http.StatusNotImplemented)
|
||||
}
|
||||
|
||||
// 获取请求体
|
||||
var completionRequest types.CompletionRequest
|
||||
err := common.UnmarshalBodyReusable(c, &completionRequest)
|
||||
if err != nil {
|
||||
return types.ErrorWrapper(err, "bind_request_body_failed", http.StatusBadRequest)
|
||||
}
|
||||
|
||||
// 检查模型映射
|
||||
isModelMapped := false
|
||||
modelMap, err := parseModelMapping(c.GetString("model_mapping"))
|
||||
if err != nil {
|
||||
return types.ErrorWrapper(err, "unmarshal_model_mapping_failed", http.StatusInternalServerError)
|
||||
}
|
||||
if modelMap != nil && modelMap[completionRequest.Model] != "" {
|
||||
completionRequest.Model = modelMap[completionRequest.Model]
|
||||
isModelMapped = true
|
||||
}
|
||||
|
||||
// 开始计算Tokens
|
||||
var promptTokens int
|
||||
promptTokens = common.CountTokenInput(completionRequest.Prompt, completionRequest.Model)
|
||||
|
||||
// 计算预付费配额
|
||||
quotaInfo := &QuotaInfo{
|
||||
modelName: completionRequest.Model,
|
||||
promptTokens: promptTokens,
|
||||
userId: userId,
|
||||
channelId: channelId,
|
||||
tokenId: tokenId,
|
||||
}
|
||||
quotaInfo.initQuotaInfo(group)
|
||||
quota_err := quotaInfo.preQuotaConsumption()
|
||||
if quota_err != nil {
|
||||
return quota_err
|
||||
}
|
||||
|
||||
usage, openAIErrorWithStatusCode := completionProvider.CompleteResponse(&completionRequest, isModelMapped, promptTokens)
|
||||
|
||||
if openAIErrorWithStatusCode != nil {
|
||||
if quotaInfo.preConsumedQuota != 0 {
|
||||
go func(ctx context.Context) {
|
||||
// return pre-consumed quota
|
||||
err := model.PostConsumeTokenQuota(tokenId, -quotaInfo.preConsumedQuota)
|
||||
if err != nil {
|
||||
common.LogError(ctx, "error return pre-consumed quota: "+err.Error())
|
||||
}
|
||||
}(c.Request.Context())
|
||||
}
|
||||
return openAIErrorWithStatusCode
|
||||
}
|
||||
|
||||
tokenName := c.GetString("token_name")
|
||||
defer func(ctx context.Context) {
|
||||
go func() {
|
||||
err = quotaInfo.completedQuotaConsumption(usage, tokenName, ctx)
|
||||
if err != nil {
|
||||
common.LogError(ctx, err.Error())
|
||||
}
|
||||
}()
|
||||
}(c.Request.Context())
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func GetCompletionProvider(channelType int, c *gin.Context) providers.CompletionProviderAction {
|
||||
switch channelType {
|
||||
case common.ChannelTypeOpenAI:
|
||||
return providers.CreateOpenAIProvider(c, "")
|
||||
case common.ChannelTypeAzure:
|
||||
return providers.CreateAzureProvider(c)
|
||||
}
|
||||
|
||||
baseURL := common.ChannelBaseURLs[channelType]
|
||||
if c.GetString("base_url") != "" {
|
||||
baseURL = c.GetString("base_url")
|
||||
}
|
||||
|
||||
if baseURL != "" {
|
||||
return providers.CreateOpenAIProvider(c, baseURL)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
@@ -1,117 +0,0 @@
|
||||
package controller
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/model"
|
||||
"one-api/providers"
|
||||
"one-api/types"
|
||||
|
||||
"github.com/gin-gonic/gin"
|
||||
)
|
||||
|
||||
func relayEmbeddingsHelper(c *gin.Context) *types.OpenAIErrorWithStatusCode {
|
||||
|
||||
// 获取请求参数
|
||||
channelType := c.GetInt("channel")
|
||||
channelId := c.GetInt("channel_id")
|
||||
tokenId := c.GetInt("token_id")
|
||||
userId := c.GetInt("id")
|
||||
// consumeQuota := c.GetBool("consume_quota")
|
||||
group := c.GetString("group")
|
||||
|
||||
// 获取 Provider
|
||||
embeddingsProvider := GetEmbeddingsProvider(channelType, c)
|
||||
if embeddingsProvider == nil {
|
||||
return types.ErrorWrapper(errors.New("API not implemented"), "api_not_implemented", http.StatusNotImplemented)
|
||||
}
|
||||
|
||||
// 获取请求体
|
||||
var embeddingsRequest types.EmbeddingRequest
|
||||
err := common.UnmarshalBodyReusable(c, &embeddingsRequest)
|
||||
if err != nil {
|
||||
return types.ErrorWrapper(err, "bind_request_body_failed", http.StatusBadRequest)
|
||||
}
|
||||
|
||||
// 检查模型映射
|
||||
isModelMapped := false
|
||||
modelMap, err := parseModelMapping(c.GetString("model_mapping"))
|
||||
if err != nil {
|
||||
return types.ErrorWrapper(err, "unmarshal_model_mapping_failed", http.StatusInternalServerError)
|
||||
}
|
||||
if modelMap != nil && modelMap[embeddingsRequest.Model] != "" {
|
||||
embeddingsRequest.Model = modelMap[embeddingsRequest.Model]
|
||||
isModelMapped = true
|
||||
}
|
||||
|
||||
// 开始计算Tokens
|
||||
var promptTokens int
|
||||
promptTokens = common.CountTokenInput(embeddingsRequest.Input, embeddingsRequest.Model)
|
||||
|
||||
// 计算预付费配额
|
||||
quotaInfo := &QuotaInfo{
|
||||
modelName: embeddingsRequest.Model,
|
||||
promptTokens: promptTokens,
|
||||
userId: userId,
|
||||
channelId: channelId,
|
||||
tokenId: tokenId,
|
||||
}
|
||||
quotaInfo.initQuotaInfo(group)
|
||||
quota_err := quotaInfo.preQuotaConsumption()
|
||||
if quota_err != nil {
|
||||
return quota_err
|
||||
}
|
||||
|
||||
usage, openAIErrorWithStatusCode := embeddingsProvider.EmbeddingsResponse(&embeddingsRequest, isModelMapped, promptTokens)
|
||||
|
||||
if openAIErrorWithStatusCode != nil {
|
||||
if quotaInfo.preConsumedQuota != 0 {
|
||||
go func(ctx context.Context) {
|
||||
// return pre-consumed quota
|
||||
err := model.PostConsumeTokenQuota(tokenId, -quotaInfo.preConsumedQuota)
|
||||
if err != nil {
|
||||
common.LogError(ctx, "error return pre-consumed quota: "+err.Error())
|
||||
}
|
||||
}(c.Request.Context())
|
||||
}
|
||||
return openAIErrorWithStatusCode
|
||||
}
|
||||
|
||||
tokenName := c.GetString("token_name")
|
||||
defer func(ctx context.Context) {
|
||||
go func() {
|
||||
err = quotaInfo.completedQuotaConsumption(usage, tokenName, ctx)
|
||||
if err != nil {
|
||||
common.LogError(ctx, err.Error())
|
||||
}
|
||||
}()
|
||||
}(c.Request.Context())
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func GetEmbeddingsProvider(channelType int, c *gin.Context) providers.EmbeddingsProviderAction {
|
||||
switch channelType {
|
||||
case common.ChannelTypeOpenAI:
|
||||
return providers.CreateOpenAIProvider(c, "")
|
||||
case common.ChannelTypeAzure:
|
||||
return providers.CreateAzureProvider(c)
|
||||
case common.ChannelTypeAli:
|
||||
return providers.CreateAliAIProvider(c)
|
||||
case common.ChannelTypeBaidu:
|
||||
return providers.CreateBaiduProvider(c)
|
||||
}
|
||||
|
||||
baseURL := common.ChannelBaseURLs[channelType]
|
||||
if c.GetString("base_url") != "" {
|
||||
baseURL = c.GetString("base_url")
|
||||
}
|
||||
|
||||
if baseURL != "" {
|
||||
return providers.CreateOpenAIProvider(c, baseURL)
|
||||
}
|
||||
|
||||
return nil
|
||||
}
|
||||
160
controller/relay-text.go
Normal file
160
controller/relay-text.go
Normal file
@@ -0,0 +1,160 @@
|
||||
package controller
|
||||
|
||||
import (
|
||||
"context"
|
||||
"errors"
|
||||
"net/http"
|
||||
"one-api/common"
|
||||
"one-api/model"
|
||||
"one-api/providers"
|
||||
providers_base "one-api/providers/base"
|
||||
"one-api/types"
|
||||
|
||||
"github.com/gin-gonic/gin"
|
||||
)
|
||||
|
||||
func relayTextHelper(c *gin.Context, relayMode int) *types.OpenAIErrorWithStatusCode {
|
||||
// 获取请求参数
|
||||
channelType := c.GetInt("channel")
|
||||
channelId := c.GetInt("channel_id")
|
||||
tokenId := c.GetInt("token_id")
|
||||
userId := c.GetInt("id")
|
||||
group := c.GetString("group")
|
||||
|
||||
// 获取 Provider
|
||||
provider := providers.GetProvider(channelType, c)
|
||||
if provider == nil {
|
||||
return types.ErrorWrapper(errors.New("channel not implemented"), "channel_not_implemented", http.StatusNotImplemented)
|
||||
}
|
||||
|
||||
modelMap, err := parseModelMapping(c.GetString("model_mapping"))
|
||||
if err != nil {
|
||||
return types.ErrorWrapper(err, "unmarshal_model_mapping_failed", http.StatusInternalServerError)
|
||||
}
|
||||
|
||||
var promptTokens int
|
||||
quotaInfo := &QuotaInfo{
|
||||
modelName: "",
|
||||
promptTokens: promptTokens,
|
||||
userId: userId,
|
||||
channelId: channelId,
|
||||
tokenId: tokenId,
|
||||
}
|
||||
|
||||
var usage *types.Usage
|
||||
var openAIErrorWithStatusCode *types.OpenAIErrorWithStatusCode
|
||||
|
||||
switch relayMode {
|
||||
case RelayModeChatCompletions:
|
||||
usage, openAIErrorWithStatusCode = handleChatCompletions(c, provider, modelMap, quotaInfo, group)
|
||||
case RelayModeCompletions:
|
||||
usage, openAIErrorWithStatusCode = handleCompletions(c, provider, modelMap, quotaInfo, group)
|
||||
case RelayModeEmbeddings:
|
||||
usage, openAIErrorWithStatusCode = handleEmbeddings(c, provider, modelMap, quotaInfo, group)
|
||||
default:
|
||||
return types.ErrorWrapper(errors.New("invalid relay mode"), "invalid_relay_mode", http.StatusBadRequest)
|
||||
}
|
||||
|
||||
if openAIErrorWithStatusCode != nil {
|
||||
if quotaInfo.preConsumedQuota != 0 {
|
||||
go func(ctx context.Context) {
|
||||
// return pre-consumed quota
|
||||
err := model.PostConsumeTokenQuota(tokenId, -quotaInfo.preConsumedQuota)
|
||||
if err != nil {
|
||||
common.LogError(ctx, "error return pre-consumed quota: "+err.Error())
|
||||
}
|
||||
}(c.Request.Context())
|
||||
}
|
||||
return openAIErrorWithStatusCode
|
||||
}
|
||||
|
||||
tokenName := c.GetString("token_name")
|
||||
defer func(ctx context.Context) {
|
||||
go func() {
|
||||
err = quotaInfo.completedQuotaConsumption(usage, tokenName, ctx)
|
||||
if err != nil {
|
||||
common.LogError(ctx, err.Error())
|
||||
}
|
||||
}()
|
||||
}(c.Request.Context())
|
||||
|
||||
return nil
|
||||
}
|
||||
|
||||
func handleChatCompletions(c *gin.Context, provider providers_base.ProviderInterface, modelMap map[string]string, quotaInfo *QuotaInfo, group string) (*types.Usage, *types.OpenAIErrorWithStatusCode) {
|
||||
var chatRequest types.ChatCompletionRequest
|
||||
isModelMapped := false
|
||||
chatProvider, ok := provider.(providers_base.ChatInterface)
|
||||
if !ok {
|
||||
return nil, types.ErrorWrapper(errors.New("channel not implemented"), "channel_not_implemented", http.StatusNotImplemented)
|
||||
}
|
||||
err := common.UnmarshalBodyReusable(c, &chatRequest)
|
||||
if err != nil {
|
||||
return nil, types.ErrorWrapper(err, "bind_request_body_failed", http.StatusBadRequest)
|
||||
}
|
||||
if modelMap != nil && modelMap[chatRequest.Model] != "" {
|
||||
chatRequest.Model = modelMap[chatRequest.Model]
|
||||
isModelMapped = true
|
||||
}
|
||||
promptTokens := common.CountTokenMessages(chatRequest.Messages, chatRequest.Model)
|
||||
|
||||
quotaInfo.modelName = chatRequest.Model
|
||||
quotaInfo.initQuotaInfo(group)
|
||||
quota_err := quotaInfo.preQuotaConsumption()
|
||||
if quota_err != nil {
|
||||
return nil, quota_err
|
||||
}
|
||||
return chatProvider.ChatAction(&chatRequest, isModelMapped, promptTokens)
|
||||
}
|
||||
|
||||
func handleCompletions(c *gin.Context, provider providers_base.ProviderInterface, modelMap map[string]string, quotaInfo *QuotaInfo, group string) (*types.Usage, *types.OpenAIErrorWithStatusCode) {
|
||||
var completionRequest types.CompletionRequest
|
||||
isModelMapped := false
|
||||
completionProvider, ok := provider.(providers_base.CompletionInterface)
|
||||
if !ok {
|
||||
return nil, types.ErrorWrapper(errors.New("channel not implemented"), "channel_not_implemented", http.StatusNotImplemented)
|
||||
}
|
||||
err := common.UnmarshalBodyReusable(c, &completionRequest)
|
||||
if err != nil {
|
||||
return nil, types.ErrorWrapper(err, "bind_request_body_failed", http.StatusBadRequest)
|
||||
}
|
||||
if modelMap != nil && modelMap[completionRequest.Model] != "" {
|
||||
completionRequest.Model = modelMap[completionRequest.Model]
|
||||
isModelMapped = true
|
||||
}
|
||||
promptTokens := common.CountTokenInput(completionRequest.Prompt, completionRequest.Model)
|
||||
|
||||
quotaInfo.modelName = completionRequest.Model
|
||||
quotaInfo.initQuotaInfo(group)
|
||||
quota_err := quotaInfo.preQuotaConsumption()
|
||||
if quota_err != nil {
|
||||
return nil, quota_err
|
||||
}
|
||||
return completionProvider.CompleteAction(&completionRequest, isModelMapped, promptTokens)
|
||||
}
|
||||
|
||||
func handleEmbeddings(c *gin.Context, provider providers_base.ProviderInterface, modelMap map[string]string, quotaInfo *QuotaInfo, group string) (*types.Usage, *types.OpenAIErrorWithStatusCode) {
|
||||
var embeddingsRequest types.EmbeddingRequest
|
||||
isModelMapped := false
|
||||
embeddingsProvider, ok := provider.(providers_base.EmbeddingsInterface)
|
||||
if !ok {
|
||||
return nil, types.ErrorWrapper(errors.New("channel not implemented"), "channel_not_implemented", http.StatusNotImplemented)
|
||||
}
|
||||
err := common.UnmarshalBodyReusable(c, &embeddingsRequest)
|
||||
if err != nil {
|
||||
return nil, types.ErrorWrapper(err, "bind_request_body_failed", http.StatusBadRequest)
|
||||
}
|
||||
if modelMap != nil && modelMap[embeddingsRequest.Model] != "" {
|
||||
embeddingsRequest.Model = modelMap[embeddingsRequest.Model]
|
||||
isModelMapped = true
|
||||
}
|
||||
promptTokens := common.CountTokenInput(embeddingsRequest.Input, embeddingsRequest.Model)
|
||||
|
||||
quotaInfo.modelName = embeddingsRequest.Model
|
||||
quotaInfo.initQuotaInfo(group)
|
||||
quota_err := quotaInfo.preQuotaConsumption()
|
||||
if quota_err != nil {
|
||||
return nil, quota_err
|
||||
}
|
||||
return embeddingsProvider.EmbeddingsAction(&embeddingsRequest, isModelMapped, promptTokens)
|
||||
}
|
||||
@@ -237,19 +237,19 @@ type CompletionsStreamResponse struct {
|
||||
func Relay(c *gin.Context) {
|
||||
var err *types.OpenAIErrorWithStatusCode
|
||||
|
||||
// relayMode := RelayModeUnknown
|
||||
relayMode := RelayModeUnknown
|
||||
if strings.HasPrefix(c.Request.URL.Path, "/v1/chat/completions") {
|
||||
err = relayChatHelper(c)
|
||||
// relayMode = RelayModeChatCompletions
|
||||
// err = relayChatHelper(c)
|
||||
relayMode = RelayModeChatCompletions
|
||||
} else if strings.HasPrefix(c.Request.URL.Path, "/v1/completions") {
|
||||
err = relayCompletionHelper(c)
|
||||
// relayMode = RelayModeCompletions
|
||||
// err = relayCompletionHelper(c)
|
||||
relayMode = RelayModeCompletions
|
||||
} else if strings.HasPrefix(c.Request.URL.Path, "/v1/embeddings") {
|
||||
err = relayEmbeddingsHelper(c)
|
||||
// err = relayEmbeddingsHelper(c)
|
||||
relayMode = RelayModeEmbeddings
|
||||
} else if strings.HasSuffix(c.Request.URL.Path, "embeddings") {
|
||||
relayMode = RelayModeEmbeddings
|
||||
}
|
||||
// relayMode = RelayModeEmbeddings
|
||||
// } else if strings.HasSuffix(c.Request.URL.Path, "embeddings") {
|
||||
// relayMode = RelayModeEmbeddings
|
||||
// } else if strings.HasPrefix(c.Request.URL.Path, "/v1/moderations") {
|
||||
// relayMode = RelayModeModerations
|
||||
// } else if strings.HasPrefix(c.Request.URL.Path, "/v1/images/generations") {
|
||||
@@ -263,7 +263,7 @@ func Relay(c *gin.Context) {
|
||||
// } else if strings.HasPrefix(c.Request.URL.Path, "/v1/audio/translations") {
|
||||
// relayMode = RelayModeAudioTranslation
|
||||
// }
|
||||
// switch relayMode {
|
||||
switch relayMode {
|
||||
// case RelayModeImagesGenerations:
|
||||
// err = relayImageHelper(c, relayMode)
|
||||
// case RelayModeAudioSpeech:
|
||||
@@ -272,9 +272,9 @@ func Relay(c *gin.Context) {
|
||||
// fallthrough
|
||||
// case RelayModeAudioTranscription:
|
||||
// err = relayAudioHelper(c, relayMode)
|
||||
// default:
|
||||
// err = relayTextHelper(c, relayMode)
|
||||
// }
|
||||
default:
|
||||
err = relayTextHelper(c, relayMode)
|
||||
}
|
||||
if err != nil {
|
||||
requestId := c.GetString(common.RequestIdKey)
|
||||
retryTimesStr := c.Query("retry")
|
||||
|
||||
Reference in New Issue
Block a user