feat: add claude-3.7-sonnet model support

This commit is contained in:
wozulong
2025-02-25 19:54:30 +08:00
parent b648816add
commit 7c0b93c38c
6 changed files with 195 additions and 140 deletions

View File

@@ -86,6 +86,9 @@ var defaultModelRatio = map[string]float64{
"claude-instant-1": 0.4, // $0.8 / 1M tokens
"claude-2.0": 4, // $8 / 1M tokens
"claude-2.1": 4, // $8 / 1M tokens
"claude-3-7-sonnet-20250219": 1.5,
"claude-3-7-sonnet-20250219-thinking": 1.5,
"claude-3-5-haiku-20241022": 0.4,
"claude-3-haiku-20240307": 0.125, // $0.25 / 1M tokens
"claude-3-5-sonnet-20240620": 1.5, // $3 / 1M tokens
"claude-3-5-sonnet-20241022": 1.5, // $3 / 1M tokens

View File

@@ -36,6 +36,13 @@ type GeneralOpenAIRequest struct {
Dimensions int `json:"dimensions,omitempty"`
ParallelToolCalls bool `json:"parallel_tool_calls,omitempty"`
EncodingFormat any `json:"encoding_format,omitempty"`
Thinking *Thinking `json:"thinking,omitempty"`
}
type Thinking struct {
Type string `json:"type"`
BudgetTokens int `json:"budget_tokens"`
}
type OpenAITools struct {
@@ -79,6 +86,7 @@ func (r GeneralOpenAIRequest) ParseInput() []string {
type Message struct {
Role string `json:"role"`
Content json.RawMessage `json:"content"`
ReasoningContent *string `json:"reasoning_content,omitempty"`
Name *string `json:"name,omitempty"`
ToolCalls any `json:"tool_calls,omitempty"`
ToolCallId string `json:"tool_call_id,omitempty"`

View File

@@ -63,6 +63,7 @@ type ChatCompletionsStreamResponseChoice struct {
type ChatCompletionsStreamResponseChoiceDelta struct {
Content *string `json:"content,omitempty"`
ReasoningContent *string `json:"reasoning_content,omitempty"`
Role string `json:"role,omitempty"`
ToolCalls []ToolCall `json:"tool_calls,omitempty"`
}

View File

@@ -8,8 +8,11 @@ var ModelList = []string{
"claude-3-sonnet-20240229",
"claude-3-opus-20240229",
"claude-3-haiku-20240307",
"claude-3-5-haiku-20241022",
"claude-3-5-sonnet-20240620",
"claude-3-5-sonnet-20241022",
"claude-3-7-sonnet-20250219",
"claude-3-7-sonnet-20250219-thinking",
}
var ChannelName = "claude"

View File

@@ -1,5 +1,7 @@
package claude
import "one-api/dto"
type ClaudeMetadata struct {
UserId string `json:"user_id"`
}
@@ -11,6 +13,9 @@ type ClaudeMediaMessage struct {
Usage *ClaudeUsage `json:"usage,omitempty"`
StopReason *string `json:"stop_reason,omitempty"`
PartialJson string `json:"partial_json,omitempty"`
Thinking string `json:"thinking,omitempty"`
Signature string `json:"signature,omitempty"`
Delta string `json:"delta,omitempty"`
// tool_calls
Id string `json:"id,omitempty"`
Name string `json:"name,omitempty"`
@@ -61,6 +66,7 @@ type ClaudeRequest struct {
Stream bool `json:"stream,omitempty"`
Tools []Tool `json:"tools,omitempty"`
ToolChoice any `json:"tool_choice,omitempty"`
Thinking *dto.Thinking `json:"thinking,omitempty"`
}
type ClaudeError struct {

View File

@@ -62,6 +62,24 @@ func RequestOpenAI2ClaudeComplete(textRequest dto.GeneralOpenAIRequest) *ClaudeR
func RequestOpenAI2ClaudeMessage(textRequest dto.GeneralOpenAIRequest) (*ClaudeRequest, error) {
claudeTools := make([]Tool, 0, len(textRequest.Tools))
if strings.HasSuffix(textRequest.Model, "-thinking") {
textRequest.Model = strings.TrimSuffix(textRequest.Model, "-thinking")
if textRequest.MaxTokens == 0 {
textRequest.MaxTokens = 4096
} else if textRequest.MaxTokens < 1280 {
textRequest.MaxTokens = 1280
}
textRequest.TopP = 0
textRequest.TopK = 0
textRequest.Temperature = 0
textRequest.Thinking = &dto.Thinking{
Type: "enabled",
BudgetTokens: int(float64(textRequest.MaxTokens) * 0.8),
}
}
for _, tool := range textRequest.Tools {
claudeTool := Tool{
Name: tool.Function.Name,
@@ -113,6 +131,7 @@ func RequestOpenAI2ClaudeMessage(textRequest dto.GeneralOpenAIRequest) (*ClaudeR
Stream: textRequest.Stream,
Tools: claudeTools,
ToolChoice: textRequest.ToolChoice,
Thinking: textRequest.Thinking,
}
if claudeRequest.MaxTokens == 0 {
claudeRequest.MaxTokens = 4096
@@ -334,12 +353,19 @@ func StreamResponseClaude2OpenAI(reqMode int, claudeResponse *ClaudeResponse) (*
if claudeResponse.Delta != nil {
choice.Index = claudeResponse.Index
choice.Delta.SetContentString(claudeResponse.Delta.Text)
if claudeResponse.Delta.Type == "input_json_delta" {
switch claudeResponse.Delta.Type {
case "input_json_delta":
tools = append(tools, dto.ToolCall{
Function: dto.FunctionCall{
Arguments: claudeResponse.Delta.PartialJson,
},
})
case "signature_delta":
reasoningContent := "\n"
choice.Delta.ReasoningContent = &reasoningContent
case "thinking_delta":
reasoningContent := claudeResponse.Delta.Thinking
choice.Delta.ReasoningContent = &reasoningContent
}
}
} else if claudeResponse.Type == "message_delta" {
@@ -377,6 +403,8 @@ func ResponseClaude2OpenAI(reqMode int, claudeResponse *ClaudeResponse) *dto.Ope
if len(claudeResponse.Content) > 0 {
responseText = claudeResponse.Content[0].Text
}
reasoningContent := ""
tools := make([]dto.ToolCall, 0)
if reqMode == RequestModeCompletion {
content, _ := json.Marshal(strings.TrimPrefix(claudeResponse.Completion, " "))
@@ -393,7 +421,8 @@ func ResponseClaude2OpenAI(reqMode int, claudeResponse *ClaudeResponse) *dto.Ope
} else {
fullTextResponse.Id = claudeResponse.Id
for _, message := range claudeResponse.Content {
if message.Type == "tool_use" {
switch message.Type {
case "tool_use":
args, _ := json.Marshal(message.Input)
tools = append(tools, dto.ToolCall{
ID: message.Id,
@@ -403,6 +432,10 @@ func ResponseClaude2OpenAI(reqMode int, claudeResponse *ClaudeResponse) *dto.Ope
Arguments: string(args),
},
})
case "thinking":
reasoningContent = message.Thinking
case "text":
responseText = message.Text
}
}
}
@@ -417,6 +450,7 @@ func ResponseClaude2OpenAI(reqMode int, claudeResponse *ClaudeResponse) *dto.Ope
if len(tools) > 0 {
choice.Message.ToolCalls = tools
}
choice.Message.ReasoningContent = &reasoningContent
fullTextResponse.Model = claudeResponse.Model
choices = append(choices, choice)
fullTextResponse.Choices = choices