From 5425b5bfc3d144c205e2950d4bb537ed3cc2103e Mon Sep 17 00:00:00 2001 From: CaIon <1808837298@qq.com> Date: Sun, 28 Apr 2024 20:45:34 +0800 Subject: [PATCH] fix: aws claude --- relay/channel/claude/dto.go | 19 ++++++++++--------- relay/channel/claude/relay-claude.go | 5 ++--- 2 files changed, 12 insertions(+), 12 deletions(-) diff --git a/relay/channel/claude/dto.go b/relay/channel/claude/dto.go index 2d13e46..47f0c3b 100644 --- a/relay/channel/claude/dto.go +++ b/relay/channel/claude/dto.go @@ -24,15 +24,16 @@ type ClaudeMessage struct { } type ClaudeRequest struct { - Model string `json:"model"` - Prompt string `json:"prompt,omitempty"` - System string `json:"system,omitempty"` - Messages []ClaudeMessage `json:"messages,omitempty"` - MaxTokens uint `json:"max_tokens,omitempty"` - StopSequences []string `json:"stop_sequences,omitempty"` - Temperature float64 `json:"temperature,omitempty"` - TopP float64 `json:"top_p,omitempty"` - TopK int `json:"top_k,omitempty"` + Model string `json:"model"` + Prompt string `json:"prompt,omitempty"` + System string `json:"system,omitempty"` + Messages []ClaudeMessage `json:"messages,omitempty"` + MaxTokens uint `json:"max_tokens,omitempty"` + MaxTokensToSample uint `json:"max_tokens_to_sample,omitempty"` + StopSequences []string `json:"stop_sequences,omitempty"` + Temperature float64 `json:"temperature,omitempty"` + TopP float64 `json:"top_p,omitempty"` + TopK int `json:"top_k,omitempty"` //ClaudeMetadata `json:"metadata,omitempty"` Stream bool `json:"stream,omitempty"` } diff --git a/relay/channel/claude/relay-claude.go b/relay/channel/claude/relay-claude.go index 07767a4..d1d8e6e 100644 --- a/relay/channel/claude/relay-claude.go +++ b/relay/channel/claude/relay-claude.go @@ -30,15 +30,14 @@ func RequestOpenAI2ClaudeComplete(textRequest dto.GeneralOpenAIRequest) *ClaudeR claudeRequest := ClaudeRequest{ Model: textRequest.Model, Prompt: "", - MaxTokens: textRequest.MaxTokens, StopSequences: nil, Temperature: textRequest.Temperature, TopP: textRequest.TopP, TopK: textRequest.TopK, Stream: textRequest.Stream, } - if claudeRequest.MaxTokens == 0 { - claudeRequest.MaxTokens = 4096 + if claudeRequest.MaxTokensToSample == 0 { + claudeRequest.MaxTokensToSample = 4096 } prompt := "" for _, message := range textRequest.Messages {