From 678928cafd95b17639c7b46d3fdbeef546f9c6b5 Mon Sep 17 00:00:00 2001 From: Sakurasan <1173092237@qq.com> Date: Sun, 13 Aug 2023 22:55:04 +0800 Subject: [PATCH 1/3] claude api support --- pkg/claude/claude.go | 202 +++++++++++++++++++++++++++++++++++++++++++ router/router.go | 24 +++++ 2 files changed, 226 insertions(+) create mode 100644 pkg/claude/claude.go diff --git a/pkg/claude/claude.go b/pkg/claude/claude.go new file mode 100644 index 0000000..7ac2ae3 --- /dev/null +++ b/pkg/claude/claude.go @@ -0,0 +1,202 @@ +/* +https://docs.anthropic.com/claude/reference/complete_post + +curl --request POST \ + --url https://api.anthropic.com/v1/complete \ + --header "anthropic-version: 2023-06-01" \ + --header "content-type: application/json" \ + --header "x-api-key: $ANTHROPIC_API_KEY" \ + --data ' +{ + "model": "claude-2", + "prompt": "\n\nHuman: Hello, world!\n\nAssistant:", + "max_tokens_to_sample": 256, + "stream": true +} +' + +{"completion":" Hello! Nice to meet you.","stop_reason":"stop_sequence","model":"claude-2.0","stop":"\n\nHuman:","log_id":"727bded01002627057967d02b3d557a01aa73266849b62f5aa0b97dec1247ed3"} + +event: completion +data: {"completion":"","stop_reason":"stop_sequence","model":"claude-2.0","stop":"\n\nHuman:","log_id":"dfd42341ad08856ff01811885fb8640a1bf977551d8331f81fe9a6c8182c6c63"} + +# Model Pricing + +Claude Instant |100,000 tokens |Prompt $1.63/million tokens |Completion $5.51/million tokens + +Claude 2 |100,000 tokens |Prompt $11.02/million tokens |Completion $32.68/million tokens +*Claude 1 is still accessible and offered at the same price as Claude 2. + +*/ + +// package anthropic +package claude + +import ( + "bufio" + "bytes" + "encoding/json" + "fmt" + "io" + "log" + "net/http" + "net/http/httputil" + "net/url" + "opencatd-open/store" + "strings" + + "github.com/gin-gonic/gin" +) + +type MessageModule struct { + Assistant string // returned data (do not modify) + Human string // input content +} + +type CompleteRequest struct { + Model string `json:"model,omitempty"` //* + Prompt string `json:"prompt,omitempty"` //* + MaxTokensToSample int `json:"max_Tokens_To_Sample,omitempty"` //* + StopSequences string `json:"stop_Sequences,omitempty"` + Temperature int `json:"temperature,omitempty"` + TopP int `json:"top_P,omitempty"` + TopK int `json:"top_K,omitempty"` + Stream bool `json:"stream,omitempty"` + Metadata struct { + UserId string `json:"user_Id,omitempty"` + } `json:"metadata,omitempty"` +} + +type CompleteResponse struct { + Completion string `json:"completion"` + StopReason string `json:"stop_reason"` + Model string `json:"model"` + Stop string `json:"stop"` + LogID string `json:"log_id"` +} + +func Create() { + url := "https://api.anthropic.com/v1/complete" + + payload := strings.NewReader("{\"model\":\"claude-2\",\"prompt\":\"\\n\\nHuman: Hello, world!\\n\\nAssistant:\",\"max_tokens_to_sample\":256}") + + req, _ := http.NewRequest("POST", url, payload) + + req.Header.Add("accept", "application/json") + req.Header.Add("anthropic-version", "2023-06-01") + req.Header.Add("x-api-key", "$ANTHROPIC_API_KEY") + req.Header.Add("content-type", "application/json") + + res, _ := http.DefaultClient.Do(req) + + defer res.Body.Close() + body, _ := io.ReadAll(res.Body) + + fmt.Println(string(body)) +} + +func ClaudeProxy(c *gin.Context) { + var chatlog store.Tokens + var complete CompleteRequest + + byteBody, _ := io.ReadAll(c.Request.Body) + c.Request.Body = io.NopCloser(bytes.NewBuffer(byteBody)) + + if err := json.Unmarshal(byteBody, &complete); err != nil { + c.AbortWithError(http.StatusBadRequest, err) + return + } + + key, err := store.SelectKeyCache("anthropic") + if err != nil { + c.JSON(http.StatusInternalServerError, gin.H{ + "error": gin.H{ + "message": err.Error(), + }, + }) + return + } + + chatlog.Model = complete.Model + + token, _ := c.Get("localuser") + + lu, err := store.GetUserByToken(token.(string)) + if err != nil { + c.JSON(http.StatusInternalServerError, gin.H{ + "error": gin.H{ + "message": err.Error(), + }, + }) + return + } + chatlog.UserID = int(lu.ID) + + // todo calc prompt token + + if key.EndPoint == "" { + key.EndPoint = "https://api.anthropic.com" + } + targetUrl, _ := url.ParseRequestURI(key.EndPoint + c.Request.URL.String()) + + proxy := httputil.NewSingleHostReverseProxy(targetUrl) + proxy.Director = func(req *http.Request) { + req.Host = targetUrl.Host + req.URL.Scheme = targetUrl.Scheme + req.URL.Host = targetUrl.Host + + req.Header.Set("anthropic-version", "2023-06-01") + req.Header.Set("content-type", "application/json") + req.Header.Set("x-api-key", key.Key) + } + + proxy.ModifyResponse = func(resp *http.Response) error { + if resp.StatusCode != http.StatusOK { + return nil + } + var byteResp []byte + byteResp, _ = io.ReadAll(resp.Body) + resp.Body = io.NopCloser(bytes.NewBuffer(byteResp)) + if complete.Stream != true { + var complete_resp CompleteResponse + + if err := json.Unmarshal(byteResp, &complete_resp); err != nil { + log.Println(err) + return nil + } + } else { + var completion string + for { + line, err := bufio.NewReader(bytes.NewBuffer(byteResp)).ReadString('\n') + if err != nil { + if strings.HasPrefix(line, "data:") { + line = strings.TrimSpace(strings.TrimPrefix(line, "data:")) + if strings.HasSuffix(line, "[DONE]") { + break + } + line = strings.TrimSpace(line) + var complete_resp CompleteResponse + if err := json.Unmarshal([]byte(line), &complete_resp); err != nil { + log.Println(err) + break + } + completion += line + } + } + } + log.Println("completion:", completion) + } + chatlog.TotalTokens = chatlog.PromptCount + chatlog.CompletionCount + + // todo calc cost + + if err := store.Record(&chatlog); err != nil { + log.Println(err) + } + if err := store.SumDaily(chatlog.UserID); err != nil { + log.Println(err) + } + return nil + } + proxy.ServeHTTP(c.Writer, c.Request) +} diff --git a/router/router.go b/router/router.go index 05eb743..7cc7108 100644 --- a/router/router.go +++ b/router/router.go @@ -298,6 +298,30 @@ func HandleAddKey(c *gin.Context) { }}) return } + } else if strings.HasPrefix(body.Name, "anthropic.") { + keynames := strings.Split(body.Name, ".") + if len(keynames) < 2 { + c.JSON(http.StatusBadRequest, gin.H{"error": gin.H{ + "message": "Invalid Key Name", + }}) + return + } + if body.Endpoint == "" { + body.Endpoint = "https://api.anthropic.com" + } + k := &store.Key{ + ApiType: "anthropic", + Name: body.Name, + Key: body.Key, + ResourceNmae: keynames[1], + EndPoint: body.Endpoint, + } + if err := store.CreateKey(k); err != nil { + c.JSON(http.StatusInternalServerError, gin.H{"error": gin.H{ + "message": err.Error(), + }}) + return + } } else { if body.ApiType == "" { if err := store.AddKey("openai", body.Key, body.Name); err != nil { From 545147abe055cd398843df7cee734e26219fab98 Mon Sep 17 00:00:00 2001 From: Sakurasan <1173092237@qq.com> Date: Sat, 16 Sep 2023 21:18:11 +0800 Subject: [PATCH 2/3] Squashed commit of feat/claude --- pkg/azureopenai/azureopenai.go | 2 + pkg/claude/claude.go | 204 +++++++++++++++++++++++++++++++-- router/router.go | 101 ++++++++++------ 3 files changed, 265 insertions(+), 42 deletions(-) diff --git a/pkg/azureopenai/azureopenai.go b/pkg/azureopenai/azureopenai.go index 8f091e8..c5a8647 100644 --- a/pkg/azureopenai/azureopenai.go +++ b/pkg/azureopenai/azureopenai.go @@ -10,6 +10,8 @@ curl $AZURE_OPENAI_ENDPOINT/openai/deployments/gpt-35-turbo/chat/completions?api "messages": [{"role": "user", "content": "你好"}] }' +https://learn.microsoft.com/zh-cn/rest/api/cognitiveservices/azureopenaistable/models/list?tabs=HTTP + curl $AZURE_OPENAI_ENDPOINT/openai/deployments?api-version=2022-12-01 \ -H "Content-Type: application/json" \ -H "api-key: $AZURE_OPENAI_KEY" \ diff --git a/pkg/claude/claude.go b/pkg/claude/claude.go index 7ac2ae3..cbce93e 100644 --- a/pkg/claude/claude.go +++ b/pkg/claude/claude.go @@ -44,8 +44,15 @@ import ( "net/url" "opencatd-open/store" "strings" + "sync" + "time" "github.com/gin-gonic/gin" + "github.com/sashabaranov/go-openai" +) + +var ( + ClaudeUrl = "https://api.anthropic.com/v1/complete" ) type MessageModule struct { @@ -56,11 +63,11 @@ type MessageModule struct { type CompleteRequest struct { Model string `json:"model,omitempty"` //* Prompt string `json:"prompt,omitempty"` //* - MaxTokensToSample int `json:"max_Tokens_To_Sample,omitempty"` //* - StopSequences string `json:"stop_Sequences,omitempty"` + MaxTokensToSample int `json:"max_tokens_to_sample,omitempty"` //* + StopSequences string `json:"stop_sequences,omitempty"` Temperature int `json:"temperature,omitempty"` - TopP int `json:"top_P,omitempty"` - TopK int `json:"top_K,omitempty"` + TopP int `json:"top_p,omitempty"` + TopK int `json:"top_k,omitempty"` Stream bool `json:"stream,omitempty"` Metadata struct { UserId string `json:"user_Id,omitempty"` @@ -76,11 +83,17 @@ type CompleteResponse struct { } func Create() { - url := "https://api.anthropic.com/v1/complete" + complet := CompleteRequest{ + Model: "claude-2", + Prompt: "Human: Hello, world!\\n\\nAssistant:", + Stream: true, + } + var payload *bytes.Buffer + json.NewEncoder(payload).Encode(complet) - payload := strings.NewReader("{\"model\":\"claude-2\",\"prompt\":\"\\n\\nHuman: Hello, world!\\n\\nAssistant:\",\"max_tokens_to_sample\":256}") + // payload := strings.NewReader("{\"model\":\"claude-2\",\"prompt\":\"\\n\\nHuman: Hello, world!\\n\\nAssistant:\",\"max_tokens_to_sample\":256}") - req, _ := http.NewRequest("POST", url, payload) + req, _ := http.NewRequest("POST", ClaudeUrl, payload) req.Header.Add("accept", "application/json") req.Header.Add("anthropic-version", "2023-06-01") @@ -90,9 +103,24 @@ func Create() { res, _ := http.DefaultClient.Do(req) defer res.Body.Close() - body, _ := io.ReadAll(res.Body) + // body, _ := io.ReadAll(res.Body) - fmt.Println(string(body)) + // fmt.Println(string(body)) + reader := bufio.NewReader(res.Body) + for { + line, err := reader.ReadString('\n') + if err == nil { + if strings.HasPrefix(line, "data:") { + fmt.Println(line) + // var result CompleteResponse + // json.Unmarshal() + } else { + continue + } + } else { + break + } + } } func ClaudeProxy(c *gin.Context) { @@ -200,3 +228,161 @@ func ClaudeProxy(c *gin.Context) { } proxy.ServeHTTP(c.Writer, c.Request) } + +func TransReq(chatreq *openai.ChatCompletionRequest) (*bytes.Buffer, error) { + transReq := CompleteRequest{ + Model: chatreq.Model, + Temperature: int(chatreq.Temperature), + TopP: int(chatreq.TopP), + Stream: chatreq.Stream, + MaxTokensToSample: chatreq.MaxTokens, + } + if transReq.MaxTokensToSample == 0 { + transReq.MaxTokensToSample = 1000000 + } + var prompt string + for _, msg := range chatreq.Messages { + switch msg.Role { + case "system": + prompt += fmt.Sprintf("\n\nHuman:%s", msg.Content) + case "user": + prompt += fmt.Sprintf("\n\nHuman:%s", msg.Content) + case "assistant": + prompt += fmt.Sprintf("\n\nAssistant:%s", msg.Content) + } + } + transReq.Prompt = prompt + "\n\nAssistant:" + var payload = bytes.NewBuffer(nil) + if err := json.NewEncoder(payload).Encode(transReq); err != nil { + return nil, err + } + return payload, nil +} + +func TransRsp(c *gin.Context, isStream bool, reader *bufio.Reader) { + if !isStream { + var completersp CompleteResponse + var chatrsp openai.ChatCompletionResponse + json.NewDecoder(reader).Decode(&completersp) + chatrsp.Model = completersp.Model + chatrsp.ID = completersp.LogID + chatrsp.Object = "chat.completion" + chatrsp.Created = time.Now().Unix() + choice := openai.ChatCompletionChoice{ + Index: 0, + FinishReason: "stop", + Message: openai.ChatCompletionMessage{ + Role: "assistant", + Content: completersp.Completion, + }, + } + chatrsp.Choices = append(chatrsp.Choices, choice) + var payload *bytes.Buffer + if err := json.NewEncoder(payload).Encode(chatrsp); err != nil { + c.JSON(http.StatusInternalServerError, gin.H{ + "error": gin.H{ + "message": err.Error(), + }, + }) + return + } + c.JSON(http.StatusOK, payload) + return + } else { + var ( + wg sync.WaitGroup + dataChan = make(chan string) + stopChan = make(chan bool) + ) + wg.Add(2) + go func() { + defer wg.Done() + for { + line, err := reader.ReadString('\n') + if err == nil { + if strings.HasPrefix(line, "data: ") { + var result CompleteResponse + json.NewDecoder(strings.NewReader(line[6:])).Decode(&result) + if result.StopReason == "" { + if result.Completion != "" { + chatrsp := openai.ChatCompletionStreamResponse{ + ID: result.LogID, + Model: result.Model, + Object: "chat.completion", + Created: time.Now().Unix(), + } + choice := openai.ChatCompletionStreamChoice{ + Delta: openai.ChatCompletionStreamChoiceDelta{ + Role: "assistant", + Content: result.Completion, + }, + FinishReason: "", + } + chatrsp.Choices = append(chatrsp.Choices, choice) + bytedate, _ := json.Marshal(chatrsp) + dataChan <- string(bytedate) + } + } else { + chatrsp := openai.ChatCompletionStreamResponse{ + ID: result.LogID, + Model: result.Model, + Object: "chat.completion", + Created: time.Now().Unix(), + } + choice := openai.ChatCompletionStreamChoice{ + Delta: openai.ChatCompletionStreamChoiceDelta{ + Role: "assistant", + Content: result.Completion, + }, + } + choice.FinishReason = openai.FinishReason(TranslatestopReason(result.StopReason)) + chatrsp.Choices = append(chatrsp.Choices, choice) + bytedate, _ := json.Marshal(chatrsp) + dataChan <- string(bytedate) + dataChan <- "[DONE]" + break + } + } else { + continue + } + } else { + break + } + } + + close(dataChan) + stopChan <- true + close(stopChan) + }() + + go func() { + defer wg.Done() + Loop: + for { + select { + case data := <-dataChan: + if data != "" { + c.Writer.WriteString("data: " + data) + c.Writer.WriteString("\n\n") + c.Writer.Flush() + } + case <-stopChan: + break Loop + } + } + }() + wg.Wait() + } +} + +// claude -> openai +func TranslatestopReason(reason string) string { + switch reason { + case "stop_sequence": + return "stop" + case "max_tokens": + return "length" + default: + return reason + } +} diff --git a/router/router.go b/router/router.go index 7cc7108..300e6c2 100644 --- a/router/router.go +++ b/router/router.go @@ -16,6 +16,7 @@ import ( "net/http/httputil" "net/url" "opencatd-open/pkg/azureopenai" + "opencatd-open/pkg/claude" "opencatd-open/store" "os" "path/filepath" @@ -286,7 +287,7 @@ func HandleAddKey(c *gin.Context) { return } k := &store.Key{ - ApiType: "azure_openai", + ApiType: "azure", Name: body.Name, Key: body.Key, ResourceNmae: keynames[1], @@ -298,7 +299,7 @@ func HandleAddKey(c *gin.Context) { }}) return } - } else if strings.HasPrefix(body.Name, "anthropic.") { + } else if strings.HasPrefix(body.Name, "claude.") { keynames := strings.Split(body.Name, ".") if len(keynames) < 2 { c.JSON(http.StatusBadRequest, gin.H{"error": gin.H{ @@ -310,7 +311,8 @@ func HandleAddKey(c *gin.Context) { body.Endpoint = "https://api.anthropic.com" } k := &store.Key{ - ApiType: "anthropic", + // ApiType: "anthropic", + ApiType: "claude", Name: body.Name, Key: body.Key, ResourceNmae: keynames[1], @@ -459,6 +461,7 @@ func HandleProy(c *gin.Context) { chatreq = openai.ChatCompletionRequest{} chatres = openai.ChatCompletionResponse{} chatlog store.Tokens + onekey store.Key pre_prompt string req *http.Request err error @@ -469,6 +472,10 @@ func HandleProy(c *gin.Context) { localuser = store.IsExistAuthCache(auth[7:]) c.Set("localuser", auth[7:]) } + if c.Request.URL.Path == "/v1/complete" { + claude.ClaudeProxy(c) + return + } if c.Request.URL.Path == "/v1/audio/transcriptions" { WhisperProxy(c) return @@ -481,12 +488,12 @@ func HandleProy(c *gin.Context) { }}) return } - onekey := store.FromKeyCacheRandomItemKey() if err := c.BindJSON(&chatreq); err != nil { c.AbortWithError(http.StatusBadRequest, err) return } + chatlog.Model = chatreq.Model for _, m := range chatreq.Messages { pre_prompt += m.Content + "\n" @@ -498,8 +505,28 @@ func HandleProy(c *gin.Context) { var body bytes.Buffer json.NewEncoder(&body).Encode(chatreq) + + if strings.HasPrefix(chatreq.Model, "claude-") { + onekey, err = store.SelectKeyCache("claude") + if err != nil { + c.AbortWithError(http.StatusForbidden, err) + } + } else { + onekey = store.FromKeyCacheRandomItemKey() + } + // 创建 API 请求 switch onekey.ApiType { + case "claude": + payload, _ := claude.TransReq(&chatreq) + buildurl := "https://api.anthropic.com/v1/complete" + req, err = http.NewRequest("POST", buildurl, payload) + req.Header.Add("accept", "application/json") + req.Header.Add("anthropic-version", "2023-06-01") + req.Header.Add("x-api-key", onekey.Key) + req.Header.Add("content-type", "application/json") + case "azure": + fallthrough case "azure_openai": var buildurl string var apiVersion = "2023-05-15" @@ -533,7 +560,7 @@ func HandleProy(c *gin.Context) { req, err = http.NewRequest(c.Request.Method, baseUrl+c.Request.RequestURI, c.Request.Body) if err != nil { log.Println(err) - c.JSON(http.StatusOK, gin.H{"error": err.Error()}) + c.JSON(http.StatusBadRequest, gin.H{"error": err.Error()}) return } req.Header = c.Request.Header @@ -542,7 +569,7 @@ func HandleProy(c *gin.Context) { resp, err := client.Do(req) if err != nil { log.Println(err) - c.JSON(http.StatusOK, gin.H{"error": err.Error()}) + c.JSON(http.StatusBadRequest, gin.H{"error": err.Error()}) return } defer resp.Body.Close() @@ -574,14 +601,42 @@ func HandleProy(c *gin.Context) { reader := bufio.NewReader(resp.Body) if resp.StatusCode == 200 && localuser { - if isStream { - contentCh := fetchResponseContent(c, reader) - var buffer bytes.Buffer - for content := range contentCh { - buffer.WriteString(content) + switch onekey.ApiType { + case "claude": + claude.TransRsp(c, isStream, reader) + return + case "openai", "azure", "azure_openai": + fallthrough + default: + if isStream { + contentCh := fetchResponseContent(c, reader) + var buffer bytes.Buffer + for content := range contentCh { + buffer.WriteString(content) + } + chatlog.CompletionCount = NumTokensFromStr(buffer.String(), chatreq.Model) + chatlog.TotalTokens = chatlog.PromptCount + chatlog.CompletionCount + chatlog.Cost = fmt.Sprintf("%.6f", Cost(chatlog.Model, chatlog.PromptCount, chatlog.CompletionCount)) + if err := store.Record(&chatlog); err != nil { + log.Println(err) + } + if err := store.SumDaily(chatlog.UserID); err != nil { + log.Println(err) + } + return } - chatlog.CompletionCount = NumTokensFromStr(buffer.String(), chatreq.Model) - chatlog.TotalTokens = chatlog.PromptCount + chatlog.CompletionCount + res, err := io.ReadAll(reader) + if err != nil { + c.JSON(http.StatusUnauthorized, gin.H{"error": gin.H{ + "message": err.Error(), + }}) + return + } + reader = bufio.NewReader(bytes.NewBuffer(res)) + json.NewDecoder(bytes.NewBuffer(res)).Decode(&chatres) + chatlog.PromptCount = chatres.Usage.PromptTokens + chatlog.CompletionCount = chatres.Usage.CompletionTokens + chatlog.TotalTokens = chatres.Usage.TotalTokens chatlog.Cost = fmt.Sprintf("%.6f", Cost(chatlog.Model, chatlog.PromptCount, chatlog.CompletionCount)) if err := store.Record(&chatlog); err != nil { log.Println(err) @@ -589,26 +644,6 @@ func HandleProy(c *gin.Context) { if err := store.SumDaily(chatlog.UserID); err != nil { log.Println(err) } - return - } - res, err := io.ReadAll(reader) - if err != nil { - c.JSON(http.StatusUnauthorized, gin.H{"error": gin.H{ - "message": err.Error(), - }}) - return - } - reader = bufio.NewReader(bytes.NewBuffer(res)) - json.NewDecoder(bytes.NewBuffer(res)).Decode(&chatres) - chatlog.PromptCount = chatres.Usage.PromptTokens - chatlog.CompletionCount = chatres.Usage.CompletionTokens - chatlog.TotalTokens = chatres.Usage.TotalTokens - chatlog.Cost = fmt.Sprintf("%.6f", Cost(chatlog.Model, chatlog.PromptCount, chatlog.CompletionCount)) - if err := store.Record(&chatlog); err != nil { - log.Println(err) - } - if err := store.SumDaily(chatlog.UserID); err != nil { - log.Println(err) } } From aa1d00cd55ecbe12cd007c248e5006e9a1e98504 Mon Sep 17 00:00:00 2001 From: Sakurasan <1173092237@qq.com> Date: Sat, 16 Sep 2023 21:18:32 +0800 Subject: [PATCH 3/3] claude doc --- README.md | 1 + doc/API.md | 7 +++++-- doc/azure.md | 4 ++++ 3 files changed, 10 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 4d492ff..5f5e586 100644 --- a/README.md +++ b/README.md @@ -15,6 +15,7 @@ OpenCat for Team的开源实现 | 任务 | 完成情况 | | --- | --- | |[Azure OpenAI](./doc/azure.md) | ✅| +|[Claude](./doc/azure.md) | ~~✅~~| | ... | ... | diff --git a/doc/API.md b/doc/API.md index fb3a8c0..f90c6cb 100644 --- a/doc/API.md +++ b/doc/API.md @@ -181,7 +181,7 @@ Req: } ``` -api_type:不传的话默认为“openai”;当前可选值[openai,azure_openai] +api_type:不传的话默认为“openai”;当前可选值[openai,azure,claude] endpoint: 当 api_type 为 azure_openai时传入(目前暂未使用) Resp: @@ -236,4 +236,7 @@ Resp: "totalUnit" : 55 } ] -``` \ No newline at end of file +``` + +## Whisper接口 +### 与openai一致 diff --git a/doc/azure.md b/doc/azure.md index 0b05d52..3f3f1b4 100644 --- a/doc/azure.md +++ b/doc/azure.md @@ -19,3 +19,7 @@ - [AMA(问天)](http://bytemyth.com/ama) 使用方式 - ![](azure_ama.png) - 每个 team server 用户旁边有一个复制按钮,点击后,把复制的链接粘贴到浏览器,可以一键设置 + +## Claude + +- opencat 添加Claude api, key name以 "claude.key名称",即("Api类型.Key名称")