add log
This commit is contained in:
@@ -60,7 +60,7 @@ func (h *Proxy) ChatHandler(c *gin.Context) {
|
||||
c.SSEvent("", data)
|
||||
}
|
||||
}
|
||||
go func() {
|
||||
|
||||
llmusage := llm.GetTokenUsage()
|
||||
|
||||
cost := tokenizer.Cost(llmusage.Model, llmusage.PromptTokens+llmusage.ToolsTokens, llmusage.CompletionTokens)
|
||||
@@ -76,6 +76,5 @@ func (h *Proxy) ChatHandler(c *gin.Context) {
|
||||
}
|
||||
h.SendUsage(&usage)
|
||||
defer fmt.Println("cost:", cost, "prompt_tokens:", llmusage.PromptTokens, "completion_tokens:", llmusage.CompletionTokens, "total_tokens:", llmusage.TotalTokens)
|
||||
}()
|
||||
|
||||
}
|
||||
|
||||
+1
-1
@@ -13,7 +13,7 @@ type LLM interface {
|
||||
|
||||
type llm struct {
|
||||
ApiKey *model.ApiKey
|
||||
Usage *model.Usage
|
||||
Usage *TokenUsage
|
||||
tools any // TODO
|
||||
Messages []any // TODO
|
||||
llm LLM
|
||||
|
||||
@@ -218,6 +218,7 @@ func (o *OpenAICompatible) StreamChat(ctx context.Context, chatReq llm.ChatReque
|
||||
// case output <- &streamResp:
|
||||
// }
|
||||
}
|
||||
fmt.Println("llm usage:", o.tokenUsage.Model, o.tokenUsage.PromptTokens, o.tokenUsage.CompletionTokens, o.tokenUsage.TotalTokens)
|
||||
}()
|
||||
return output, nil
|
||||
}
|
||||
|
||||
Reference in New Issue
Block a user