fix: use Gemini response metadata for token counting (#11226)
This commit is contained in:
@@ -254,6 +254,10 @@ class GoogleLargeLanguageModel(LargeLanguageModel):
|
|||||||
assistant_prompt_message = AssistantPromptMessage(content=response.text)
|
assistant_prompt_message = AssistantPromptMessage(content=response.text)
|
||||||
|
|
||||||
# calculate num tokens
|
# calculate num tokens
|
||||||
|
if response.usage_metadata:
|
||||||
|
prompt_tokens = response.usage_metadata.prompt_token_count
|
||||||
|
completion_tokens = response.usage_metadata.candidates_token_count
|
||||||
|
else:
|
||||||
prompt_tokens = self.get_num_tokens(model, credentials, prompt_messages)
|
prompt_tokens = self.get_num_tokens(model, credentials, prompt_messages)
|
||||||
completion_tokens = self.get_num_tokens(model, credentials, [assistant_prompt_message])
|
completion_tokens = self.get_num_tokens(model, credentials, [assistant_prompt_message])
|
||||||
|
|
||||||
|
Reference in New Issue
Block a user