From da90303450345708649186756904de4d9a5035dd Mon Sep 17 00:00:00 2001 From: Nate Date: Wed, 29 Oct 2025 16:00:13 -0700 Subject: [PATCH] Add cache_write_tokens and cache_read_tokens to Anthropic prompt token details --- packages/openai-adapters/src/apis/Anthropic.ts | 6 +++++- 1 file changed, 5 insertions(+), 1 deletion(-) diff --git a/packages/openai-adapters/src/apis/Anthropic.ts b/packages/openai-adapters/src/apis/Anthropic.ts index abfe53446b9..0a7bc573cc5 100644 --- a/packages/openai-adapters/src/apis/Anthropic.ts +++ b/packages/openai-adapters/src/apis/Anthropic.ts @@ -366,9 +366,13 @@ export class AnthropicApi implements BaseLlmApi { const startEvent = rawEvent as RawMessageStartEvent; usage.prompt_tokens = startEvent.message.usage?.input_tokens ?? 0; usage.prompt_tokens_details = { + cache_write_tokens: + startEvent.message.usage?.cache_creation_input_tokens ?? 0, + cache_read_tokens: + startEvent.message.usage?.cache_read_input_tokens ?? 0, cached_tokens: startEvent.message.usage?.cache_read_input_tokens ?? 0, - }; + } as any; break; case "message_delta": const deltaEvent = rawEvent as RawMessageDeltaEvent;