chore: Track all token usage metrics (#916)

This commit is contained in:
Devansh Jain
2026-02-11 12:53:05 -08:00
committed by GitHub
parent c8ca88275a
commit 94376a3233
7 changed files with 171 additions and 26 deletions

View File

@@ -9353,6 +9353,11 @@ ${SYSTEM_REMINDER_CLOSE}
step_count: stats.usage.stepCount,
prompt_tokens: stats.usage.promptTokens,
completion_tokens: stats.usage.completionTokens,
total_tokens: stats.usage.totalTokens,
cached_input_tokens: stats.usage.cachedInputTokens,
cache_write_tokens: stats.usage.cacheWriteTokens,
reasoning_tokens: stats.usage.reasoningTokens,
context_tokens: stats.usage.contextTokens,
};
})(),
agent_info: {

View File

@@ -47,6 +47,12 @@ export function formatUsageStats({
`Total duration (API): ${formatDuration(stats.totalApiMs)}`,
`Total duration (wall): ${formatDuration(stats.totalWallMs)}`,
`Session usage: ${stats.usage.stepCount} steps, ${formatCompact(stats.usage.promptTokens)} input, ${formatCompact(stats.usage.completionTokens)} output`,
`Token details: ${formatCompact(stats.usage.totalTokens)} total, ${formatCompact(stats.usage.cachedInputTokens)} cached_input, ${formatCompact(stats.usage.cacheWriteTokens)} cache_write, ${formatCompact(stats.usage.reasoningTokens)} reasoning`,
...(stats.usage.contextTokens !== undefined
? [
`Latest context: ${formatCompact(stats.usage.contextTokens)} tokens`,
]
: []),
"",
];

View File

@@ -220,8 +220,10 @@ export type Buffers = {
promptTokens: number;
completionTokens: number;
totalTokens: number;
cachedTokens: number;
cachedInputTokens: number;
cacheWriteTokens: number;
reasoningTokens: number;
contextTokens?: number;
stepCount: number;
};
// Aggressive static promotion: split streaming content at paragraph boundaries
@@ -249,7 +251,8 @@ export function createBuffers(agentId?: string): Buffers {
promptTokens: 0,
completionTokens: 0,
totalTokens: 0,
cachedTokens: 0,
cachedInputTokens: 0,
cacheWriteTokens: 0,
reasoningTokens: 0,
stepCount: 0,
},
@@ -807,10 +810,40 @@ export function onChunk(
if (chunk.total_tokens !== undefined) {
b.usage.totalTokens += chunk.total_tokens;
}
if (
chunk.cached_input_tokens !== undefined &&
chunk.cached_input_tokens !== null
) {
b.usage.cachedInputTokens += chunk.cached_input_tokens;
}
if (
chunk.cache_write_tokens !== undefined &&
chunk.cache_write_tokens !== null
) {
b.usage.cacheWriteTokens += chunk.cache_write_tokens;
}
if (
chunk.reasoning_tokens !== undefined &&
chunk.reasoning_tokens !== null
) {
b.usage.reasoningTokens += chunk.reasoning_tokens;
}
const usageChunk = chunk as typeof chunk & {
context_tokens?: number | null;
};
if (
usageChunk.context_tokens !== undefined &&
usageChunk.context_tokens !== null
) {
// context_tokens is a snapshot metric, not additive.
b.usage.contextTokens = usageChunk.context_tokens;
}
// Use context_tokens from SDK (estimate of tokens in context window)
if (ctx) {
const usageChunk = chunk as typeof chunk & { context_tokens?: number };
if (usageChunk.context_tokens !== undefined) {
if (
usageChunk.context_tokens !== undefined &&
usageChunk.context_tokens !== null
) {
ctx.lastContextTokens = usageChunk.context_tokens;
// Track history for time-series display
const compacted = ctx.pendingCompaction;