chore: Track all token usage metrics (#916)
This commit is contained in:
@@ -9353,6 +9353,11 @@ ${SYSTEM_REMINDER_CLOSE}
|
||||
step_count: stats.usage.stepCount,
|
||||
prompt_tokens: stats.usage.promptTokens,
|
||||
completion_tokens: stats.usage.completionTokens,
|
||||
total_tokens: stats.usage.totalTokens,
|
||||
cached_input_tokens: stats.usage.cachedInputTokens,
|
||||
cache_write_tokens: stats.usage.cacheWriteTokens,
|
||||
reasoning_tokens: stats.usage.reasoningTokens,
|
||||
context_tokens: stats.usage.contextTokens,
|
||||
};
|
||||
})(),
|
||||
agent_info: {
|
||||
|
||||
@@ -47,6 +47,12 @@ export function formatUsageStats({
|
||||
`Total duration (API): ${formatDuration(stats.totalApiMs)}`,
|
||||
`Total duration (wall): ${formatDuration(stats.totalWallMs)}`,
|
||||
`Session usage: ${stats.usage.stepCount} steps, ${formatCompact(stats.usage.promptTokens)} input, ${formatCompact(stats.usage.completionTokens)} output`,
|
||||
`Token details: ${formatCompact(stats.usage.totalTokens)} total, ${formatCompact(stats.usage.cachedInputTokens)} cached_input, ${formatCompact(stats.usage.cacheWriteTokens)} cache_write, ${formatCompact(stats.usage.reasoningTokens)} reasoning`,
|
||||
...(stats.usage.contextTokens !== undefined
|
||||
? [
|
||||
`Latest context: ${formatCompact(stats.usage.contextTokens)} tokens`,
|
||||
]
|
||||
: []),
|
||||
"",
|
||||
];
|
||||
|
||||
|
||||
@@ -220,8 +220,10 @@ export type Buffers = {
|
||||
promptTokens: number;
|
||||
completionTokens: number;
|
||||
totalTokens: number;
|
||||
cachedTokens: number;
|
||||
cachedInputTokens: number;
|
||||
cacheWriteTokens: number;
|
||||
reasoningTokens: number;
|
||||
contextTokens?: number;
|
||||
stepCount: number;
|
||||
};
|
||||
// Aggressive static promotion: split streaming content at paragraph boundaries
|
||||
@@ -249,7 +251,8 @@ export function createBuffers(agentId?: string): Buffers {
|
||||
promptTokens: 0,
|
||||
completionTokens: 0,
|
||||
totalTokens: 0,
|
||||
cachedTokens: 0,
|
||||
cachedInputTokens: 0,
|
||||
cacheWriteTokens: 0,
|
||||
reasoningTokens: 0,
|
||||
stepCount: 0,
|
||||
},
|
||||
@@ -807,10 +810,40 @@ export function onChunk(
|
||||
if (chunk.total_tokens !== undefined) {
|
||||
b.usage.totalTokens += chunk.total_tokens;
|
||||
}
|
||||
if (
|
||||
chunk.cached_input_tokens !== undefined &&
|
||||
chunk.cached_input_tokens !== null
|
||||
) {
|
||||
b.usage.cachedInputTokens += chunk.cached_input_tokens;
|
||||
}
|
||||
if (
|
||||
chunk.cache_write_tokens !== undefined &&
|
||||
chunk.cache_write_tokens !== null
|
||||
) {
|
||||
b.usage.cacheWriteTokens += chunk.cache_write_tokens;
|
||||
}
|
||||
if (
|
||||
chunk.reasoning_tokens !== undefined &&
|
||||
chunk.reasoning_tokens !== null
|
||||
) {
|
||||
b.usage.reasoningTokens += chunk.reasoning_tokens;
|
||||
}
|
||||
const usageChunk = chunk as typeof chunk & {
|
||||
context_tokens?: number | null;
|
||||
};
|
||||
if (
|
||||
usageChunk.context_tokens !== undefined &&
|
||||
usageChunk.context_tokens !== null
|
||||
) {
|
||||
// context_tokens is a snapshot metric, not additive.
|
||||
b.usage.contextTokens = usageChunk.context_tokens;
|
||||
}
|
||||
// Use context_tokens from SDK (estimate of tokens in context window)
|
||||
if (ctx) {
|
||||
const usageChunk = chunk as typeof chunk & { context_tokens?: number };
|
||||
if (usageChunk.context_tokens !== undefined) {
|
||||
if (
|
||||
usageChunk.context_tokens !== undefined &&
|
||||
usageChunk.context_tokens !== null
|
||||
) {
|
||||
ctx.lastContextTokens = usageChunk.context_tokens;
|
||||
// Track history for time-series display
|
||||
const compacted = ctx.pendingCompaction;
|
||||
|
||||
Reference in New Issue
Block a user