262 lines
9.4 KiB
TypeScript
262 lines
9.4 KiB
TypeScript
import type { BetaUsage as Usage } from '@anthropic-ai/sdk/resources/beta/messages/messages.mjs'
|
|
import { roughTokenCountEstimationForMessages } from '../services/tokenEstimation.js'
|
|
import type { AssistantMessage, Message } from '../types/message.js'
|
|
import { SYNTHETIC_MESSAGES, SYNTHETIC_MODEL } from './messages.js'
|
|
import { jsonStringify } from './slowOperations.js'
|
|
|
|
export function getTokenUsage(message: Message): Usage | undefined {
|
|
if (
|
|
message?.type === 'assistant' &&
|
|
'usage' in message.message &&
|
|
!(
|
|
message.message.content[0]?.type === 'text' &&
|
|
SYNTHETIC_MESSAGES.has(message.message.content[0].text)
|
|
) &&
|
|
message.message.model !== SYNTHETIC_MODEL
|
|
) {
|
|
return message.message.usage
|
|
}
|
|
return undefined
|
|
}
|
|
|
|
/**
|
|
* Get the API response id for an assistant message with real (non-synthetic) usage.
|
|
* Used to identify split assistant records that came from the same API response —
|
|
* when parallel tool calls are streamed, each content block becomes a separate
|
|
* AssistantMessage record, but they all share the same message.id.
|
|
*/
|
|
function getAssistantMessageId(message: Message): string | undefined {
|
|
if (
|
|
message?.type === 'assistant' &&
|
|
'id' in message.message &&
|
|
message.message.model !== SYNTHETIC_MODEL
|
|
) {
|
|
return message.message.id
|
|
}
|
|
return undefined
|
|
}
|
|
|
|
/**
|
|
* Calculate total context window tokens from an API response's usage data.
|
|
* Includes input_tokens + cache tokens + output_tokens.
|
|
*
|
|
* This represents the full context size at the time of that API call.
|
|
* Use tokenCountWithEstimation() when you need context size from messages.
|
|
*/
|
|
export function getTokenCountFromUsage(usage: Usage): number {
|
|
return (
|
|
usage.input_tokens +
|
|
(usage.cache_creation_input_tokens ?? 0) +
|
|
(usage.cache_read_input_tokens ?? 0) +
|
|
usage.output_tokens
|
|
)
|
|
}
|
|
|
|
export function tokenCountFromLastAPIResponse(messages: Message[]): number {
|
|
let i = messages.length - 1
|
|
while (i >= 0) {
|
|
const message = messages[i]
|
|
const usage = message ? getTokenUsage(message) : undefined
|
|
if (usage) {
|
|
return getTokenCountFromUsage(usage)
|
|
}
|
|
i--
|
|
}
|
|
return 0
|
|
}
|
|
|
|
/**
|
|
* Final context window size from the last API response's usage.iterations[-1].
|
|
* Used for task_budget.remaining computation across compaction boundaries —
|
|
* the server's budget countdown is context-based, so remaining decrements by
|
|
* the pre-compact final window, not billing spend. See monorepo
|
|
* api/api/sampling/prompt/renderer.py:292 for the server-side computation.
|
|
*
|
|
* Falls back to top-level input_tokens + output_tokens when iterations is
|
|
* absent (no server-side tool loops, so top-level usage IS the final window).
|
|
* Both paths exclude cache tokens to match #304930's formula.
|
|
*/
|
|
export function finalContextTokensFromLastResponse(
|
|
messages: Message[],
|
|
): number {
|
|
let i = messages.length - 1
|
|
while (i >= 0) {
|
|
const message = messages[i]
|
|
const usage = message ? getTokenUsage(message) : undefined
|
|
if (usage) {
|
|
// Stainless types don't include iterations yet — cast like advisor.ts:43
|
|
const iterations = (
|
|
usage as {
|
|
iterations?: Array<{
|
|
input_tokens: number
|
|
output_tokens: number
|
|
}> | null
|
|
}
|
|
).iterations
|
|
if (iterations && iterations.length > 0) {
|
|
const last = iterations.at(-1)!
|
|
return last.input_tokens + last.output_tokens
|
|
}
|
|
// No iterations → no server tool loop → top-level usage IS the final
|
|
// window. Match the iterations path's formula (input + output, no cache)
|
|
// rather than getTokenCountFromUsage — #304930 defines final window as
|
|
// non-cache input + output. Whether the server's budget countdown
|
|
// (renderer.py:292 calculate_context_tokens) counts cache the same way
|
|
// is an open question; aligning with the iterations path keeps the two
|
|
// branches consistent until that's resolved.
|
|
return usage.input_tokens + usage.output_tokens
|
|
}
|
|
i--
|
|
}
|
|
return 0
|
|
}
|
|
|
|
/**
|
|
* Get only the output_tokens from the last API response.
|
|
* This excludes input context (system prompt, tools, prior messages).
|
|
*
|
|
* WARNING: Do NOT use this for threshold comparisons (autocompact, session memory).
|
|
* Use tokenCountWithEstimation() instead, which measures full context size.
|
|
* This function is only useful for measuring how many tokens Claude generated
|
|
* in a single response, not how full the context window is.
|
|
*/
|
|
export function messageTokenCountFromLastAPIResponse(
|
|
messages: Message[],
|
|
): number {
|
|
let i = messages.length - 1
|
|
while (i >= 0) {
|
|
const message = messages[i]
|
|
const usage = message ? getTokenUsage(message) : undefined
|
|
if (usage) {
|
|
return usage.output_tokens
|
|
}
|
|
i--
|
|
}
|
|
return 0
|
|
}
|
|
|
|
export function getCurrentUsage(messages: Message[]): {
|
|
input_tokens: number
|
|
output_tokens: number
|
|
cache_creation_input_tokens: number
|
|
cache_read_input_tokens: number
|
|
} | null {
|
|
for (let i = messages.length - 1; i >= 0; i--) {
|
|
const message = messages[i]
|
|
const usage = message ? getTokenUsage(message) : undefined
|
|
if (usage) {
|
|
return {
|
|
input_tokens: usage.input_tokens,
|
|
output_tokens: usage.output_tokens,
|
|
cache_creation_input_tokens: usage.cache_creation_input_tokens ?? 0,
|
|
cache_read_input_tokens: usage.cache_read_input_tokens ?? 0,
|
|
}
|
|
}
|
|
}
|
|
return null
|
|
}
|
|
|
|
export function doesMostRecentAssistantMessageExceed200k(
|
|
messages: Message[],
|
|
): boolean {
|
|
const THRESHOLD = 200_000
|
|
|
|
const lastAsst = messages.findLast(m => m.type === 'assistant')
|
|
if (!lastAsst) return false
|
|
const usage = getTokenUsage(lastAsst)
|
|
return usage ? getTokenCountFromUsage(usage) > THRESHOLD : false
|
|
}
|
|
|
|
/**
|
|
* Calculate the character content length of an assistant message.
|
|
* Used for spinner token estimation (characters / 4 ≈ tokens).
|
|
* This is used when subagent streaming events are filtered out and we
|
|
* need to count content from completed messages instead.
|
|
*
|
|
* Counts the same content that handleMessageFromStream would count via deltas:
|
|
* - text (text_delta)
|
|
* - thinking (thinking_delta)
|
|
* - redacted_thinking data
|
|
* - tool_use input (input_json_delta)
|
|
* Note: signature_delta is excluded from streaming counts (not model output).
|
|
*/
|
|
export function getAssistantMessageContentLength(
|
|
message: AssistantMessage,
|
|
): number {
|
|
let contentLength = 0
|
|
for (const block of message.message.content) {
|
|
if (block.type === 'text') {
|
|
contentLength += block.text.length
|
|
} else if (block.type === 'thinking') {
|
|
contentLength += block.thinking.length
|
|
} else if (block.type === 'redacted_thinking') {
|
|
contentLength += block.data.length
|
|
} else if (block.type === 'tool_use') {
|
|
contentLength += jsonStringify(block.input).length
|
|
}
|
|
}
|
|
return contentLength
|
|
}
|
|
|
|
/**
|
|
* Get the current context window size in tokens.
|
|
*
|
|
* This is the CANONICAL function for measuring context size when checking
|
|
* thresholds (autocompact, session memory init, etc.). Uses the last API
|
|
* response's token count (input + output + cache) plus estimates for any
|
|
* messages added since.
|
|
*
|
|
* Always use this instead of:
|
|
* - Cumulative token counting (which double-counts as context grows)
|
|
* - messageTokenCountFromLastAPIResponse (which only counts output_tokens)
|
|
* - tokenCountFromLastAPIResponse (which doesn't estimate new messages)
|
|
*
|
|
* Implementation note on parallel tool calls: when the model makes multiple
|
|
* tool calls in one response, the streaming code emits a SEPARATE assistant
|
|
* record per content block (all sharing the same message.id and usage), and
|
|
* the query loop interleaves each tool_result immediately after its tool_use.
|
|
* So the messages array looks like:
|
|
* [..., assistant(id=A), user(result), assistant(id=A), user(result), ...]
|
|
* If we stop at the LAST assistant record, we only estimate the one tool_result
|
|
* after it and miss all the earlier interleaved tool_results — which will ALL
|
|
* be in the next API request. To avoid undercounting, after finding a usage-
|
|
* bearing record we walk back to the FIRST sibling with the same message.id
|
|
* so every interleaved tool_result is included in the rough estimate.
|
|
*/
|
|
export function tokenCountWithEstimation(messages: readonly Message[]): number {
|
|
let i = messages.length - 1
|
|
while (i >= 0) {
|
|
const message = messages[i]
|
|
const usage = message ? getTokenUsage(message) : undefined
|
|
if (message && usage) {
|
|
// Walk back past any earlier sibling records split from the same API
|
|
// response (same message.id) so interleaved tool_results between them
|
|
// are included in the estimation slice.
|
|
const responseId = getAssistantMessageId(message)
|
|
if (responseId) {
|
|
let j = i - 1
|
|
while (j >= 0) {
|
|
const prior = messages[j]
|
|
const priorId = prior ? getAssistantMessageId(prior) : undefined
|
|
if (priorId === responseId) {
|
|
// Earlier split of the same API response — anchor here instead.
|
|
i = j
|
|
} else if (priorId !== undefined) {
|
|
// Hit a different API response — stop walking.
|
|
break
|
|
}
|
|
// priorId === undefined: a user/tool_result/attachment message,
|
|
// possibly interleaved between splits — keep walking.
|
|
j--
|
|
}
|
|
}
|
|
return (
|
|
getTokenCountFromUsage(usage) +
|
|
roughTokenCountEstimationForMessages(messages.slice(i + 1))
|
|
)
|
|
}
|
|
i--
|
|
}
|
|
return roughTokenCountEstimationForMessages(messages)
|
|
}
|