fix: reasoning ui for auto models (#1316)

This commit is contained in:
cthomas
2026-03-09 19:32:15 -07:00
committed by GitHub
parent 4a261a2ae8
commit b9c4b3068f
2 changed files with 5 additions and 3 deletions

View File

@@ -1562,8 +1562,10 @@ export default function App({
}, [currentModelLabel, derivedReasoningEffort, llmConfig]);
const currentModelProvider = llmConfig?.provider_name ?? null;
const currentReasoningEffort: ModelReasoningEffort | null =
derivedReasoningEffort ??
inferReasoningEffortFromModelPreset(currentModelId, currentModelLabel);
currentModelLabel?.startsWith("letta/auto")
? null
: (derivedReasoningEffort ??
inferReasoningEffortFromModelPreset(currentModelId, currentModelLabel));
// Billing tier for conditional UI and error context (fetched once on mount)
const [billingTier, setBillingTier] = useState<string | null>(null);

View File

@@ -2544,7 +2544,7 @@ async function handleIncomingMessage(
onStatusChange?.("processing", connectionId);
}
let messagesToSend: Array<MessageCreate | ApprovalCreate> = [];
const messagesToSend: Array<MessageCreate | ApprovalCreate> = [];
let turnToolContextId: string | null = null;
let queuedInterruptedToolCallIds: string[] = [];