fix: reasoning ui for auto models (#1316)
This commit is contained in:
@@ -1562,8 +1562,10 @@ export default function App({
|
||||
}, [currentModelLabel, derivedReasoningEffort, llmConfig]);
|
||||
const currentModelProvider = llmConfig?.provider_name ?? null;
|
||||
const currentReasoningEffort: ModelReasoningEffort | null =
|
||||
derivedReasoningEffort ??
|
||||
inferReasoningEffortFromModelPreset(currentModelId, currentModelLabel);
|
||||
currentModelLabel?.startsWith("letta/auto")
|
||||
? null
|
||||
: (derivedReasoningEffort ??
|
||||
inferReasoningEffortFromModelPreset(currentModelId, currentModelLabel));
|
||||
|
||||
// Billing tier for conditional UI and error context (fetched once on mount)
|
||||
const [billingTier, setBillingTier] = useState<string | null>(null);
|
||||
|
||||
@@ -2544,7 +2544,7 @@ async function handleIncomingMessage(
|
||||
onStatusChange?.("processing", connectionId);
|
||||
}
|
||||
|
||||
let messagesToSend: Array<MessageCreate | ApprovalCreate> = [];
|
||||
const messagesToSend: Array<MessageCreate | ApprovalCreate> = [];
|
||||
let turnToolContextId: string | null = null;
|
||||
let queuedInterruptedToolCallIds: string[] = [];
|
||||
|
||||
|
||||
Reference in New Issue
Block a user