Files
letta-server/letta
jnjpng eb4a0daabd fix: allow explicit null for max_tokens on GPT-5 models (#9562)
The Pydantic validator `set_model_specific_defaults` was checking
`values.get("max_tokens") is None`, which matched both "field not
provided" and "field explicitly set to null". This meant users could
not disable the max output tokens limit for GPT-5/GPT-4.1 models -
the validator would always override null with a default value during
request deserialization.

Changed to `"max_tokens" not in values` so that an explicit
`max_tokens: null` is preserved while still applying defaults when
the field is omitted entirely.
2026-02-24 10:55:12 -08:00
..
2025-10-07 17:50:45 -07:00
2025-09-17 15:47:40 -07:00
2025-09-17 15:47:40 -07:00
2025-09-17 15:47:40 -07:00
2025-09-17 15:47:40 -07:00
2025-10-07 17:50:45 -07:00
2025-09-17 15:47:40 -07:00
2026-02-24 10:55:08 -08:00
2025-09-17 15:47:40 -07:00
2025-09-17 15:47:40 -07:00