feat: xhigh reasoning for gpt-5.2 (#6735)

This commit is contained in:
Ari Webb
2025-12-11 17:13:06 -08:00
committed by Caren Thomas
parent bffb9064b8
commit fecf503ad9
7 changed files with 105 additions and 74 deletions

View File

@@ -17856,7 +17856,8 @@
"minimal",
"low",
"medium",
"high"
"high",
"xhigh"
],
"nullable": true
},
@@ -21043,7 +21044,8 @@
"additionalProperties": true,
"type": "object",
"required": ["type", "url_citation"],
"title": "Annotation"
"title": "Annotation",
"description": "A URL citation when using web search."
},
"AnnotationURLCitation": {
"properties": {
@@ -21067,7 +21069,8 @@
"additionalProperties": true,
"type": "object",
"required": ["end_index", "start_index", "title", "url"],
"title": "AnnotationURLCitation"
"title": "AnnotationURLCitation",
"description": "A URL citation when using web search."
},
"AnthropicModelSettings": {
"properties": {
@@ -21963,7 +21966,8 @@
},
"type": "object",
"required": ["id"],
"title": "Audio"
"title": "Audio",
"description": "Data about a previous audio response from the model.\n[Learn more](https://platform.openai.com/docs/guides/audio)."
},
"AuthRequest": {
"properties": {
@@ -23369,7 +23373,8 @@
"additionalProperties": true,
"type": "object",
"required": ["id", "choices", "created", "model", "object"],
"title": "ChatCompletion"
"title": "ChatCompletion",
"description": "Represents a chat completion response returned by model, based on the provided input."
},
"ChatCompletionAssistantMessageParam": {
"properties": {
@@ -23454,7 +23459,8 @@
},
"type": "object",
"required": ["role"],
"title": "ChatCompletionAssistantMessageParam"
"title": "ChatCompletionAssistantMessageParam",
"description": "Messages sent by the model in response to user messages."
},
"ChatCompletionAudio": {
"properties": {
@@ -23478,7 +23484,8 @@
"additionalProperties": true,
"type": "object",
"required": ["id", "data", "expires_at", "transcript"],
"title": "ChatCompletionAudio"
"title": "ChatCompletionAudio",
"description": "If the audio output modality is requested, this object contains data\nabout the audio response from the model. [Learn more](https://platform.openai.com/docs/guides/audio)."
},
"ChatCompletionContentPartImageParam": {
"properties": {
@@ -23493,7 +23500,8 @@
},
"type": "object",
"required": ["image_url", "type"],
"title": "ChatCompletionContentPartImageParam"
"title": "ChatCompletionContentPartImageParam",
"description": "Learn about [image inputs](https://platform.openai.com/docs/guides/vision)."
},
"ChatCompletionContentPartInputAudioParam": {
"properties": {
@@ -23508,7 +23516,8 @@
},
"type": "object",
"required": ["input_audio", "type"],
"title": "ChatCompletionContentPartInputAudioParam"
"title": "ChatCompletionContentPartInputAudioParam",
"description": "Learn about [audio inputs](https://platform.openai.com/docs/guides/audio)."
},
"ChatCompletionContentPartRefusalParam": {
"properties": {
@@ -23540,7 +23549,8 @@
},
"type": "object",
"required": ["text", "type"],
"title": "ChatCompletionContentPartTextParam"
"title": "ChatCompletionContentPartTextParam",
"description": "Learn about [text inputs](https://platform.openai.com/docs/guides/text-generation)."
},
"ChatCompletionDeveloperMessageParam": {
"properties": {
@@ -23570,7 +23580,8 @@
},
"type": "object",
"required": ["content", "role"],
"title": "ChatCompletionDeveloperMessageParam"
"title": "ChatCompletionDeveloperMessageParam",
"description": "Developer-provided instructions that the model should follow, regardless of\nmessages sent by the user. With o1 models and newer, `developer` messages\nreplace the previous `system` messages."
},
"ChatCompletionFunctionMessageParam": {
"properties": {
@@ -23687,7 +23698,8 @@
"additionalProperties": true,
"type": "object",
"required": ["role"],
"title": "ChatCompletionMessage"
"title": "ChatCompletionMessage",
"description": "A chat completion message generated by the model."
},
"ChatCompletionMessageCustomToolCall": {
"properties": {
@@ -23707,7 +23719,8 @@
"additionalProperties": true,
"type": "object",
"required": ["id", "custom", "type"],
"title": "ChatCompletionMessageCustomToolCall"
"title": "ChatCompletionMessageCustomToolCall",
"description": "A call to a custom tool created by the model."
},
"ChatCompletionMessageCustomToolCallParam": {
"properties": {
@@ -23726,7 +23739,8 @@
},
"type": "object",
"required": ["id", "custom", "type"],
"title": "ChatCompletionMessageCustomToolCallParam"
"title": "ChatCompletionMessageCustomToolCallParam",
"description": "A call to a custom tool created by the model."
},
"ChatCompletionMessageFunctionToolCall-Input": {
"properties": {
@@ -23746,7 +23760,8 @@
"additionalProperties": true,
"type": "object",
"required": ["id", "function", "type"],
"title": "ChatCompletionMessageFunctionToolCall"
"title": "ChatCompletionMessageFunctionToolCall",
"description": "A call to a function tool created by the model."
},
"ChatCompletionMessageFunctionToolCall-Output": {
"properties": {
@@ -23766,7 +23781,8 @@
"additionalProperties": true,
"type": "object",
"required": ["id", "function", "type"],
"title": "ChatCompletionMessageFunctionToolCall"
"title": "ChatCompletionMessageFunctionToolCall",
"description": "A call to a function tool created by the model."
},
"ChatCompletionMessageFunctionToolCallParam": {
"properties": {
@@ -23785,7 +23801,8 @@
},
"type": "object",
"required": ["id", "function", "type"],
"title": "ChatCompletionMessageFunctionToolCallParam"
"title": "ChatCompletionMessageFunctionToolCallParam",
"description": "A call to a function tool created by the model."
},
"ChatCompletionRequest": {
"properties": {
@@ -23980,7 +23997,8 @@
},
"type": "object",
"required": ["content", "role"],
"title": "ChatCompletionSystemMessageParam"
"title": "ChatCompletionSystemMessageParam",
"description": "Developer-provided instructions that the model should follow, regardless of\nmessages sent by the user. With o1 models and newer, use `developer` messages\nfor this purpose instead."
},
"ChatCompletionTokenLogprob": {
"properties": {
@@ -24090,7 +24108,8 @@
},
"type": "object",
"required": ["content", "role"],
"title": "ChatCompletionUserMessageParam"
"title": "ChatCompletionUserMessageParam",
"description": "Messages sent by an end user, containing prompts or additional context\ninformation."
},
"ChildToolRule": {
"properties": {
@@ -24238,7 +24257,8 @@
},
"additionalProperties": true,
"type": "object",
"title": "ChoiceLogprobs"
"title": "ChoiceLogprobs",
"description": "Log probability information for the choice."
},
"CodeInput": {
"properties": {
@@ -24367,7 +24387,8 @@
},
"additionalProperties": true,
"type": "object",
"title": "CompletionTokensDetails"
"title": "CompletionTokensDetails",
"description": "Breakdown of tokens used in a completion."
},
"CompletionUsage": {
"properties": {
@@ -24407,7 +24428,8 @@
"additionalProperties": true,
"type": "object",
"required": ["completion_tokens", "prompt_tokens", "total_tokens"],
"title": "CompletionUsage"
"title": "CompletionUsage",
"description": "Usage statistics for the completion request."
},
"ConditionalToolRule": {
"properties": {
@@ -25888,7 +25910,8 @@
},
"type": "object",
"required": ["input", "name"],
"title": "Custom"
"title": "Custom",
"description": "The custom tool that the model called."
},
"Custom-Output": {
"properties": {
@@ -25904,7 +25927,8 @@
"additionalProperties": true,
"type": "object",
"required": ["input", "name"],
"title": "Custom"
"title": "Custom",
"description": "The custom tool that the model called."
},
"DeepseekModelSettings": {
"properties": {
@@ -26621,7 +26645,8 @@
},
"type": "object",
"required": ["file", "type"],
"title": "File"
"title": "File",
"description": "Learn about [file inputs](https://platform.openai.com/docs/guides/text) for text generation."
},
"FileAgentSchema": {
"properties": {
@@ -27460,7 +27485,8 @@
"additionalProperties": true,
"type": "object",
"required": ["arguments", "name"],
"title": "Function"
"title": "Function",
"description": "The function that the model called."
},
"FunctionCall-Input": {
"properties": {
@@ -27475,7 +27501,8 @@
},
"type": "object",
"required": ["arguments", "name"],
"title": "FunctionCall"
"title": "FunctionCall",
"description": "Deprecated and replaced by `tool_calls`.\n\nThe name and arguments of a function that should be called, as generated by the model."
},
"FunctionCall-Output": {
"properties": {
@@ -27491,7 +27518,8 @@
"additionalProperties": true,
"type": "object",
"required": ["arguments", "name"],
"title": "FunctionCall"
"title": "FunctionCall",
"description": "Deprecated and replaced by `tool_calls`.\n\nThe name and arguments of a function that should be called, as generated by the model."
},
"FunctionDefinition": {
"properties": {
@@ -30331,7 +30359,7 @@
"anyOf": [
{
"type": "string",
"enum": ["none", "minimal", "low", "medium", "high"]
"enum": ["none", "minimal", "low", "medium", "high", "xhigh"]
},
{
"type": "null"
@@ -32572,7 +32600,7 @@
"anyOf": [
{
"type": "string",
"enum": ["none", "minimal", "low", "medium", "high"]
"enum": ["none", "minimal", "low", "medium", "high", "xhigh"]
},
{
"type": "null"
@@ -32905,7 +32933,7 @@
"properties": {
"reasoning_effort": {
"type": "string",
"enum": ["none", "minimal", "low", "medium", "high"],
"enum": ["none", "minimal", "low", "medium", "high", "xhigh"],
"title": "Reasoning Effort",
"description": "The reasoning effort to use when generating text reasoning models",
"default": "minimal"
@@ -33563,7 +33591,8 @@
},
"additionalProperties": true,
"type": "object",
"title": "PromptTokensDetails"
"title": "PromptTokensDetails",
"description": "Breakdown of tokens used in the prompt."
},
"Provider": {
"properties": {
@@ -41495,7 +41524,8 @@
"additionalProperties": true,
"type": "object",
"required": ["arguments", "name"],
"title": "Function"
"title": "Function",
"description": "The function that the model called."
},
"openai__types__chat__chat_completion_message_function_tool_call_param__Function": {
"properties": {
@@ -41510,7 +41540,8 @@
},
"type": "object",
"required": ["arguments", "name"],
"title": "Function"
"title": "Function",
"description": "The function that the model called."
},
"LettaMessageUnion": {
"oneOf": [

View File

@@ -70,7 +70,7 @@ class LLMConfig(BaseModel):
enable_reasoner: bool = Field(
True, description="Whether or not the model should use extended thinking if it is a 'reasoning' style model"
)
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high"]] = Field(
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] = Field(
None,
description="The reasoning effort to use when generating text reasoning models",
)

View File

@@ -65,7 +65,7 @@ class Model(LLMConfig, ModelBase):
description="Deprecated: Whether or not the model should use extended thinking if it is a 'reasoning' style model.",
deprecated=True,
)
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high"]] = Field(
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] = Field(
None, description="Deprecated: The reasoning effort to use when generating text reasoning models.", deprecated=True
)
max_reasoning_tokens: int = Field(0, description="Deprecated: Configurable thinking budget for extended thinking.", deprecated=True)
@@ -210,7 +210,7 @@ class ModelSettings(BaseModel):
class OpenAIReasoning(BaseModel):
reasoning_effort: Literal["none", "minimal", "low", "medium", "high"] = Field(
reasoning_effort: Literal["none", "minimal", "low", "medium", "high", "xhigh"] = Field(
"minimal", description="The reasoning effort to use when generating text reasoning models"
)

View File

@@ -146,7 +146,7 @@ class ChatCompletionRequest(BaseModel):
parallel_tool_calls: Optional[bool] = None
instructions: Optional[str] = None
verbosity: Optional[Literal["low", "medium", "high"]] = None # For verbosity control in GPT-5 models
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high"]] = (
reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]] = (
None # For reasoning effort control in reasoning models
)

View File

@@ -44,7 +44,7 @@ dependencies = [
"llama-index-embeddings-openai>=0.3.1",
"anthropic>=0.75.0",
"letta-client>=1.3.1",
"openai>=1.99.9",
"openai>=2.11.0",
"opentelemetry-api==1.30.0",
"opentelemetry-sdk==1.30.0",
"opentelemetry-instrumentation-requests==0.51b0",

View File

@@ -367,7 +367,7 @@ def test_reasoning_toggle_by_provider(
expected_enable_reasoner: bool,
expected_put_inner_thoughts_in_kwargs: bool,
expected_max_reasoning_tokens: int,
expected_reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high"]],
expected_reasoning_effort: Optional[Literal["none", "minimal", "low", "medium", "high", "xhigh"]],
):
model_endpoint_type, model = handle.split("/")
config = LLMConfig(

68
uv.lock generated
View File

@@ -2535,7 +2535,7 @@ requires-dist = [
{ name = "modal", marker = "extra == 'modal'", specifier = ">=1.1.0" },
{ name = "nltk", specifier = ">=3.8.1" },
{ name = "numpy", specifier = ">=2.1.0" },
{ name = "openai", specifier = ">=1.99.9" },
{ name = "openai", specifier = ">=2.11.0" },
{ name = "opentelemetry-api", specifier = "==1.30.0" },
{ name = "opentelemetry-exporter-otlp", specifier = "==1.30.0" },
{ name = "opentelemetry-instrumentation-requests", specifier = "==0.51b0" },
@@ -2648,7 +2648,7 @@ wheels = [
[[package]]
name = "llama-index"
version = "0.13.3"
version = "0.14.10"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "llama-index-cli" },
@@ -2660,28 +2660,28 @@ dependencies = [
{ name = "llama-index-readers-llama-parse" },
{ name = "nltk" },
]
sdist = { url = "https://files.pythonhosted.org/packages/ee/21/eea15746c34cdf5cbae1685a6f3ee48c3bbcd92a113f62bb33ad5526aaf6/llama_index-0.13.3.tar.gz", hash = "sha256:2a0dd99aae2f9e736b02463134e4c605200d5fc1f1f4b0fb1ca9fc52905f1f4d", size = 8030, upload-time = "2025-08-22T20:07:12.803Z" }
sdist = { url = "https://files.pythonhosted.org/packages/f9/ab/96ed543df874118cc7d01490bda0ff6bdd4f781cb511731786b2605d3f54/llama_index-0.14.10.tar.gz", hash = "sha256:49f539e4c68ddf77bb2a10f80f49bd950de3a91fe0e3ecd3b2d55a6fbc05c31e", size = 8461, upload-time = "2025-12-04T19:45:57.788Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/32/c2/c872d9dc9d906e6c0c6ff00b7cf885a755e29cb849bad9707370950702ec/llama_index-0.13.3-py3-none-any.whl", hash = "sha256:492ab1a8bac02f53cd804f15c82ccbd6118ad5a3c61dba279a87f1d480a624d3", size = 7030, upload-time = "2025-08-22T20:07:10.934Z" },
{ url = "https://files.pythonhosted.org/packages/56/50/81b8f6b9d4f7a329981ef9a6a9e9322df1a02702eebd64fd1862aa77f902/llama_index-0.14.10-py3-none-any.whl", hash = "sha256:6f03a2a8a82c059705ba54b92d4e96c8d8ba4ce70103b07825e55110452aec84", size = 7459, upload-time = "2025-12-04T19:45:56.547Z" },
]
[[package]]
name = "llama-index-cli"
version = "0.5.0"
version = "0.5.3"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "llama-index-core" },
{ name = "llama-index-embeddings-openai" },
{ name = "llama-index-llms-openai" },
]
sdist = { url = "https://files.pythonhosted.org/packages/22/5e/549765c3f1afba26a1462ffb442ffaa674dae00710780a35900e0c16b3a3/llama_index_cli-0.5.0.tar.gz", hash = "sha256:2eb9426232e8d89ffdf0fa6784ff8da09449d920d71d0fcc81d07be93cf9369f", size = 24799, upload-time = "2025-07-30T20:02:30.774Z" }
sdist = { url = "https://files.pythonhosted.org/packages/67/84/41e820efffbe327c38228d3b37fe42512a37e0c3ee4ff6bf97a394e9577a/llama_index_cli-0.5.3.tar.gz", hash = "sha256:ebaf39e785efbfa8d50d837f60cb0f95125c04bf73ed1f92092a2a5f506172f8", size = 24821, upload-time = "2025-09-29T18:03:10.798Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/b7/97/61044f804328a2e12a6609c442fb69674545592da3e89bb5144f713861ab/llama_index_cli-0.5.0-py3-none-any.whl", hash = "sha256:e331ca98005c370bfe58800fa5eed8b10061d0b9c656b84a1f5f6168733a2a7b", size = 28159, upload-time = "2025-07-30T20:02:29.992Z" },
{ url = "https://files.pythonhosted.org/packages/54/81/b7b3778aa8662913760fbbee77578daf4407aeaa677ccbf0125c4cfa2e67/llama_index_cli-0.5.3-py3-none-any.whl", hash = "sha256:7deb1e953e582bd885443881ce8bd6ab2817b594fef00079dce9993c47d990f7", size = 28173, upload-time = "2025-09-29T18:03:10.024Z" },
]
[[package]]
name = "llama-index-core"
version = "0.13.3"
version = "0.14.10"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "aiohttp" },
@@ -2712,36 +2712,36 @@ dependencies = [
{ name = "typing-inspect" },
{ name = "wrapt" },
]
sdist = { url = "https://files.pythonhosted.org/packages/28/89/2954d6150835c33bad5300e405f8087fbc0a55fd885dccf885f7263d3500/llama_index_core-0.13.3.tar.gz", hash = "sha256:bd7fffa7e6793b2c76fe55eb9e9336250b54e94fdee3fd017b354dbb0d6309a1", size = 7231354, upload-time = "2025-08-22T15:47:11.4Z" }
sdist = { url = "https://files.pythonhosted.org/packages/6b/77/a1aa39cca5b32e3801f28e877add27c9c0daf4b3dc087ac9f31df53337b5/llama_index_core-0.14.10.tar.gz", hash = "sha256:b930aac3161a5e145f4003e81ba94cbef062228a0d3ef39752e03471b0cb386b", size = 11582499, upload-time = "2025-12-04T19:45:24.093Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/c2/70/1ed5e7c7b5b95b600b1c4d967ab74b27ce377af796999dc2f9bd42d3169d/llama_index_core-0.13.3-py3-none-any.whl", hash = "sha256:0d87f817f6f306ffbe56a4bb8c85f574d5fe5142e9f0e2793c93a33eafed54c5", size = 7574514, upload-time = "2025-08-22T15:47:04.205Z" },
{ url = "https://files.pythonhosted.org/packages/a4/fa/98618331a22f8fedd050af0a0096916fdcec39bbff82bdd8ae7693536866/llama_index_core-0.14.10-py3-none-any.whl", hash = "sha256:ea0d563c777a7209664594588e3f0885a4784a1ae39dddaa28526867c79a3d65", size = 11925643, upload-time = "2025-12-04T19:45:21.585Z" },
]
[[package]]
name = "llama-index-embeddings-openai"
version = "0.5.0"
version = "0.5.1"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "llama-index-core" },
{ name = "openai" },
]
sdist = { url = "https://files.pythonhosted.org/packages/26/6a/80ed46993c6827786cdec4f6b553f3f4e5fc8741c31e8903c694833d24bf/llama_index_embeddings_openai-0.5.0.tar.gz", hash = "sha256:ac587839a111089ea8a6255f9214016d7a813b383bbbbf9207799be1100758eb", size = 7019, upload-time = "2025-07-30T19:55:05.699Z" }
sdist = { url = "https://files.pythonhosted.org/packages/10/36/90336d054a5061a3f5bc17ac2c18ef63d9d84c55c14d557de484e811ea4d/llama_index_embeddings_openai-0.5.1.tar.gz", hash = "sha256:1c89867a48b0d0daa3d2d44f5e76b394b2b2ef9935932daf921b9e77939ccda8", size = 7020, upload-time = "2025-09-08T20:17:44.681Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/01/21/65f13a385292d7d573dfde472da7daff5f779345d60c5c3e274142ec8ba2/llama_index_embeddings_openai-0.5.0-py3-none-any.whl", hash = "sha256:d817edb22e3ff475e8cd1833faf1147028986bc1d688f7894ef947558864b728", size = 7009, upload-time = "2025-07-30T19:55:04.86Z" },
{ url = "https://files.pythonhosted.org/packages/23/4a/8ab11026cf8deff8f555aa73919be0bac48332683111e5fc4290f352dc50/llama_index_embeddings_openai-0.5.1-py3-none-any.whl", hash = "sha256:a2fcda3398bbd987b5ce3f02367caee8e84a56b930fdf43cc1d059aa9fd20ca5", size = 7011, upload-time = "2025-09-08T20:17:44.015Z" },
]
[[package]]
name = "llama-index-indices-managed-llama-cloud"
version = "0.9.2"
version = "0.9.4"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "deprecated" },
{ name = "llama-cloud" },
{ name = "llama-index-core" },
]
sdist = { url = "https://files.pythonhosted.org/packages/8f/5c/bd229ed367fa26ea9ddb18972bc84ac021014bf741de454e8b0534bd16fd/llama_index_indices_managed_llama_cloud-0.9.2.tar.gz", hash = "sha256:19af55da8f1218d80390fcbe5cdfef6100acc755b0177a4077a924c31a2fb345", size = 15060, upload-time = "2025-08-17T22:02:49.885Z" }
sdist = { url = "https://files.pythonhosted.org/packages/61/4a/79044fcb3209583d1ffe0c2a7c19dddfb657a03faeb9fe0cf5a74027e646/llama_index_indices_managed_llama_cloud-0.9.4.tar.gz", hash = "sha256:b5e00752ab30564abf19c57595a2107f5697c3b03b085817b4fca84a38ebbd59", size = 15146, upload-time = "2025-09-08T20:29:58.673Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/e6/de/4ec2e278866d6d308fc3acaee223fcf580d787637e04c574a3b5fde6259e/llama_index_indices_managed_llama_cloud-0.9.2-py3-none-any.whl", hash = "sha256:d139602d1c268ae4ce2003838e857c5ef5527fb1adc78add2d9ce18fef4b4bfa", size = 16904, upload-time = "2025-08-17T22:02:48.707Z" },
{ url = "https://files.pythonhosted.org/packages/a6/6a/0e33245df06afc9766c46a1fe92687be8a09da5d0d0128bc08d84a9f5efa/llama_index_indices_managed_llama_cloud-0.9.4-py3-none-any.whl", hash = "sha256:535a08811046803ca6ab7f8e9d510e926aa5306608b02201ad3d9d21701383bc", size = 17005, upload-time = "2025-09-08T20:29:57.876Z" },
]
[[package]]
@@ -2759,20 +2759,20 @@ wheels = [
[[package]]
name = "llama-index-llms-openai"
version = "0.5.4"
version = "0.6.10"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "llama-index-core" },
{ name = "openai" },
]
sdist = { url = "https://files.pythonhosted.org/packages/20/59/4c414d79a21189d9db6de58ecbc297cd0f5ea121803b836bd134c67dd7a3/llama_index_llms_openai-0.5.4.tar.gz", hash = "sha256:9e36b6d2fc5f056b00ee655901b3bb7e7060b23f7b19439889fb78d696340f54", size = 24230, upload-time = "2025-08-16T22:41:17.408Z" }
sdist = { url = "https://files.pythonhosted.org/packages/fb/36/78642da224528d9cc49fbebdc4571b57390764a36c63ef1ca4769e5b3f4d/llama_index_llms_openai-0.6.10.tar.gz", hash = "sha256:bce2621cb650d8fed69eda2712b804e6eb34559fcaeba92c54a3b5b3b3c69161", size = 25543, upload-time = "2025-11-27T22:25:34.014Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/ad/62/aec65450b8d7ba723fa557884ac34d94b2b8f3876a54249c05d240a2be6c/llama_index_llms_openai-0.5.4-py3-none-any.whl", hash = "sha256:8d42fbfa56b5f281ad0dfcb2915916c188b5876625f9f8d27016b7dc4366cc24", size = 25357, upload-time = "2025-08-16T22:41:16.472Z" },
{ url = "https://files.pythonhosted.org/packages/78/0f/f2972ac7787ba3e0d0128e445d82d4d4b1642837411645194c767d9a362f/llama_index_llms_openai-0.6.10-py3-none-any.whl", hash = "sha256:51bdc5855c56e3b303278a3172c1bfeeef9cdeb896772bd6666ad209890ea7d8", size = 26548, upload-time = "2025-11-27T22:25:35.102Z" },
]
[[package]]
name = "llama-index-readers-file"
version = "0.5.2"
version = "0.5.5"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "beautifulsoup4" },
@@ -2782,36 +2782,36 @@ dependencies = [
{ name = "pypdf" },
{ name = "striprtf" },
]
sdist = { url = "https://files.pythonhosted.org/packages/e7/9c/d86710937b3d14c51bd02fc67f1f4babd77ac3710817fc230778020cd102/llama_index_readers_file-0.5.2.tar.gz", hash = "sha256:049d971ac4c936edbf4832915ba7128cfee8f5ead435266792b71edd87f5305c", size = 32517, upload-time = "2025-08-16T22:25:52.86Z" }
sdist = { url = "https://files.pythonhosted.org/packages/99/13/5bb486c57832efdd8dddc28fc04535f74dcc6ea07847b1ab2075a9d9c6e8/llama_index_readers_file-0.5.5.tar.gz", hash = "sha256:024b841cfdf6d37e4e24cea6c89edfac5f0eaffcbcfc4904e4b0b169847c10e5", size = 32563, upload-time = "2025-11-21T16:02:33.062Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/96/b9/8d55b916901df08b55c0e7994363fcea41c02b7fbe8ccba07e4c2ad57f9c/llama_index_readers_file-0.5.2-py3-none-any.whl", hash = "sha256:c85e20022f3da3f635d38401863243c557d4e9ad6872acaafc9281e41a0c9460", size = 51780, upload-time = "2025-08-16T22:25:51.635Z" },
{ url = "https://files.pythonhosted.org/packages/19/f6/0dc2a15f8d821721cc958873c8c83367d7fbbd411c512d0262b64c055ce3/llama_index_readers_file-0.5.5-py3-none-any.whl", hash = "sha256:948adef8d0aff839c99c95ae876c3ae57a58b319e3dd0862d68932dbd408a022", size = 51826, upload-time = "2025-11-21T16:02:31.807Z" },
]
[[package]]
name = "llama-index-readers-llama-parse"
version = "0.5.0"
version = "0.5.1"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "llama-index-core" },
{ name = "llama-parse" },
]
sdist = { url = "https://files.pythonhosted.org/packages/be/77/6ca0b2fcdf2ad6c5a36e4374b3380a1e65d216a07f874f27f341c69e2c55/llama_index_readers_llama_parse-0.5.0.tar.gz", hash = "sha256:891b21fb63fe1fe722e23cfa263a74d9a7354e5d8d7a01f2d4040a52f8d8feef", size = 3858, upload-time = "2025-07-30T21:11:30.906Z" }
sdist = { url = "https://files.pythonhosted.org/packages/b3/77/5bfaab20e6ec8428dbf2352e18be550c957602723d69383908176b5686cd/llama_index_readers_llama_parse-0.5.1.tar.gz", hash = "sha256:2b78b73faa933e30e6c69df351e4e9f36dfe2ae142e2ab3969ddd2ac48930e37", size = 3858, upload-time = "2025-09-08T20:41:29.201Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/66/df/e0f96794fcc1edb2aaef8c3d1a862940c9f27ea39f179cbc2d82b061f277/llama_index_readers_llama_parse-0.5.0-py3-none-any.whl", hash = "sha256:e63ebf2248c4a726b8a1f7b029c90383d82cdc142942b54dbf287d1f3aee6d75", size = 3201, upload-time = "2025-07-30T21:11:29.885Z" },
{ url = "https://files.pythonhosted.org/packages/68/81/52410c7245dcbf1a54756a9ce3892cdd167ec0b884d696de1304ca3f452e/llama_index_readers_llama_parse-0.5.1-py3-none-any.whl", hash = "sha256:0d41450ed29b0c49c024e206ef6c8e662b1854e77a1c5faefed3b958be54f880", size = 3203, upload-time = "2025-09-08T20:41:28.438Z" },
]
[[package]]
name = "llama-index-workflows"
version = "1.3.0"
version = "2.11.5"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "llama-index-instrumentation" },
{ name = "pydantic" },
{ name = "typing-extensions" },
]
sdist = { url = "https://files.pythonhosted.org/packages/00/dc/54fd5dec0ad3c65f3e8a520db7a3024141b71cd41660d0baca3cd6b18707/llama_index_workflows-1.3.0.tar.gz", hash = "sha256:9c1688e237efad384f16485af71c6f9456a2eb6d85bf61ff49e5717f10ff286d", size = 1040839, upload-time = "2025-08-07T09:11:00.307Z" }
sdist = { url = "https://files.pythonhosted.org/packages/d4/14/b15613ece1d7bfbe3f3b1a7f76cf9eaa603b55dd6b5921e81de2ee21f580/llama_index_workflows-2.11.5.tar.gz", hash = "sha256:defba6c5169a9c986fe06dd9f3ee5e1a6cc5f58c776605c0458baa0a55f07643", size = 70743, upload-time = "2025-11-24T18:37:59.187Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/3b/cf/0c50bc6e5c4fb7913f5682a0d26a60b976533dd8a87a5dbd84f617c6f1ab/llama_index_workflows-1.3.0-py3-none-any.whl", hash = "sha256:328cc25d92b014ef527f105a2f2088c0924fff0494e53d93decb951f14fbfe47", size = 42527, upload-time = "2025-08-07T09:10:59.155Z" },
{ url = "https://files.pythonhosted.org/packages/90/ac/e911594a2f10445717ea45b61b3a93f3bb91594320745fe1bb796c2dc87a/llama_index_workflows-2.11.5-py3-none-any.whl", hash = "sha256:3c5a419129114bb0b1bd83b88aa5f653f84181b2e39e33473e8747ec6e88538e", size = 91982, upload-time = "2025-11-24T18:37:58.265Z" },
]
[[package]]
@@ -3544,7 +3544,7 @@ wheels = [
[[package]]
name = "openai"
version = "1.102.0"
version = "2.11.0"
source = { registry = "https://pypi.org/simple" }
dependencies = [
{ name = "anyio" },
@@ -3556,9 +3556,9 @@ dependencies = [
{ name = "tqdm" },
{ name = "typing-extensions" },
]
sdist = { url = "https://files.pythonhosted.org/packages/07/55/da5598ed5c6bdd9939633854049cddc5cbac0da938dfcfcb3c6b119c16c0/openai-1.102.0.tar.gz", hash = "sha256:2e0153bcd64a6523071e90211cbfca1f2bbc5ceedd0993ba932a5869f93b7fc9", size = 519027, upload-time = "2025-08-26T20:50:29.397Z" }
sdist = { url = "https://files.pythonhosted.org/packages/f4/8c/aa6aea6072f985ace9d6515046b9088ff00c157f9654da0c7b1e129d9506/openai-2.11.0.tar.gz", hash = "sha256:b3da01d92eda31524930b6ec9d7167c535e843918d7ba8a76b1c38f1104f321e", size = 624540, upload-time = "2025-12-11T19:11:58.539Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/bd/0d/c9e7016d82c53c5b5e23e2bad36daebb8921ed44f69c0a985c6529a35106/openai-1.102.0-py3-none-any.whl", hash = "sha256:d751a7e95e222b5325306362ad02a7aa96e1fab3ed05b5888ce1c7ca63451345", size = 812015, upload-time = "2025-08-26T20:50:27.219Z" },
{ url = "https://files.pythonhosted.org/packages/e5/f1/d9251b565fce9f8daeb45611e3e0d2f7f248429e40908dcee3b6fe1b5944/openai-2.11.0-py3-none-any.whl", hash = "sha256:21189da44d2e3d027b08c7a920ba4454b8b7d6d30ae7e64d9de11dbe946d4faa", size = 1064131, upload-time = "2025-12-11T19:11:56.816Z" },
]
[[package]]
@@ -4482,11 +4482,11 @@ wheels = [
[[package]]
name = "pypdf"
version = "6.0.0"
version = "6.4.1"
source = { registry = "https://pypi.org/simple" }
sdist = { url = "https://files.pythonhosted.org/packages/20/ac/a300a03c3b34967c050677ccb16e7a4b65607ee5df9d51e8b6d713de4098/pypdf-6.0.0.tar.gz", hash = "sha256:282a99d2cc94a84a3a3159f0d9358c0af53f85b4d28d76ea38b96e9e5ac2a08d", size = 5033827, upload-time = "2025-08-11T14:22:02.352Z" }
sdist = { url = "https://files.pythonhosted.org/packages/0c/e0/57f914ae9fedbc91fe3ebe74b78c88903943ec9c232b6da15947bb3bf8ab/pypdf-6.4.1.tar.gz", hash = "sha256:36eb0b52730fc3077d2b8d4122751e696d46af9ef9e5383db492df1ab0cc4647", size = 5275322, upload-time = "2025-12-07T14:19:27.922Z" }
wheels = [
{ url = "https://files.pythonhosted.org/packages/2c/83/2cacc506eb322bb31b747bc06ccb82cc9aa03e19ee9c1245e538e49d52be/pypdf-6.0.0-py3-none-any.whl", hash = "sha256:56ea60100ce9f11fc3eec4f359da15e9aec3821b036c1f06d2b660d35683abb8", size = 310465, upload-time = "2025-08-11T14:22:00.481Z" },
{ url = "https://files.pythonhosted.org/packages/db/ef/68c0f473d8b8764b23f199450dfa035e6f2206e67e9bde5dd695bab9bdf0/pypdf-6.4.1-py3-none-any.whl", hash = "sha256:1782ee0766f0b77defc305f1eb2bafe738a2ef6313f3f3d2ee85b4542ba7e535", size = 328325, upload-time = "2025-12-07T14:19:26.286Z" },
]
[[package]]