feat: set default temperature to 1.0 [LET-6920] (#8618)

* temp 1

* stage

* update core tests
This commit is contained in:
Kevin Lin
2026-01-16 14:58:21 -08:00
committed by Sarah Wooders
parent 8872c2b5d3
commit e5ed8ca0e8
27 changed files with 28 additions and 28 deletions

View File

@@ -6,7 +6,7 @@
"context_window": 256000,
"put_inner_thoughts_in_kwargs": true,
"handle": "openai/gpt-4.1",
"temperature": 0.7,
"temperature": 1.0,
"max_tokens": 8192,
"enable_reasoner": false,
"max_reasoning_tokens": 0

View File

@@ -2,7 +2,7 @@
"handle": "azure/gpt-4o-mini",
"model_settings": {
"provider_type": "azure",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -2,7 +2,7 @@
"handle": "bedrock/arn:aws:bedrock:us-east-1:474668403324:inference-profile/us.anthropic.claude-sonnet-4-20250514-v1:0",
"model_settings": {
"provider_type": "bedrock",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -2,7 +2,7 @@
"handle": "google_vertex/gemini-2.5-flash",
"model_settings": {
"provider_type": "google_vertex",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 65536,
"parallel_tool_calls": false,
"thinking_config": {

View File

@@ -2,7 +2,7 @@
"handle": "google_vertex/gemini-2.5-pro",
"model_settings": {
"provider_type": "google_vertex",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 65536,
"parallel_tool_calls": false,
"thinking_config": {

View File

@@ -2,7 +2,7 @@
"handle": "google_ai/gemini-2.5-pro",
"model_settings": {
"provider_type": "google_ai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 65536,
"parallel_tool_calls": false,
"thinking_config": {

View File

@@ -2,7 +2,7 @@
"handle": "groq/qwen/qwen3-32b",
"model_settings": {
"provider_type": "groq",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -2,7 +2,7 @@
"handle": "ollama/qwen2.5:7b",
"model_settings": {
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -2,7 +2,7 @@
"handle": "openai/gpt-4.1-2025-04-14",
"model_settings": {
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -2,7 +2,7 @@
"handle": "openai/gpt-4o-mini",
"model_settings": {
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -2,7 +2,7 @@
"handle": "openai/o1",
"model_settings": {
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false,
"reasoning": {

View File

@@ -2,7 +2,7 @@
"handle": "openai/o3",
"model_settings": {
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false,
"reasoning": {

View File

@@ -2,7 +2,7 @@
"handle": "openai/o4-mini",
"model_settings": {
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false,
"reasoning": {

View File

@@ -2,7 +2,7 @@
"handle": "together/Qwen/Qwen2.5-72B-Instruct-Turbo",
"model_settings": {
"provider_type": "together",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -2,7 +2,7 @@
"handle": "zai/glm-4.6",
"model_settings": {
"provider_type": "zai",
"temperature": 0.7,
"temperature": 1.0,
"max_output_tokens": 4096,
"parallel_tool_calls": false
}

View File

@@ -14,7 +14,7 @@ AGENTS_CREATE_PARAMS = [
"parallel_tool_calls": False,
"strict": False,
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"reasoning": {"reasoning_effort": "minimal"},
"response_format": None,
}
@@ -34,7 +34,7 @@ AGENTS_UPDATE_PARAMS = [
"parallel_tool_calls": False,
"strict": False,
"provider_type": "openai",
"temperature": 0.7,
"temperature": 1.0,
"reasoning": {"reasoning_effort": "minimal"},
"response_format": None,
}

File diff suppressed because one or more lines are too long

View File

@@ -81,7 +81,7 @@
"context_window": 30000,
"put_inner_thoughts_in_kwargs": false,
"handle": "anthropic/claude-sonnet-4-20250514",
"temperature": 0.7,
"temperature": 1.0,
"max_tokens": 8192,
"enable_reasoner": true,
"reasoning_effort": null,

File diff suppressed because one or more lines are too long

View File

@@ -47,7 +47,7 @@
"context_window": 32000,
"put_inner_thoughts_in_kwargs": false,
"handle": "anthropic/claude-3.5-sonnet",
"temperature": 0.7,
"temperature": 1.0,
"max_tokens": 8192,
"enable_reasoner": false,
"reasoning_effort": null,

View File

@@ -25,7 +25,7 @@
"context_window": 30000,
"put_inner_thoughts_in_kwargs": true,
"handle": "anthropic/claude-sonnet-4-20250514",
"temperature": 0.7,
"temperature": 1.0,
"max_tokens": 8192,
"enable_reasoner": false,
"reasoning_effort": null,

File diff suppressed because one or more lines are too long

File diff suppressed because one or more lines are too long

View File

@@ -115,7 +115,7 @@
"context_window": 180000,
"put_inner_thoughts_in_kwargs": false,
"handle": "anthropic/claude-sonnet-4-5-20250929",
"temperature": 0.7,
"temperature": 1.0,
"max_tokens": 8192,
"enable_reasoner": true,
"reasoning_effort": null,

View File

@@ -59,7 +59,7 @@
"context_window": 32000,
"put_inner_thoughts_in_kwargs": true,
"handle": "openai/gpt-4o-mini",
"temperature": 0.7,
"temperature": 1.0,
"max_tokens": 4096,
"enable_reasoner": true,
"reasoning_effort": null,

View File

@@ -58,7 +58,7 @@
"context_window": 32000,
"put_inner_thoughts_in_kwargs": true,
"handle": "openai/gpt-4.1-mini",
"temperature": 0.7,
"temperature": 1.0,
"max_tokens": null,
"enable_reasoner": true,
"reasoning_effort": null,