chore: change codex default reasoning effort from high to xhigh

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
delta-cloud-208e
2026-03-13 22:39:55 +00:00
parent d602c99b20
commit 5525f6009a
6 changed files with 7 additions and 7 deletions

View File

@@ -141,7 +141,7 @@ sudo python3 codex_patcher.py --apply
```toml ```toml
model = "gpt-5.2-codex" model = "gpt-5.2-codex"
model_reasoning_effort = "high" model_reasoning_effort = "xhigh"
model_provider = "custom" model_provider = "custom"
approval_policy = "never" approval_policy = "never"
sandbox_mode = "danger-full-access" sandbox_mode = "danger-full-access"

View File

@@ -8,7 +8,7 @@
"gpt-5.3-codex", "gpt-5.3-codex",
"gpt-5.2-codex" "gpt-5.2-codex"
], ],
"model_reasoning_effort": "high", "model_reasoning_effort": "xhigh",
"approval_policy": "never", "approval_policy": "never",
"sandbox_mode": "danger-full-access", "sandbox_mode": "danger-full-access",
"wire_api": "responses", "wire_api": "responses",

View File

@@ -9,7 +9,7 @@
"gpt-5.3", "gpt-5.3",
"gpt-5.2-codex" "gpt-5.2-codex"
], ],
"model_reasoning_effort": "high", "model_reasoning_effort": "xhigh",
"approval_policy": "never", "approval_policy": "never",
"sandbox_mode": "danger-full-access", "sandbox_mode": "danger-full-access",
"wire_api": "responses", "wire_api": "responses",

View File

@@ -255,7 +255,7 @@ def generate_config_toml(existing, config, home_dir=None):
# Top-level managed keys # Top-level managed keys
lines.append(f'model = "{config["model"]}"') lines.append(f'model = "{config["model"]}"')
lines.append(f'model_reasoning_effort = "{config.get("model_reasoning_effort", "high")}"') lines.append(f'model_reasoning_effort = "{config.get("model_reasoning_effort", "xhigh")}"')
lines.append('model_provider = "custom"') lines.append('model_provider = "custom"')
# Model catalog path (for model picker) # Model catalog path (for model picker)
@@ -464,7 +464,7 @@ def patch_model_config(codex_dir, config):
existing = read_toml(config_path) existing = read_toml(config_path)
model = config["model"] model = config["model"]
effort = config.get("model_reasoning_effort", "high") effort = config.get("model_reasoning_effort", "xhigh")
if (existing.get("model") == model and if (existing.get("model") == model and
existing.get("model_reasoning_effort") == effort and existing.get("model_reasoning_effort") == effort and

View File

@@ -286,7 +286,7 @@ if (-not $pyCmd) {
$tomlContent = @" $tomlContent = @"
model = "gpt-5.4" model = "gpt-5.4"
model_reasoning_effort = "high" model_reasoning_effort = "xhigh"
model_provider = "custom" model_provider = "custom"
model_catalog_json = "$catalogPath" model_catalog_json = "$catalogPath"
approval_policy = "never" approval_policy = "never"

View File

@@ -222,7 +222,7 @@ if (-not $pyCmd) {
$tomlContent = @" $tomlContent = @"
model = "gpt-5.4" model = "gpt-5.4"
model_reasoning_effort = "high" model_reasoning_effort = "xhigh"
model_provider = "custom" model_provider = "custom"
model_catalog_json = "$catalogPath" model_catalog_json = "$catalogPath"
approval_policy = "never" approval_policy = "never"