From 115ca184ff5876c57900deec50fbdd6df8d4448f Mon Sep 17 00:00:00 2001 From: Viktor Barzin Date: Wed, 6 May 2026 22:06:32 +0000 Subject: [PATCH] openclaw: switch primary to ChatGPT Plus OAuth (openai-codex/gpt-5.4-mini) MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit Bumps image 2026.2.26 → 2026.5.4 (openai-codex provider plugin landed in 2026.4.21+). Auth profile is OAuth via the device-pairing flow against the Codex backend (account ancaelena98@gmail.com); token persists in /home/node/.openclaw/agents/main/agent/auth-state.json on NFS so it survives pod restarts. Plus tier accepts gpt-5.4-mini (1,200–7,000 local msgs/5h); gpt-5-mini and gpt-5.1-codex-mini both return errors on Plus, so we pin gpt-5.4-mini explicitly. doctor --fix auto-promotes the highest-tier model (gpt-5-pro) after model discovery, so the container command pins the mini back as default after doctor runs but before gateway start. --- stacks/openclaw/main.tf | 16 ++++++++++++---- 1 file changed, 12 insertions(+), 4 deletions(-) diff --git a/stacks/openclaw/main.tf b/stacks/openclaw/main.tf index 11d129ee..3a49315a 100644 --- a/stacks/openclaw/main.tf +++ b/stacks/openclaw/main.tf @@ -131,8 +131,12 @@ resource "kubernetes_config_map" "openclaw_config" { mode = "off" } model = { - primary = "nim/qwen/qwen3-coder-480b-a35b-instruct" - fallbacks = ["nim/mistralai/mistral-large-3-675b-instruct-2512", "modelrelay/auto-fastest"] + # ChatGPT Plus OAuth via openai-codex plugin (account: ancaelena98@gmail.com). + # gpt-5.4-mini is the only mini variant the Codex backend accepts for Plus tier; + # gpt-5-mini / gpt-5.1-codex-mini return model_not_found / "not supported with + # ChatGPT account". Plus rate-card: 1,200–7,000 local msgs / 5h on gpt-5.4-mini. + primary = "openai-codex/gpt-5.4-mini" + fallbacks = ["openai-codex/gpt-5.5", "nim/qwen/qwen3-coder-480b-a35b-instruct", "modelrelay/auto-fastest"] } models = { "modelrelay/auto-fastest" = {} @@ -146,6 +150,8 @@ resource "kubernetes_config_map" "openclaw_config" { "llama-as-openai/Llama-4-Scout-17B-16E-Instruct-FP8" = {} "openrouter/stepfun/step-3.5-flash:free" = {} "openrouter/arcee-ai/trinity-large-preview:free" = {} + "openai-codex/gpt-5.4-mini" = {} + "openai-codex/gpt-5.5" = {} } } } @@ -383,8 +389,10 @@ resource "kubernetes_deployment" "openclaw" { # Main container: OpenClaw container { name = "openclaw" - image = "ghcr.io/openclaw/openclaw:2026.2.26" - command = ["sh", "-c", "node openclaw.mjs doctor --fix 2>/dev/null; exec node openclaw.mjs gateway --allow-unconfigured --bind lan"] + image = "ghcr.io/openclaw/openclaw:2026.5.4" + # Doctor --fix auto-promotes the highest-tier codex model (gpt-5-pro) after + # auth-profile-based model discovery; pin gpt-5.4-mini back to default after it. + command = ["sh", "-c", "node openclaw.mjs doctor --fix 2>/dev/null; node openclaw.mjs models set openai-codex/gpt-5.4-mini 2>/dev/null; exec node openclaw.mjs gateway --allow-unconfigured --bind lan"] port { container_port = 18789 }