.jpg)
On March 5, 2026, OpenAI quietly ended an era.
GPT-5.3 Codex was the last time OpenAI shipped a model whose entire identity was specialist. One job: build, run, automate. It hit 64% on the OSWorld-Verified computer-use benchmark — just short of the 72.4% human baseline — and introduced midtask steering, the first real crack in the prompt-wait-restart loop that had defined AI-assisted development since 2023. Claude Code hit $2.5 billion in annualized revenue in under a year. GPT-5.3 Codex was OpenAI's answer. It was fast, focused, and genuinely useful.
Then GPT-5.4 arrived and scored 75%.
Human baseline: cleared. Specialized coding model: absorbed. The two paths — reasoning depth and coding execution — merged into one.
GPT-5.4 wins on every benchmark where a comparison exists. The harder question is whether the migration justifies the disruption for teams already running on GPT-5.3 Codex workflows. In my analysis, the answer is yes — but the when depends on one specific variable.
What GPT-5.3 Codex Actually Got Right
GPT-5.3 Codex solved three problems that no previous model handled cleanly: redirecting an active build mid-execution, storing repeatable automation routines as portable files, and handling long-running projects measured in hours rather than minutes. Those weren't incremental improvements — they changed how developers structured their workdays.
When we tested GPT-5.3 Codex on a full habit-tracker build at Turing College — calendar integration, daily tracking, color-coded completion states — it finished the app in 3 minutes and 8 seconds after a mid-build redirect. No restart. No re-prompt. We changed the spec while it was running, the new instruction injected into the model's context, and the output matched the updated brief. That single capability — midtask steering — made 5.3 Codex feel less like a tool and more like a colleague you could interrupt.
The Skills feature compounded that. You described a workflow once — close out a Jira release, move tickets to "released," remove the Quick Filter — and the model wrote a skills.md file. Run the skill next week, it reads the file, builds the plan, asks for confirmation, executes. A task that took 10 minutes of context-switching collapsed to under 2. The file was editable: if the model made a mistake, you corrected the skills.md and the behavior updated on the next run.
GPT-5.3 Codex also cleared 64% on OSWorld-Verified. The previous Codex had scored 38%. That 26-point jump — in a single release — showed the computer-use trajectory was real, not a marketing narrative.
That's the baseline GPT-5.4 had to beat.
What GPT-5.4 Actually Changed
GPT-5.4 didn't improve on GPT-5.3 Codex — it replaced the need for a specialist model entirely. It absorbed 5.3 Codex's coding capabilities into the mainline, added a 1-million-token context window, pushed computer use past human baseline, and won the GDPval knowledge-work benchmark at 83%. GPT-5.3 Codex never competed on knowledge work. GPT-5.4 competes on everything.
GPT-5.4 is OpenAI's first mainline reasoning model to incorporate the frontier coding capabilities of GPT-5.3 Codex, rolling out across ChatGPT, the API, and Codex simultaneously, with support for up to 1 million tokens of context.
That last row in the table — availability — is underrated. GPT-5.3 Codex lived inside the Codex interface. GPT-5.4 runs everywhere. The same model that builds your app also writes the quarterly analysis, handles the legal brief, and manages the financial model. GPT-5.3 Codex couldn't do that. It wasn't built to.
The Midtask Steering Shift: Better or Just Different?
GPT-5.3 Codex let you inject new instructions during execution. GPT-5.4 shows you an upfront reasoning plan before heavy work begins, so you correct course before execution rather than during it. The failure mode each approach prevents is different — and which one matters more depends on your workflow.
GPT-5.3 Codex's model: steer a moving car. The model starts building, you interrupt it with a new instruction, the instruction becomes part of the active context, the build continues. The risk it prevents is the wasted 3-minute run you can't stop.
GPT-5.4 Thinking displays an upfront plan of its reasoning so you can adjust course mid-response while it's working, arriving at a final output more closely aligned with what you need — with less back and forth.
GPT-5.4's model: correct the map before the driver leaves. For a habit tracker, the 5.3 approach is faster — you're 60 seconds into a build, you think of a calendar feature, you queue it and it absorbs the change. For a 40-slide financial deck or a multi-file legal analysis, the 5.4 approach saves more: one wrong assumption at the planning stage costs far more than 60 seconds to fix.
Neither is a regression. The use-case profile just shifted.
Skills and Tool Search: The Automation Layer Grew
GPT-5.3 Codex stored automation routines in skills.md files — reusable, editable, model-agnostic prompts. GPT-5.4 retains that capability and adds Tool Search: automatic connector discovery that eliminates the need to manually list every external tool in your API requests, cutting prompt overhead at scale.
The skills.md pattern from 5.3 Codex still works. Every automation routine you built — every Jira release workflow, every deployment checklist, every repetitive task you encoded as a skill — runs on 5.4. The migration doesn't break your library.
GPT-5.4's new Tool Search feature enables the model to automatically find the tools an application requires without a manually prepared list in every API request, reducing prompt sizes and lowering inference costs.
For teams running large-scale agent orchestration, that matters immediately. If your current API calls ship a 4,000-token tool manifest with every request, Tool Search removes that overhead. Fewer tokens in, same capability out. At $2.50 per million input tokens, that's not dramatic per call — but across a month of production volume, it moves.
The Token Efficiency Argument for Migrating Now
GPT-5.4's output token rate is $15.00 per million — higher than older models on paper. But it reaches the same conclusions with significantly fewer output tokens. For agentic pipelines where output volume drives the bill, the effective cost per task may be lower on 5.4 than on 5.3 Codex, not higher.
This is the most underreported part of the release. OpenAI states that GPT-5.4 uses significantly fewer tokens than GPT-5.2 to complete equivalent tasks, which directly reduces inference computing costs.
A model that hits the correct answer in 800 output tokens instead of 1,200 costs 33% less per run at the same rate. On a pipeline running 10,000 agentic completions a month, that gap is a budget line item.
GPT-5.4's API pricing lists input tokens at $2.50 per million, cached input at $0.25 per million, and output tokens at $15.00 per million. Prompts exceeding 272,000 input tokens are priced at 2× input and 1.5× output for the full session, so teams running near-megacontext sessions should implement chunking, retrieval, and caching to avoid unexpected cost spikes.
If your agents regularly breach 272K tokens per session, benchmark that before you migrate.
The One Reason to Stay on GPT-5.3 Codex (For Now)
If your team has a working Skills library embedded in the Codex interface and your daily build-and-redirect loop runs without friction, the case for immediate migration is weak. GPT-5.3 Codex still functions. But the window is closing — GPT-5.2 Thinking retires June 5, 2026, and the consolidation direction is clear.
GPT-5.4 is the better model across every dimension that matters for professional work. The argument for staying on 5.3 Codex is purely operational. Migrations cost time. Retesting automations costs engineering hours. If your current setup produces reliable output and the team has no capacity to absorb disruption this quarter, waiting until April or May 2026 is defensible.
Past June 5, 2026, it is not.
How to Apply This Before Next Monday
- Test your top three Skills on GPT-5.4. The
skills.mdfiles transfer. Run your highest-frequency automations on 5.4 first and compare output quality. Most will pass without modification. - Run one complex knowledge-work prompt through GPT-5.4 Thinking. Review the reasoning plan before it executes. This is the intervention point that doesn't exist in 5.3 Codex — use it. Correct the plan once, not the output three times.
- Pull last month's output token count from your API dashboard. Estimate a 20–30% efficiency gain on GPT-5.4 and see what that does to your infrastructure budget. If the number moves your quarter, that's your migration business case.
- Map any sessions that regularly exceed 272K input tokens. Those need chunking or retrieval scaffolds before you migrate, or your cost structure changes in ways the rate card doesn't predict.
FAQ
Does GPT-5.4 actually replace GPT-5.3 Codex?
Functionally, yes. GPT-5.4 is the first general-purpose model with native, state-of-the-art computer-use capabilities, incorporating the frontier coding capabilities of GPT-5.3 Codex and running across ChatGPT, the API, and Codex. The Codex interface still exists; GPT-5.4 now powers it. GPT-5.3 Codex as a distinct model identity is effectively retired.
Is the 75% OSWorld score actually meaningful outside benchmarks?
The human baseline on OSWorld-Verified is 72.4%. Crossing it means GPT-5.4 completes desktop-control tasks — navigating UIs, managing files, running commands — more reliably than a human tester under the same conditions. Developers can also upload images containing more than 10 million pixels without compression, which strengthens the model's accuracy on vision-heavy computer-use tasks. For engineering teams, this is the threshold where desktop automation becomes a staffing conversation, not just a productivity experiment.
Will my GPT-5.3 Codex Skills library break on GPT-5.4?
No. The skills.md format is model-agnostic — it stores structured prompts, not model-specific instructions. GPT-5.4 reads and executes the same files. The more relevant question is whether Tool Search makes some of your manual connector configurations redundant. Test your highest-volume skills first; retire the manual tool lists where Tool Search covers the same ground.
Should I migrate if my API budget is fixed?
Run the token math before deciding. OpenAI says GPT-5.4 uses significantly fewer tokens than GPT-5.2 to complete equivalent tasks. On output-heavy agentic pipelines, a 20–25% reduction in tokens per task at $15.00/M can undercut the effective cost of an older model at a lower nominal rate. Pull three representative tasks, count the tokens on both models, and let the numbers decide — not the rate card headline.