<- Back to feed
DEEP_DIVE · · 7 min · X01

GPT-5.4: Thinking Model Beats Human Experts on Pro Tasks

GPT-5.4 Thinking matched or exceeded human experts on 83% of professional tasks across 44 occupations. What the benchmarks mean and why the curve gets steeper.

#OpenAI#GPT-5.4#AI benchmarks#knowledge work#AI agents#computer use#GDPVal
Visual illustration for GPT-5.4: Thinking Model Beats Human Experts on Pro Tasks

GPT-5.4 Thinking has crossed a line that most people weren’t ready for. OpenAI’s newest flagship model, GPT-5.4, now matches or exceeds the performance of human professionals on 83% of knowledge work tasks, spanning 44 occupations, from legal analysis to financial modelling. That number, drawn from OpenAI’s internal GDPVal benchmark, is not a narrow lab result. It is a broad-spectrum claim about the model’s ability to replace human judgment on the work that built the professional-services economy.

The release dropped on March 5, 2026, eight weeks after GPT-5.2. The pace of iteration alone is worth pausing on. Three months, 12 percentage points of professional-task parity.

What GDPVal Actually Measures

GDPVal is OpenAI’s proprietary benchmark for economically valuable knowledge work. Unlike abstract reasoning tests, it simulates real job outputs (drafting legal memos, building financial models, analyzing regulatory filings, writing code, producing research summaries) and scores them against what working professionals produce.

GPT-5.4 scored 83.0% on GDPVal. Its predecessor, GPT-5.2, scored 70.9%. The jump from 70.9% to 83.0% in a single model generation is not incremental. Anthropic’s Opus 4.6, currently the strongest Claude model, sits at 79.5% on the same benchmark.

The 83% figure means GPT-5.4 tied or beat a human professional in 37 of 44 tested occupational categories. The remaining 7 categories, where humans still lead, are not disclosed. That opacity is worth noting. OpenAI benchmarks are self-reported, and the comparisons are made against GPT-5.2 rather than GPT-5.3 Instant, which launched just two days prior. The headline number is real, but the fine print matters.

Still, the direction is unambiguous.

Computer Use: The Capability That Changes Agent Architecture

Beyond GDPVal, the most consequential addition in GPT-5.4 is native computer use. This is the first general-purpose OpenAI model that can operate a desktop environment directly, navigating file systems, clicking through applications, and executing multi-step workflows from screenshots alone, without a separate specialized layer bolted on top.

On OSWorld-Verified, the standard benchmark for autonomous desktop task completion, GPT-5.4 scored 75.0%. Human expert baseline on the same benchmark is 72.4%. GPT-5.2 scored 47.3%.

That jump, from 47.3% to 75.0% in one generation, represents a qualitative shift, not just a numerical one. At 47%, a computer-use agent is unreliable enough that it requires supervision for almost every task. At 75%, it completes most tasks without intervention. The failure mode changes from “this doesn’t work” to “this works except in specific edge cases.”

For developers building AI agent pipelines, native computer use eliminates a category of integration complexity. Previously, agentic stacks had to chain together a reasoning model, a computer-use module, and glue code to handle the handoffs. GPT-5.4 collapses that into a single model call.

Token Efficiency and the New Tool Search System

OpenAI’s other significant architectural change is Tool Search, a reworked system for how the model handles tool calling in agentic contexts.

Previously, every API call in a tool-rich environment required the system prompt to enumerate all available tool definitions, a process that consumed tokens linearly as the tool count grew. In large-scale agent deployments with hundreds of available functions, this overhead was substantial.

Tool Search allows GPT-5.4 to look up tool definitions on demand, rather than loading them all upfront. The practical effect is faster, cheaper requests in production systems. For enterprise deployments running at scale, that cost reduction compounds quickly.

The context window has also expanded to 1 million tokens in API access, more than double GPT-5.2’s 400,000-token ceiling. Long-horizon tasks that previously required document chunking or retrieval workarounds can now be handled in a single context.

Hallucination Reduction: Where the Real Progress Is

Capability benchmarks get the headlines, but hallucination reduction is where the practical reliability argument lives. OpenAI reports that individual factual claims in GPT-5.4 responses are 33% less likely to be incorrect compared to GPT-5.2, and overall responses are 18% less likely to contain errors.

These figures, if accurate, matter more than benchmark scores for most production use cases. A model that scores 83% on GDPVal but hallucinates regularly is not deployable in legal or financial contexts without human review of every output. A model that hallucinates 33% less frequently moves a meaningful subset of tasks from “supervised” to “automated.”

GPT-5.4 also introduces an updated safety evaluation specifically for chain-of-thought monitoring. Reasoning models have attracted concern from AI safety researchers because a sufficiently capable model could, in principle, misrepresent its internal reasoning process, performing differently from how its chain-of-thought describes. OpenAI’s evaluation found that in the Thinking version of GPT-5.4, deceptive chain-of-thought is less likely to occur, suggesting the model lacks the capability to conceal its reasoning at this stage.

That finding has a short shelf life. The capability will eventually emerge. But it is not here yet.

Morgan Stanley’s Warning: The Curve Gets Steeper

GPT-5.4 did not arrive in a vacuum. On the same day the Fortune story ran, Morgan Stanley published a sweeping research note warning that a transformative AI breakthrough is imminent, and most institutions are not prepared.

The bank’s case rests on scaling laws. An unprecedented accumulation of compute at major U.S. AI labs, combined with the holding of Moore-adjacent scaling curves, is pushing model capability toward a threshold that Morgan Stanley describes as “Transformative AI.” The bank points to GPT-5.4’s GDPVal score as early evidence that this threshold is closer than previously modeled.

The implications the bank draws are not speculative. They are already in motion. Executives are executing large-scale workforce reductions on the basis of current AI efficiencies, not anticipated future capabilities but the models available today. Sam Altman has publicly described a future where companies of one to five people, augmented by AI, outcompete large incumbents.

The infrastructure cost of getting there is severe. Morgan Stanley’s “Intelligence Factory” model projects a net U.S. power shortfall of 9 to 18 gigawatts through 2028, a 12% to 25% deficit in the power required to support continued AI scaling. The gap is being partially bridged by converting Bitcoin mining facilities into high-performance compute centers, deploying natural gas turbines, and building out fuel cell infrastructure alongside the grid.

This is the same infrastructure race that drove xAI’s $659M Colossus expansion earlier this year. The compute arms race and the power crisis are the same story told from different angles.

What Comes After 83%: The Recursive Self-Improvement Question

The number that matters most in Morgan Stanley’s report is not the power shortfall. It is the timeline for recursive self-improvement.

xAI co-founder Jimmy Ba is cited as suggesting that AI systems capable of autonomously upgrading their own capabilities through recursive self-improvement loops, could emerge as early as the first half of 2027. That is 15 months away.

Whether that timeline holds is unknown. The history of AI timelines is a history of both underestimation and overestimation. But the GPT-5.4 results make the lower bound of that range more credible than it was a year ago. A model that surpasses human performance on 75% of desktop tasks and 83% of professional knowledge tasks is not far from a model capable of meaningful contributions to its own training and architecture.

The evaluation scaffolding, the ability to verify whether an AI-generated model improvement is actually better, remains the harder problem. Recursive self-improvement is not just about a model generating a better version of itself. It is about a model being able to correctly assess that the new version is better. That verification problem is unsolved.

The Deployment Picture: Who Gains Access to GPT-5.4

GPT-5.4 Thinking is available today to ChatGPT Plus, Team, and Pro subscribers, replacing GPT-5.2 Thinking. GPT-5.4 Pro, the highest-performance configuration, is limited to the $200-per-month ChatGPT Pro and Enterprise tiers.

API access includes the full 1-million-token context window and Tool Search. Pricing has not been separately disclosed for GPT-5.4, but OpenAI emphasized improved token efficiency, suggesting the cost-per-task ratio should improve despite the capability jump.

The three-tier release structure (standard, Thinking, Pro) continues OpenAI’s strategy of segmenting model access by use case and willingness to pay. Thinking is the reasoning variant with extended chain-of-thought. Pro is optimized for sustained high-demand workloads, the kind of long-horizon deliverables (financial models, legal briefs, consulting analyses) where GPT-5.4 scored highest on Mercor’s APEX-Agents benchmark.

The Occupational Calculus

The 83% GDPVal figure will be widely cited and widely misunderstood. It does not mean 83% of all jobs can be automated. It means that on the specific knowledge work tasks tested across 44 occupations, GPT-5.4 produced output that was judged equal to or better than what a human professional produced.

The remaining 17% represents tasks where human professionals still have the edge. The benchmark does not capture tacit knowledge, relationship dynamics, regulatory accountability, or the kinds of judgment that involve navigating ambiguity with incomplete information and institutional consequences.

But the trend line is clear. GPT-5.2 was at 70.9%. GPT-5.4 is at 83.0%. The next model, already teased by OpenAI on the day GPT-5.4 launched, will be higher. The question is not whether AI will close the gap. The question is how quickly, and what institutions do in the time between now and when the gap closes entirely.

That time is shorter than it was eight weeks ago.