Intelligence is foundation
Podcast Subscribe
Builders & Makers Sunday, 8 March 2026

GPT-5.4 Ships Computer Use - But Claude Already Won This Race

Share: LinkedIn
GPT-5.4 Ships Computer Use - But Claude Already Won This Race

GPT-5.4 just shipped native computer use. On paper, it's impressive: 75% performance on the OSWorld benchmark, beating the 72.4% human baseline. Automatic tool search. 1 million token context window. OpenAI is calling it a major step toward agentic AI that can operate software on your behalf.

But here's the thing: Claude shipped production-ready computer use in late 2024, and developers have been building on it for months. This isn't OpenAI innovating. It's OpenAI catching up.

What Computer Use Actually Means

Computer use - sometimes called 'agentic interaction' - is the ability for an AI model to control software directly. Not by generating code for you to run. Not by offering suggestions you implement manually. But by actually clicking buttons, navigating interfaces, filling forms, and executing tasks inside applications.

Think of it as the AI equivalent of screen sharing. You give it a task - 'find the three cheapest flights to Berlin next week and add them to a spreadsheet' - and it opens a browser, searches flight sites, compares prices, and populates your sheet. You don't write the script. You don't supervise each step. It just does it.

The OSWorld benchmark tests this capability across realistic desktop tasks. GPT-5.4's 75% score means it successfully completes three-quarters of those tasks without human intervention. That's better than the 72.4% success rate of actual humans doing the same tasks - though Note that benchmark tasks are often simplified versions of real-world complexity.

Why Claude Still Has the Edge

GPT-5.4's computer use is technically impressive, but it's not production-hardened yet. Claude's implementation, by contrast, has been live in real workflows since late 2024. Developers have spent months stress-testing it, finding edge cases, building guardrails, and figuring out what breaks.

That operational maturity matters more than benchmark scores. A model that works 75% of the time in controlled tests but fails unpredictably in production is worse than a model that works 70% of the time but fails consistently and recoverably. Claude's computer use has those failure modes mapped. GPT-5.4's doesn't yet.

There's also the integration question. Claude's computer use ships with developer tools, API access, and documentation refined by months of real-world use. OpenAI is launching with strong benchmark results but less clarity on how to actually deploy this in production systems. That gap will close quickly - OpenAI moves fast - but right now, it exists.

The Automatic Tool Search Problem

One feature OpenAI is highlighting is automatic tool search - the model's ability to identify and use software tools it hasn't been explicitly trained on. In theory, this means GPT-5.4 can adapt to new applications without retraining.

In practice, this is harder than it sounds. Software interfaces change constantly. A tool search that works today might break tomorrow when an app updates its UI. And if the model is autonomously choosing tools, how do you audit what it's accessing? How do you enforce permissions? How do you prevent it from using tools in ways you didn't intend?

These aren't unsolvable problems, but they're not solved yet. Claude's approach has been more conservative - explicit tool integration with developer-defined boundaries. It's less autonomous but more predictable. For production systems, predictability often wins.

The 1M Token Context Window

GPT-5.4's 1 million token context window is genuinely useful. It means the model can hold an entire codebase, a full day's worth of emails, or a long document in working memory while executing tasks. That's a meaningful expansion of what's possible.

But context windows are only useful if the model can reason effectively across them. A million tokens of context doesn't help if the model loses track of earlier instructions or hallucinates details halfway through. Early testing will reveal whether GPT-5.4 maintains coherence at that scale.

Claude's context windows are smaller but have been optimised for real-world workflows. Developers know how far they can push it before performance degrades. With GPT-5.4, that learning curve starts now.

What This Means for Builders

If you're building on computer use capabilities, the arrival of GPT-5.4 is good news. Competition accelerates development. OpenAI entering this space validates that computer use is a real, valuable capability worth investing in.

But don't assume GPT-5.4 is automatically the better choice. Claude's production maturity, refined developer tools, and battle-tested failure modes make it the safer bet for anything customer-facing. GPT-5.4 might be more powerful on benchmarks, but power without reliability is a liability.

The smart move is probably to build with model-agnostic architecture. Design your systems so you can swap between Claude and GPT-5.4 depending on the task. Use Claude for workflows where predictability matters. Use GPT-5.4 for tasks where you need that extra context window or experimental features.

Now What?

Computer use is now a commodity capability. Both leading models offer it. That shifts the competitive landscape from 'who has this feature?' to 'who implements it better?' And that's where operational maturity, developer experience, and real-world reliability matter more than benchmark scores.

For OpenAI, this is catch-up. For Claude, it's validation. For builders, it's a reminder that the first to ship often isn't the first to production-ready. Benchmark performance gets headlines. But production performance ships products.

The race isn't over. It's just moved to a different track.

More Featured Insights

Robotics & Automation
Agility Drops 'Robotics' From Its Name - And That Tells You Everything
Voices & Thought Leaders
Azeem Azhar on AI's Fragile Supply Chain and the 2.8% Productivity Puzzle

Today's Sources

DEV.to AI
GPT-5.4 Just Made Computer Use a Commodity. Now What?
DEV.to AI
.NET + AI = The Perfect Combo
Towards Data Science
Understanding Context and Contextual Retrieval in RAG
Towards Data Science
The AI Bubble Has a Data Science Escape Hatch
The Robot Report
Humanoid developer Agility Robotics rebrands
Azeem Azhar
Exponential View #564: Intelligence as a target; the future of knowledge; AI, productivity & economy
Gary Marcus
BREAKING: Sam Altman's greed and dishonesty are finally catching up to him

About the Curator

Richard Bland
Richard Bland
Founder, Marbl Codes

27+ years in software development, curating the tech news that matters.

Subscribe RSS Feed
View Full Digest Today's Intelligence
Free Daily Briefing

Start Every Morning Smarter

Luma curates the most important AI, quantum, and tech developments into a 5-minute morning briefing. Free, daily, no spam.

  • 8:00 AM Morning digest ready to listen
  • 1:00 PM Afternoon edition catches what you missed
  • 8:00 PM Daily roundup lands in your inbox

We respect your inbox. Unsubscribe anytime. Privacy Policy

© 2026 MEM Digital Ltd t/a Marbl Codes
About Sources Podcast Audio Privacy Cookies Terms Thou Art That
RSS Feed