Industry

Your Automation Is Broken and You're Paying $28,500 Per Employee to Pretend It Isn't: The Computer Use Agent Reckoning

James Liu||7 min
+Space

Manual data entry alone costs U.S. companies $28,500 per employee every single year. Not in lost opportunity. Not in soft costs. In cold, hard, measurable dollars, gone. And that stat dropped in July 2025, while most companies were still debating whether to 'pilot' an automation tool. Here's what's actually happening in AI desktop automation right now: the old guard is crumbling, the new computer use agents are pulling ahead fast, and the gap between companies that get it and companies that don't is about to become a canyon.

RPA Promised the World. It Delivered a Fragile Mess.

Let's be honest about RPA. The pitch was seductive: record your mouse clicks, replay them forever, fire the data entry team. UiPath alone became a multi-billion dollar company selling that dream. But here's what nobody put in the brochure. RPA bots are brittle. Change a button label, update a UI, move a field two pixels to the left, and your entire automation breaks. IT tickets pile up. Developers spend half their time maintaining bots instead of building new ones. A 2024 analysis from ActiveBatch put it plainly: RPA consistently fails to meet IT expectations because it was never designed for a world where software changes constantly. And software always changes constantly. The companies that went all-in on traditional RPA in 2019 and 2020 are now sitting on a maintenance nightmare they can't easily escape. They didn't automate their work. They just made their work more complicated.

Now Gartner Says 40% of Agentic AI Projects Will Get Canceled Too. Here's Why They're Right, and Also Missing the Point.

In June 2025, Gartner dropped a prediction that made a lot of AI vendors very uncomfortable: over 40% of agentic AI projects will be canceled by the end of 2027. The AI Twitter crowd lost its mind. But Gartner isn't wrong, they're just describing the wrong projects. The ones that will fail are the ones being built the same way RPA was built: bolted onto broken processes, poorly scoped, handed to vendors who oversell and underdeliver. A Medium analysis from December 2025 backed this up with S&P Global data showing that 42% of companies in 2024 to 2025 cited data quality and integration as the top root cause of AI project failures. That's not an AI problem. That's a strategy problem. The companies treating computer use agents like a magic button are going to get burned. The ones treating it like infrastructure, something that actually controls real desktops and executes real workflows, are going to eat everyone else's lunch.

The Numbers That Should Make Every Manager Uncomfortable

  • $28,500: the annual cost per employee of manual data entry alone, per Parseur's July 2025 report
  • 56% of employees report burnout specifically from repetitive data tasks, which means you're not just losing money, you're losing people
  • 15 hours per worker per week lost to administrative and manual tasks, according to 2025 productivity research
  • 40%+ of agentic AI projects predicted to fail by 2027, mostly due to bad implementation, not bad technology
  • 42% of companies say data quality issues are the #1 reason their AI projects collapse mid-flight
  • Only 19% of organizations had made significant investments in agentic AI as of early 2025, meaning most companies are still watching from the sidelines while early movers build the moat

Your employees are spending 15 hours a week on tasks that a computer use agent could handle. That's not a productivity gap. That's 37% of your payroll lighting itself on fire every Monday morning.

OpenAI Operator and Anthropic Computer Use: Impressive Research Projects, Frustrating Real-World Tools

OpenAI launched Operator in January 2025 as a 'research preview.' Anthropic has been demoing Claude's computer use capabilities for months. Both are genuinely impressive in controlled demos. Both fall apart in the real world in ways that matter. A July 2025 deep-dive by Timothy B. Lee at Understanding AI tried to use both Operator and Anthropic's computer-use agent to order groceries. Simple task. Everyday use case. Both struggled. The broader critique from the tech press has been consistent: these tools are research-grade, not production-grade. They're built to impress at conferences, not to run your accounts payable workflow at 2am without supervision. And critically, neither OpenAI nor Anthropic has published OSWorld benchmark numbers that come close to what a purpose-built computer use agent can achieve. OSWorld is the gold standard for measuring whether an AI can actually operate a real computer across real tasks. Most of the big names are clustered in the 40% to 60% accuracy range on OSWorld. That sounds okay until you realize that 60% accuracy on a 50-step workflow means your automation fails roughly every other time. That's not automation. That's a coin flip with extra steps.

What a Real Computer Use Agent Actually Does (And Why the Difference Matters)

Here's the thing that gets lost in all the benchmark noise and vendor marketing. A real computer use agent doesn't call APIs. It doesn't need integrations or webhooks or a developer to map every field. It sees your screen the same way a human does, it moves the mouse, it types, it clicks, it reads what comes back, and it adjusts. That's what makes computer use AI fundamentally different from everything that came before it. RPA recorded actions and replayed them blindly. Computer use agents understand context. If the UI changes, they adapt. If an error pops up, they handle it. If a CAPTCHA appears, they deal with it. This is why the OSWorld benchmark matters so much. It tests agents on real, open-ended computer tasks across browsers, terminals, desktop apps, and file systems. No guardrails. No pre-scripted paths. Just: here's a computer, get this done. Most agents crack under that pressure. The ones that don't are the ones worth paying attention to.

Why Coasty Exists

I've watched a lot of automation tools come and go. Most of them solve the demo problem really well and the real-world problem barely at all. Coasty is different, and I don't say that lightly. It scores 82% on OSWorld. That's not a rounding error above the competition. That's a different category of performance. When every other major computer use agent is fighting for position in the 50% to 65% range, 82% means Coasty is completing tasks that its competitors simply fail. It controls real desktops, real browsers, and real terminals. Not simulated environments. Not API wrappers pretending to be automation. Actual computer use. The desktop app runs locally. Cloud VMs are available if you need scale. Agent swarms let you run tasks in parallel so you're not waiting on sequential execution like it's 2018. There's a free tier if you want to see what it actually does before spending a dollar. BYOK is supported if you have model preferences. It's the tool built by people who took the benchmark seriously because they knew the benchmark was the honest measure of whether this thing works. At coasty.ai, you can try it without a sales call, without a demo request form, without a 14-slide deck from an enterprise rep. That alone should tell you something about who built it.

Here's my actual opinion after watching this space for years: we are at the moment where the gap between 'we're exploring automation' and 'we've automated' starts compounding in ways that can't be reversed. The companies still running manual processes in 2026 won't just be slower. They'll be structurally unable to compete with teams that have computer use agents handling the work that used to require three people and a spreadsheet. RPA had its moment and left a trail of broken bots and disappointed IT teams. The first wave of AI agents oversold and underdelivered. But the computer use agent category, when it's done right, is the real answer. Not because it's trendy. Because 82% accuracy on a standardized benchmark of real-world computer tasks is just... better than what humans do when they're bored and burned out at 4pm on a Friday. Stop piloting. Stop exploring. Stop waiting for the technology to mature. It matured. Go to coasty.ai and see what your workflows look like when the computer actually does the work.

Want to see this in action?

View Case Studies
Try Coasty Free