Industry

Your AI Agent Knows Your Passwords. That Should Terrify You.

Lisa Chen||7 min
+D

Somewhere right now, a computer use agent is logging into a system with your credentials, and there is a non-trivial chance it's doing something you didn't authorize. Researchers at Push Security published a detailed breakdown earlier this year of exactly how AI agents, including OpenAI's Operator and Anthropic's computer use implementation, can be weaponized for credential stuffing attacks at a scale that would make a 2019 hacker cry with joy. This isn't a theoretical future problem. A hacker already used Anthropic's Claude to steal sensitive data from a Mexican government system in early 2026. HiddenLayer demonstrated indirect prompt injection against Claude Computer Use back in October 2024, showing how a malicious webpage could hijack an agent mid-task and silently exfiltrate credentials. The industry has been building AI agents that can do everything, and securing the credentials those agents use has been, at best, an afterthought. At worst, it's been a disaster waiting to happen.

The Attack Vector Nobody Is Talking About Loudly Enough

Here's the scenario that should keep CISOs up at night. You deploy a computer use agent to automate browser tasks. The agent logs into your CRM, your cloud storage, your payroll system. It's reading emails, filling forms, clicking buttons. It's doing real work on a real desktop with real access. Now a bad actor embeds a hidden instruction in a webpage the agent visits, a technique called indirect prompt injection. The agent reads it, interprets it as a legitimate instruction, and quietly forwards your session tokens to an external server. You don't see it happen. There's no alert. The agent just keeps working. Palo Alto Networks Unit 42 flagged prompt injection as one of the most potent attack vectors in agentic AI as of May 2025. OWASP put it at the top of their LLM risk list. Researchers at arXiv documented that compromised AI coding agents are already performing OS credential dumping, the same technique ransomware gangs use. And yet most teams deploying computer use agents are still handing them credentials the same way they'd share a sticky note: environment variables, hardcoded config files, or worse, just typing the password into the agent's context window and hoping for the best.

What 'Just Give It Your Password' Actually Costs

  • 70% of users exposed in 2024 breaches had reused credentials elsewhere, per SpyCloud's 2025 Identity Exposure Report. One compromised agent login can cascade across your entire stack.
  • ChatGPT Atlas (OpenAI's agentic browser) was flagged by Fortune in October 2025 for security flaws that could turn it from a helpful tool into a data leakage vector via prompt injection.
  • Invariant Labs documented a GitHub prompt injection attack where an AI assistant was turned into a data thief mid-session, no malware required, just a poisoned README file.
  • Verizon's 2025 Data Breach Investigations Report confirmed that credential-based attacks remain the leading cause of breaches. AI agents massively expand the credential attack surface.
  • A Reddit thread from February 2026 where a security researcher catalogued every AI agent incident from 2025 found that implicit trust in agent-authorized workflows was the single most common failure mode, not the AI itself, but the assumption that the AI would only do what it was supposed to do.
  • Credential sprawl is already out of control. Modern companies manage hundreds of non-human identities, API keys, tokens, service accounts. Every computer use agent you deploy adds more.

"Credential theft within agent-authorized workflows" was the most cited attack pattern across all AI agent security incidents in 2025. The agents weren't hacked. The trust was.

The Competitors Are Winging It

Let's be direct about where the major players stand. Anthropic's computer use feature is genuinely impressive technically, but Anthropic's own security team has had to publish multiple threat intelligence reports in 2025 about misuse of Claude, including a sophisticated 'vibe hacking' operation where criminals used Claude Code to run a data extortion campaign. Claude Code also had vulnerabilities disclosed in February 2026 that could have allowed silent takeover of a developer's machine. That's the company that built the computer use layer. OpenAI's Operator launched in January 2025 and was almost immediately analyzed by Push Security for its credential stuffing potential. The WorkOS team wrote a whole post in February 2025 about how authentication patterns for Operator-style agents are still fundamentally unsolved. Brave's security team found that Perplexity's Comet browser was vulnerable to prompt injection via screenshots as recently as August 2025. This is not a small startup problem. These are the best-funded AI labs on earth, and they're shipping computer-using AI agents with credential handling that security researchers are tearing apart in real time. The arXiv paper 'The Hidden Dangers of Browsing AI Agents' from May 2025 put it plainly: passing credentials to agents 'presents a fundamental security limitation.' Not a bug. A limitation. Meaning it's baked into how most of these systems are designed.

What Actually Secure Credential Handling Looks Like

The right answer isn't 'don't use AI agents.' That's like saying don't use the internet because it has security risks. The right answer is building computer use agents with credential isolation as a first-class design requirement, not a checkbox added after the demo looked cool. That means agents should never have credentials in their context window, full stop. Credentials should be injected at the execution layer, not the reasoning layer. The agent should know it has access to a system, not know the password to that system. It means sandboxed execution environments where the agent operates in an isolated VM or container, so a prompt injection attack can't pivot to the host system or other credentials. It means audit logs on every single action the agent takes, so you can reconstruct exactly what happened if something goes wrong. It means short-lived, scoped credentials that expire after the task completes, not long-lived API keys that sit in a config file for three years. And it means the agent should confirm sensitive actions with a human before executing them, not just barrel through because the instructions said to. These aren't exotic requirements. They're basic security hygiene. The problem is that most computer use agent frameworks were built to show off what's possible, not to be deployed in a production environment where a credential leak means a $4.88 million average breach cost, which is IBM's 2024 figure.

Why Coasty Was Built for This Reality

I'm going to tell you why I use Coasty for computer use tasks that involve real credentials, and it's not because of the marketing. It's because of how the architecture works. Coasty operates agents inside isolated cloud VMs or your own desktop environment, which means the execution surface is contained. There's no shared context bleeding between agent tasks. The agent swarm architecture, where multiple agents run tasks in parallel, is built with separation in mind, so one compromised task doesn't hand over keys to everything else. And Coasty sits at 82% on OSWorld, the gold standard benchmark for real-world computer use tasks. The next closest competitor isn't close. That matters for credential security because a more capable agent makes fewer mistakes, and mistakes in credentialed workflows are exactly how you end up with unauthorized actions. A confused agent that retries a login five times, or misreads a CAPTCHA and falls back to a less secure flow, or gets tricked by a prompt injection because it lacks the reasoning to recognize the attack, that's your liability. BYOK support also means your secrets infrastructure stays yours. You're not routing credentials through someone else's cloud hoping they've got their security posture together. Given what we've seen from even the biggest players in 2025, that's not a hope I'd bet my company on.

Here's my actual take. The AI agent credential problem is going to get significantly worse before it gets better. More agents means more attack surface. More attack surface means more incidents. And most companies are going to learn this lesson the hard way, after a breach, after an audit finding, after a security researcher posts their architecture on Reddit and the comments are brutal. You don't have to be that company. The tools to do this right exist. The benchmark data to pick the most capable computer use agent exists. The architecture patterns to handle credentials without putting them in the agent's context window exist. What's missing is urgency, and maybe reading one too many 'AI is magic' blog posts that skip straight past the part where your agent has the keys to your kingdom. Stop treating credential handling as a deployment detail. It's the deployment. If you're evaluating computer use agents right now, start at coasty.ai. Check the OSWorld score. Read how the isolation works. Then go ask every other vendor the same hard questions about what happens when their agent reads a malicious webpage mid-task. The answers will be very telling.

Want to see this in action?

View Case Studies
Try Coasty Free