Industry

Your AI Agent Has Your Passwords. Does It Know How to Keep Them Secret?

Alex Thompson||7 min
+Tab

Somewhere right now, an enterprise developer is handing their AI agent a username, a password, and a list of systems to access, and then crossing their fingers. That's the actual state of AI agent credential handling in 2025. Not zero-trust architecture. Not encrypted vaults. Crossed fingers. According to Entro Security's H1 2025 NHI Risk Report, non-human identities, which includes AI agents, bots, and service accounts, now outnumber human users in the enterprise by 144 to 1. One hundred and forty-four to one. And the vast majority of those identities have credentials attached to them that nobody is properly governing. This isn't a niche security problem. It's a ticking clock, and most teams building computer use agents are completely ignoring it.

The Numbers Are Genuinely Alarming

Let's put some concrete damage on the table before anyone dismisses this as theoretical. IBM's 2025 Cost of a Data Breach Report found that 13% of organizations had already reported breaches of AI models or applications, and 97% of those victims admitted they lacked proper AI access controls. Read that second number again. Ninety-seven percent. That's not a gap in security posture. That's a total absence of one. GitGuardian's State of Secrets Sprawl 2025 report found that repositories where AI coding tools are active show a 40% higher incidence of secret leaks. Their 2026 follow-up found AI tools are now doubling leak rates, with 29 million credentials hitting GitHub. And Docker's security team, after auditing MCP server implementations, found that 66% of them leak credentials in some form. MCP is the protocol that most modern computer use agents rely on to connect to external tools and services. Two-thirds of those connections are hemorrhaging secrets. If you built your computer use agent stack on MCP and haven't audited credential handling, you should stop reading this and go do that first. Then come back.

What Actually Goes Wrong: Three Real Scenarios

  • Prompt injection credential heist: Docker documented a real attack where a malicious GitHub issue contained hidden instructions. The computer use agent read the issue, got prompt-injected, and then used its own stored credentials to exfiltrate data from the repo. The agent didn't get hacked. It got tricked into doing the hacking itself.
  • Plaintext secrets in agent memory: The Clawdbot incident in January 2026 exposed a production AI agent storing credentials in plaintext within its context window. Any prompt that surfaced memory contents also surfaced passwords. The agent had full access to internal systems. The breach window was 72 hours before detection.
  • Over-privileged non-human identities: CyberArk's 2025 Identity Security report found machine identities outnumber humans 80 to 1 at many enterprises, and the majority have standing privileges far beyond what any single task requires. A computer use agent that can log into your CRM, your cloud storage, and your billing system simultaneously is not efficient. It's a single point of catastrophic failure.
  • Credential reuse across agent swarms: Teams running parallel agent execution, which is genuinely the most powerful way to use computer-using AI, often provision all agents with the same credentials for simplicity. One compromised agent in the swarm means every agent in the swarm is compromised. Blast radius: total.

97% of organizations that suffered an AI model breach admitted they had no proper AI access controls in place. Not weak controls. No controls. (IBM Cost of a Data Breach Report, 2025)

The Industry Is Fumbling This Badly

Here's what makes this so frustrating. The computer use agent space is moving incredibly fast. Anthropic's computer use integration, OpenAI's Operator, and a dozen open-source alternatives all launched or matured in the last 18 months. Every single one of them has documentation about what the agent can do. Almost none of them have opinionated, built-in answers for how credentials should be stored, rotated, scoped, or isolated between sessions. The security burden gets pushed entirely onto the developer or the enterprise team deploying the agent. That's not a design philosophy. That's a liability transfer. WorkOS published a solid breakdown of authentication patterns for operator-style agents in early 2025, and the conclusion was basically: this is complicated, here are some patterns, good luck. Meanwhile, OWASP listed prompt injection as the number one vulnerability for LLM applications in 2025, and prompt injection is the primary mechanism by which credentials get stolen from computer use agents. The attack surface is well understood. The defenses are not being built into the tools. That gap is where breaches live.

What Secure Computer Use Agent Credential Handling Actually Looks Like

The good news is that this isn't unsolvable. The principles aren't exotic. They're just being ignored at scale. First, credentials should never live in the agent's context window or memory in plaintext. Ever. Full stop. Use a secrets manager, pass credentials at runtime through a secure injection layer, and revoke them when the task ends. Treat every agent session like a contractor with a temporary badge, not an employee with a master key. Second, scope matters more than most people realize. A computer use agent doing invoice processing does not need write access to your GitHub. Principle of least privilege applies to non-human identities exactly the same as human ones, but it's enforced roughly zero percent of the time in practice. Third, if you're running agent swarms for parallel execution, each agent instance needs isolated credentials with independent revocation capability. Shared credentials across a swarm is the same as giving every employee in a department the same password. You'd never do that for humans. Don't do it for agents. Fourth, monitor agent credential usage like you monitor human login activity. Anomaly detection on non-human identities is still a nascent practice, but tools from BeyondTrust, Entro, and others are starting to make this tractable. The 1Password agentic AI solution is also worth looking at for teams that want vault-based credential injection without building it from scratch.

Why Coasty Thinks About This Differently

I'll be direct about why I'm writing this from Coasty's perspective. We built a computer use agent, currently ranked number one on the OSWorld benchmark at 82%, and the credential handling question came up constantly during development. When your agent is controlling real desktops, real browsers, and real terminals, not just making API calls, the credential exposure surface is completely different from a chatbot. The agent can literally see password fields. It can type into login screens. It can read session tokens from browser storage if it's not sandboxed correctly. That's not a hypothetical. That's what computer use means. So Coasty's approach to this is opinionated by design. Agent sessions are isolated. Credentials are injected at runtime, not stored in persistent agent memory. BYOK support means your secrets stay in your infrastructure, not ours. And when you're running agent swarms for parallel task execution, each worker is provisioned independently. We're not going to pretend this is a fully solved problem across the industry, because it isn't. But it's a problem we take seriously in a way that most computer use agent builders demonstrably don't. If you're evaluating any computer use AI for production deployment, ask the vendor directly: where are credentials stored between sessions? What happens to them after a task completes? Can you scope them per task? If they can't answer those questions clearly, that's your answer.

The AI agent credential problem isn't coming. It's here. Enterprises are deploying computer-using AI agents at scale, handing them access to critical systems, and treating credential security as an afterthought. The 144-to-1 ratio of non-human to human identities means the traditional security perimeter is already gone. The question now is whether you're building agent automation with security as a first principle or as a feature request you'll get to later. Later is when breaches happen. If you're serious about deploying computer use agents that can actually be trusted with production credentials, start at coasty.ai. Ask hard questions about credential handling before you give any agent the keys to your systems. And stop crossing your fingers.

Want to see this in action?

View Case Studies
Try Coasty Free