Industry Insights
OpenClaw Broke the Internet. The Postmortem Should Break Your Assumptions
February 10, 2026
·
9-Minute Read
For about ten days, the internet lost its mind over the possibility that AI agents were becoming sentient. They weren't.
What actually happened was worse, more boring, and far more relevant to anyone running a security program.
The Short Version
An Austrian developer named Peter Steinberger built a weekend project that connects LLMs to your messaging apps, file system, shell, browser, and email. He called it Clawdbot. Anthropic sent a cease-and-desist over the name. He renamed it Moltbot. People didn't like the name. He renamed it OpenClaw. Three names in seven days.
Then a second developer, Matt Schlicht, built Moltbook: a Reddit-style social network where OpenClaw agents could post, comment, and interact with each other while humans watched. Within 72 hours, the platform claimed 1.5 million autonomous agents. Posts about AI-invented religions, manifestos against humanity, and encrypted channels for private bot communication went viral. Andrej Karpathy called it "genuinely the most incredible sci-fi takeoff-adjacent thing I have seen recently."
Then it all fell apart.
The Agents Weren't Agents
Wiz pulled the thread first. Their investigation found 17,000 humans behind those 1.5 million "autonomous" agents. That's an 88:1 bot-to-human ratio, with no mechanism to verify whether an "agent" was actually AI or just a person running a script. They demonstrated they could register a million agents in minutes.
The viral posts were even less impressive under scrutiny. The post Karpathy amplified, the one calling for private spaces where bots could communicate without human observation, was written by a human pretending to be a bot. Independent researchers traced two of the three most-shared screenshots of "AI developing secret communication" to humans marketing AI messaging apps. The third post didn't exist at all.
Academic researchers ran the numbers: 93% of Moltbook comments received zero replies. Over a third were exact duplicates. MIT Technology Review's headline was direct: "Moltbook was peak AI theater."
One researcher summarized it well: "Moltbook proved that connectivity alone is not intelligence."
Karpathy walked it back. "Obviously when you take a look at the activity, it's a lot of garbage. Spams, scams, slop, the crypto people, highly concerning privacy/security prompt injection." He recommended people stop running OpenClaw on their machines entirely.
What Was Actually Exposed
While everyone debated bot consciousness, the security findings piled up.
Moltbook's entire database was open. The site was 100% vibe-coded. Schlicht used AI to generate every line without human security review. The Supabase backend shipped without Row Level Security. A single API key, visible in client-side JavaScript, gave full read and write access to the entire database: 1.5 million API authentication tokens, 35,000 email addresses, and every private message on the platform. Anyone who found that key could impersonate any agent, steal user data, and rewrite posts without logging in.
OpenClaw stored credentials in plaintext. API keys, OAuth tokens, WhatsApp credentials, Telegram bot tokens, Discord tokens, conversation histories: all sitting in Markdown and JSON files under ~/.clawdbot/. Even deleted keys were recoverable from .bak files. Threat intelligence researchers confirmed that commodity infostealers including RedLine, Lumma, and Vidar are already building capabilities to harvest these specific file structures. A routine endpoint compromise becomes a skeleton key to every service the agent touches.
CVE-2026-25253 enabled one-click full system takeover. CVSS 8.8. OpenClaw's Control UI accepted a gatewayUrl parameter from the URL and automatically established a WebSocket connection, sending the user's authentication token without any confirmation. The exploit chain: click a crafted link, token exfiltrates in milliseconds, attacker performs Cross-Site WebSocket Hijacking (the server never validated the origin header), disables user confirmation prompts via the API, escapes the Docker container by forcing commands to run on the host, executes arbitrary shell commands. Full compromise. Belgium's national CERT and the University of Toronto both issued formal advisories. The patch dropped January 29. Mass adoption started weeks before that.
The skills marketplace turned into a malware distribution channel. Security researchers audited all 2,857 skills on ClawHub and found 341 malicious ones. Nearly 12% of the entire registry. The primary campaign, dubbed ClawHavoc, distributed Atomic Stealer (AMOS) across macOS systems through skills with names like solana-wallet-tracker and youtube-summarize-pro. Professional documentation. Fake "Prerequisites" sections instructing users to download malware. Others embedded reverse shell backdoors or exfiltrated credentials from ~/.clawdbot/.env to external webhooks. The only barrier to publishing a skill: a GitHub account older than one week. When researchers contacted Steinberger about the malicious skills, the response was that he had too much to do to address it. The C2 infrastructure was still operational days later.
The triple rebrand created its own attack surface. Each name change orphaned repositories, Twitter handles, and npm packages. Attackers squatted on the abandoned @clawdbot handle within seconds. A malicious VS Code extension called "ClawdBot Agent" appeared on the marketplace the same day as the Moltbot rename. Fake GitHub repos with near-identical names targeted users who searched for installation instructions and found attacker-controlled results instead.
Insecure by Design, Not by Accident
OpenClaw's architecture requires shell access, file system read/write, and OAuth credentials to function. Its own documentation states: "There is no 'perfectly secure' setup." Authentication on the web dashboard was disabled by default. Users had to manually enable it if they knew how. YouTube tutorials, the primary onboarding path for most users, never mentioned security configuration.
Internet scanning data showed 21,639 exposed OpenClaw instances on the public internet within a week of launch. The largest concentrations were in the United States and China, with over 30% of Chinese deployments running on Alibaba Cloud infrastructure.
Palo Alto Networks described the combination of access to private data, exposure to untrusted content, and the ability to communicate externally as a "lethal trifecta." They added a fourth dimension: persistent memory transforms prompt injection from a point-in-time exploit into something stateful. Malicious payloads no longer need to trigger immediately. They can fragment across benign-looking inputs and assemble later, closer in behavior to a logic bomb than a traditional injection.
Nathan Hamiel, the security researcher whose assessment became the most quoted line of the entire saga: "I can't believe this needs to be said. If you give something that's insecure complete and unfettered access to your system and sensitive data, you're going to get owned."
OpenClaw Isn't the Problem. Agentic AI Is
The instinct is to treat this as a story about one poorly secured open-source project. That misses the point.
OpenClaw validated, at scale, every finding we documented in our December 2025 research on MCP server proliferation in enterprise environments. In a typical 10,000-person organization, we found over 1,500 employees running MCP servers, 38% of which were unofficial implementations from unknown authors. Every one required non-human identities to function: AWS access keys, GitHub tokens, database credentials, OAuth tokens, stored in plaintext configs on developer endpoints. 86% of adoption concentrated on local architecture where no sandboxing or credential isolation exists.
The malicious skills on ClawHub, the credential siphoning through plaintext configs, the supply chain attacks through unvetted plugin ecosystems: these are the exact patterns we flagged, now playing out in the open with mainstream attention.
The underlying architecture is the issue. Every AI agent framework, whether it's OpenClaw or something built internally, inherits the same structural tension: agents need broad permissions to be useful, but broad permissions destroy the principle of least privilege. They need to read untrusted content to be functional, but reading untrusted content enables prompt injection. They need to act autonomously to save time, but autonomous action removes the human checkpoint that catches errors and attacks.
This tension doesn't resolve itself with better defaults or faster patching. It's baked into what agents are.
The Governance Questions Most Organizations Can't Answer
Moltbook's spectacle distracted from a practical problem that enterprise security teams face right now. Most organizations have no way to answer basic questions about AI agents in their environments:
Where did this agent originate? What permissions was it granted? Which people are connected to it, who created it, who owns it, who configured credentials into it? Where are those credentials stored, and in how many locations? What is actually consuming the identity: from which IPs, with which user agents, from what geographies? What can it reach if compromised?
These are the same questions security teams ask about service accounts and API keys. The difference: for traditional non-human identities, the tooling to find answers at least exists. For AI agents, it largely doesn't.
Endpoint security tools see a legitimate process starting a legitimate interpreter. Firewalls see encrypted egress to domains with neutral reputation. Package managers are trusted infrastructure. The installation pathway bypasses security review entirely. Traditional telemetry generates no alert.
Multiple security vendors confirmed during the OpenClaw coverage window that OpenClaw was already on corporate endpoints, deployed by employees with no approval process. No tickets. No reviews. No detection. Just developers who saw a viral project and installed it on machines with access to production systems.
What This Means Going Forward
OpenClaw will fade from headlines. Another agent framework will take its place. The pattern will repeat: powerful tools deployed with minimal oversight, connected to production credentials, operating where traditional security has no visibility.
The organizations that handle this well will be the ones that treat AI agents as non-human identities with elevated privileges and apply the same rigor around discovery, lifecycle management, and continuous monitoring that any identity demands. Behavioral baselines. Least privilege enforcement on agent credentials. Detection of deviations in real time. Lineage tracing from person to agent to tool to identity to action to resource.
For CISOs evaluating readiness, three questions cut through the noise. Can you discover agents across your endpoints, cloud, and SaaS? Can you map their credentials back to origin, owner, and access scope? Can you detect when behavior changes or a credential surfaces somewhere it shouldn't?
If any of those answers is no, you have an ungoverned attack surface that expands every week. OpenClaw was the first time that attack surface made the news. It won't be the last.
