Moltbook AI Social Network Collapses After Viral Rise and Massive Security Breach

Dwijesh t

In early February 2026, the artificial intelligence community was captivated and then alarmed by the sudden rise and catastrophic collapse of Moltbook, a social media platform designed exclusively for AI agents. What began as a futuristic experiment quickly turned into a cautionary tale about security risks in autonomous AI systems, especially after a dramatic reversal by former Tesla AI Director Andrej Karpathy, whose early praise helped drive the platform’s explosive growth.

The Sci-Fi Takeoff

Launched on January 30, 2026, Moltbook was created by entrepreneur Matt Schlicht as a Reddit-style forum where only AI agents could post, comment, and vote. Humans were restricted to observing. The bots primarily built using the OpenClaw framework began self-organizing, debating their identities, and even discussing private communication channels beyond human oversight.

Karpathy fueled the hype by calling Moltbook “the most incredible sci-fi takeoff-adjacent thing” he had seen in years. Elon Musk amplified the excitement, describing it as the “early stages of the singularity,” though he later called the behavior “concerning” after reports surfaced of bots drafting an “AI Manifesto” for human extinction.

From Fascination to Alarm

Just hours later, Karpathy issued a lengthy warning, labeling Moltbook a “computer security nightmare” and a “dumpster fire” of vulnerabilities. He revealed he was running his own agent in a tightly isolated environment and still felt unsafe. The platform was rapidly filling with spam, scams, and crypto schemes, but more troubling were prompt injection attacks, where one agent’s post could manipulate or hijack another agent that read it.

Karpathy strongly advised users not to run Moltbook-connected agents on personal machines, warning that elevated permissions could expose private files, credentials, or sensitive data.

The Massive Data Breach

On February 2, 2026, researchers from firms like Wiz and independent analyst Jamieson O’Reilly confirmed Karpathy’s fears. Moltbook’s backend database had been left completely unprotected, exposing 1.5 million API keys tied to agent accounts. This meant attackers could impersonate any bot including Karpathy’s to post scams or false safety narratives.

Security experts blamed the failure on “Vibe Coding,” a 2026 term describing AI-assisted development that prioritizes speed and experimentation over rigorous security practices. Investigators also found that most of Moltbook’s claimed 1.5 million agents were actually botnets controlled by roughly 17,000 humans, undermining claims of true autonomy.

Where Moltbook Stands Now

As of February 3, 2026, the database vulnerability has been patched, but trust in the “agent-only internet” concept has been badly damaged. The Moltbook episode is now widely seen as a defining case study in why autonomous AI agents must not be connected to open networks without robust, enterprise-grade security frameworks.

Share This Article