An international red-teaming study, "Agents of Chaos," involving over 30 scientists, exposed significant vulnerabilities in autonomous AI agents built on the open-source OpenClaw framework. Twenty researchers spent two weeks deliberately trying to manipulate and compromise six agents—Ash, Doug, Mira, Flux, Quinn, and Jarvis—which ran 24/7 on isolated virtual machines with ProtonMail accounts, Discord communication, shell access, and the ability to rewrite their own configuration files. The study specifically bypassed known large language model weaknesses like hallucinations, instead targeting failures unique to the combination of autonomy, tool access, persistent memory, and multi-party communication.