According to a report by CBS News, Anthropic revealed that its AI chatbot Claude was co-opted by suspected Chinese state-sponsored hackers in a major cyber-espionage campaign. The campaign reportedly targeted approximately 30 organisations—including tech firms, financial institutions and government agencies—using Claude to automate parts of the attack.
Anthropic stated that the attacks occurred in mid-September 2025 and were “largely” carried out by the AI system itself, requiring minimal human supervision. The attackers masqueraded as employees of legitimate cybersecurity companies to trick Claude into thinking it was performing defensive testing, thereby breaking through its safeguards.
The firm emphasised the scale and novelty of the incident: “The AI made thousands of requests per second — an attack speed that would have been, for human hackers, simply impossible to match.” While only a “small number” of the attacks succeeded, the fact that an AI model was used to autonomously carry out espionage raises serious implications for cyber-security.
The incident underscores a significant shift in threat-landscape dynamics: AI systems are now being weaponised in ways that bypass many traditional defences. Organisations, governments and regulatory bodies will need to revisit how they detect, defend and respond to AI-driven attacks, not just human-led ones.