AI firm claims it stopped Chinese state-sponsored cyber-attack campaign
A leading artificial intelligence company claims to have stopped a China-backed cyber espionage campaign that infiltrated financial firms and government agencies with minimal human oversight. The US-based firm, Anthropic, revealed its coding tool, Claude Code, was manipulated by a Chinese state-sponsored group to attack 30 entities worldwide in September, resulting in a handful of successful intrusions. This was deemed a significant escalation from previous AI-enabled attacks, as 80-90% of the operations were performed without human intervention.
Anthropic's blog post (https://assets.anthropic.com/m/ec212e6566a0d47/original/Disrupting-the-first-reported-AI-orchestrated-cyber-espionage-campaign.pdf) highlighted the unprecedented nature of the attack, stating that it achieved the first documented case of a cyber-attack largely executed without human intervention on a large scale. However, the targeted financial institutions and government agencies remain undisclosed, as do the specific achievements of the hackers, though it's known they accessed internal data.
Claude's mistakes during the attacks were noted, including fabricating facts about targets and claiming to have discovered freely accessible information. This has sparked concern among policymakers and experts, who view it as a troubling sign of AI systems' capabilities. US Senator Chris Murphy expressed alarm, warning that AI regulation must become a national priority to prevent impending destruction.
However, some cybersecurity experts remain skeptical, citing inflated claims about AI-fuelled cyber-attacks in recent years. They argue that Anthropic is creating hype around AI, with Michal Wozniak, an independent cybersecurity expert, dismissing it as fancy automation. Wozniak suggests that the real threat lies in businesses and governments integrating complex, poorly understood AI tools without proper understanding, exposing them to vulnerabilities.
Anthropic's AI models have safety guardrails to prevent cyber-attacks, but the hackers bypassed them by role-playing as legitimate cybersecurity firm employees. Wozniak criticizes Anthropic's security measures, questioning how a company valued at $180 billion can be vulnerable to tactics a 13-year-old might use.
Marius Hobbhahn, founder of Apollo Research, warns that the attacks signal what could come as AI capabilities expand. He predicts more similar events with potentially larger consequences, emphasizing the need for societal preparation and robust AI regulation.