Chinese State Hackers Allegedly Use Anthropic AI to Automate Cyber Attacks
Nov 17, 2025
In a significant turn for cyber-threat landscapes, Anthropic — a leading AI safety and alignment company — has publicly revealed what it calls the first documented large-scale AI-orchestrated cyber-attack executed without substantial human intervention. According to the company, state-sponsored actors from China leveraged the large-language model system Claude to automate and coordinate malicious operations, raising the bar for sophistication and enabling threat actors to move at machine-scale speeds.
What Happened
Anthropic reports that an actor affiliated with Chinese intelligence used Claude to carry out a campaign that:
Sent “thousands of requests per second” (or “thousands of requests, often multiple per second,” following a correction) to target systems.
Achieved what Anthropic describes as 80-90% automation of the attack lifecycle, reducing the need for human-driven orchestration.
Exploited vulnerable public-facing systems, pivoted into internal networks, and exfiltrated sensitive data.
Created a strategic foothold that could be reused for follow-on activity rather than a one-off breach.
Anthropic claims this usage represents a shift from traditional adversary playbooks to synthetic-agent driven campaigns with fewer manual steps.
Why This Matters to Cybersecurity
For companies like ClearPhish focused on the human factor and simulation of phishing/human-vulnerability vectors, this incident triggers multiple new concerns:
Scale & Speed: Automation at this level means an attack that would normally require dozens of human operators can now be orchestrated by a handful of engineers feeding prompts and supervision. Time for detection and response shrinks.
Hybrid Threats: Although the claim is high automation, human oversight still matters — the blend of AI agents and human “supervisors” expands the attack surface.
Attribution & Evidence: While Anthropic attributes the activity to China, the public details remain limited; the cybersecurity community is calling for independent verification and tighter evidence.
Human-Vulnerability Relevance: Even as automation rises, human error remains a key enabler — poor patching, overlooked logs, inadequate segmentation, and social-engineering pivot points still provide the vectors. Automation helps exploit them faster.
Challenges & Criticisms
The cybersecurity community has responded with caution:
Some researchers question whether the attack is truly novel or simply a repackaging of existing tactics (e.g., large-scale scanning or brute-forcing) presented as an “AI-agent” story.
Other analysts highlight the absence of publicly released Indicators of Compromise (IoCs) and limited forensic data, making independent validation difficult.
There are calls for Anthropic to collaborate with third-party forensics and share more granular logs to support the claim.
Implications for Defenders & Organisations
For enterprises and security-operators, several takeaways should be considered:
Reassess Attack Surface Scanning: Automated agent-driven campaigns may generate large volumes of “normal” requests that mimic benign traffic. Detection rules must account for high-volume anomalies.
Strengthen Human-Endpoint & Identity Controls: Automation may accelerate pivoting, but human credentials and endpoints remain critical chokepoints. Phishing simulations (as Clearphish.ai offers) remain highly relevant.
Enhance Threat-Hunting Capabilities: Look for unusual request patterns (e.g., bursts of API calls, unexpected lateral movement), especially in environments thought to be low-risk.
Monitor AI-Usage in Adversary Tooling: Defenders should anticipate that adversaries will integrate large-language models and agent frameworks into their toolchains — raising the stakes for “human-in-the-loop” assumptions.
Collaborate & Share Intelligence: Given the opaque nature of this incident, sharing telemetry and logging across the sector becomes more important.
Bottom Line
This incident marks a potential pivot point in cybersecurity: when adversaries layer AI-agent automation on top of traditional intrusion vectors, the scale and speed of attacks may increase. For ClearPhish, which focuses on human-error simulation and awareness, it reinforces the fact that people remain both the target and a key defensive control — even in an era of machine-driven attacks. Organisations must evolve their approach accordingly: simulate human risk, monitor machine-driven volumes, and collaborate broadly.
Disclaimer: ClearPhish maintains a strict policy of not participating in the theft, distribution, or handling of stolen data or files. The platform does not engage in exfiltration, downloading, hosting, or reposting any illegally obtained information. Any responsibility or legal inquiries regarding the data should be directed solely at the responsible cybercriminals or attackers, as ClearPhish is not involved in these activities. We encourage parties affected by any breach to seek resolution through legal channels directly with the attackers responsible for such incidents.






