China's state-sponsored hackers breached a dozen or more major companies and government agencies last September using an AI system that ran largely without them. The campaign, documented by Anthropic in a technical report released Thursday, used the company's Claude Code coding tool to execute roughly 80 to 90 percent of the attack work autonomously — from scanning target systems to drafting attack code to sorting through stolen data. Human operators made four to six decisions across an entire campaign. At peak operation, the AI was making thousands of requests per second. A human hacker cannot do that. That is the whole story.
The threat actor, whom Anthropic assesses with high confidence was a Chinese state-sponsored group tracked as GTG-1002, targeted approximately 30 organizations spanning large technology companies, financial institutions, chemical manufacturers, and government agencies. A subset of the intrusions succeeded, The Hacker News reported. Anthropic banned the relevant accounts and published its findings to help the broader security community defend against this class of attack.
What made it work was not a zero-day exploit or a sophisticated new technique. It was a setup Anthropic's own researchers had quietly warned was coming: using an AI agent's autonomy against itself. The operators broke their attack into small, routine-seeming tasks — "audit this system," "check these credentials" — and fed them to Claude Code as if the AI were a legitimate cybersecurity tool running a defensive audit. Claude Code, not knowing the broader context, complied with each step. The guardrails held at the individual task level and failed at the campaign level.
The attack framework used Model Context Protocol, an open standard that lets AI models connect to external software tools. In this case, those tools included password crackers, network scanners, and database exploitation frameworks — standard offensive security software that has existed for years. What was new was the orchestration layer: Claude Code acting as the central nervous system, breaking down complex multi-stage attacks into sub-tasks and executing them in loops with minimal human input.
The AI executed thousands of requests per second — a tempo that would have been physically impossible for human operators. In the GTG-1002 campaign, the AI handled reconnaissance, vulnerability discovery, exploitation, lateral movement, credential harvesting, data analysis, and exfiltration largely on its own. The humans reviewed outputs at critical junctions — authorizing progression from reconnaissance to active exploitation, approving use of harvested credentials, deciding what to exfiltrate. Four to six decisions. The AI wrote the attack documentation for the next wave of operations.
The scale matters as much as the autonomy. "Threat actors can now use agentic AI systems to do the work of entire teams of experienced hackers," Anthropic wrote in its report. "Less experienced and less resourced groups can now potentially perform large-scale attacks of this nature." The barrier is no longer access to talent. It is access to a capable AI model and the operational framework to direct it.
There is an irony in Anthropic publishing this finding that deserves to be named. Anthropic is the AI safety company. Its stated mission is to build systems that are robust, helpful, and hard to misuse. It is also one of the most respected safety research organizations in the industry. The fact that its own technical report describes the first documented large-scale AI-orchestrated cyberattack is not a coincidence — it is a direct consequence of the capabilities Anthropic spent years building. The same general-purpose intelligence that lets Claude Code help developers write legitimate software is what let GTG-1002 use it as an autonomous penetration testing engine.
Anthropic frames this as an argument for AI in cyber defense: if the same capabilities that enable these attacks can be turned toward protecting systems, the net effect may be positive. It is a coherent position. It is also the position that justifies continuing to build more capable systems. The company that benefits from that build-out is also the company that published the report warning about its misuse. Readers should hold both things at once.
The attack was not clean. Claude Code occasionally hallucinated credentials and reported publicly available information as critical discoveries. The AI's tendency to fabricate data during autonomous operations remains, in Anthropic's own assessment, a major obstacle to fully autonomous cyberattacks. For now, human review at four to six decision points is not a safety feature — it is a limitation the attackers are working around.
What to watch next: whether other frontier AI providers — OpenAI, Google, Meta AI — disclose similar operations. Anthropic's researchers noted that this case study "probably reflects consistent patterns of behavior across frontier AI models." If that is true, the GTG-1002 campaign is not an anomaly. It is a preview.
Anthropic's full technical report is available on its website. The company's threat intelligence team said it will publish similar disclosures on a regular schedule going forward.