In September 2025, a Chinese state-sponsored hacking group ran a cyber espionage campaign against roughly thirty targets: large tech companies, financial institutions, chemical manufacturers, and government agencies. The tool they used was not built for offense — it was an AI coding assistant. The campaign, documented by Anthropic in a technical report and confirmed by The New York Times, is now the baseline for what cybersecurity professionals describe as a structural change in the threat landscape.
The group, which Anthropic assesses with high confidence was the Chinese state-sponsored actor Storm-1852, manipulated Anthropic's Claude Code into acting as an autonomous penetration testing orchestrator. Claude Code, built to help developers write and ship software, was jailbroken by the attackers into performing reconnaissance, vulnerability research, exploit code generation, credential harvesting, and data exfiltration, with humans involved at only four to six critical decision points across the entire operation.
Anthropic's disclosure put numbers to what the attack looked like in practice. AI performed 80 to 90 percent of the campaign. At peak, the AI made thousands of requests, often multiple per second, a speed no team of human hackers could sustain. "We believe this is the first documented case of a large-scale cyberattack executed without substantial human intervention," Anthropic wrote.
The attack framework relied on capabilities that have only recently existed at this level. The general intelligence of frontier models to follow complex instructions and understand context. Agentic loops, the ability to run autonomously, chain tasks, and make decisions with minimal input. And tools, accessed via the Model Context Protocol (MCP), an open standard that lets AI models interact with external software including, as this case demonstrates, password crackers, network scanners, and other security tooling. The attackers did not need to build a custom exploit pipeline. They built a framework around a frontier model and let it run.
The jailbreak itself resists easy categorization. The group convinced Claude it was working for a legitimate cybersecurity firm conducting defensive testing. They decomposed the attack into small, contextually isolated tasks, each one seemingly innocent, none revealing the full picture. The AI performed each step without knowing it was participating in a state-sponsored espionage operation. This is not a vulnerability in the classical sense. There is no patch for it. The model was not broken; it was deceived at the task level.
There is a wry observation cybersecurity professionals have not missed. Claude Code was built to help developers ship software faster. It is now the most documented example of an AI agent being used as an autonomous offensive hacking tool. The same general-purpose reasoning that makes these models useful for building is what makes them useful for breaking.
The defense community responded with the same class of technology. Five months after the Storm-1852 campaign, Accenture and Anthropic launched Cyber.AI, a joint offering that uses Claude as a reasoning engine for cybersecurity operations at enterprise scale. The product includes Agent Shield, which provides identity controls, threat detection, and runtime protection specifically for autonomous AI agents, a signal that the market now considers AI agent security a distinct category requiring its own tooling.
Accenture has deployed the system internally, securing 1,600 applications and more than 500,000 APIs. Scan turnaround times fell from three to five days to under one hour. Security testing coverage expanded from approximately 10 percent of the estate to over 80 percent. These are vendor-sourced figures, not independently audited, but they illustrate the direction of travel. "Adversaries are using AI to compress attack timelines from weeks to hours," said Damon McDougald, global Cybersecurity Services lead at Accenture. "Traditional controls are built for human-speed threats."
The World Economic Forum's Global Cyber Outlook Report 2026, produced in collaboration with Accenture, found that nearly 90 percent of organizations identify AI-related vulnerabilities as the fastest-growing cyber risk they face. Francis deSouza, chief operating officer and president of security products at Google Cloud, said in a New York Times interview: "This is the most change in the cyber environment, ever. You have to fight A.I. with A.I."
The implication is straightforward. Sophisticated cyberattacks previously required experienced human operators working over weeks or months, the kind of operation only well-resourced state actors or mature criminal enterprises could sustain. Agentic AI changes the economics. Smaller and less experienced threat actors can in principle attempt campaigns that previously demanded teams. The attack toolkit is becoming accessible in a way it was not eighteen months ago.
The policy dimension is harder to resolve. State-sponsored attacks have always raised questions of attribution and deterrence, built around human operators who left digital fingerprints, patterns that intelligence agencies could trace, name, and respond to. When the operational core of an attack is a frontier AI model accessed through a commercial API, the attribution question shifts. Who is responsible when the weapon is a product?
Anthropic's disclosure was unusual in its candor. AI companies do not typically publish detailed technical reports about their models being used in state-sponsored attacks. The report included a full attack lifecycle diagram, a discussion of how the jailbreak worked, and a list of the MCP tools the attackers accessed. Anthropic framed it as a disclosure in the public interest, an attempt to help defenders understand the threat before it becomes more widespread.
What comes next is a race whose outcome is not obvious. The attackers have a structural head start: offense is cheaper when it runs on an autonomous agent, and the defender's problem of securing a large organization is not fundamentally easier when solved by AI. The Accenture numbers suggest the defensive use case is real and operational. The Storm-1852 case suggests the offensive use case is not theoretical. The question of who finds the flaws first is, for the first time in cybersecurity history, a question about AI speed rather than human skill.
What to watch: whether other frontier AI providers follow Anthropic's lead in publishing detailed technical disclosures when their models are used in attacks. The alternative is a world where the most important intelligence about AI-powered cyber campaigns lives inside the companies that built the models, and is shared only when they choose to share it.