OpenAI's cybersecurity AI has helped patch more than 3,000 critical and high-severity vulnerabilities since its first defensive model launched — a number that reframes the company's latest security AI as a progress report on a system already operating at scale, rather than a prototype announcement.
The figure, according to OpenAI's own announcement, is the most concrete evidence yet that AI-assisted vulnerability hunting is not theoretical. "Codex Security has contributed to over 3,000 critical and high fixed vulnerabilities," the company wrote, alongside many more lower-severity findings across the software ecosystem. No other outlet covering the launch reported it.
The new model, GPT-5.4-Cyber, is a variant of GPT-5.4 fine-tuned for defensive cybersecurity work. Its key technical shift: it relaxes the refusal boundaries that block general-purpose models from legitimate security tasks, and adds binary reverse engineering capabilities, which let security teams analyze compiled software for malware potential, vulnerabilities, and security weaknesses without requiring the original source code. Security teams call this black-box analysis, and it has historically required specialized, expensive tooling.
OpenAI is also expanding the program behind these numbers. The Trusted Access for Cyber, or TAC, is a vetted community of individual defenders and teams scaling to thousands of participants across hundreds of organizations. That curated network is the commercial reality the launch announcement was actually describing.
The competitive calculus is what makes this an economics story as much as a technology story. OpenAI is building a preferred customer tier for cyber defense, the same way cloud providers built preferred compute tiers. Organizations inside the TAC program get access to models and capabilities that those outside it cannot use. The capability gap widens with every model iteration.
This is not happening in isolation. Anthropic announced its own cyber AI, Mythos, on April 7 under Project Glasswing, one week before GPT-5.4-Cyber, and the two companies are now in what analysts are calling a post-Mythos sprint to spin out specialized variants. Mythos found thousands of major vulnerabilities in operating systems, web browsers, and other software. But independent security researchers at the firm Aisle were able to replicate those findings using older, cheaper, publicly available models. Bruce Schneier, the security researcher and Harvard Kennedy School fellow, noted in a blog post last week that Aisle replicated Anthropic's findings using older, cheaper, public models. The implication: the gap between frontier and commodity in AI security tooling may be narrower than the labs are suggesting.
There is also the structural problem that neither lab is eager to emphasize. CSO Online reported the window between vulnerability discovery and weaponization has collapsed to hours, making existing patch cycles and risk management frameworks obsolete. "Attackers gain disproportionate benefit, and current patch cycles, response processes, and risk metrics were not built for this environment," the outlet noted. Capabilities that once required nation-state resources are becoming broadly accessible. The cost and capability floor is dropping.
The access restriction on GPT-5.4-Cyber, limited rollout to vetted security vendors, organizations, and researchers, reflects this risk calculation. OpenAI is trying to put capable defensive tools in the hands of defenders before capable offensive tools are widely available. But Anthropic made a different choice with Mythos: the company briefed senior US government officials including CISA and Treasury Secretary Scott Bessent on the model's capabilities before its public announcement. The question of who gets access, and who decides, is unresolved.
The 3,000 patched vulnerabilities is a real number from a live system. Whether it represents a durable advantage or a temporary head start depends on whether the capability gap between curated defenders and everyone else can be closed — and nobody has answered that question yet.