Anthropic announced a $100 million initiative to use frontier AI for open-source security. The announcement described Mythos, an unreleased model, autonomously finding thousands of zero-day vulnerabilities including a 27-year-old bug in OpenBSD and a 16-year-old bug in FFmpeg. What the announcement did not say: the bottleneck is not finding vulnerabilities. It is processing them.
The Mythos research post (Anthropic) describes a model that achieved 595 crashes at tiers 1 and 2 of the OSS-Fuzz benchmark, reached full control flow hijack on ten fully patched targets at tier 5, and in one case chained together four browser vulnerabilities with a JIT heap spray to escape both renderer and OS sandboxes. Expert contractors hired by Anthropic agreed with Mythos's severity assessments exactly in 89 percent of 198 manually reviewed vulnerability reports. Fewer than 1 percent of vulnerabilities found by Mythos have been patched by maintainers, because the responsible disclosure and triage process is long and labor-intensive. These results are real. The question is what they mean.
Independent researchers at AISLE, who have been running an AI vulnerability discovery system against live targets since mid-2025 and have discovered 180 validated CVEs across OpenSSL, curl, and other critical infrastructure, tested the Mythos claim directly. According to their published results, they took the specific vulnerabilities Anthropic showcased, isolated the relevant code, and ran it through a range of small, open-weights models. Eight out of eight models detected the FreeBSD NFS stack buffer overflow, including a mixture-of-experts model with only 3.6 billion active parameters costing $0.11 per million tokens. A separate 5.1 billion active parameter model recovered the full chain of the OpenBSD SACK vulnerability that Mythos identified, matching its severity assessment.
The finding is not that Mythos is unimpressive. It is that detection capability is commoditizing faster than the infrastructure to act on it. AISLE's own production data illustrates the gap. According to their LessWrong post, they discovered 12 out of 12 zero-day vulnerabilities in an OpenSSL security release in January 2026, and 13 out of 14 across all of 2025. The OpenSSL CTO Tomas Mraz said in a statement carried in that post: "We appreciate the high quality of the reports and their constructive collaboration throughout the remediation." The contrast with what happened to curl's bug bounty program is instructive.
Daniel Stenberg, the maintainer of curl, cancelled the program's bug bounty on January 31, 2026, citing an unmanageable flood of submissions. Approximately 20 percent were AI-generated noise, and only about 5 percent of 2025 submissions were genuine vulnerabilities. The small curl security team could not keep pace with the volume. Stenberg wrote that he hoped ending the bounty would "remove the incentive for people to submit crap and non-well researched reports." The program had paid out over $90,000 for 81 genuine vulnerabilities since 2019. It was killed by noise, not by lack of signal.
Anthropic's response is Project Glasswing, announced April 7, 2026. The consortium includes Amazon Web Services, Apple, Broadcom, Cisco, CrowdStrike, Google, JPMorganChase, the Linux Foundation, Microsoft, Nvidia, and Palo Alto Networks. Anthropic committed up to $100 million in usage credits for Mythos Preview and $4 million in direct donations to open-source security organizations. The premise: if AI can find vulnerabilities faster than human researchers, the solution is to give defenders better AI.
The premise is correct. The implementation may not be the bottleneck. The Glasswing announcement describes Mythos's scaffold in detail: launch a container, prompt the model to scan files, hypothesize and test vulnerabilities, use AddressSanitizer as a crash oracle, rank files by attack surface, run validation. AISLE has described a similar architecture. Their results suggest that the same detection quality can be achieved with smaller, cheaper models, and that the differentiator is not the model but the targeting, the iterative deepening, the validation, and the maintainer trust that determines whether a report results in an accepted patch.
That last part is where the gap lives. AISLE discovered 12 out of 12 OpenSSL vulnerabilities and earned a statement from the CTO praising their quality and collaboration. Stenberg's team was drowning in noise from a different class of AI-generated submissions, including some that AISLE had reported to curl, where the team had found five genuine CVEs. The distinction is not whether AI found the bug. It is whether the report arrives in a form a human maintainer can use, at a pace they can absorb, with enough context to trust and act on it.
The discovery capability is commoditizing. The triage and remediation infrastructure is not. Project Glasswing is building on the right problem. Whether it is building on the right part of the problem is an open question.
What happens next is a test of whether the open-source security ecosystem can absorb what AI can now produce. If every major software project is flooded with high-quality vulnerability reports from multiple AI systems, the bottleneck shifts from discovery to patch review, disclosure coordination, and maintainer capacity. The same dynamic that killed curl's bug bounty, but at scale. Glasswing's $4 million in direct donations to open-source security organizations is small relative to the labor required. The more consequential question is whether the consortium's members treat maintainer acceptance as the metric, not just vulnerabilities found.
As Anthropic noted in its own post, ten years after the DARPA Cyber Grand Challenge, frontier AI is becoming competitive with the best humans at vulnerability discovery. The transition period, as they described it, may be tumultuous. The tumultuous part is not the discovery. It is everything after.