The number everyone is quoting from the Stanford HAI AI Index 2026 is 77.3 percent — AI agents handling real-world computing tasks, up from 20 percent a year ago. It sounds like the agents have arrived.
The number nobody is quoting is 38.78 percent. That is how well the best AI multi-agent system — multiple AI systems working together in a pipeline — scored on PaperArena, a benchmark designed to test whether AI agents can actually reason about science. On that same test, human PhD experts scored 83.5 percent. The gap between those two numbers — the benchmark AI agents are acing and the one they are failing — is 45 percentage points, and it is the real story in this year's Stanford HAI report.
"Agents are wonderful, but we are still far from a place where we understand how to use them effectively," Yolanda Gil told Nature. Gil is a computer scientist at the University of Southern California and a lead author of the report.
The gap between what AI agents demo and what they actually deliver on scientific reasoning tasks is 45 percentage points. The Stanford HAI AI Index 2026 contains both numbers. Most coverage has focused on the first one. The second one is the story.
PaperArena was built to test the kind of work a PhD student does when they are actually learning their field, not just retrieving facts. It presents AI agents with real scientific papers and tasks requiring synthesis across multiple documents — following experimental logic, identifying methodological weaknesses, interpreting conflicting results.
On the hardest subset of those tasks, the best multi-agent system scored 18.47 percent, according to the PaperArena paper on arXiv. The researchers describe the results as "far below" the human expert baseline. Gemini 2.5 Pro dropped nearly 20 accuracy points from its own average when tasks required harder cross-paper synthesis.
Separate data from ReplicationBench, which tests whether frontier models can replicate scientific findings from astrophysics papers, shows scores below 20 percent, per the Stanford HAI report. On that benchmark, AI cannot yet reliably replicate the science it reads.
This does not mean AI agents are useless for science. SWE-bench, which tests coding tasks, rose from roughly 60 percent to near 100 percent in a single year, according to IEEE Spectrum — a genuine capability jump. Agents are genuinely better at terminal tasks, software debugging, and structured information retrieval than they were 12 months ago. The Stanford HAI report notes that AI mentions in scientific publications have grown 30-fold since 2010, and that AI tools have produced measurable productivity gains in customer support and software development, in the range of 14 to 26 percent.
But scientific reasoning — the kind that requires holding a complex hypothesis in mind while checking it against evidence spread across multiple papers — is not the same as retrieving a GitHub commit or following a known debugging path. The benchmarks are testing different things. The 77.3 percent figure is real. It just does not mean what people are using it to mean.
The pressure here falls on research institutions that have built strategies around AI agents handling multistep scientific workflows — literature review automation, hypothesis generation, experimental design assistance. Those institutions are not seeing the 77.3 percent number in their labs. They are seeing something closer to 38.78 percent on the tasks that actually matter for discovery. A 45-point accuracy gap is not an incremental improvement problem. It is a strategy mismatch.
The employment data adds a second pressure point. Entry-level software developer employment in the United States, for workers aged 22 to 25, has dropped nearly 20 percent since its peak in late 2022, according to the Stanford Digital Economy Lab, as reported by MIT Technology Review — during a period when AI coding tools have seen their largest adoption growth. The SWE-bench gains are real. They are also displacing the workers the gains were supposed to augment. The Stanford HAI report notes that 73 percent of US AI experts are positive about AI job impact, compared with 23 percent of the general public.
Global corporate AI investment reached $581.7 billion in 2025, up 130 percent from the prior year, per the Stanford HAI data. The models themselves are converging — the performance gap between the top US labs and the best Chinese labs has collapsed to roughly 2.7 percentage points as of March 2026. The capabilities are there. The question is whether the benchmarks being cited are the ones that actually tell you whether a given AI agent will do the scientific work you need it to do.
The honest answer, based on PaperArena: probably not yet, at least not for the hard cases. Gil put it plainly in the report's overview. "We are still far from a place where we understand how to use them effectively."
The benchmark euphoria is real. The science is not there yet.