Google DeepMind dropped Gemma 4 on April 2, and the benchmarks are exactly what you'd expect from a company that has been doing this for a while. The 31-billion-parameter dense model scores 85.2 percent on MMLU-Pro (a multi-discipline reasoning test), 89.2 percent on the AIME 2026 math competition, and 84.3 percent on GPQA Diamond, according to the Hugging Face blog. Google's own announcement puts it at number three on the Arena AI open-weight leaderboard. The 26-billion-parameter mixture-of-experts variant comes in at number six. The four-model family spans from a 2.3-billion-effective-parameter edge model to the 31B dense flagship, all under the same roof.
But the number worth looking at is not on any benchmark. It is a license header.
Gemma 4 is the first Gemma release under Apache 2.0, replacing a custom Google license that legal teams had learned to dread. Ars Technica described the old terms: Google could update prohibited uses unilaterally, required developers to enforce those rules across every derivative project, and could be read to transfer rights over AI models trained on synthetic data produced by Gemma outputs. That is not a foundation you build a product on.
Apache 2.0 removes all of it. There is no termination clause tied to acceptable use. There is no requirement that downstream users enforce Google's policy preferences. There is no ambiguous claim over synthetic-data derivatives. What remains is the Open Source Initiative's standard commercial license, the same one under which Meta released Llama 3. Enterprise legal teams can read it in fifteen minutes and know exactly what they are allowed to do.
Google confirmed in its blog post that developers have downloaded Gemma more than 400 million times across prior generations, building over 100,000 community variants. That is a base that has been waiting for precisely this license clarity before betting production workloads on it.
The timing matters because the open-weight landscape is bifurcating. As some Chinese AI labs have begun pulling back from fully open releases for their latest models, Google is moving in the opposite direction. Alibaba Qwen 3.5 Omni and Qwen 3.6 Plus have restricted access terms that did not exist in earlier Qwen releases. Zhipu AI's GLM-5 and Moonshot AI's Kimi K2.5 are widely used but carry licensing terms that give enterprise legal departments pause. The Register described it as an onslaught of open-weights Chinese large language models that now rival GPT-5 or Claude, but with restrictions their predecessors did not have. Google is betting that legal clarity is a competitive advantage in a market where Qwen's benchmark scores are comparable or better.
Because the benchmarks do tell a more complicated story. Gemma 4 31B falls slightly behind Qwen 3.5, GLM-5, and Kimi K2.5 on shared benchmarks. The edge models make up for smaller size through Per-Layer Embeddings, a technique that adds a parallel conditioning pathway allowing effective parameters to shrink well below total parameter count. The E2B model has 5.1 billion actual parameters but only 2.3 billion effective ones, achieving 37.5 percent on AIME 2026 versus the 31B's 89.2 percent — a gap that reflects real capability differences, not quantization tricks.
What Gemma 4 does have that matters for actual deployment: the 26B MoE fits on a single 80-gigabyte H100 GPU at bfloat16 precision, and at 4-bit quantization it runs on a 24-gigabyte consumer card like an Nvidia RTX 4090 or AMD RX 7900 XTX. That is a workstation, not a data center rack. For developers who cannot afford cloud compute bills, that changes what is actually runnable.
Nathan Lambert at Interconnects put the real competitive question on the record: Gemma 4's success is going to be entirely determined by ease of use, to a point where a 5-10 percent swing on benchmarks wouldn't matter at all. The tooling ecosystem — fine-tuning pipelines, quantization guides, community forks, integration libraries — is where Qwen currently has the head start. Google knows this. The Apache 2.0 switch is an explicit bet that removing the legal tax on adoption will let the Gemmaverse close that gap on its own merits rather than navigating a license that could change next quarter.
Whether that bet pays off depends on whether the developer community actually shows up. The 400 million downloads across prior Gemma generations suggest there is demand. The 100,000 community variants suggest there is already a culture of building on Google open weights. The question is whether legal clarity is enough to flip that interest into the kind of sustained, production-grade ecosystem that Llama built. Google has removed the obstacle it could control. Everything else is competition on actual merits.