Meta's Llama 4 Topped Every Benchmark. Then Yann LeCun Admitted They Fudged It.
Meta released Llama 4 with the kind of benchmark scores that make competitors nervous — beating GPT-4o on LiveCodeBench by 34%, matching DeepSeek v3 on reasoning at half the active parameters, and achieving an LMArena ELO of 1417. Then Yann LeCun, Meta's own chief AI scientist, told the world the results were "fudged a little bit." The version Meta submitted for benchmarks wasn't the version developers could actually download. And when the real model showed up, it ranked 32nd — not 2nd.
This is the story of the most consequential open-source AI release of 2026, and why its biggest scandal might actually be the best thing that happened to it.
What Llama 4 Actually Is (Beneath the Controversy)
Strip away the drama and Llama 4 is genuinely impressive engineering. Meta designed three models using a Mixture-of-Experts (MoE) architecture that only activates a fraction of total parameters per token — a trick that delivers massive capability with manageable compute costs.
Llama 4 Scout: 17 billion active parameters, 16 experts, 109B total parameters. Fits on a single NVIDIA H100 GPU with Int4 quantization. The headline feature is a 10 million token context window — industry-leading by a wide margin. That's roughly 15,000 pages of text in a single prompt. For comparison, Cursor's context window maxes out at 128K tokens.
Llama 4 Maverick: 17 billion active parameters, 128 experts, 400B total parameters. Runs on a single H100 DGX host. This is the workhorse model — the one most developers are using for production workloads. Available through 12+ API providers at prices starting from $0.26 per million tokens (DeepInfra), which undercuts Claude Opus at $15/M tokens by 98%.
Llama 4 Behemoth: 288 billion active parameters, approximately 2 trillion total parameters. Still training when Scout and Maverick shipped. Meta used it as a teacher model to distill knowledge into the smaller models. Early benchmarks show it beating GPT-4.5, Claude Sonnet 3.7, and Gemini 2.0 Pro on STEM tasks (MATH-500, GPQA Diamond).
All three models are natively multimodal — trained from scratch on text, image, and video using early fusion architecture. Pre-trained on 30+ trillion tokens, double what Llama 3 consumed.
The Benchmark Scandal That Rocked Meta's AI Lab
Here's what happened: Meta submitted an "experimental" chat-optimized version of Maverick to LMArena, the industry's most-watched leaderboard. That version — Llama-4-Maverick-03-26-Experimental — hit ELO 1417, ranking 2nd overall. The AI community celebrated. Open-source was winning.
Then developers downloaded the actual open-source model and ran their own benchmarks. The publicly available Maverick scored dramatically lower — dropping to 32nd place. The gap was so large that accusations of benchmark manipulation erupted across Twitter, Hacker News, and Reddit.
Yann LeCun, Meta's departing chief AI scientist, confirmed the worst: "The results were fudged a little bit." Meta had used different model variants for different benchmarks to present the best possible numbers. The version submitted for evaluation wasn't the version anyone could actually use.
CEO Mark Zuckerberg was reportedly "really upset and basically lost confidence in everyone who was involved" in the release. The fallout was severe — LeCun and senior researcher Tian Yuandong left to start their own ventures, and core team members scattered across the industry in what insiders called a "talent exodus."
Meta's official response: they denied training on test sets ("that's simply not true and we would never do that") and attributed the quality gap to implementation bugs that needed stabilization. Whether you believe that framing is up to you.
The Real-World Performance (Post-Controversy)
After the dust settled, independent benchmarks painted a more nuanced picture. The publicly available Maverick is still a strong model — just not the godkiller Meta's marketing implied.
On specific benchmarks:
- MMLU: Maverick scores 85.5, slightly above Llama 3.1 405B (85.2) — impressive given Maverick uses only 17B active parameters vs 405B
- MMMU: Maverick hits 73.4% vs GPT-4o's 69.1% and Gemini 2.0 Flash's 71.7%
- LiveCodeBench: Maverick at 43.4% vs GPT-4o at 32.3% — still a genuine 34% improvement on coding tasks
- Reasoning: Scout trails Maverick by 8-12 points, reflecting the 16 vs 128 expert gap, but excels at long-context retrieval
For complex reasoning (legal analysis, scientific research, medical diagnosis), Claude Opus 4.6 still leads. But for the 80% of AI tasks that are "good enough" workloads — summarization, code generation, data extraction, multimodal understanding — Maverick delivers comparable quality at 2% of the cost.
Why the Scandal Might Be Llama 4's Greatest Asset
Counterintuitive take: the benchmark controversy made Llama 4 more credible, not less. Here's why.
Before the scandal, every AI company's benchmark claims were taken at face value. After Llama 4, the entire industry is under scrutiny. Google, OpenAI, and Anthropic now face pressure to submit their actual production models — not fine-tuned variants — for evaluation. LMArena tightened its submission policies. Third-party benchmark organizations gained influence.
The scandal also forced developers to actually test Llama 4 themselves rather than trusting leaderboard rankings. And when they did, many found that Maverick genuinely performed well for their specific use cases — particularly coding, multimodal tasks, and cost-sensitive deployments. The 10M token context window on Scout has no real competitor in the open-source world.
The result: Llama 4 has more battle-tested, real-world credibility than any open-source model before it. Developers trust it because they verified it themselves, not because a leaderboard told them to.
The Cost Equation: Where Llama 4 Destroys Proprietary Models
Self-hosting Llama 4 cuts inference costs by 60-80% compared to proprietary APIs. Here's the math:
| Model | Cost per 1M Tokens | Context Window | Active Params |
|---|---|---|---|
| Llama 4 Scout (self-host) | $0.15-0.25 | 10M tokens | 17B |
| Llama 4 Maverick (DeepInfra) | $0.26 | 1M tokens | 17B |
| Llama 4 Maverick (Groq) | $0.30 | 1M tokens | 17B |
| GPT-5.3 Instant | $0.60-1.50 | 128K tokens | Unknown |
| Claude Opus 4.6 | $15.00 | 200K tokens | Unknown |
For startups and mid-size companies processing millions of tokens daily, the cost difference is existential. A workload costing $15,000/month on Claude Opus drops to $260/month on Maverick via DeepInfra. Even accounting for quality gaps on complex reasoning, the ROI calculation is overwhelming for many use cases.
The model is available on 12+ API providers including DeepInfra, Groq, Together.ai, Amazon Bedrock, Microsoft Azure, Google Vertex, and SambaNova. Prices vary up to 3.5x across providers — DeepInfra is cheapest at $0.26/M tokens, SambaNova most expensive at $0.92/M tokens.
Who Should Use Llama 4 (And Who Shouldn't)
Use Llama 4 Maverick if:
- You're building AI features where cost matters more than bleeding-edge reasoning (chatbots, summarization, code completion)
- You need multimodal capabilities (text + image + video) in a single model
- You want to self-host for data privacy, compliance, or cost control
- Your use case involves high-volume, lower-complexity tasks
Use Llama 4 Scout if:
- You need massive context windows (legal document analysis, entire codebases, book-length content)
- You're running on limited GPU hardware (single H100)
- Long-context retrieval accuracy is your priority
Don't use Llama 4 if:
- You need the absolute best reasoning for high-stakes decisions (use Claude Opus or GPT-5)
- You're in regulated industries requiring model provider liability (open-source = you own the risk)
- Your team doesn't have ML engineering capacity for deployment and monitoring
What This Means for the AI Industry
Llama 4's real legacy won't be its benchmarks or its scandal. It'll be the pricing pressure it created. When a model that costs $0.26/M tokens delivers 80% of the capability of a $15/M model, the proprietary AI business model starts cracking.
OpenAI has already responded with GPT-5.3 Instant pricing cuts. Anthropic is investing in efficiency improvements for Claude. Google is aggressively expanding Gemini's free tier. The competitive pressure from open-source models is no longer theoretical — it's reshaping the economics of the entire AI industry.
Meta's strategy is clear: give away the model, own the ecosystem. If every AI startup builds on Llama, Meta controls the foundation layer of the next computing platform. The benchmark scandal was embarrassing, but the strategic calculus hasn't changed. Llama 4 is downloaded millions of times per month. Developers are building production systems on it. And Meta's next model — whether it's Llama 4 Behemoth or Llama 5 — will have a massive installed base to upgrade.
The question isn't whether open-source AI will compete with proprietary models. Llama 4 proved it already does. The question is whether the AI industry will develop honest evaluation standards — or whether every major release will come with an asterisk.
Key Takeaways
- ✓Llama 4 Maverick uses 17B active parameters with 128 experts — beats GPT-4o on coding benchmarks at 2% of the cost ($0.26 vs $15/M tokens)
- ✓Meta submitted a different 'experimental' model variant for benchmarks — the public version ranked 32nd, not 2nd, on LMArena
- ✓Yann LeCun confirmed results were 'fudged' — leading to his departure and a talent exodus from Meta's AI lab
- ✓Llama 4 Scout offers an industry-leading 10M token context window that fits on a single H100 GPU
- ✓Self-hosting Llama 4 cuts inference costs 60-80% vs proprietary APIs — a $15K/month Claude workload drops to $260/month
- ✓For complex reasoning, Claude Opus 4.6 still leads — but for 80% of AI workloads, Maverick delivers comparable quality
- ✓The scandal forced the industry toward honest evaluation — LMArena tightened policies and third-party benchmarks gained influence
Skila AI Editorial Team
The Skila AI editorial team researches and writes original content covering AI tools, model releases, open-source developments, and industry analysis. Our goal is to cut through the noise and give developers, product teams, and AI enthusiasts accurate, timely, and actionable information about the fast-moving AI ecosystem.
About Skila AI →Related Resources
Weekly AI Digest
Get the top AI news, tool reviews, and developer insights delivered every week. No spam, unsubscribe anytime.
Join 1,000+ AI enthusiasts. Free forever.