Llama 4 Maverick was released April 5, 2025 as a 17B active parameter mixture-of-experts model (400B total, 128 experts). Scored 92.2% on MMLU, outperforming GPT-4o and Gemini 2.0 Flash. Featured a 1M token context window and native multimodal support. Could be deployed on a single H100 DGX host.
Benchmarks6
Reasoning
1 benchmarkCoding
1 benchmarkPercentile among tested models:Top 25%50-75%25-50%Bottom 25%
Llama Family5
| Model | Tier | Released | Input $/MTok |
|---|---|---|---|
| Llama 4 Scout | — | 2025-04-05 | — |
| Llama 3.3 | — | 2024-12-06 | — |
| Llama 3.1 | — | 2024-07-23 | — |
| Llama 3 | — | 2024-04-18 | — |
| Llama 2 | — | 2023-07-18 | — |
Details
Model FamilyLlama
Generation4
Release Date2025-04-05
Parameters400B
Context Window1M tokens
Open WeightYes
Modalitytext, image
Capabilities2
tool-usevision
Sources1
Tags
llamametaopen-weightmixture-of-expertsfrontier-ai