Llama 4 Maverick

MetabalancedOpen Source

Meta's mixture-of-experts model with 17B active parameters and 128 experts. Strong multimodal and multilingual performance.

Context Window

1M

tokens

Max Output

33K

tokens

Input Price

$0.2

per 1M tokens

Output Price

$0.6

per 1M tokens

Speed

135

tokens/sec

Released

Apr 2025

2025-04-05

Blended Cost

$0.40

per 1M tokens

Value Score

200.0

quality per $

Capabilities

ChatVisionCode GenerationFunction Calling

Benchmarks

Quality Index
80
MMLU Pro
86.1
HumanEval (Coding)
85.4
MATH
74.8
Arena ELO
1260

Open Source — Licensed under Llama 4 Community

About Llama 4 Maverick

Llama 4 Maverick is a balanced AI model by Meta, released on April 5, 2025. It supports a context window of 1000K tokens and can generate up to 33K output tokens.

At $0.2 per million input tokens and $0.6 per million output tokens, its blended cost of $0.40/1M tokens makes it one of the most affordable models available. Its value score of 200.0 reflects the balance of quality and cost.

Llama 4 Maverick is available as an open-source model under the Llama 4 Community license, meaning you can self-host it for predictable costs or use it through API providers like Swfte Connect.

Using Llama 4 Maverick with Swfte

Access Llama 4 Maverick through Swfte Connect, our unified LLM gateway. Connect gives you a single API for 50+ models, with automatic routing, cost optimization, and fallback handling. You can also try Llama 4 Maverick in our AI Playground before integrating.