Llama 4 Maverick pricing
Llama 4 Maverick 17B Instruct (128E) is a high-capacity multimodal language model from Meta, built on a mixture-of-experts (MoE) architecture with 128 experts and 17 billion active parameters per forward... This page tracks 2 listings in total. Highlighted lows are $0.150 per million input and $0.600 per million output (see table for which seller matches each).
Pricing across providers
Every row is a seller of Llama 4 Maverick with token pricing we track. The cheapest input in this snapshot is from Openrouter. The bar chart shows the same input and output dollars per million for a quick scan.
| Provider | Input / 1M | Output / 1M | Cached input | Batch |
|---|---|---|---|---|
O Openrouter | $0.150 | $0.600 | — | — |
VA Vercel Ai Gateway | $0.200 | $0.600 | — | — |
Input vs output · per provider
Cost calculator
Pick any of the providers above and type how many tokens you expect per day, week, or year. We turn that into rough dollar totals for Llama 4 Maverick.
Provider
0.015000¢ / req
0.030000¢ / req
Model specifications
Quick spec sheet for Llama 4 Maverick before you dive back into pricing. Reported under Meta.
- Context window
- 131,072 tokens
- Max output
- 8,192 tokens
- Vision (images)
- Yes
- Tool / function calling
- Yes
- Streaming
- No
- Released
- Apr 2025
- Primary provider
- Meta
- Model family
- N/A
Compare Llama 4 Maverick
Open a pair page to see Llama 4 Maverick next to another model with a shared provider matrix. 6 shortcuts below.
Locked
Compare with
Pick a model on both sides.
Popular Llama 4 Maverick comparisons
- Llama 4 Maverick vs Llama 3.1 70B
Compare pricing side by side
- Llama 4 Maverick vs Llama 3.1 8B
Compare pricing side by side
- Llama 4 Maverick vs GPT-4o
Compare pricing side by side
- Llama 4 Maverick vs GPT-4o mini
Compare pricing side by side
- Llama 4 Maverick vs Claude Sonnet 4.6
Compare pricing side by side
- Llama 4 Maverick vs Gemini 2.0 Flash
Compare pricing side by side
Frequently asked questions
Answers pull from the same numbers you see on this page. The short model note from our index: Llama 4 Maverick 17B Instruct (128E) is a high-capacity multimodal language model from Meta, built on a mixture-of-experts (MoE) architecture with 128 experts and 17 billion active parameters per forward...
Also from Meta
Other models by Meta with live pricing in our catalog.