Llama 4 Maverick

Llama 4 Maverick is a text model from Vercel AI Gateway with a context window of 131K tokens and max output of 8K tokens. Pricing starts at 0.20 per million input tokens and 0.60 per million output tokens.

Capabilities

Vision Function Calling Reasoning JSON Schema System Messages Web Search Prompt Caching Audio Input Audio Output

Specifications

Model Keyvercel_ai_gateway/meta/llama-4-maverick
ProviderVercel AI Gateway
Provider IDvercel_ai_gateway
ModeText
Canonical Namellama-maverick-4
Context Window131K tokens
Max Output8K tokens

Pricing

TypePer 1K TokensPer 1M Tokens
Input Tokens0.0002000.200
Output Tokens0.0006000.600

Benchmarks

Intelligence Index18.4#100
Coding Index15.6#94
Math Index19.3#102
MMLU-Pro0.8#42
GPQA0.7#75
HLE0.0#113
LiveCodeBench0.4#81
AIME0.4#40
IFBench0.4#58
Time to First Token0.53s#137
SciCode0.3#82
MATH-5000.9#44
AIME 20250.2#102
LCR0.5#57
TerminalBench Hard0.1#87
TAU20.2#129