Meta Llama 3.1 8B Instruct
Meta Llama 3.1 8B Instruct is a text model from Nebius with a context window of 128K tokens and max output of 128K tokens. Pricing starts at 0.02 per million input tokens and 0.06 per million output tokens (cheapest at Vertex AI (Llama)).
Capabilities
✗ Vision✓ Function Calling✗ Reasoning✗ JSON Schema✗ System Messages✗ Web Search✗ Prompt Caching✗ Audio Input✗ Audio Output
Specifications
| Model Key | nebius/meta-llama/Meta-Llama-3.1-8B-Instruct |
| Provider | Nebius |
| Provider ID | nebius |
| Mode | Text |
| Canonical Name | llama-3.1-8b |
| Context Window | 128K tokens |
| Max Output | 128K tokens |
Pricing
| Type | Per 1K Tokens | Per 1M Tokens |
|---|---|---|
| Input Tokens | 0.000020 | 0.020 |
| Output Tokens | 0.000060 | 0.060 |
Price Comparison by Provider
Compare prices for Meta Llama 3.1 8B Instruct across different providers. The same model may be available through multiple providers at different price points.
All Variants
All available versions, regions, and API endpoints for Meta Llama 3.1 8B Instruct.
Model Key | Provider | Mode | Input Price, $ | Output Price, $ | Context | Max Output | Vision | Functions |
|---|---|---|---|---|---|---|---|---|
| meta.llama3-1-8b-instruct-v1:0 | Text | 0.220 | 0.220 | 128K | 2K | no | yes | |
| us.meta.llama3-1-8b-instruct-v1:0 | Text | 0.220 | 0.220 | 128K | 2K | no | yes | |
| azure_ai/Meta-Llama-3.1-8B-Instruct | Text | 0.300 | 0.610 | 128K | 2K | no | no | |
| cerebras/llama3.1-8b | Text | 0.100 | 0.100 | 128K | 128K | no | yes | |
| databricks/databricks-meta-llama-3-1-8b-instruct | Text | 0.150 | 0.450 | 200K | 128K | no | no | |
| deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct | Text | 0.030 | 0.050 | 131K | 131K | no | yes | |
| deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo | Text | 0.020 | 0.030 | 131K | 131K | no | yes | |
| fireworks_ai/accounts/fireworks/models/llama-v3p1-8b-instruct | Text | 0.100 | 0.100 | 16K | 16K | no | no | |
| friendliai/meta-llama-3.1-8b-instruct | Text | 0.100 | 0.100 | 8K | 8K | no | yes | |
| groq/llama-3.1-8b-instant | Text | 0.050 | 0.080 | 128K | 8K | no | yes | |
| hyperbolic/meta-llama/Meta-Llama-3.1-8B-Instruct | Text | 0.120 | 0.300 | 33K | 33K | no | yes | |
| lambda_ai/llama3.1-8b-instruct | Text | 0.025 | 0.040 | 131K | 131K | no | yes | |
| llamagate/llama-3.1-8b | LlamaGate | Text | 0.030 | 0.050 | 131K | 8K | no | yes |
| nebius/meta-llama/Meta-Llama-3.1-8B-Instruct | Nebius | Text | 0.020 | 0.060 | 128K | 128K | no | yes |
| novita/meta-llama/llama-3.1-8b-instruct | Text | 0.020 | 0.050 | 16K | 16K | no | no | |
| nscale/meta-llama/Llama-3.1-8B-Instruct | Text | 0.030 | 0.030 | N/A | N/A | no | no | |
| ollama/llama3.1 | Text | N/A | N/A | 8K | 8K | no | yes | |
| ovhcloud/Llama-3.1-8B-Instruct | Text | 0.100 | 0.100 | 131K | 131K | no | yes | |
| perplexity/llama-3.1-8b-instruct | Text | 0.200 | 0.200 | 131K | 131K | no | no | |
| sambanova/Meta-Llama-3.1-8B-Instruct | Text | 0.100 | 0.200 | 16K | 16K | no | yes | |
| snowflake/llama3.1-8b | Text | N/A | N/A | 128K | 8K | no | no | |
| together_ai/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo | Text | 0.180 | 0.180 | N/A | N/A | no | yes | |
| vercel_ai_gateway/meta/llama-3.1-8b | Text | 0.050 | 0.080 | 131K | 131K | no | yes | |
| vertex_ai/meta/llama-3.1-8b-instruct-maas | Text | N/A | N/A | 128K | 2K | yes | no | |
| wandb/meta-llama/Llama-3.1-8B-Instruct | Text | 0.022 | 0.022 | 128K | 128K | no | no |