Llama 3.1 8B Instruct Maas
Llama 3.1 8B Instruct Maas is a text model from
Vertex AI (Llama) with a context window of 128K tokens and max output of 2K tokens.
Capabilities
✓ Vision✗ Function Calling✗ Reasoning✗ JSON Schema✓ System Messages✗ Web Search✗ Prompt Caching✗ Audio Input✗ Audio Output
Specifications
| Model Key | vertex_ai/meta/llama-3.1-8b-instruct-maas |
| Provider | |
| Provider ID | vertex_ai-llama_models |
| Mode | Text |
| Canonical Name | llama-3.1-8b |
| Context Window | 128K tokens |
| Max Output | 2K tokens |
Pricing
| Type | Per 1K Tokens | Per 1M Tokens |
|---|---|---|
| Input Tokens | N/A | N/A |
| Output Tokens | N/A | N/A |
Price Comparison by Provider
Compare prices for Llama 3.1 8B Instruct Maas across different providers. The same model may be available through multiple providers at different price points.
All Variants
All available versions, regions, and API endpoints for Llama 3.1 8B Instruct Maas.
Model Key | Provider | Mode | Input Price, $ | Output Price, $ | Context | Max Output | Vision | Functions |
|---|---|---|---|---|---|---|---|---|
| meta.llama3-1-8b-instruct-v1:0 | Text | 0.220 | 0.220 | 128K | 2K | no | yes | |
| us.meta.llama3-1-8b-instruct-v1:0 | Text | 0.220 | 0.220 | 128K | 2K | no | yes | |
| azure_ai/Meta-Llama-3.1-8B-Instruct | Text | 0.300 | 0.610 | 128K | 2K | no | no | |
| cerebras/llama3.1-8b | Text | 0.100 | 0.100 | 128K | 128K | no | yes | |
| databricks/databricks-meta-llama-3-1-8b-instruct | Text | 0.150 | 0.450 | 200K | 128K | no | no | |
| deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct | Text | 0.030 | 0.050 | 131K | 131K | no | yes | |
| deepinfra/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo | Text | 0.020 | 0.030 | 131K | 131K | no | yes | |
| fireworks_ai/accounts/fireworks/models/llama-v3p1-8b-instruct | Text | 0.100 | 0.100 | 16K | 16K | no | no | |
| friendliai/meta-llama-3.1-8b-instruct | Text | 0.100 | 0.100 | 8K | 8K | no | yes | |
| groq/llama-3.1-8b-instant | Text | 0.050 | 0.080 | 128K | 8K | no | yes | |
| hyperbolic/meta-llama/Meta-Llama-3.1-8B-Instruct | Text | 0.120 | 0.300 | 33K | 33K | no | yes | |
| lambda_ai/llama3.1-8b-instruct | Text | 0.025 | 0.040 | 131K | 131K | no | yes | |
| llamagate/llama-3.1-8b | LlamaGate | Text | 0.030 | 0.050 | 131K | 8K | no | yes |
| nebius/meta-llama/Meta-Llama-3.1-8B-Instruct | Nebius | Text | 0.020 | 0.060 | 128K | 128K | no | yes |
| novita/meta-llama/llama-3.1-8b-instruct | Text | 0.020 | 0.050 | 16K | 16K | no | no | |
| nscale/meta-llama/Llama-3.1-8B-Instruct | Text | 0.030 | 0.030 | N/A | N/A | no | no | |
| ollama/llama3.1 | Text | N/A | N/A | 8K | 8K | no | yes | |
| ovhcloud/Llama-3.1-8B-Instruct | Text | 0.100 | 0.100 | 131K | 131K | no | yes | |
| perplexity/llama-3.1-8b-instruct | Text | 0.200 | 0.200 | 131K | 131K | no | no | |
| sambanova/Meta-Llama-3.1-8B-Instruct | Text | 0.100 | 0.200 | 16K | 16K | no | yes | |
| snowflake/llama3.1-8b | Text | N/A | N/A | 128K | 8K | no | no | |
| together_ai/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo | Text | 0.180 | 0.180 | N/A | N/A | no | yes | |
| vercel_ai_gateway/meta/llama-3.1-8b | Text | 0.050 | 0.080 | 131K | 131K | no | yes | |
| vertex_ai/meta/llama-3.1-8b-instruct-maas | Text | N/A | N/A | 128K | 2K | yes | no | |
| wandb/meta-llama/Llama-3.1-8B-Instruct | Text | 0.022 | 0.022 | 128K | 128K | no | no |