Llama 3.3 70B
Meta (via providers)Llama 3.3 70B by Meta (via providers) costs $0.88/M input, $0.88/M output with 128K context window. Updated February 2026. Compare with GPT-5, Claude, Gemini & 40+ models.
Input Price
$0.88
per 1M tokens
Output Price
$0.88
per 1M tokens
Context Window
128K
tokens
Specifications
| Provider | Meta (via providers) |
| Model ID | llama-3-3-70b |
| Input Price | $0.88 / 1M tokens |
| Output Price | $0.88 / 1M tokens |
| Context Window | 128K tokens |
| Max Output | 4K tokens |
| Capabilities | textfunction_calling |
| Release Date | 2024-12 |
| Notes | Open-source. Pricing varies by hosting provider. Shown: typical cloud API price. |
Monthly Cost Estimates
Estimated monthly costs based on different daily usage levels (assuming 50% input / 50% output split).
| Daily Tokens | Monthly Cost | Annual Cost |
|---|---|---|
| 10K | $0.26 | $3.17 |
| 50K | $1.32 | $15.84 |
| 100K | $2.64 | $31.68 |
| 500K | $13.20 | $158.40 |
| 1.0M | $26.40 | $316.80 |
About Llama 3.3 70B
Llama 3.3 70B is a large language model by Meta (via providers). It features a 128K token context window with up to 4K tokens of output per request. The model supports 2 capabilities: text, function_calling.
At $0.88 per million input tokens and $0.88 per million output tokens, Llama 3.3 70B is positioned as a cost-effective option in the Meta (via providers) lineup. Use our Token Counter to estimate how many tokens your prompts use, and our Pricing Calculator to compare costs across all models.
Other Meta (via providers) Models
Similar Price Range
Related Tools
FAQ
How much does Llama 3.3 70B cost?
Llama 3.3 70B costs $0.88 per million input tokens and $0.88 per million output tokens. For a typical workload of 100K tokens/day, expect approximately $3.96/month.
What is Llama 3.3 70B's context window?
Llama 3.3 70B supports a context window of 128K tokens. This means your combined input prompt and output response can be up to 128K tokens. The maximum output per response is 4K tokens.
Is Llama 3.3 70B good for my use case?
Llama 3.3 70B supports text, function_calling. As a budget-friendly model, it works well for high-volume tasks like classification, summarization, and simple generation. Use our Pricing Calculator to compare with alternatives.