AI Model Pricing Directory
Browse pricing for 40+ AI models including GPT-5, Claude, Gemini, DeepSeek, Llama & more. Compare input/output costs, context windows, and capabilities.
Last updated: 2026-03-31 · 42 models from 12 providers
42
Models
12
Providers
$0.04
Cheapest Input/M
2.0M
Max Context
Need to estimate costs?
Use our interactive Pricing Calculator to compare models side-by-side and estimate monthly costs.
OpenAI
11 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| GPT-5.4 Latest flagship iteration with improved reasoning and computer use. | $2.50 | $15.00 | 1.1M |
| GPT-5 | $1.25 | $10.00 | 400K |
| GPT-5 Mini | $0.25 | $2.00 | 400K |
| GPT-5 Nano | $0.05 | $0.40 | 128K |
| GPT-4o | $2.50 | $10.00 | 128K |
| GPT-4o mini | $0.15 | $0.60 | 128K |
| o3-pro Highest reasoning capability for elite tasks. | $20.00 | $80.00 | 200K |
| o3 Standard reasoning model. | $2.00 | $8.00 | 200K |
| o3-mini Fast reasoning model for coding. | $1.10 | $4.40 | 200K |
| o1 Legacy reasoning model. | $15.00 | $60.00 | 200K |
| GPT-5.3-Codex Agentic coding model optimized for code generation and multi-step tasks. 25% faster than GPT-5.2 with 200K context window and 32K max output. Best for autonomous coding workflows. | $2.00 | $10.00 | 200K |
Anthropic
3 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Claude Opus 4.6 Most capable model. 1M context window is currently beta for eligible orgs. 80.8% SWE-bench. | $5.00 | $25.00 | 1.0M |
| Claude Sonnet 4.6 1M context window is currently beta for eligible orgs. 79.6% SWE-bench. Best value flagship. | $3.00 | $15.00 | 1.0M |
| Claude Haiku 4.5 Fastest Claude model. 200K context, 64K max output. | $1.00 | $5.00 | 200K |
| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Gemini 3.1 Pro Flagship reasoning score among Google models. 2M+ token context. Pricing >200K: $4 input, $18 output. | $2.00 | $12.00 | 2.0M |
| Gemini 3.1 Flash-Lite Fastest Gemini model. Optimized for high-throughput, multimodal tasks. | $0.25 | $1.50 | 1.0M |
| Gemini 2.5 Pro Long context >200K: $2.50 input, $15.00 output per 1M. | $1.25 | $10.00 | 2.0M |
| Gemini 2.5 Flash | $0.30 | $2.50 | 1.0M |
| Gemini 2.5 Flash-Lite | $0.10 | $0.40 | 1.0M |
| Gemini 2.0 Flash Deprecated. Scheduled for shutdown on June 1, 2026. | $0.10 | $0.40 | 1.0M |
xAI
4 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Grok 3 | $3.00 | $15.00 | 131K |
| Grok 3 Mini | $0.30 | $0.50 | 131K |
| Grok 4 xAI flagship reasoning model. Strong performance on benchmarks with 256K context. | $3.00 | $15.00 | 256K |
| Grok 4 Fast Fast reasoning variant of Grok 4 with 2M context window. Extremely cost-efficient at $0.20/M input. | $0.20 | $0.50 | 2.0M |
Meta (via providers)
2 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Llama 3.3 70B Open-source. Pricing varies by hosting provider. Shown: typical cloud API price. | $0.88 | $0.88 | 128K |
| Llama 3.1 405B Open-source. Pricing varies by hosting provider. | $3.50 | $3.50 | 128K |
Mistral
3 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Mistral Large 3 | $2.00 | $6.00 | 128K |
| Mistral Medium 3 | $1.00 | $3.00 | 128K |
| Mistral Small 3.1 | $0.20 | $0.60 | 128K |
DeepSeek
3 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| DeepSeek V3.2 | $0.27 | $1.10 | 128K |
| DeepSeek R1 Reasoning model with chain-of-thought. | $0.55 | $2.19 | 128K |
| DeepSeek V4 1T params, hybrid reasoning + generation. Open-weight. | $0.30 | $0.50 | 1.0M |
Moonshot AI
1 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Kimi K2.5 76.8% SWE-bench. Strongest open-source model. Agent Swarm support. | $0.60 | $2.00 | 128K |
Alibaba
2 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Qwen 2.5 72B Open-source. Competitive with Llama 3.3 70B. | $0.40 | $1.20 | 128K |
| Qwen 2.5 Coder 32B Code-specialized. Open-source. | $0.20 | $0.60 | 128K |
Cohere
2 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Command R+ Enterprise RAG-optimized with citation grounding. | $2.50 | $10.00 | 128K |
| Command R | $0.15 | $0.60 | 128K |
AI21 Labs
2 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Jamba 1.5 Large Mamba-based SSM architecture. 256K context. | $2.00 | $8.00 | 256K |
| Jamba 1.5 Mini | $0.20 | $0.40 | 256K |
Amazon
3 models| Model | Input $/M | Output $/M | Context |
|---|---|---|---|
| Amazon Nova Pro Via AWS Bedrock. 300K context. | $0.80 | $3.20 | 300K |
| Amazon Nova Lite | $0.06 | $0.24 | 300K |
| Amazon Nova Micro Text-only. Lowest cost option on Bedrock. | $0.04 | $0.14 | 128K |