Complete pricing breakdown for all 0 Mistral API models. Compare input and output costs per million tokens for Mistral Large, Mistral Medium, Mistral Small, Codestral, and every other model. Includes a cost calculator and side-by-side comparison with OpenAI and Anthropic.
| Model | Input $/1M | Output $/1M |
|---|
See how Mistral API pricing stacks up against OpenAI (GPT) and Anthropic (Claude) models. All prices in USD per million tokens.
| Model | In | Out |
|---|
| Model | In | Out |
|---|---|---|
| gpt-oss-120b (free) | Free | Free |
| gpt-oss-20b (free) | Free | Free |
| Sora | Free | Free |
| gpt-oss-20b | $0.030 | $0.140 |
| gpt-oss-120b | $0.039 | $0.190 |
| gpt-oss-safeguard-20b | $0.075 | $0.300 |
| GPT-5 Nano | $0.050 | $0.400 |
| GPT-4.1 Nano | $0.100 | $0.400 |
| Model | In | Out |
|---|---|---|
| Claude 3 Haiku | $0.250 | $1.25 |
| Claude 3.5 Haiku | $0.800 | $4.00 |
| Claude Haiku 4.5 | $1.00 | $5.00 |
| Claude Sonnet 4.6 | $3.00 | $15.00 |
| Claude Sonnet 4.5 | $3.00 | $15.00 |
| Claude Sonnet 4 | $3.00 | $15.00 |
| Claude 3.7 Sonnet | $3.00 | $15.00 |
| Claude 3.7 Sonnet (thinking) | $3.00 | $15.00 |
Estimated daily and monthly costs for common usage patterns. Assumes an average of ~1,000 input tokens and ~500 output tokens per request.
Note: Actual costs vary with prompt length, response length, and batch processing. Mistral offers competitive pricing and batch API discounts for high-volume usage. Try the interactive calculator for custom estimates.
Mistral charges per token, not per request. A token is roughly 3/4 of a word. The sentence "Hello, how are you?" is about 6 tokens. Prices are quoted per million tokens. Input tokens (your prompts) are cheaper than output tokens (the model's response) because output generation requires more computation.
Mistral Large is the most powerful model for complex reasoning, multilingual tasks, and code generation. Mistral Small offers an excellent balance of performance and cost for most production use cases. Codestral is specialized for code completion and generation, supporting 80+ programming languages with optimized performance.
Mistral is known for releasing open-weight models that can be self-hosted. While the API provides managed access with pay-per-token pricing, you can also download and run models like Mistral 7B, Mixtral, and others on your own infrastructure, potentially reducing costs for high-volume workloads.
Start with Mistral Small for simple tasks and only upgrade to Large when the task demands it. Use shorter prompts and set appropriate max_tokens limits to control output costs. For high-volume workloads, consider self-hosting open-weight models or using batch API endpoints for discounted pricing.
Compare with GPT-4o, o3, and all OpenAI model costs.
Compare with Claude Opus 4, Sonnet 4, and all Anthropic models.
Compare with Gemini 2.5 Pro, Flash, and all Google models.
Find the most affordable models across all providers.
Mistral Large pricing is Free/1M input tokens and Free/1M output tokens. Mistral Large is Mistral AI's most capable model, designed for complex reasoning, multilingual tasks, and code generation.
Mistral offers 0 free models via their API. Mistral provides competitive pricing with some of the most affordable models in the industry. Mistral Small is one of the cheapest high-quality options for production workloads, while Mistral Large offers premium capabilities at a fraction of the cost of comparable models from competitors.
Pricing data is currently being updated. Check back soon for the latest Mistral model costs.
Mistral and OpenAI target different price points. Mistral's average output price is Free/1M tokens across 0 paid models. OpenAI offers 62 models with varying price points. Mistral models are generally more affordable than OpenAI equivalents, especially for multilingual and code-generation use cases. For the most accurate comparison, see our side-by-side pricing table above.
Mistral charges per token for API usage. Tokens are the basic units of text -- roughly 3/4 of a word. Pricing is split into input tokens (your prompts and context) and output tokens (the model's response). Output tokens are typically more expensive because they require more computation. Prices are quoted per million tokens.
Codestral is Mistral's specialized code generation model, optimized for programming tasks like code completion, refactoring, and generation. It offers competitive pricing compared to general-purpose models and excels at code-related tasks with support for 80+ programming languages. It's available via the Mistral API and through various IDE integrations.