Complete pricing breakdown for all 0 Mistral API models. Compare input and output costs per million tokens for Mistral Large, Mistral Medium, Mistral Small, Codestral, and every other model. Includes a cost calculator and side-by-side comparison with OpenAI and Anthropic.
| Model | Input $/1M | Output $/1M |
|---|
See how Mistral API pricing stacks up against OpenAI (GPT) and Anthropic (Claude) models. All prices in USD per million tokens.
| Model | In | Out |
|---|
| Model | In | Out |
|---|---|---|
| gpt-oss-120b (free) | Free | Free |
| gpt-oss-20b (free) | Free | Free |
| Sora | Free | Free |
| gpt-oss-20b | $0.030 | $0.140 |
| gpt-oss-120b | $0.039 | $0.180 |
| gpt-oss-safeguard-20b | $0.075 | $0.300 |
| GPT-5 Nano | $0.050 | $0.400 |
| GPT-4.1 Nano | $0.100 | $0.400 |
| Model | In | Out |
|---|---|---|
| Claude 3 Haiku | $0.250 | $1.25 |
| Claude 3.5 Haiku | $0.800 | $4.00 |
| Claude Haiku 4.5 | $1.00 | $5.00 |
| Claude Sonnet 4.6 | $3.00 | $15.00 |
| Claude Sonnet 4.5 | $3.00 | $15.00 |
| Claude Sonnet 4 | $3.00 | $15.00 |
| Claude 3.7 Sonnet | $3.00 | $15.00 |
| Claude 3.7 Sonnet (thinking) | $3.00 | $15.00 |
Mistral API cost projections at various request volumes. Based on ~1,000 input tokens and ~500 output tokens per request. Self-hosting open-weight models eliminates these per-token costs.
Note: Actual costs vary with prompt length, response length, and batch processing. Mistral offers competitive pricing and batch API discounts for high-volume usage. Try the interactive calculator for custom estimates.
Mistral uses a SentencePiece tokenizer shared across their model family. Token counts are consistent whether you use Mistral Large or Small - only the per-token price changes. Prices are quoted per million tokens. As an open-weight provider, Mistral also lets you self-host models to eliminate per-token API costs entirely for high-volume workloads.
Mistral Large is the most powerful model for complex reasoning, multilingual tasks, and code generation. Mistral Small offers an excellent balance of performance and cost for most production use cases. Codestral is specialized for code completion and generation, supporting 80+ programming languages with optimized performance.
Mistral is known for releasing open-weight models that can be self-hosted. While the API provides managed access with pay-per-token pricing, you can also download and run models like Mistral 7B, Mixtral, and others on your own infrastructure, potentially reducing costs for high-volume workloads.
Start with Mistral Small for simple tasks and only upgrade to Large when the task demands it. Use shorter prompts and set appropriate max_tokens limits to control output costs. For high-volume workloads, consider self-hosting open-weight models or using batch API endpoints for discounted pricing.
Compare with GPT-4o, o3, and all OpenAI model costs.
Compare with Claude Opus 4, Sonnet 4, and all Anthropic models.
Compare with Gemini 2.5 Pro, Flash, and all Google models.
Find the most affordable models across all providers.
Mistral Large pricing is Free/1M input tokens and Free/1M output tokens. Mistral Large is Mistral AI's most capable model, designed for complex reasoning, multilingual tasks, and code generation.
Mistral offers 0 free models via their API. Mistral provides competitive pricing with some of the most affordable models in the industry. Mistral Small is one of the cheapest high-quality options for production workloads, while Mistral Large offers premium capabilities at a fraction of the cost of comparable models from competitors.
Pricing data is currently being updated. Check back soon for the latest Mistral model costs.
Mistral and OpenAI target different price points. Mistral's average output price is Free/1M tokens across 0 paid models. OpenAI offers 67 models with varying price points. Mistral models are generally more affordable than OpenAI equivalents, especially for multilingual and code-generation use cases. For the most accurate comparison, see our side-by-side pricing table above.
Mistral charges per token with pricing that reflects their model hierarchy - from the affordable Mistral Small to the premium Mistral Large. As an open-weight provider, their API pricing competes with self-hosting costs. Codestral is priced separately for code-specific workloads. All models share the same tokenizer, so token counts are consistent across the lineup.
Codestral is Mistral's specialized code generation model, optimized for programming tasks like code completion, refactoring, and generation. It offers competitive pricing compared to general-purpose models and excels at code-related tasks with support for 80+ programming languages. It's available via the Mistral API and through various IDE integrations.