API Provider3 active models1 organizationmistral.ai

Mistral AI: API pricing, performance & models

Mistral AI hosts 3 active AI models, with input pricing from $0.15 per 1M tokens, averaging 46 tok/s output throughput, with up to 262K context window. Compare Mistral AI's API pricing, latency, and feature support against other LLM providers.

First-party only
3Active
Pricing
$0.150/MFrom
$0.325/MAvg
Performance
46tok/sThroughput
1.70sLatency
262KMax

Catalog

Type
Price
5 models
Model
Voxtral Mini
Mistral Small 4
Mistral Small 4
Mistral Large 3 (675B Instruct 2512)
Mistral Large 3 (675B Instruct 2512)

FAQ

Common questions about Mistral AI.

What is Mistral AI?

Mistral AI is an API provider that hosts large language models. Active models: 3; From (input): $0.15 / 1M tok; Avg throughput: 46 tok/s; Avg latency: 1.70 s; Max context: 262K.

How many models does Mistral AI offer?

Mistral AI currently serves 3 active models out of 17 historical offerings on LLM Stats.

What is Mistral AI's API pricing?

Mistral AI input pricing starts from $0.15 per 1M tokens, with the most expensive offering at $0.5 per 1M tokens. See the Pricing tab above for the full per-model breakdown.

How fast is Mistral AI?

Mistral AI averages 46 output tokens per second across its catalog, with average latency of 1.70s. Per-model performance is shown in the Performance tab.

Does Mistral AI support multimodal models?

Yes. Mistral AI's catalog includes 2 vision-capable models. See the Models and Capabilities tabs for the full per-model breakdown.

Whose models does Mistral AI host?

Mistral AI hosts models from Mistral AI. See the Models tab for the full catalog grouped by creator.

How do I start using Mistral AI?

Sign up at https://mistral.ai to get an API key, then call Mistral AI's API directly from your application. Use the Pricing and Performance tabs above to pick the right model for your latency, cost, and context-window requirements.