Model Comparison

Magistral Medium vs Mistral Small 3 24B Instruct

Magistral Medium significantly outperforms across most benchmarks.

Performance Benchmarks

Comparative analysis across standard metrics

1 benchmarks

Magistral Medium outperforms in 1 benchmarks (GPQA), while Mistral Small 3 24B Instruct is better at 0 benchmarks.

Magistral Medium significantly outperforms across most benchmarks.

Sun May 17 2026 • llm-stats.com

Arena Performance

Human preference votes

Model Size

Parameter count comparison

0.0M diff

Mistral Small 3 24B Instruct has 0.0B more parameters than Magistral Medium, making it 0.0% larger.

Mistral AI
Magistral Medium
24.0Bparameters
Mistral AI
Mistral Small 3 24B Instruct
24.0Bparameters
24.0B
Magistral Medium
24.0B
Mistral Small 3 24B Instruct

Context Window

Maximum input and output token capacity

Only Mistral Small 3 24B Instruct specifies input context (32,000 tokens). Only Mistral Small 3 24B Instruct specifies output context (32,000 tokens).

Mistral AI
Magistral Medium
Input- tokens
Output- tokens
Mistral AI
Mistral Small 3 24B Instruct
Input32,000 tokens
Output32,000 tokens
Sun May 17 2026 • llm-stats.com

Input Capabilities

Supported data types and modalities

Magistral Medium supports multimodal inputs, whereas Mistral Small 3 24B Instruct does not.

Magistral Medium can handle both text and other forms of data like images, making it suitable for multimodal applications.

Magistral Medium

Text
Images
Audio
Video

Mistral Small 3 24B Instruct

Text
Images
Audio
Video

License

Usage and distribution terms

Both models are licensed under Apache 2.0.

Both models share the same licensing terms, providing consistent usage rights.

Magistral Medium

Apache 2.0

Open weights

Mistral Small 3 24B Instruct

Apache 2.0

Open weights

Release Timeline

When each model was launched

Magistral Medium was released on 2025-06-10, while Mistral Small 3 24B Instruct was released on 2025-01-30.

Magistral Medium is 4 months newer than Mistral Small 3 24B Instruct.

Magistral Medium

Jun 10, 2025

11 months ago

4mo newer
Mistral Small 3 24B Instruct

Jan 30, 2025

1.3 years ago

Knowledge Cutoff

When training data ends

Magistral Medium has a knowledge cutoff of 2025-06-01, while Mistral Small 3 24B Instruct has a cutoff of 2023-10-01.

Magistral Medium has more recent training data (up to 2025-06-01), making it potentially better informed about events through that date compared to Mistral Small 3 24B Instruct (2023-10-01).

Magistral Medium

Jun 2025

1.7 yr newer
Mistral Small 3 24B Instruct

Oct 2023

Outputs Comparison

Notice missing or incorrect data?Start an Issue discussion

Key Takeaways

Supports multimodal inputs
Higher GPQA score (70.8% vs 45.3%)
Larger context window (32,000 tokens)

Detailed Comparison

AI Model Comparison Table
Feature
Mistral AI
Magistral Medium
Mistral AI
Mistral Small 3 24B Instruct

FAQ

Common questions about Magistral Medium vs Mistral Small 3 24B Instruct.

Which is better, Magistral Medium or Mistral Small 3 24B Instruct?

Magistral Medium significantly outperforms across most benchmarks. Magistral Medium is made by Mistral AI and Mistral Small 3 24B Instruct is made by Mistral AI. The best choice depends on your use case — compare their benchmark scores, pricing, and capabilities above.

How does Magistral Medium compare to Mistral Small 3 24B Instruct in benchmarks?

Magistral Medium scores AIME 2024: 73.6%, GPQA: 70.8%, AIME 2025: 64.9%, LiveCodeBench: 50.3%, Aider-Polyglot: 47.1%. Mistral Small 3 24B Instruct scores Arena Hard: 87.6%, HumanEval: 84.8%, MT-Bench: 83.5%, IFEval: 82.9%, MATH: 70.6%.

What are the context window sizes for Magistral Medium and Mistral Small 3 24B Instruct?

Magistral Medium supports an unknown number of tokens and Mistral Small 3 24B Instruct supports 32K tokens. A larger context window lets you process longer documents, conversations, or codebases in a single request.

What are the main differences between Magistral Medium and Mistral Small 3 24B Instruct?

Key differences include multimodal support (yes vs no). See the full comparison above for benchmark-by-benchmark results.