Back to Calculator0
Magistral Medium
Enterprise-tier version of Magistral reasoning family. More capable than Magistral Small for complex multi-step tasks in math, law, finance, and code. Available via Mistral API and Le Chat; not open-weight.
Model Specifications
ArchitectureTEXT
Parameters-
Familymagistral
VRAM (Q4)Unknown
Share this Model
Send this model's specs directly to your community.
Similar Models
Magistral Small
24BMistral AI's first open-weight reasoning model — 24B parameters under Apache 2.0. Chain-of-thought reasoning in 20+ languages, 70.7% on AIME2024. Fits on a single RTX 4090 or 32GB MacBook.
Llama 3.3 70B
70.55BRefined Llama 3 with superior following
Llama 3.2 3B
3.21BMobile-optimized small model
Related Guides
How much VRAM do you really need?
A complete breakdown of quantization levels and VRAM overhead for running local models.
Best GPUs for Machine Learning in 2026
Comparing NVIDIA and AMD options for the best speed-to-dollar ratio.
GGUF vs EXL2 vs AWQ
Understanding local AI formats and which one to pick for your specific hardware.