Back to Calculator
Llama 4 Maverick
HotMeta's flagship open-weight MoE with 128 experts and 1M context, distilled from Llama 4 Behemoth
Model Specifications
ArchitectureVISION
Parameters400B
Familyllama
VRAM (Q4)200.0GB
Mixture of ExpertsActive inference parameters: 17B.
Available in BF16 and FP8 quantized weights
Share this Model
Send this model's specs directly to your community.
Similar Models
Related Guides
How much VRAM do you really need?
A complete breakdown of quantization levels and VRAM overhead for running local models.
Best GPUs for Machine Learning in 2026
Comparing NVIDIA and AMD options for the best speed-to-dollar ratio.
GGUF vs EXL2 vs AWQ
Understanding local AI formats and which one to pick for your specific hardware.