LocalOps
Back to Calculator

Llama 4 Scout

Hot

Meta's first open-weight natively multimodal MoE model with 16 experts and an industry-leading 10M token context window

Model Specifications

ArchitectureVISION
Parameters109B
Familyllama
VRAM (Q4)54.5GB
Mixture of ExpertsActive inference parameters: 17B.
Fits on a single H100 GPU with int4 quantization
#meta#moe#multimodal#long-context#trendingSource

Share this Model

Send this model's specs directly to your community.

Post