Back to Calculator
GLM-4.6V
Z.ai's VLM with 106B total / 12B active MoE parameters. SOTA visual understanding at its scale, 128K multimodal context, native Function Call support for visual-to-action workflows.
Model Specifications
ArchitectureVISION
Parameters106B
Familyglm
VRAM (Q4)53.0GB
Mixture of ExpertsActive inference parameters: 12B.
50% cheaper API pricing vs GLM-4.5V. Native function calling allows images to trigger APIs directly — key for visual agent pipelines.
Share this Model
Send this model's specs directly to your community.
Similar Models
Related Guides
How much VRAM do you really need?
A complete breakdown of quantization levels and VRAM overhead for running local models.
Best GPUs for Machine Learning in 2026
Comparing NVIDIA and AMD options for the best speed-to-dollar ratio.
GGUF vs EXL2 vs AWQ
Understanding local AI formats and which one to pick for your specific hardware.