LocalOps
Back to Calculator

GLM Z1 Rumination 32B

Z.ai's deep-reasoning "rumination" model at 32B — designed for extended chain-of-thought with multiple self-reflection passes. Open-source under Apache 2.0.

Model Specifications

ArchitectureTEXT
Parameters32B
Familyglm
VRAM (Q4)16.0GB
Rumination mode enables extended internal reasoning; slower but more thorough than standard Z1-32B.
#zhipu#reasoning#deep-thinking#apache2#open-sourceSource

Estimated Quantization Sizes

FormatPrecisionEst. VRAMRecommendation
FP16 / BF1616-bit64.0 GBUncompressed Base
Q8_0High8-bit32.0 GBNear Lossless
Q6_K6-bit24.0 GBExcellent Balance
Q4_K_MPopular4-bit16.0 GBStandard Use

Share this Model

Send this model's specs directly to your community.

Post