Back to CalculatorDeploy Now
nanowhale
A 100M parameter language model from HuggingFaceTB.
Specifications
SourceArchitectureLLM
Parameters100B
Familynanowhale
VRAM (Q4)50.0G
text-generationsmallhuggingfacetb
Build your Local Rig
Ready to run locally? Shop top-tier GPUs on Amazon for the best performance.
Instant Cloud GPUs
Running out of VRAM? Rent a high-end H100 or RTX 4090 on RunPod and deploy in seconds.
Quantization Estimates
| Format | VRAM Need | Tier |
|---|---|---|
| FP16 | 200.0 GB | Full Precision |
| Q8_0 | 100.0 GB | High |
| Q6_K | 85.0 GB | Excellent |
| Q5_K_M | 70.0 GB | Great |
| Q4_K_M | 50.0 GB | Sweet Spot |
| Q2_K | 30.0 GB | Emergency |
Share this Model
Send these specs directly to your community.
Similar Models
No similar models found.