Trinity Nano
Arcee AI's smallest Trinity model. 6B sparse MoE with 1B active params per token. Fully post-trained for web apps and agent tasks. Apache 2.0.
Specifications
SourceBuild your Local Rig
Ready to run locally? Shop top-tier GPUs on Amazon for the best performance.
Instant Cloud GPUs
Running out of VRAM? Rent a high-end H100 or RTX 4090 on RunPod and deploy in seconds.
Quantization Estimates
| Format | VRAM Need | Tier |
|---|---|---|
| FP16 | 12.0 GB | Full Precision |
| Q8_0 | 6.0 GB | High |
| Q6_K | 5.1 GB | Excellent |
| Q5_K_M | 4.2 GB | Great |
| Q4_K_M | 3.0 GB | Sweet Spot |
| Q2_K | 1.8 GB | Emergency |
Share this Model
Send these specs directly to your community.
Similar Models
Trinity Mini
26BArcee AI's 26B sparse MoE with 3B active params per token. Fully post-trained reasoning model for web apps and agentic tasks. Apache 2.0.
Trinity Large
400BArcee AI's frontier 400B sparse MoE with 13B active params per token. One of the largest Apache 2.0 models ever from a US lab. Excels at creative writing, chat, tool use, and long-horizon agentic workflows.
Trinity Large Thinking
398BArcee AI's reasoning-focused Trinity variant. 400B sparse MoE, 13B active per token, designed for complex multi-step reasoning, math, and long-horizon agents. Apache 2.0.