Getting started
Installation
Install dependencies for using Zen models
Installation
Requirements
- Python 3.10+
- PyTorch 2.2+
- CUDA 12.1+ (for GPU inference)
Install Transformers
pip install torch transformers accelerateInstall vLLM (Production)
pip install vllmInstall MLX (Apple Silicon)
pip install mlx mlx-lmInstall SGLang
pip install sglang[all]GPU Memory Requirements
| Model | Minimum VRAM | Recommended |
|---|---|---|
| zen-nano | 2GB | 4GB |
| zen-coder-4b | 8GB | 16GB |
| zen-coder-flash | 24GB | 48GB |
| zen-max | 160GB | 320GB |
zen-coder-flash uses MoE architecture with 31B total params but only 3B active, making it efficient for its capability level.