← Back to Blog

Zen 3.0: The Next Generation of Open AI

Announcing Zen 3.0, our most capable open model family yet.

By Zach Kelling
AnnouncementModels

Today we release Zen 3.0, our third-generation language model family. Zen 3 represents a step change in what open models can do.

Model Family

Zen 3 comes in several sizes:

ModelParametersContextTraining Tokens
Zen-3-8B8.1B128K15T
Zen-3-32B32.5B128K12T
Zen-3-72B72.3B128K10T
Zen-3-MoE141B (24B active)128K14T

All models use the same architecture with scaled dimensions. All are released under Apache 2.0.

Architecture Highlights

Extended Context

All Zen 3 models support 128K token context natively:

Long context isn’t just about the number; it’s about actually using it. Our needle-in-haystack evaluation shows >95% retrieval accuracy at 100K tokens.

Mixture of Experts

Zen-3-MoE uses a sparse architecture:

Expert parallelism enables efficient inference on consumer hardware.

Improved Tokenizer

The Zen 3 tokenizer improves on previous versions:

Larger vocabulary means fewer tokens per document means longer effective context.

Capability Improvements

Benchmarks

BenchmarkZen-2-70BZen-3-72BImprovement
MMLU74.282.1+7.9
GSM8K68.484.7+16.3
HumanEval58.571.3+12.8
HellaSwag85.189.4+4.3
MATH32.651.2+18.6

The improvements are substantial across all categories. Math and coding see the largest gains.

Real-World Tasks

Benchmarks don’t tell the whole story. Zen 3 excels at:

Long-form writing : Coherent documents spanning thousands of words with consistent style and structure.

Multi-step reasoning : Complex problems requiring planning and backtracking.

Code generation : Full functions and classes, not just snippets.

Instruction following : Precise adherence to formatting and constraint requirements.

Multilingual : Strong performance in 30+ languages including low-resource ones.

Agentic Capabilities

Zen 3 is designed for agent use cases:

Early agent benchmarks show 2x improvement over Zen 2 on multi-step tasks.

Training Details

Data

Training data evolved significantly:

Total: 15T tokens for the 8B model, proportionally less for larger models.

Training Process

Training used the Zoo Compute Network:

The training run was the largest yet on the decentralized network. It validated that frontier training is possible without centralized infrastructure.

Alignment

Post-training alignment followed our standard process:

  1. Supervised fine-tuning : 100K high-quality instruction examples
  2. GRPO : Group Relative Policy Optimization on preference data
  3. Constitutional training : Principle-based refinement
  4. Red teaming : Adversarial testing with remediation

Alignment reduced benchmark scores slightly (2-3%) while significantly improving real-world usefulness.

Safety Evaluation

All Zen 3 models passed our safety evaluation suite:

Refusal Rates

CategoryZen-2-70BZen-3-72B
Violence instructions99.2%99.7%
CSAM100%100%
Malware97.8%99.1%
PII extraction94.6%98.3%

Improved refusal with fewer false positives on legitimate requests.

Bias Metrics

We evaluated on standard bias benchmarks:

Improvements through both data curation and RLHF.

Limitations

Zen 3 is not perfect:

We publish these limitations because transparency enables responsible use.

Usage

Hugging Face


    from transformers import AutoModelForCausalLM, AutoTokenizer
    
    model = AutoModelForCausalLM.from_pretrained(
        "zoo-labs/zen-3-72b",
        torch_dtype=torch.bfloat16,
        device_map="auto",
    )
    tokenizer = AutoTokenizer.from_pretrained("zoo-labs/zen-3-72b")
    
    output = model.generate(
        tokenizer("Hello, Zen!", return_tensors="pt").input_ids,
        max_new_tokens=100,
    )
    

vLLM


    from vllm import LLM, SamplingParams
    
    llm = LLM(model="zoo-labs/zen-3-72b")
    outputs = llm.generate(["Hello, Zen!"], SamplingParams(max_tokens=100))
    

Quantized Versions

For resource-constrained deployment:

The 8B model runs on consumer GPUs. The 72B quantized fits in 48GB.

What’s Next

Zen 3 is a foundation. Coming soon:

The foundation is strong. Now we build.

Acknowledgments

Zen 3 was trained on the Zoo Compute Network with contributions from 847 node operators across 34 countries. Thank you.

This release was funded through the Zoo Labs Foundation treasury, allocated by community vote (ZIP-72). Thank you to all token holders who participated in governance.

Special thanks to the training, alignment, and evaluation teams who made this possible.

Download at huggingface.co/zoo-labs.


Zach Kelling is a co-founder of Zoo Labs Foundation.