Four years into Zen’s development, it’s worth stepping back to assess where we are and where we’re going. Open AI development has made remarkable progress. It also faces significant challenges. Here’s my honest assessment.
What We’ve Achieved
Competitive Models
Open models now match or exceed proprietary alternatives in many domains:
- Coding: Open models lead on HumanEval and MBPP
- Reasoning: Competitive on GSM8K and MATH
- General knowledge: Within 5% on MMLU
- Multilingual: Often superior for non-English languages
The capability gap that seemed insurmountable in 2021 has largely closed for models under 100B parameters.
Real Adoption
Open models power real applications:
- Millions of API calls daily through community inference providers
- Thousands of fine-tuned variants for specialized tasks
- Integration into major development tools
- Deployment in production systems worldwide
This isn’t just research anymore. It’s infrastructure.
Community Scale
The open AI community has grown enormously:
- Hundreds of contributors to major projects
- Thousands of researchers building on open foundations
- Millions of users accessing open models
- Billions of inference requests monthly
Network effects are real. Each contribution makes the ecosystem more valuable.
Persistent Challenges
Compute Concentration
Training frontier models requires massive compute. This compute remains concentrated:
- Top 5 cloud providers control 80%+ of AI-grade compute
- Chip manufacturing bottlenecks persist
- Training costs continue rising with scale
The Zoo Compute Network helps but isn’t yet large enough for frontier training. We need 10x current capacity.
Data Moats
Large organizations have accumulated unique datasets:
- Proprietary user interactions
- Licensed content
- Internal documents
- Real-world feedback signals
Synthetic data and creative commons content help but don’t fully substitute. Data remains a key differentiator.
Coordination Costs
Open development is harder to coordinate than corporate efforts:
- Decisions require consensus
- Resources are distributed
- Timelines are uncertain
- Accountability is diffuse
DAOs and governance frameworks help but add overhead. Closed organizations can simply decide and execute.
Safety Complexity
As models become more capable, safety becomes more critical:
- Evaluation is hard (what are we even measuring?)
- Mitigation often reduces capability
- Adversarial robustness remains elusive
- Dual-use concerns complicate openness
We can’t ignore safety. We also can’t let it become a barrier to open development.
The Path Forward
Frontier Open Models
Training truly frontier models openly requires:
- Scaled compute network: 100K+ GPU network with reliable coordination
- Novel architectures: Efficiency improvements that reduce compute needs
- Better data: Synthetic data, simulation, and permissioned partnerships
- Sustained funding: Multi-year commitments to training runs
This is achievable. It requires coordination the community hasn’t yet demonstrated.
Infrastructure Maturity
Open infrastructure needs to mature:
- Training: Training Gym is solid but needs scaling features
- Inference: Efficient serving remains fragmented
- Fine-tuning: Needs better tooling for non-experts
- Evaluation: Benchmarks don’t capture real-world performance
Each layer needs investment. Each layer enables the next.
Governance Evolution
Current governance works for small decisions. Bigger decisions need:
- Faster processes: Weeks, not months
- Expert input: Technical decisions need technical judgment
- Accountability: Approved decisions must be implemented
- Representation: Diverse stakeholders, not just token holders
ZIPs are a start. We need to evolve.
Safety by Design
Open development can lead on safety:
- Transparency: Open models can be audited
- Diversity: Multiple approaches reduce correlated failures
- Iteration: Faster feedback loops on safety interventions
- Community: More eyes on potential problems
We should demonstrate that openness enhances safety, not undermines it.
Predictions for 2025
Concrete predictions to be evaluated:
- Open models will match GPT-4 class on major benchmarks (>50% confidence)
- Zoo Compute Network will reach 50K GPUs (60% confidence)
- At least one open model will be in top-3 on major leaderboards (70% confidence)
- ZIPs will process >100 proposals (80% confidence)
- Foundation funding will exceed $50M (40% confidence)
I’ll revisit these next November.
A Personal Note
When we started Zen, open AI development was a niche concern. Most assumed the future belonged to well-resourced corporations. That assumption was wrong.
Open development has proven viable. Community coordination works. Distributed resources aggregate. Transparency accelerates progress.
But viability isn’t victory. The next phase requires:
- Sustained commitment through multi-year projects
- Patience with governance overhead
- Investment in unsexy infrastructure
- Honesty about limitations and failures
We’ve shown open AI development is possible. Now we need to show it’s preferable.
The future of AI is still being written. Let’s write it together.
Zach Kelling is a co-founder of Zoo Labs Foundation.