🪷 Zen LM
Models

zen-embedding

Foundation embedding model for search and retrieval.

zen-embedding

Foundation Embeddings

The foundation embedding model for search and retrieval. Produces 3072-dimensional dense vectors optimized for semantic similarity, enabling high-quality semantic search, RAG retrieval, and document clustering across an 8K token context window.

Specifications

PropertyValue
Model IDzen-embedding
Dimensions3,072
ArchitectureEmbedding
Context Window8K tokens
Tierpro
StatusAvailable
HuggingFacezenlm/zen-embedding

Capabilities

  • 3072-dimensional dense vector embeddings
  • Semantic search by meaning, not just keywords
  • RAG pipeline document retrieval
  • Document clustering and deduplication
  • Classification feature generation
  • Cosine similarity scoring for ranking

API Usage

curl https://api.hanzo.ai/v1/embeddings \
  -H "Authorization: Bearer $HANZO_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "zen-embedding",
    "input": "Zen LM is a family of frontier AI models"
  }'
from hanzoai import Hanzo

client = Hanzo(api_key="hk-your-api-key")

response = client.embeddings.create(
    model="zen-embedding",
    input=["Hello world", "Zen LM models"],
)

for embedding in response.data:
    print(f"Vector dim: {len(embedding.embedding)}")

HuggingFace Usage

from transformers import AutoModel, AutoTokenizer
import torch

tokenizer = AutoTokenizer.from_pretrained("zenlm/zen-embedding")
model = AutoModel.from_pretrained("zenlm/zen-embedding")

inputs = tokenizer("Zen LM is a family of frontier AI models",
                   return_tensors="pt", truncation=True)
with torch.no_grad():
    outputs = model(**inputs)

# Mean pool over token embeddings
embeddings = outputs.last_hidden_state.mean(dim=1)
print(f"Embedding shape: {embeddings.shape}")  # [1, 3072]

Response Format

{
  "object": "list",
  "data": [
    {
      "object": "embedding",
      "index": 0,
      "embedding": [0.0023, -0.0091, 0.0152, "..."]
    }
  ],
  "model": "zen-embedding",
  "usage": {
    "prompt_tokens": 8,
    "total_tokens": 8
  }
}

Try It

Open in Hanzo Chat

Resources

See Also

On this page