zen-nano-0.6b

Lightweight 0.6B parameter edge model optimized for mobile and edge devices

Model Details

  • Developed by: Zen Research Authors
  • Organization: Zen Research DAO under Zoo Labs Inc (501(c)(3) Non-Profit)
  • Location: San Francisco, California, USA
  • Model type: language-model
  • Architecture: Qwen3-0.6B
  • Parameters: 0.6B
  • License: Apache 2.0
  • Training: Trained with Zen Gym
  • Inference: Optimized for Zen Engine

🌟 Zen AI Ecosystem

This model is part of the Zen Research hypermodal AI family - the world's most comprehensive open-source AI ecosystem.

Complete Model Family

Language Models:

3D & World Generation:

  • zen-3d - Controllable 3D asset generation
  • zen-voyager - Camera-controlled world exploration
  • zen-world - Large-scale world simulation

Video Generation:

Audio Generation:

Infrastructure:

Usage

Quick Start

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("zenlm/zen-nano-0.6b")
tokenizer = AutoTokenizer.from_pretrained("zenlm/zen-nano-0.6b")

messages = [{"role": "user", "content": "Explain quantum computing"}]
inputs = tokenizer.apply_chat_template(messages, return_tensors="pt")
outputs = model.generate(inputs, max_new_tokens=200)
print(tokenizer.decode(outputs[0]))

With Zen Engine

# High-performance inference (44K tokens/sec on M3 Max)
zen-engine serve --model zenlm/zen-nano-0.6b --port 3690
# OpenAI-compatible API
from openai import OpenAI

client = OpenAI(base_url="http://localhost:3690/v1")
response = client.chat.completions.create(
    model="zenlm/zen-nano-0.6b",
    messages=[{"role": "user", "content": "Hello!"}]
)

Training

Fine-tune with Zen Gym:

git clone https://github.com/zenlm/zen-gym
cd zen-gym

# LoRA fine-tuning
llamafactory-cli train --config configs/zen_lora.yaml \
    --model_name_or_path zenlm/zen-nano-0.6b

# GRPO reinforcement learning (40-60% memory reduction)
llamafactory-cli train --config configs/zen_grpo.yaml \
    --model_name_or_path zenlm/zen-nano-0.6b

Supported methods: LoRA, QLoRA, DoRA, GRPO, GSPO, DPO, PPO, KTO, ORPO, SimPO, Unsloth

Performance

  • Speed: 44K tokens/sec (M3 Max), 35K tokens/sec (RTX 4090)
  • Memory: 0.4GB (Q2_K) to 1.2GB (F16)
  • Edge: Optimized for mobile and IoT devices
  • Formats: PyTorch, MLX, GGUF (Q2_K to F16)

Ethical Considerations

  • Open Research: Released under Apache 2.0 for maximum accessibility
  • Environmental Impact: Optimized for eco-friendly deployment
  • Transparency: Full training details and model architecture disclosed
  • Safety: Comprehensive testing and evaluation
  • Non-Profit: Developed by Zoo Labs Inc (501(c)(3)) for public benefit

Citation

@misc{zenzennano0.6b2025,
  title={zen-nano-0.6b: Lightweight 0.6B parameter edge model optimized for mobile and edge devices},
  author={Zen Research Authors},
  year={2025},
  publisher={Zoo Labs Inc},
  organization={Zen Research DAO},
  url={https://huggingface.co/zenlm/zen-nano-0.6b}
}

Links

License

Apache License 2.0

Copyright 2025 Zen Research Authors


Zen Research - Building open, eco-friendly AI for everyone 🌱

Downloads last month
237
Safetensors
Model size
596M params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for zenlm/zen-nano-0.6b

Quantizations
2 models