Zen4 Mini (MLX)
MLX quantization of Zen4 Mini for efficient inference on Apple Silicon.
Model Details
| Property | Value |
|---|---|
| Parameters | Zen4 Mini |
| Format | MLX 4-bit quantized |
| Architecture | Zen4 |
| License | Apache 2.0 |
| Authors | Zen LM Authors |
Usage
from mlx_lm import load, generate
model, tokenizer = load("zenlm/zen4-mini-MLX")
response = generate(model, tokenizer, prompt="Hello, how can I help you?", max_tokens=512)
print(response)
About
Zen4 Mini is a compact, efficient language model from the Zen4 family, optimized for Apple Silicon via MLX quantization.
Developed by: Zen LM Authors
- Downloads last month
- 34
Model size
0.6B params
Tensor type
BF16
·
U32 ·
Hardware compatibility
Log In to add your hardware
4-bit
Model tree for zenlm/zen4-mini-MLX
Base model
zenlm/zen4-mini