aquif-moe
Collection
Our first MoE models. It's a series of entry-level models using the granitemoe architecture, finetuned by aquif AI.
•
2 items
•
Updated
aquif-moe-400m is our compact Mixture of Experts (MoE) model, with only 400 million active parameters. It offers impressive performance-per-VRAM efficiency, making it a strong choice for resource-limited setups.
aquif-moe-400m
aquif-moe-400m delivers solid performance across multiple benchmarks, especially for its size:
Benchmark | aquif-moe (0.4b) | Qwen 2.5 (0.5b) | Gemma 3 (1b) |
---|---|---|---|
MMLU | 26.6 | 45.4 | 26.5 |
HumanEval | 32.3 | 22.0 | 8.1 |
GSM8K | 33.9 | 36.0 | 6.1 |
Average | 30.9 | 34.4 | 11.3 |
aquif-moe-400m excels in VRAM efficiency:
Model | Average Performance | VRAM (GB) | Performance per VRAM |
---|---|---|---|
aquif-moe | 30.9 | 0.4 | 77.3 |
Qwen 2.5 | 34.4 | 0.6 | 57.3 |
Gemma 3 | 11.3 | 1.0 | 11.3 |
To run via Ollama:
ollama run aquiffoo/aquif-moe-400m
``´