Arcee Trinity Mini

Trinity Nano Base

Trinity Nano is an Arcee AI 6B MoE model with 1B active parameters. It is the small-sized model in our new Trinity family, a series of open-weight models for enterprise and tinkerers alike.

This base model pre fine tuning, and so is not suitable for chatting, and should be trained for your specific domain before use.

Trinity Nano is trained on 10T tokens gathered and curated through a key partnership with Datology, building upon the excellent dataset we used on AFM-4.5B with additional math and code.

Training was performed on a cluster of 512 H200 GPUs powered by Prime Intellect using HSDP parallelism.

More details, including key architecture decisions, can be found on our blog here


Model Details

  • Model Architecture: AfmoeForCausalLM
  • Parameters: 6B, 1B active
  • Experts: 128 total, 8 active, 1 shared
  • Context length: 128k
  • Training Tokens: 10T
  • License: Apache 2.0

Benchmarks

πŸ”’ Math & Reasoning

Benchmark Score
GSM8K 58.4%
Minerva Math 500 36.0%
DROP (0-shot) 4.5%
DROP (5-shot) 63.6%

πŸ’» Code Generation

Benchmark Pass@1 Pass@10
HumanEval (3-shot, bpb) 36.3% (bpb) -
HumanEval+ (temp 0.8) 31.7% -
MBPP+ 44.7% -

🧠 Knowledge & Reasoning

Benchmark 5-shot 0-shot
ARC-Challenge 84.0% 78.2%
ARC-Easy 94.8% 91.2%
CommonsenseQA 74.9% 62.7%
OpenBookQA 82.2% 75.2%
WinoGrande 72.8% 68.0%
MMLU 67.7% 64.2%
MMLU Pro 35.8% 27.7%
AGI Eval (English) 51.8% -
BBH (CoT) 50.4% 7.6%

πŸ“˜ Understanding & QA

Benchmark Score
BoolQ (5-shot) 84.3%
HellaSwag (5-shot) 77.4%
PIQA (5-shot) 82.2%
SciQ (5-shot) 93.2%
Social IQA (5-shot) 73.0%
Powered by Datology

License

Trinity-Mini-Base is released under the Apache-2.0 license.

Downloads last month
172
Safetensors
Model size
6B params
Tensor type
BF16
Β·
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for arcee-ai/Trinity-Nano-Base

Finetunes
1 model
Quantizations
3 models

Collection including arcee-ai/Trinity-Nano-Base