TARS-1B
TARS-1B is a 1 billion parameter Liquid Neural Network language model built from scratch without using transformers or attention mechanisms. The architecture is designed for continuous-time reasoning and efficient generalization.
TARS:
- Architecture: Non-transformer, non-attention LNN (Liquid Neural Network)
- Training Tokens: Only 300 million tokens (Still training)
📊 Benchmarks
Benchmark | Accuracy |
---|---|
MMLU-Redux (Math) | 25.4% |
ARC-Easy | 32.95% |
PIQA | 51.36% |
BoolQ | 66.00% |
⚠️ These results were achieved with only 300M tokens of pretraining, making them exceptionally efficient compared to transformer-based models.
🚧 Training Status
We are still actively pretraining the model to reach higher performance on reasoning, coding, and complex tasks. Expect significant improvements as we scale toward billions of tokens.
🧪 Research Direction
TARS-1B is part of a broader research effort at SILX AI to build language models that generalize better with minimal compute using biologically inspired architectures.
- Downloads last month
- 70
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
🙋
Ask for provider support