view reply Great article, thanks for sharing. I am wondering why the finetuning was done on Instruct models rather than the base models?
view article Article Fine-tuning LLMs to 1.58bit: extreme quantization made easy By medmekk and 5 others • Sep 18, 2024 • 261
BitNet Collection 🔥BitNet family of large language models (1-bit LLMs). • 7 items • Updated May 1 • 47
Running 2.84k 2.84k The Ultra-Scale Playbook 🌌 The ultimate guide to training LLM on large GPU Clusters