view article Article Supercharge Edge AI with High Accuracy Reasoning Using Llama Nemotron Nano 4B By nvidia and 3 others • 8 days ago • 6
Efficient Hybrid Language Model Compression through Group-Aware SSM Pruning Paper • 2504.11409 • Published Apr 15 • 10
Hymba: A Hybrid-head Architecture for Small Language Models Paper • 2411.13676 • Published Nov 20, 2024 • 46
MaskLLM: Learnable Semi-Structured Sparsity for Large Language Models Paper • 2409.17481 • Published Sep 26, 2024 • 48
Minitron Collection A family of compressed models obtained via pruning and knowledge distillation • 12 items • Updated 1 day ago • 61
LLM Pruning and Distillation in Practice: The Minitron Approach Paper • 2408.11796 • Published Aug 21, 2024 • 59