ARMOR: High-Performance Semi-Structured Pruning via Adaptive Matrix Factorization
Abstract
ARMOR, a novel post-training pruning algorithm, enhances the performance of large language models by factorizing weight matrices into sparse cores and block diagonal matrices, achieving better accuracy and memory efficiency compared to conventional pruning methods.
Large language models (LLMs) present significant deployment challenges due to their immense computational and memory requirements. While semi-structured pruning, particularly 2:4 sparsity, offers a path to practical hardware acceleration, existing methods often incur substantial performance degradation. To bridge this gap, we introduce ARMOR: (Adaptive Representation with Matrix-factORization), a novel one-shot post-training pruning algorithm. Instead of directly pruning weights, ARMOR factorizes each weight matrix into a 2:4 sparse core wrapped by two low-overhead, block diagonal matrices. These wrappers act as efficient pre and post-transformation error correctors, offering greater flexibility to preserve model quality compared to conventional 2:4 pruning techniques. The sparse core and block diagonal wrappers are chosen through a block coordinate descent algorithm that minimizes a layer-wise proxy loss. We theoretically prove this optimization is guaranteed to converge to a solution with a proxy loss less than or equal to state-of-the-art pruning algorithms. Experiments on Llama (Touvron et al., 2023; Dubey et al., 2024) and Qwen (Yang et al., 2025) model families demonstrate that ARMOR consistently and significantly outperforms state-of-the-art 2:4 pruning methods across a wide range of downstream tasks and perplexity evaluations. ARMOR achieves this superior performance while retaining the inference speedups and substantial memory usage reductions of 2:4 pruning, establishing a more effective trade-off between model compression and task accuracy
Community
ARMOR is a new semi-structured pruning method that introduces a matrix factorization consisting of a 2:4 sparse core surrounded by two lightweight block diagonal matrices. We introduce a one-shot compression algorithm using this factorization, and show that it significantly and consistently outperforms existing one-shot semi-structured pruning algorithms on an industry standard suite of benchmarks.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- PATCH: Learnable Tile-level Hybrid Sparsity for LLMs (2025)
- NIRVANA: Structured pruning reimagined for large language models compression (2025)
- UniPruning: Unifying Local Metric and Global Feedback for Scalable Sparse LLMs (2025)
- CAST: Continuous and Differentiable Semi-Structured Sparsity-Aware Training for Large Language Models (2025)
- PTQTP: Post-Training Quantization to Trit-Planes for Large Language Models (2025)
- Lightweight error mitigation strategies for post-training N:M activation sparsity in LLMs (2025)
- COSPADI: Compressing LLMs via Calibration-Guided Sparse Dictionary Learning (2025)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment:
@librarian-bot
recommend
Models citing this paper 0
No model linking this paper
Datasets citing this paper 0
No dataset linking this paper
Spaces citing this paper 0
No Space linking this paper
Collections including this paper 0
No Collection including this paper