Cache-to-Cache: Direct Semantic Communication Between Large Language Models Paper • 2510.03215 • Published 29 days ago • 93
QeRL: Beyond Efficiency -- Quantization-enhanced Reinforcement Learning for LLMs Paper • 2510.11696 • Published 19 days ago • 169
R2R Collection Collections for paper "R2R: Efficiently Navigating Divergent Reasoning Paths with Small-Large Model Token Routing" • 5 items • Updated 17 days ago • 2
PAROAttention: Pattern-Aware ReOrdering for Efficient Sparse and Quantized Attention in Visual Generation Models Paper • 2506.16054 • Published Jun 19 • 60
MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression Paper • 2406.14909 • Published Jun 21, 2024 • 16
A Survey on Efficient Inference for Large Language Models Paper • 2404.14294 • Published Apr 22, 2024 • 3
Distilled Decoding 1: One-step Sampling of Image Auto-regressive Models with Flow Matching Paper • 2412.17153 • Published Dec 22, 2024 • 39
R2R: Efficiently Navigating Divergent Reasoning Paths with Small-Large Model Token Routing Paper • 2505.21600 • Published May 27 • 70