Token-Efficient Long Video Understanding for Multimodal LLMs Paper • 2503.04130 • Published 4 days ago • 66
From Hours to Minutes: Lossless Acceleration of Ultra Long Sequence Generation up to 100K Tokens Paper • 2502.18890 • Published 12 days ago • 23
Phi-4-Mini Technical Report: Compact yet Powerful Multimodal Language Models via Mixture-of-LoRAs Paper • 2503.01743 • Published 7 days ago • 65
Multi-Turn Code Generation Through Single-Step Rewards Paper • 2502.20380 • Published 11 days ago • 29
Chain of Draft: Thinking Faster by Writing Less Paper • 2502.18600 • Published 12 days ago • 44
Running 2.15k 2.15k The Ultra-Scale Playbook 🌌 The ultimate guide to training LLM on large GPU Clusters
MoBA: Mixture of Block Attention for Long-Context LLMs Paper • 2502.13189 • Published 20 days ago • 14
LightThinker: Thinking Step-by-Step Compression Paper • 2502.15589 • Published 17 days ago • 26
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling Paper • 2502.06703 • Published 28 days ago • 142
SFT Memorizes, RL Generalizes: A Comparative Study of Foundation Model Post-training Paper • 2501.17161 • Published Jan 28 • 108
Towards General-Purpose Model-Free Reinforcement Learning Paper • 2501.16142 • Published Jan 27 • 26
Kimi k1.5: Scaling Reinforcement Learning with LLMs Paper • 2501.12599 • Published Jan 22 • 101
VideoLLaMA 3: Frontier Multimodal Foundation Models for Image and Video Understanding Paper • 2501.13106 • Published Jan 22 • 84
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning Paper • 2501.12948 • Published Jan 22 • 341