Personalize Anything for Free with Diffusion Transformer Paper • 2503.12590 • Published 12 days ago • 41
FlowTok: Flowing Seamlessly Across Text and Image Tokens Paper • 2503.10772 • Published 15 days ago • 17
Kolmogorov-Arnold Attention: Is Learnable Attention Better For Vision Transformers? Paper • 2503.10632 • Published 15 days ago • 12
PLADIS: Pushing the Limits of Attention in Diffusion Models at Inference Time by Leveraging Sparsity Paper • 2503.07677 • Published 19 days ago • 81
Automated Movie Generation via Multi-Agent CoT Planning Paper • 2503.07314 • Published 18 days ago • 42
ObjectMover: Generative Object Movement with Video Prior Paper • 2503.08037 • Published 18 days ago • 4
One-step Diffusion Models with f-Divergence Distribution Matching Paper • 2502.15681 • Published Feb 21 • 7
Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention Paper • 2502.11089 • Published Feb 16 • 150
I Think, Therefore I Diffuse: Enabling Multimodal In-Context Reasoning in Diffusion Models Paper • 2502.10458 • Published Feb 12 • 34
Small Models Struggle to Learn from Strong Reasoners Paper • 2502.12143 • Published Feb 17 • 33
MME-CoT: Benchmarking Chain-of-Thought in Large Multimodal Models for Reasoning Quality, Robustness, and Efficiency Paper • 2502.09621 • Published Feb 13 • 27
FlashVideo:Flowing Fidelity to Detail for Efficient High-Resolution Video Generation Paper • 2502.05179 • Published Feb 7 • 24
TransPixar: Advancing Text-to-Video Generation with Transparency Paper • 2501.03006 • Published Jan 6 • 26
STAR: Spatial-Temporal Augmentation with Text-to-Video Models for Real-World Video Super-Resolution Paper • 2501.02976 • Published Jan 6 • 55
VideoAnydoor: High-fidelity Video Object Insertion with Precise Motion Control Paper • 2501.01427 • Published Jan 2 • 54