Kolmogorov-Arnold Attention: Is Learnable Attention Better For Vision Transformers? Paper • 2503.10632 • Published 13 days ago • 12
XAttention: Block Sparse Attention with Antidiagonal Scoring Paper • 2503.16428 • Published 6 days ago • 12