Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention Paper • 2502.11089 • Published Feb 16 • 165
view article Article Fine-tuning 20B LLMs with RLHF on a 24GB consumer GPU By edbeeching and 5 others • Mar 9, 2023 • 61
Internal Consistency and Self-Feedback in Large Language Models: A Survey Paper • 2407.14507 • Published Jul 19, 2024 • 47