Long Context
updated
FLM-101B: An Open LLM and How to Train It with $100K Budget
Paper
• 2309.03852
• Published
• 45
Extending LLMs' Context Window with 100 Samples
Paper
• 2401.07004
• Published
• 16
LongAgent: Scaling Language Models to 128k Context through Multi-Agent
Collaboration
Paper
• 2402.11550
• Published
• 19
The What, Why, and How of Context Length Extension Techniques in Large
Language Models -- A Detailed Survey
Paper
• 2401.07872
• Published
• 2
A Human-Inspired Reading Agent with Gist Memory of Very Long Contexts
Paper
• 2402.09727
• Published
• 38
Data Engineering for Scaling Language Models to 128K Context
Paper
• 2402.10171
• Published
• 25
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs
Miss
Paper
• 2402.10790
• Published
• 42
LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning
Paper
• 2401.01325
• Published
• 27
RetrievalAttention: Accelerating Long-Context LLM Inference via Vector
Retrieval
Paper
• 2409.10516
• Published
• 43
Needle Threading: Can LLMs Follow Threads through Near-Million-Scale
Haystacks?
Paper
• 2411.05000
• Published
• 22
Context Is What You Need: The Maximum Effective Context Window for Real
World Limits of LLMs
Paper
• 2509.21361
• Published
• 11