floom
's Collections
Long-context
updated
InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding
Extremely Long Sequences with Training-Free Memory
Paper
•
2402.04617
•
Published
•
4
BurstAttention: An Efficient Distributed Attention Framework for
Extremely Long Sequences
Paper
•
2403.09347
•
Published
•
20
Resonance RoPE: Improving Context Length Generalization of Large
Language Models
Paper
•
2403.00071
•
Published
•
22
Training-Free Long-Context Scaling of Large Language Models
Paper
•
2402.17463
•
Published
•
19
Same Task, More Tokens: the Impact of Input Length on the Reasoning
Performance of Large Language Models
Paper
•
2402.14848
•
Published
•
18
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens
Paper
•
2402.13753
•
Published
•
114
LongAgent: Scaling Language Models to 128k Context through Multi-Agent
Collaboration
Paper
•
2402.11550
•
Published
•
16
In Search of Needles in a 10M Haystack: Recurrent Memory Finds What LLMs
Miss
Paper
•
2402.10790
•
Published
•
41
Make Your LLM Fully Utilize the Context
Paper
•
2404.16811
•
Published
•
52