Collections
Discover the best community collections!
Collections including paper arxiv:2406.15319
-
Instruction Pre-Training: Language Models are Supervised Multitask Learners
Paper • 2406.14491 • Published • 95 -
Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
Paper • 2405.21060 • Published • 68 -
Perplexed by Perplexity: Perplexity-Based Data Pruning With Small Reference Models
Paper • 2405.20541 • Published • 24 -
MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark
Paper • 2406.01574 • Published • 49
-
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
An Information Bottleneck Perspective for Effective Noise Filtering on Retrieval-Augmented Generation
Paper • 2406.01549 • Published -
Retrieval Oriented Masking Pre-training Language Model for Dense Passage Retrieval
Paper • 2210.15133 • Published -
HippoRAG: Neurobiologically Inspired Long-Term Memory for Large Language Models
Paper • 2405.14831 • Published • 4
-
MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via Dynamic Sparse Attention
Paper • 2407.02490 • Published • 28 -
Can Few-shot Work in Long-Context? Recycling the Context to Generate Demonstrations
Paper • 2406.13632 • Published • 5 -
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
Make Your LLM Fully Utilize the Context
Paper • 2404.16811 • Published • 55
-
MInference 1.0: Accelerating Pre-filling for Long-Context LLMs via Dynamic Sparse Attention
Paper • 2407.02490 • Published • 28 -
Can Few-shot Work in Long-Context? Recycling the Context to Generate Demonstrations
Paper • 2406.13632 • Published • 5 -
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
Make Your LLM Fully Utilize the Context
Paper • 2404.16811 • Published • 55
-
Instruction Pre-Training: Language Models are Supervised Multitask Learners
Paper • 2406.14491 • Published • 95 -
Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality
Paper • 2405.21060 • Published • 68 -
Perplexed by Perplexity: Perplexity-Based Data Pruning With Small Reference Models
Paper • 2405.20541 • Published • 24 -
MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark
Paper • 2406.01574 • Published • 49
-
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
An Information Bottleneck Perspective for Effective Noise Filtering on Retrieval-Augmented Generation
Paper • 2406.01549 • Published -
Retrieval Oriented Masking Pre-training Language Model for Dense Passage Retrieval
Paper • 2210.15133 • Published -
HippoRAG: Neurobiologically Inspired Long-Term Memory for Large Language Models
Paper • 2405.14831 • Published • 4