-
LLoCO: Learning Long Contexts Offline
Paper • 2404.07979 • Published • 23 -
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens
Paper • 2402.13753 • Published • 117 -
LongAgent: Scaling Language Models to 128k Context through Multi-Agent Collaboration
Paper • 2402.11550 • Published • 18 -
LongAlign: A Recipe for Long Context Alignment of Large Language Models
Paper • 2401.18058 • Published • 23
Collections
Discover the best community collections!
Collections including paper arxiv:2406.15319
-
Mixture-of-Agents Enhances Large Language Model Capabilities
Paper • 2406.04692 • Published • 60 -
CRAG -- Comprehensive RAG Benchmark
Paper • 2406.04744 • Published • 49 -
Boosting Large-scale Parallel Training Efficiency with C4: A Communication-Driven Approach
Paper • 2406.04594 • Published • 8 -
Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models
Paper • 2406.04271 • Published • 31
-
MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Paper • 2403.09611 • Published • 128 -
Evolutionary Optimization of Model Merging Recipes
Paper • 2403.13187 • Published • 56 -
MobileVLM V2: Faster and Stronger Baseline for Vision Language Model
Paper • 2402.03766 • Published • 15 -
LLM Agent Operating System
Paper • 2403.16971 • Published • 71
-
Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference
Paper • 2403.09636 • Published • 3 -
TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
Paper • 2404.11912 • Published • 17 -
Infinite-LLM: Efficient LLM Service for Long Context with DistAttention and Distributed KVCache
Paper • 2401.02669 • Published • 16 -
LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding
Paper • 2404.16710 • Published • 80
-
Same Task, More Tokens: the Impact of Input Length on the Reasoning Performance of Large Language Models
Paper • 2402.14848 • Published • 20 -
The Prompt Report: A Systematic Survey of Prompting Techniques
Paper • 2406.06608 • Published • 65 -
CRAG -- Comprehensive RAG Benchmark
Paper • 2406.04744 • Published • 49 -
Transformers meet Neural Algorithmic Reasoners
Paper • 2406.09308 • Published • 45
-
Similarity is Not All You Need: Endowing Retrieval Augmented Generation with Multi Layered Thoughts
Paper • 2405.19893 • Published • 32 -
From RAGs to rich parameters: Probing how language models utilize external knowledge over parametric information for factual queries
Paper • 2406.12824 • Published • 21 -
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
A Tale of Trust and Accuracy: Base vs. Instruct LLMs in RAG Systems
Paper • 2406.14972 • Published • 7
-
Toward Self-Improvement of LLMs via Imagination, Searching, and Criticizing
Paper • 2404.12253 • Published • 56 -
FlowMind: Automatic Workflow Generation with LLMs
Paper • 2404.13050 • Published • 35 -
How Far Can We Go with Practical Function-Level Program Repair?
Paper • 2404.12833 • Published • 7 -
Replacing Judges with Juries: Evaluating LLM Generations with a Panel of Diverse Models
Paper • 2404.18796 • Published • 71
-
MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries
Paper • 2401.15391 • Published • 6 -
RAFT: Adapting Language Model to Domain Specific RAG
Paper • 2403.10131 • Published • 73 -
Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation
Paper • 2404.06910 • Published • 3 -
Stylus: Automatic Adapter Selection for Diffusion Models
Paper • 2404.18928 • Published • 15
-
Generative Representational Instruction Tuning
Paper • 2402.09906 • Published • 55 -
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval
Paper • 2407.12883 • Published • 11 -
mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval
Paper • 2407.19669 • Published • 24
-
LLoCO: Learning Long Contexts Offline
Paper • 2404.07979 • Published • 23 -
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens
Paper • 2402.13753 • Published • 117 -
LongAgent: Scaling Language Models to 128k Context through Multi-Agent Collaboration
Paper • 2402.11550 • Published • 18 -
LongAlign: A Recipe for Long Context Alignment of Large Language Models
Paper • 2401.18058 • Published • 23
-
Same Task, More Tokens: the Impact of Input Length on the Reasoning Performance of Large Language Models
Paper • 2402.14848 • Published • 20 -
The Prompt Report: A Systematic Survey of Prompting Techniques
Paper • 2406.06608 • Published • 65 -
CRAG -- Comprehensive RAG Benchmark
Paper • 2406.04744 • Published • 49 -
Transformers meet Neural Algorithmic Reasoners
Paper • 2406.09308 • Published • 45
-
Mixture-of-Agents Enhances Large Language Model Capabilities
Paper • 2406.04692 • Published • 60 -
CRAG -- Comprehensive RAG Benchmark
Paper • 2406.04744 • Published • 49 -
Boosting Large-scale Parallel Training Efficiency with C4: A Communication-Driven Approach
Paper • 2406.04594 • Published • 8 -
Buffer of Thoughts: Thought-Augmented Reasoning with Large Language Models
Paper • 2406.04271 • Published • 31
-
Similarity is Not All You Need: Endowing Retrieval Augmented Generation with Multi Layered Thoughts
Paper • 2405.19893 • Published • 32 -
From RAGs to rich parameters: Probing how language models utilize external knowledge over parametric information for factual queries
Paper • 2406.12824 • Published • 21 -
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
A Tale of Trust and Accuracy: Base vs. Instruct LLMs in RAG Systems
Paper • 2406.14972 • Published • 7
-
Toward Self-Improvement of LLMs via Imagination, Searching, and Criticizing
Paper • 2404.12253 • Published • 56 -
FlowMind: Automatic Workflow Generation with LLMs
Paper • 2404.13050 • Published • 35 -
How Far Can We Go with Practical Function-Level Program Repair?
Paper • 2404.12833 • Published • 7 -
Replacing Judges with Juries: Evaluating LLM Generations with a Panel of Diverse Models
Paper • 2404.18796 • Published • 71
-
MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training
Paper • 2403.09611 • Published • 128 -
Evolutionary Optimization of Model Merging Recipes
Paper • 2403.13187 • Published • 56 -
MobileVLM V2: Faster and Stronger Baseline for Vision Language Model
Paper • 2402.03766 • Published • 15 -
LLM Agent Operating System
Paper • 2403.16971 • Published • 71
-
MultiHop-RAG: Benchmarking Retrieval-Augmented Generation for Multi-Hop Queries
Paper • 2401.15391 • Published • 6 -
RAFT: Adapting Language Model to Domain Specific RAG
Paper • 2403.10131 • Published • 73 -
Superposition Prompting: Improving and Accelerating Retrieval-Augmented Generation
Paper • 2404.06910 • Published • 3 -
Stylus: Automatic Adapter Selection for Diffusion Models
Paper • 2404.18928 • Published • 15
-
Dynamic Memory Compression: Retrofitting LLMs for Accelerated Inference
Paper • 2403.09636 • Published • 3 -
TriForce: Lossless Acceleration of Long Sequence Generation with Hierarchical Speculative Decoding
Paper • 2404.11912 • Published • 17 -
Infinite-LLM: Efficient LLM Service for Long Context with DistAttention and Distributed KVCache
Paper • 2401.02669 • Published • 16 -
LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding
Paper • 2404.16710 • Published • 80
-
Generative Representational Instruction Tuning
Paper • 2402.09906 • Published • 55 -
LongRAG: Enhancing Retrieval-Augmented Generation with Long-context LLMs
Paper • 2406.15319 • Published • 65 -
BRIGHT: A Realistic and Challenging Benchmark for Reasoning-Intensive Retrieval
Paper • 2407.12883 • Published • 11 -
mGTE: Generalized Long-Context Text Representation and Reranking Models for Multilingual Text Retrieval
Paper • 2407.19669 • Published • 24