I Have Covered All the Bases Here: Interpreting Reasoning Features in Large Language Models via Sparse Autoencoders Paper • 2503.18878 • Published Mar 24 • 118
When Less is Enough: Adaptive Token Reduction for Efficient Image Representation Paper • 2503.16660 • Published Mar 20 • 73
Combining Flow Matching and Transformers for Efficient Solution of Bayesian Inverse Problems Paper • 2503.01375 • Published Mar 3 • 5
GHOST 2.0: generative high-fidelity one shot transfer of heads Paper • 2502.18417 • Published Feb 25 • 67
LLM-Microscope: Uncovering the Hidden Role of Punctuation in Context Memory of Transformers Paper • 2502.15007 • Published Feb 20 • 175
How Much Knowledge Can You Pack into a LoRA Adapter without Harming LLM? Paper • 2502.14502 • Published Feb 20 • 90
Cramming 1568 Tokens into a Single Vector and Back Again: Exploring the Limits of Embedding Space Capacity Paper • 2502.13063 • Published Feb 18 • 70
Memory, Benchmark & Robots: A Benchmark for Solving Complex Tasks with Reinforcement Learning Paper • 2502.10550 • Published Feb 14 • 5
Memory, Benchmark & Robots: A Benchmark for Solving Complex Tasks with Reinforcement Learning Paper • 2502.10550 • Published Feb 14 • 5
Memory, Benchmark & Robots: A Benchmark for Solving Complex Tasks with Reinforcement Learning Paper • 2502.10550 • Published Feb 14 • 5 • 2
SynthDetoxM: Modern LLMs are Few-Shot Parallel Detoxification Data Annotators Paper • 2502.06394 • Published Feb 10 • 90
3DGraphLLM: Combining Semantic Graphs and Large Language Models for 3D Scene Understanding Paper • 2412.18450 • Published Dec 24, 2024 • 37