Dynamic Attention-Guided Context Decoding for Mitigating Context Faithfulness Hallucinations in Large Language Models Paper • 2501.01059 • Published Jan 2 • 1
Self-Enhanced Reasoning Training: Activating Latent Reasoning in Small Models for Enhanced Reasoning Distillation Paper • 2502.12744 • Published Feb 18 • 1
Sentinel: Attention Probing of Proxy Models for LLM Context Compression with an Understanding Perspective Paper • 2505.23277 • Published May 29 • 1
How Instruction and Reasoning Data shape Post-Training: Data Quality through the Lens of Layer-wise Gradients Paper • 2504.10766 • Published Apr 14 • 40