Well Begun is Half Done: Low-resource Preference Alignment by Weak-to-Strong Decoding Paper • 2506.07434 • Published Jun 9 • 7
TIME: A Multi-level Benchmark for Temporal Reasoning of LLMs in Real-World Scenarios Paper • 2505.12891 • Published May 19 • 2
Odysseus Navigates the Sirens' Song: Dynamic Focus Decoding for Factual and Diverse Open-Ended Text Generation Paper • 2503.08057 • Published Mar 11
Omni-MATH: A Universal Olympiad Level Mathematic Benchmark For Large Language Models Paper • 2410.07985 • Published Oct 10, 2024 • 33
Towards a Unified View of Preference Learning for Large Language Models: A Survey Paper • 2409.02795 • Published Sep 4, 2024 • 74
ICDPO: Effectively Borrowing Alignment Capability of Others via In-context Direct Preference Optimization Paper • 2402.09320 • Published Feb 14, 2024 • 6
Making Large Language Models Better Reasoners with Alignment Paper • 2309.02144 • Published Sep 5, 2023 • 2
Interacting with Non-Cooperative User: A New Paradigm for Proactive Dialogue Policy Paper • 2204.07433 • Published Apr 7, 2022
API-Bank: A Comprehensive Benchmark for Tool-Augmented LLMs Paper • 2304.08244 • Published Apr 14, 2023
Scaling Data Diversity for Fine-Tuning Language Models in Human Alignment Paper • 2403.11124 • Published Mar 17, 2024 • 1