Falcon-H1 Collection Falcon-H1 Family of Hybrid-Head Language Models, including 0.5B, 1.5B, 1.5B-Deep, 3B, 7B, and 34B (pretrained and instruction-tuned). • 37 items • Updated 3 days ago • 30
MM-PRM: Enhancing Multimodal Mathematical Reasoning with Scalable Step-Level Supervision Paper • 2505.13427 • Published 5 days ago • 24
Seek in the Dark: Reasoning via Test-Time Instance-Level Policy Gradient in Latent Space Paper • 2505.13308 • Published 5 days ago • 25
Scaling Computer-Use Grounding via User Interface Decomposition and Synthesis Paper • 2505.13227 • Published 5 days ago • 43
Delta Attention: Fast and Accurate Sparse Attention Inference by Delta Correction Paper • 2505.11254 • Published 8 days ago • 47
AdaCoT: Pareto-Optimal Adaptive Chain-of-Thought Triggering via Reinforcement Learning Paper • 2505.11896 • Published 7 days ago • 54
Incorporating brain-inspired mechanisms for multimodal learning in artificial intelligence Paper • 2505.10176 • Published 9 days ago • 3
Think Only When You Need with Large Hybrid-Reasoning Models Paper • 2505.14631 • Published 4 days ago • 18
Optimizing Anytime Reasoning via Budget Relative Policy Optimization Paper • 2505.13438 • Published 5 days ago • 33