view article Article Transformers backend integration in SGLang By marcsun13 and 4 others • Jun 23 • 52
PAROAttention: Pattern-Aware ReOrdering for Efficient Sparse and Quantized Attention in Visual Generation Models Paper • 2506.16054 • Published Jun 19 • 60
R2R Collection Collections for paper "R2R: Efficiently Navigating Divergent Reasoning Paths with Small-Large Model Token Routing" • 4 items • Updated 13 days ago • 2
Distilled Decoding 1: One-step Sampling of Image Auto-regressive Models with Flow Matching Paper • 2412.17153 • Published Dec 22, 2024 • 40
FrameFusion: Combining Similarity and Importance for Video Token Reduction on Large Visual Language Models Paper • 2501.01986 • Published Dec 30, 2024 • 1
Learning Strategic Language Agents in the Werewolf Game with Iterative Latent Space Policy Optimization Paper • 2502.04686 • Published Feb 7 • 2
VS-Bench: Evaluating VLMs for Strategic Reasoning and Decision-Making in Multi-Agent Environments Paper • 2506.02387 • Published Jun 3 • 57
DeepResearchGym: A Free, Transparent, and Reproducible Evaluation Sandbox for Deep Research Paper • 2505.19253 • Published May 25 • 29
SageAttention2++: A More Efficient Implementation of SageAttention2 Paper • 2505.21136 • Published May 27 • 47
R2R: Efficiently Navigating Divergent Reasoning Paths with Small-Large Model Token Routing Paper • 2505.21600 • Published May 27 • 71
MoA: Mixture of Sparse Attention for Automatic Large Language Model Compression Paper • 2406.14909 • Published Jun 21, 2024 • 16