Post
1650
this paper lists ways to make reasoning LLMs more efficient:
> enforce token limits per reasoning step
> route tasks to different models (small/large)
> compress reasoning chains during SFT
> reward based on reasoning length
> parallel search at test-time
and more...
@Xiaoye08 @yaful @Warrieryes
A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond (2503.21614)
> enforce token limits per reasoning step
> route tasks to different models (small/large)
> compress reasoning chains during SFT
> reward based on reasoning length
> parallel search at test-time
and more...
@Xiaoye08 @yaful @Warrieryes
A Survey of Efficient Reasoning for Large Reasoning Models: Language, Multimodality, and Beyond (2503.21614)