Search papers, labs, and topics across Lattice.
Baidu Inc., Shenzhen University
1
0
2
LLMs can slash over 80% of their chain-of-thought tokens with a minor accuracy boost, thanks to a new RL-based method that targets the "Minimal Sufficient Length" of reasoning.