Search papers, labs, and topics across Lattice.
3
0
4
Continuous diffusion LMs can rival discrete models by fixing the token-rounding bottleneck with a contextual autoregressive decoder, unlocking a fluency-diversity knob in the process.
Key contribution not extracted.
Stop wasting compute: PonderLM-3 learns to spend extra inference FLOPs only on the tokens that actually need them, outperforming fixed-step pondering methods.