Search papers, labs, and topics across Lattice.
2
0
4
9
1.58-bit LLMs are surprisingly more resilient to sparsity than their full-precision counterparts, opening new avenues for extreme compression.
LLMs trained with a novel "second-order rollout" that generates critiques in addition to responses learn more effectively from the same data, unlocking better reasoning.