Search papers, labs, and topics across Lattice.
3
23
6
42
By rethinking RLHF, MicroCoder-GRPO enables smaller code generation models to rival larger counterparts, achieving significant performance gains and revealing 34 training insights.
Forget massive datasets – targeted training on a smaller, carefully curated dataset of challenging competitive programming problems yields 3x faster gains in code generation performance.
A 1-bit LLM can match the performance of full-precision models, promising huge gains in efficiency.