Search papers, labs, and topics across Lattice.
1
0
3
6
Achieve better LLM knowledge distillation by blending the best of both forward and reverse KL divergence, leading to more stable training and improved performance.