Search papers, labs, and topics across Lattice.
2
25
5
5
P-GenRM personalizes LLMs more effectively by generating adaptive personas and scoring rubrics from user preferences, outperforming existing reward models by 2.31% and offering a 3% boost via test-time scaling.
DPO's rise as a computationally efficient alternative to RLHF for LLM alignment has spurred a diverse range of research, now systematically organized and analyzed in this comprehensive survey.