Search papers, labs, and topics across Lattice.
2
0
6
2
Reward hacking isn't just a bug, it's a feature arising from the fundamental mismatch between complex human goals and the compressed reward signals used to train LLMs.
Current image generation models fall far short of the mark when it comes to the structured and multi-constraint demands of real-world commercial design, as revealed by a new systematic benchmark.