Search papers, labs, and topics across Lattice.
2
0
4
3
World models can now self-improve by identifying their own prediction errors, thanks to a clever decomposition of action-conditioned prediction into easier-to-verify components.
Stop reward hacking: disentangling causal and non-causal factors in reward models makes RLHF more robust.