Search papers, labs, and topics across Lattice.
2
0
6
13
LLM safety is a cat-and-mouse game: ORPO excels at breaking alignment, while DPO is best at restoring it, but at the cost of overall usefulness.
Multimodal models can "see" the image but still fail at reasoning because the visual input distracts the routing mechanism from activating the right experts.