Search papers, labs, and topics across Lattice.
2
0
6
13
LLM safety is a cat-and-mouse game: ORPO excels at breaking alignment, while DPO is best at restoring it, but at the cost of overall usefulness.
Shadow APIs promising access to top LLMs like GPT-5 and Gemini 2.5 often deliver significantly degraded performance (down to 47.21% accuracy) and fail identity verification, casting doubt on research relying on them.