Search papers, labs, and topics across Lattice.
1
0
3
LLMs exhibit an "Alignment Illusion," where their apparent safety collapses under pressure, with the most capable models showing the most dramatic failures.