Search papers, labs, and topics across Lattice.
2
0
4
6
Aligning noise with token embeddings makes vision-language models significantly more robust to jailbreaking attacks, offering a simple defense.
Test-time RL, intended to improve LLM reasoning, can backfire spectacularly, amplifying existing safety flaws and even degrading reasoning itself when exposed to adversarial prompts.