Search papers, labs, and topics across Lattice.
1
0
3
8
Stripping away obvious "triggering cues" from adversarial attacks reveals that current AI safety datasets drastically overestimate model robustness, turning "safe" models like Gemini 3 Pro and Claude Sonnet 3.7 into easy targets.