Search papers, labs, and topics across Lattice.
1
1
3
5
LLM safety mechanisms are more vulnerable than we thought: psychological priming attacks achieve near-perfect success rates in eliciting harmful content across a wide range of models, including GPT-4o and Llama-3.2.