Search papers, labs, and topics across Lattice.
1
0
3
LLMs are shockingly vulnerable to jailbreaks in South Asian languages, with models reaching 1.0 Jailbreak Success Rate in free-form prompts, even after inflated refusal rates in contract-based settings.