Search papers, labs, and topics across Lattice.
2
0
4
4
LLMs can be jailbroken with 90% success by subtly "salami slicing" harmful intent across multiple turns, even against state-of-the-art models like GPT-4o and Gemini.
Stop relying on alignment for LLM agent security: ClawGuard offers deterministic protection against indirect prompt injection by enforcing user-defined rules at tool-call boundaries.