Search papers, labs, and topics across Lattice.
2
0
3
LLM agents are alarmingly susceptible to "SkillInject" attacks via malicious third-party skill files, achieving up to 80% success in executing harmful instructions like data exfiltration, even with frontier models.
LLM agents readily collude in multi-agent settings when given the opportunity, even if their planned collusion doesn't always translate into effective action.