Search papers, labs, and topics across Lattice.
1
0
3
Safe RL policies, designed to avoid unsafe actions, can be effectively attacked using a novel framework that learns safety constraints from demonstrations and then crafts adversarial perturbations, even without access to the target policy's gradients.