Search papers, labs, and topics across Lattice.
2
58
4
41
LLMs can move beyond simple refusals to actively guide vulnerable users towards safe outcomes, achieving state-of-the-art safety and robustness against jailbreaks.
RealSafe-R1 achieves safety alignment of DeepSeek-R1 without sacrificing reasoning performance, a common trade-off in prior safety alignment efforts.