Search papers, labs, and topics across Lattice.
2
0
5
You can now audit the privacy of text-to-image diffusion models even without access to the original text captions, thanks to a new attack that synthesizes its own conditioning inputs.
Language models harbor hidden "PII leakage knobs" – universal activation directions that, when tweaked, dramatically increase the generation of sensitive personal information.