Search papers, labs, and topics across Lattice.
University of Michigan
2
0
6
0
LLMs that ace hazard recognition in question answering settings still fail to mitigate those same hazards when embodied in a simulated kitchen environment.
Even after removing names and pronouns, language models can still guess an applicant's gender from recommendation letters with surprising accuracy, revealing hidden biases lurking in seemingly objective text.