Search papers, labs, and topics across Lattice.
University of Chinese Academy of Sciences, Beijing, China
4
0
6
27
Forget text-dominance: Today's Omni-modal LLMs surprisingly favor visual inputs, creating new challenges for cross-modal reasoning.
LLMs exhibit a "Utopian bias" when simulating human behavior, converging towards an unrealistic "positive average person" and failing to capture individual differences and long-tail behaviors.
LLMs trained with reinforcement learning from verifiable rewards (RLVR) become overconfident in incorrect answers, but a simple fix鈥攄ecoupling reasoning and calibration objectives鈥攃an restore proper calibration without sacrificing accuracy.
By grounding reflection in the visual artifacts of presentation slides, DeepPresenter enables agents to iteratively refine presentations in a way that internal reasoning traces alone cannot.