Search papers, labs, and topics across Lattice.
Shanghai Jiao Tong University
3
0
8
0
Omni-modal LLMs can ace captioning and QA, but AVID reveals they're surprisingly bad at spotting audio-visual inconsistencies in videos, a crucial skill for trustworthy AI.
Forget LLMs – generating realistic 3D scenes might just hinge on learning how objects relate to each other locally.
Forget hand-crafted LTV pipelines: AgentLTV uses LLM-driven agents to automatically search for and evolve high-performing models, adapting to diverse data patterns and improving prediction accuracy, especially for critical high-value and negative-LTV segments.