Search papers, labs, and topics across Lattice.
Huawei Research
2
0
4
0
Video-LLMs can be sped up by nearly 3x without sacrificing performance, simply by loosening the strict matching requirements of speculative decoding and focusing on visual-semantic relevance.
Real-time, open-ended video understanding is now possible: AURA enables VideoLLMs to proactively respond to live video streams, moving beyond simple captioning.