Search papers, labs, and topics across Lattice.
4
0
7
22
HERMES++ achieves state-of-the-art performance in both future point cloud prediction and 3D scene understanding, proving that a unified world model can outperform specialized approaches in autonomous driving.
Current vision-language-action models falter in dynamic robotic manipulation due to limited data and poor spatiotemporal reasoning, but a new dataset and architecture close the gap.
VideoLLMs can now watch and think *simultaneously*, achieving 15x faster response times and improved accuracy on video understanding tasks.
Unleashing powerful reasoning in OLLMs doesn't require expensive training data or compute – just clever guidance from existing Large Reasoning Models.