Search papers, labs, and topics across Lattice.
2
0
3
0
Squeeze your embodied AI models: DyQ-VLA cuts memory footprint by 70% and speeds up inference by 40% without sacrificing performance, all by dynamically adjusting bit-widths based on real-time kinematic data.
VLA models get a 1.73x speedup with only 5-7% overhead thanks to RAPID, a new edge-cloud collaborative inference framework that smartly handles visual noise and motion continuity.