Search papers, labs, and topics across Lattice.
2
0
6
1
VLA models get a 1.73x speedup with only 5-7% overhead thanks to RAPID, a new edge-cloud collaborative inference framework that smartly handles visual noise and motion continuity.
Attention entropy reveals exploitable sparsity in VAR models, enabling 3.4x faster image generation without sacrificing quality.