Search papers, labs, and topics across Lattice.
2
0
4
0
Ditching caches for compiler-managed data streams, Li Auto's M100 architecture achieves higher utilization than GPUs on autonomous driving tasks, hinting at a new path for efficient AI inference.
VLMs can achieve better multimodal reasoning simply by dynamically rescaling positional indices based on information density, without any training or architectural changes.