Search papers, labs, and topics across Lattice.
University of Amsterdam
2
0
5
0
Forget massive offline datasets: TCL slashes tensor program optimization time by 16x while *improving* inference latency, thanks to a Mamba-powered cost model and continual learning.
Ditch the GPU? Active Imitation Learning can tame the thermal chaos of running large foundation models on 3D-stacked CPUs, unlocking a cost-effective alternative for LFM inference.