Search papers, labs, and topics across Lattice.
1
0
2
Achieve nearly 3x faster LLM inference by intelligently splitting the workload between edge devices and the cloud, without any training.