Search papers, labs, and topics across Lattice.
1
0
3
On-device LLM inference can be sped up by an order of magnitude with a flexible TrustZone-based system that selectively protects memory and the NPU.