Search papers, labs, and topics across Lattice.
1
0
3
Mobile GPUs can now run large DNNs and multi-DNN workloads efficiently thanks to FlashMem, which slashes memory consumption by up to 8.4x and accelerates inference by up to 75x.