Search papers, labs, and topics across Lattice.
SKLP, Institute of Computing Technology, Chinese Academy of Sciences
1
0
3
3
Wafer-scale SRAM CIM can deliver up to 17x better energy efficiency for LLM inference by eliminating off-chip data movement and using token-grained pipelining.