Search papers, labs, and topics across Lattice.
University of California, San Diego
2
0
6
1
Forget GPU-centric designs: AMMA slashes attention latency by 15x and energy consumption by 7x with a memory-centric architecture for long-context LLMs.
LLMs can discover equations with significantly higher accuracy and robustness by explicitly modeling the multi-step reasoning process scientists use, rather than guessing directly from data.