Search papers, labs, and topics across Lattice.
Communication University of China
3
0
6
Nemotron 3 Super proves you can achieve comparable accuracy to existing 120B models, but with significantly higher inference throughput, by combining Mamba, Attention, and Mixture-of-Experts.
Real-time, lightweight image compression is now possible with diffusion models, thanks to a novel architecture that swaps transformers for convolutions and prioritizes compression-focused pre-training.
LLMs that ace math and physics still struggle with general reasoning, achieving only 63% accuracy on a new K-12 level benchmark.