Search papers, labs, and topics across Lattice.
3
0
8
Nemotron 3 Super proves you can achieve comparable accuracy to existing 120B models, but with significantly higher inference throughput, by combining Mamba, Attention, and Mixture-of-Experts.
LG's EXAONE 4.5 shows that strategically curating training data, particularly document-centric corpora, unlocks substantial gains in specialized tasks like document understanding and Korean contextual reasoning, even while maintaining competitive general performance.
LLM agents can appear to reason well (high entropy) while completely ignoring the input, and mutual information is a far better metric for catching this failure.