Search papers, labs, and topics across Lattice.
2
0
3
Train massive MoEs on Hopper GPUs faster and with less memory, even without native FP4 support, by cleverly quantizing activations and communication.
By strategically amplifying updates along flat directions in the loss landscape, LITE unlocks faster LLM pre-training with existing matrix-based optimizers like Muon and SOAP.