Search papers, labs, and topics across Lattice.
1
0
3
Training LLMs on heterogeneous GPUs just got a whole lot faster: AutoHet achieves up to 1.79x speedup over Megatron-LM and Whale by automatically optimizing the 3D parallelism strategy.