Search papers, labs, and topics across Lattice.
1
0
3
MLLM training gets a 1.36x speed boost with Dynamic Hybrid Parallelism (DHP), which adaptively optimizes parallelism strategies to handle the data heterogeneity that plagues multimodal datasets.