Search papers, labs, and topics across Lattice.
The University of Tokyo, Riken
1
0
3
12
NeuronMoE slashes multilingual LLM parameter counts by 40% without sacrificing performance, by cleverly allocating experts based on neuron-level language specialization rather than blunt layer-level assignments.