Towards Understanding the Mixture-of-Experts Layer in Deep Learning
–Neural Information Processing Systems
The Mixture-of-Experts (MoE) layer, a sparsely-activated model controlled by a router, has achieved great success in deep learning. However, the understanding of such architecture remains elusive. In this paper, we formally study how the MoE layer improves the performance of neural network learning and why the mixture model will not collapse into a single model. Our empirical results suggest that the cluster structure of the underlying problem and the non-linearity of the expert are pivotal to the success of MoE. This motivates us to consider a challenging classification problem with intrinsic cluster structures.
Neural Information Processing Systems
Jan-17-2025, 20:07:29 GMT
- Technology: