When Expressivity Meets Trainability: Fewer than n Neurons Can Work
–Neural Information Processing Systems
Modern neural networks are often quite wide, causing large memory and computation costs. It is thus of great interest to train a narrower network. However, training narrow neural nets remains a challenging task. We ask two theoretical questions: Can narrow networks have as strong expressivity as wide ones? If so, does the loss function exhibit a benign optimization landscape?
Neural Information Processing Systems
Dec-24-2025, 02:23:18 GMT
- Technology: