Scaling MLPs: A Tale of Inductive Bias
–Neural Information Processing Systems
In this work we revisit the most fundamental building block in deep learning, the multi-layer perceptron (MLP), and study the limits of its performance on vision tasks. Empirical insights into MLPs are important for multiple reasons. To that end, MLPs offer an ideal test bed, as they lack any vision-specific inductive bias. Surprisingly, experimental datapoints for MLPs are very difficult to find in the literature, especially when coupled with large pre-training protocols. This discrepancy between practice and theory is worrying: \textit{Do MLPs reflect the empirical advances exhibited by practical models?}
Neural Information Processing Systems
Jan-19-2025, 21:10:12 GMT
- Technology: