On Convergence and Generalization of Dropout Training
–Neural Information Processing Systems
We study dropout in two-layer neural networks with rectified linear unit (ReLU) activations. Under mild overparametrization and assuming that the limiting kernel can separate the data distribution with a positive margin, we show that the dropout training with logistic loss achieves \epsilon -suboptimality in the test error in O(1/\epsilon) iterations.
Neural Information Processing Systems
Jan-15-2025, 11:41:47 GMT
- Technology: