Review for NeurIPS paper: Practical Quasi-Newton Methods for Training Deep Neural Networks
–Neural Information Processing Systems
This paper develops a quasi-newton optimization algorithm for training fully connected neural networks. The authors develop an LBFGS like approximation of the Hessian via a block kronecker product factorization. On the experimental side they demonstrate the acceleration effect of the proposed methods for autoencoder feed-forward neural network models with either nine or thirteen layers applied to three datasets. The authors also provide some theoretical guarantees on convergence to stationarity with additional assumptions. All reviewers found the paper interesting.
Neural Information Processing Systems
Jan-22-2025, 04:14:00 GMT
- Technology: