Stochastic Quasi-Newton Optimization in Large Dimensions Including Deep Network Training
Suman, Uttam, Mamajiwala, Mariya, Saxena, Mukul, Tyagi, Ankit, Roy, Debasish
–arXiv.org Artificial Intelligence
Our proposal is on a new stochastic optimizer for non-convex and possibly non-smooth objective functions typically defined over large dimensional design spaces. Towards this, we have tried to bridge noise-assisted global search and faster local convergence, the latter being the characteristic feature of a Newton-like search. Our specific scheme -- acronymed FINDER (Filtering Informed Newton-like and Derivative-free Evolutionary Recursion), exploits the nonlinear stochastic filtering equations to arrive at a derivative-free update that has resemblance with the Newton search employing the inverse Hessian of the objective function. Following certain simplifications of the update to enable a linear scaling with dimension and a few other enhancements, we apply FINDER to a range of problems, starting with some IEEE benchmark objective functions to a couple of archetypal data-driven problems in deep networks to certain cases of physics-informed deep networks. The performance of the new method vis-\'a-vis the well-known Adam and a few others bears evidence to its promise and potentialities for large dimensional optimization problems of practical interest.
arXiv.org Artificial Intelligence
Oct-18-2024
- Country:
- Asia > India (0.28)
- North America > United States (0.28)
- Genre:
- Research Report (0.40)
- Technology: