Neural Learning in Structured Parameter Spaces - Natural Riemannian Gradient
–Neural Information Processing Systems
The parameter space of neural networks has a Riemannian metric structure. The natural Riemannian gradient should be used instead of the conventional gradient, since the former denotes the true steepest descent direction of a loss function in the Riemannian space. The behavior of the stochastic gradient learning algorithm is much more effective if the natural gradient is used. The present paper studies the information-geometrical structure of perceptrons and other networks, and prove that the online learning method based on the natural gradient is asymptotically as efficient as the optimal batch algorithm. Adaptive modification of the learning constant is proposed and analyzed in terms of the Riemannian measure and is shown to be efficient.
Neural Information Processing Systems
Dec-31-1997
- Country:
- North America > United States > Massachusetts > Middlesex County > Cambridge (0.14)
- Industry:
- Education > Educational Setting > Online (0.39)
- Technology: