Dynamics of On-Line Gradient Descent Learning for Multilayer Neural Networks
–Neural Information Processing Systems
We consider the problem of online gradient descent learning for general two-layer neural networks. An analytic solution is presented and used to investigate the role of the learning rate in controlling the evolution and convergence of the learning process. Two-layer networks with an arbitrary number of hidden units have been shown to be universal approximators [1] for such N-to-one dimensional maps. We investigate the emergence of generalization ability in an online learning scenario [2], in which the couplings are modified after the presentation of each example so as to minimize the corresponding error. The resulting changes in {J} are described as a dynamical evolution; the number of examples plays the role of time.
Neural Information Processing Systems
Dec-31-1996
- Country:
- Europe
- Denmark > Capital Region
- Copenhagen (0.04)
- United Kingdom (0.04)
- Denmark > Capital Region
- North America > United States (0.04)
- Europe
- Industry:
- Education (0.67)
- Technology: