Extended Regularization Methods for Nonconvergent Model Selection

Finnoff, W., Hergert, F., Zimmermann, H. G.

Neural Information Processing Systems 

Rep. Germany Abstract Many techniques for model selection in the field of neural networks correspond to well established statistical methods. The method of'stopped training', on the other hand, in which an oversized network is trained until the error on a further validation set of examples deteriorates,then training is stopped, is a true innovation, since model selection doesn't require convergence of the training process. Inthis paper we show that this performance can be significantly enhanced by extending the'nonconvergent model selection method' of stopped training to include dynamic topology modifications (dynamic weight pruning) and modified complexity penalty term methods in which the weighting of the penalty term is adjusted during the training process. 1 INTRODUCTION One of the central topics in the field of neural networks is that of model selection. Both the theoretical and practical side of this have been intensively investigated and a vast array of methods have been suggested to perform this task. A widely used class of techniques starts by choosing an'oversized' network architecture then either removing redundant elements based on some measure of saliency (pruning), adding a further term to the cost function penalizing complexity (penalty terms), and finally, observing the error on a further validation set of examples, then stopping training as soon as this performance begins to deteriorate (stopped training).

Similar Docs  Excel Report  more

TitleSimilaritySource
None found