Asymptotic Theory for Regularization: One-Dimensional Linear Case

Koistinen, Petri

Neural Information Processing Systems 

The generalization ability of a neural network can sometimes be improved dramatically by regularization. To analyze the improvement oneneeds more refined results than the asymptotic distribution ofthe weight vector. Here we study the simple case of one-dimensional linear regression under quadratic regularization, i.e., ridge regression. We study the random design, misspecified case, where we derive expansions for the optimal regularization parameter andthe ensuing improvement. It is possible to construct examples where it is best to use no regularization.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found