Some New Bounds on the Generalization Error of Combined Classifiers
Koltchinskii, Vladimir, Panchenko, Dmitriy, Lozano, Fernando
–Neural Information Processing Systems
In this paper we develop the method of bounding the generalization error of a classifier in terms of its margin distribution which was introduced in the recent papers of Bartlett and Schapire, Freund, Bartlett and Lee. The theory of Gaussian and empirical processes allow us to prove the margin type inequalities for the most general functional classes, the complexity of the class being measured via the so called Gaussian complexity functions. As a simple application of our results, we obtain the bounds of Schapire, Freund, Bartlett and Lee for the generalization error of boosting. We also substantially improve the results of Bartlett on bounding the generalization error of neural networks in terms of h -norms of the weights of neurons. Furthermore, under additional assumptions on the complexity of the class of hypotheses we provide some tighter bounds, which in the case of boosting improve the results of Schapire, Freund, Bartlett and Lee.
Neural Information Processing Systems
Dec-31-2001
- Country:
- North America > United States > New Mexico (0.16)
- Genre:
- Research Report > New Finding (0.34)
- Technology: