Incremental Algorithms for Hierarchical Classification
Cesa-bianchi, Nicolò, Gentile, Claudio, Tironi, Andrea, Zaniboni, Luca
–Neural Information Processing Systems
We study the problem of hierarchical classification when labels corresponding topartial and/or multiple paths in the underlying taxonomy are allowed. We introduce a new hierarchical loss function, the H-loss, implementing thesimple intuition that additional mistakes in the subtree of a mistaken class should not be charged for. Based on a probabilistic data model introduced in earlier work, we derive the Bayes-optimal classifier for the H-loss. We then empirically compare two incremental approximations ofthe Bayes-optimal classifier with a flat SVM classifier and with classifiers obtained by using hierarchical versions of the Perceptron and SVM algorithms. The experiments show that our simplest incremental approximationof the Bayes-optimal classifier performs, after just one training epoch, nearly as well as the hierarchical SVM classifier (which performs best). For the same incremental algorithm we also derive an H-loss bound showing, when data are generated by our probabilistic data model, exponentially fast convergence to the H-loss of the hierarchical classifier based on the true model parameters.
Neural Information Processing Systems
Dec-31-2005
- Country:
- Europe
- Austria > Styria
- Graz (0.04)
- Italy > Lombardy
- Milan (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- Austria > Styria
- North America > United States
- Pennsylvania > Allegheny County > Pittsburgh (0.04)
- Europe
- Technology: