The Kernel Gibbs Sampler
Graepel, Thore, Herbrich, Ralf
–Neural Information Processing Systems
We present an algorithm that samples the hypothesis space of kernel classifiers.Given a uniform prior over normalised weight vectors and a likelihood based on a model of label noise leads to a piecewise constantposterior that can be sampled by the kernel Gibbs sampler (KGS). The KGS is a Markov Chain Monte Carlo method that chooses a random direction in parameter space and samples from the resulting piecewise constant density along the line chosen. The KGS can be used as an analytical tool for the exploration of Bayesian transduction, Bayes point machines, active learning, and evidence-based model selection on small data sets that are contaminated withlabel noise. For a simple toy example we demonstrate experimentally how a Bayes point machine based on the KGS outperforms anSVM that is incapable of taking into account label noise. 1 Introduction Two great ideas have dominated recent developments in machine learning: the application ofkernel methods and the popularisation of Bayesian inference. Focusing on the task of classification, various connections between the two areas exist: kernels havelong been a part of Bayesian inference in the disguise of covariance nmctions thatcharacterise priors over functions [9].
Neural Information Processing Systems
Dec-31-2001
- Country:
- Europe > Germany
- Berlin (0.05)
- North America
- Canada > Ontario
- Toronto (0.14)
- United States
- California > San Mateo County
- San Mateo (0.04)
- Wisconsin > Dane County
- Madison (0.14)
- California > San Mateo County
- Canada > Ontario
- Oceania > Australia
- Australian Capital Territory > Canberra (0.04)
- Europe > Germany