On Last-Layer Algorithms for Classification: Decoupling Representation from Uncertainty Estimation
Brosse, Nicolas, Riquelme, Carlos, Martin, Alice, Gelly, Sylvain, Moulines, Éric
Uncertainty quantification for deep learning is a challenging open problem. Bayesian statistics offer a mathematically grounded framework to reason about uncertainties; however, approximate posteriors for modern neural networks still require prohibitive computational costs. We propose a family of algorithms which split the classification task into two stages: representation learning and uncertainty estimation. We compare four specific instances, where uncertainty estimation is performed via either an ensemble of Stochastic Gradient Descent or Stochastic Gradient Langevin Dynamics snapshots, an ensemble of bootstrapped logistic regressions, or via a number of Monte Carlo Dropout passes. We evaluate their performance in terms of \emph{selective} classification (risk-coverage), and their ability to detect out-of-distribution samples. Our experiments suggest there is limited value in adding multiple uncertainty layers to deep classifiers, and we observe that these simple methods strongly outperform a vanilla point-estimate SGD in some complex benchmarks like ImageNet.
Jan-22-2020
- Country:
- Asia
- China (0.04)
- Middle East
- Europe
- France (0.04)
- Spain > Andalusia
- Cádiz Province > Cadiz (0.04)
- Switzerland > Zürich
- Zürich (0.04)
- United Kingdom > Scotland
- City of Edinburgh > Edinburgh (0.04)
- North America
- Canada > Quebec
- Montreal (0.04)
- United States
- California > San Diego County
- San Diego (0.04)
- District of Columbia > Washington (0.04)
- Massachusetts > Middlesex County
- Cambridge (0.04)
- New York > New York County
- New York City (0.04)
- California > San Diego County
- Canada > Quebec
- Oceania > Australia
- New South Wales > Sydney (0.04)
- Asia
- Genre:
- Research Report
- Experimental Study (0.54)
- New Finding (0.66)
- Research Report
- Industry:
- Health & Medicine (0.46)
- Technology: