Measuring Stochastic Data Complexity with Boltzmann Influence Functions
Ng, Nathan, Grosse, Roger, Ghassemi, Marzyeh
–arXiv.org Artificial Intelligence
Estimating the uncertainty of a model's prediction on a test point is a crucial part of ensuring reliability and calibration under distribution shifts. A minimum description length approach to this problem uses the predictive normalized maximum likelihood (pNML) distribution, which considers every possible label for a data point, and decreases confidence in a prediction if other labels are also consistent with the model and training data. In this work we propose IF-COMP, a scalable and efficient approximation of the pNML distribution that linearizes the model with a temperature-scaled Boltzmann influence function. IF-COMP can be used to produce well-calibrated predictions on test points as well as measure complexity in both labelled and unlabelled settings. We experimentally validate IF-COMP on uncertainty calibration, mislabel detection, and OOD detection tasks, where it consistently matches or beats strong baseline methods.
arXiv.org Artificial Intelligence
Jun-4-2024
- Country:
- North America
- Canada > Ontario
- Toronto (0.14)
- United States > New York
- New York County > New York City (0.14)
- Canada > Ontario
- North America
- Genre:
- Research Report (1.00)
- Industry:
- Health & Medicine (0.68)