Beyond Top-Class Agreement: Using Divergences to Forecast Performance under Distribution Shift
Schirmer, Mona, Zhang, Dan, Nalisnick, Eric
–arXiv.org Artificial Intelligence
Knowing if a model will generalize to data 'in the wild' is crucial for safe deployment. To this end, we study model disagreement notions that consider the full predictive distribution - specifically disagreement based on Hellinger distance, Jensen-Shannon and Kullback-Leibler divergence. We find that divergence-based scores provide better test error estimates and detection rates on out-of-distribution data compared to their top-1 counterparts. Experiments involve standard vision and foundation models.
arXiv.org Artificial Intelligence
Dec-13-2023
- Country:
- Europe
- Germany > Baden-Württemberg
- Tübingen Region > Tübingen (0.04)
- Netherlands > North Holland
- Amsterdam (0.04)
- Portugal > Porto
- Porto (0.04)
- Germany > Baden-Württemberg
- Europe
- Genre:
- Research Report (0.64)
- Technology: