aeb7b30ef1d024a76f21a1d40e30c302-Paper.pdf

Neural Information Processing Systems 

Ideally, we want networks to be accurate, calibrated and confident. We show that, as opposed to the standard cross-entropy loss, focal loss [19] allows us to learn models that are already very well calibrated. When combined with temperature scaling, whilst preserving accuracy, it yields state-of-the-art calibrated models. We provide a thorough analysis of the factors causing miscalibration, and use the insights we glean from this to justify the empirically excellent performance of focal loss.

Similar Docs  Excel Report  more

TitleSimilaritySource
None found