intra order-preserving function
Intra Order-preserving Functions for Calibration of Multi-Class Neural Networks
Predicting calibrated confidence scores for multi-class deep networks is important for avoiding rare but costly mistakes. A common approach is to learn a post-hoc calibration function that transforms the output of the original network into calibrated confidence scores while maintaining the network's accuracy. However, previous post-hoc calibration techniques work only with simple calibration functions, potentially lacking sufficient representation to calibrate the complex function landscape of deep networks. In this work, we aim to learn general post-hoc calibration functions that can preserve the top-k predictions of any deep network. We call this family of functions intra order-preserving functions. We propose a new neural network architecture that represents a class of intra order-preserving functions by combining common neural network components. Additionally, we introduce order-invariant and diagonal sub-families, which can act as regularization for better generalization when the training data size is small. We show the effectiveness of the proposed method across a wide range of datasets and classifiers. Our method outperforms state-of-the-art post-hoc calibration methods, namely temperature scaling and Dirichlet calibration, in several evaluation metrics for the task.
A Missing Proofs A.1 Proof of Theorem 1, Intra Order-preserving Functions Theorem 1. A continuous function f: R
By Lemma 1, we know ˆ f is continuous and therefore w is also continuous. U w ( y), where U, w, and y are in Theorem 1. To prove Theorem 2, we first study the properties of order invariant functions in Appendix A.2.1. Theorem 1 to prove Theorem 2 in Appendix A.2.2. In fact, every order invariant function is equality-preserving.
- Oceania > Australia > New South Wales > Sydney (0.04)
- North America > United States > California (0.04)
- North America > Canada (0.04)
- Oceania > Australia > New South Wales > Sydney (0.04)
- North America > United States > California (0.04)
- North America > Canada > British Columbia > Metro Vancouver Regional District > Vancouver (0.04)
Review for NeurIPS paper: Intra Order-preserving Functions for Calibration of Multi-Class Neural Networks
Additional Feedback: The proposed methods perform very strongly in ECE, slightly better than the state-of-the-art in NLL and slightly worse in classwise-ECE. It would be good to have some explanation about why ECE and classwise-ECE give so different results. As ECE studies the calibration of only the class with the highest predicted probability and ignores other class probabilities, does it mean that the proposed method is better than the state-of-the-art in top-1 probability but slightly weaker on other classes? In the appendix provided as supplemental material, at lines 739-742 it is claimed that ECE does not suffer from the same problem that is highlighted about classwise-ECE at lines 731-738. While this is technically correct, it misses the point. Actually, ECE also suffers from essentially the same problem.
Review for NeurIPS paper: Intra Order-preserving Functions for Calibration of Multi-Class Neural Networks
The paper addresses the problem of post-training calibration in deep nets which has seen a lot of interest in the community lately. The work brings theoretical and practical contributions that are valuable to both researchers and practitioners. The proposed technique may also be useful for other problems besides post-training calibration. I would encourage the authors to try to improve the clarity of the presentation.
Intra Order-preserving Functions for Calibration of Multi-Class Neural Networks
Predicting calibrated confidence scores for multi-class deep networks is important for avoiding rare but costly mistakes. A common approach is to learn a post-hoc calibration function that transforms the output of the original network into calibrated confidence scores while maintaining the network's accuracy. However, previous post-hoc calibration techniques work only with simple calibration functions, potentially lacking sufficient representation to calibrate the complex function landscape of deep networks. In this work, we aim to learn general post-hoc calibration functions that can preserve the top-k predictions of any deep network. We call this family of functions intra order-preserving functions.
Intra Order-preserving Functions for Calibration of Multi-Class Neural Networks
Rahimi, Amir, Shaban, Amirreza, Cheng, Ching-An, Boots, Byron, Hartley, Richard
Predicting calibrated confidence scores for multi-class deep networks is important for avoiding rare but costly mistakes. A common approach is to learn a post-hoc calibration function that transforms the output of the original network into calibrated confidence scores while maintaining the network's accuracy. However, previous post-hoc calibration techniques work only with simple calibration functions, potentially lacking sufficient representation to calibrate the complex function landscape of deep networks. In this work, we aim to learn general post-hoc calibration functions that can preserve the top-k predictions of any deep network. We call this family of functions intra order-preserving functions. We propose a new neural network architecture that represents a class of intra order-preserving functions by combining common neural network components. Additionally, we introduce order-invariant and diagonal sub-families, which can act as regularization for better generalization when the training data size is small. We show the effectiveness of the proposed method across a wide range of datasets and classifiers. Our method outperforms state-of-the-art post-hoc calibration methods, namely temperature scaling and Dirichlet calibration, in multiple settings.