Goto

Collaborating Authors

 mll



219e052492f4008818b8adb6366c7ed6-Paper.pdf

Neural Information Processing Systems

Absent assumptions onthe nature of shift, the problem is underspecified. Multiple assumptions may be compatible with the same observations while implying different courses ofaction.


A Unified View of Label Shift Estimation

Neural Information Processing Systems

Under label shift, the label distribution $p(y)$ might change but the class-conditional distributions $p(x|y)$ do not. There are two dominant approaches for estimating the label marginal. BBSE, a moment-matching approach based on confusion matrices, is provably consistent and provides interpretable error bounds. However, a maximum likelihood estimation approach, which we call MLLS, dominates empirically. In this paper, we present a unified view of the two methods and the first theoretical characterization of MLLS. Our contributions include (i) consistency conditions for MLLS, which include calibration of the classifier and a confusion matrix invertibility condition that BBSE also requires; (ii) a unified framework, casting BBSE as roughly equivalent to MLLS for a particular choice of calibration method; and (iii) a decomposition of MLLS's finite-sample error into terms reflecting miscalibration and estimation error. Our analysis attributes BBSE's statistical inefficiency to a loss of information due to coarse calibration.




Reviewer # 1: We thank you for appreciating our contributions and providing valuable feedback, which will be taken

Neural Information Processing Systems

The empirical results comparing parameter tying vs. naive design are in fact reported in Table 3 of Appendix C.2; a Zhou, 2018) are related to IPVI, as you have suggested. We would like to address your comments and questions below. Regarding the necessity of parameter tying, we think overfitting is still an issue to be addressed. We provide some experimental evidence below, as you have suggested. Train/test mean log-likelihood (MLL) achieved by IPVI with and without parameter tying over 10 runs.


A Unified View of Label Shift Estimation

Neural Information Processing Systems

Under label shift, the label distribution p(y) might change but the class-conditional distributions p(x y) do not. There are two dominant approaches for estimating the label marginal. BBSE, a moment-matching approach based on confusion matrices, is provably consistent and provides interpretable error bounds. However, a maximum likelihood estimation approach, which we call MLLS, dominates empirically. In this paper, we present a unified view of the two methods and the first theoretical characterization of MLLS.


Exploring Ordinality in Text Classification: A Comparative Study of Explicit and Implicit Techniques

Kasa, Siva Rajesh, Goel, Aniket, Gupta, Karan, Roychowdhury, Sumegh, Bhanushali, Anish, Pattisapu, Nikhil, Murthy, Prasanna Srinivasa

arXiv.org Artificial Intelligence

Ordinal Classification (OC) is a widely encountered challenge in Natural Language Processing (NLP), with applications in various domains such as sentiment analysis, rating prediction, and more. Previous approaches to tackle OC have primarily focused on modifying existing or creating novel loss functions that \textbf{explicitly} account for the ordinal nature of labels. However, with the advent of Pretrained Language Models (PLMs), it became possible to tackle ordinality through the \textbf{implicit} semantics of the labels as well. This paper provides a comprehensive theoretical and empirical examination of both these approaches. Furthermore, we also offer strategic recommendations regarding the most effective approach to adopt based on specific settings.


Rethinking Loss Functions for Fact Verification

Mukobara, Yuta, Shigeto, Yutaro, Shimbo, Masashi

arXiv.org Artificial Intelligence

We explore loss functions for fact verification in the FEVER shared task. While the cross-entropy loss is a standard objective for training verdict predictors, it fails to capture the heterogeneity among the FEVER verdict classes. In this paper, we develop two task-specific objectives tailored to FEVER. Experimental results confirm that the proposed objective functions outperform the standard cross-entropy. Performance is further improved when these objectives are combined with simple class weighting, which effectively overcomes the imbalance in the training data. The souce code is available at https://github.com/yuta-mukobara/RLF-KGAT


Particle-Based Score Estimation for State Space Model Learning in Autonomous Driving

Singh, Angad, Makhlouf, Omar, Igl, Maximilian, Messias, Joao, Doucet, Arnaud, Whiteson, Shimon

arXiv.org Artificial Intelligence

Multi-object state estimation is a fundamental problem for robotic applications where a robot must interact with other moving objects. Typically, other objects' relevant state features are not directly observable, and must instead be inferred from observations. Particle filtering can perform such inference given approximate transition and observation models. However, these models are often unknown a priori, yielding a difficult parameter estimation problem since observations jointly carry transition and observation noise. In this work, we consider learning maximum-likelihood parameters using particle methods. Recent methods addressing this problem typically differentiate through time in a particle filter, which requires workarounds to the non-differentiable resampling step, that yield biased or high variance gradient estimates. By contrast, we exploit Fisher's identity to obtain a particle-based approximation of the score function (the gradient of the log likelihood) that yields a low variance estimate while only requiring stepwise differentiation through the transition and observation models. We apply our method to real data collected from autonomous vehicles (AVs) and show that it learns better models than existing techniques and is more stable in training, yielding an effective smoother for tracking the trajectories of vehicles around an AV.