Goto

Collaborating Authors

 icm


A Graphical Terminology An arbitrary graph

Neural Information Processing Systems

We refer the readers to ( Peters et al., 2017) for more detailed graphical terminology. We base our proof mostly on ( Kirsch, 2019). The first statement follows directly from the first theorem in ( Haviland, 1936). Without loss of generality, we reorder the variables according to reversed topological ordering, i.e. a Follows directly from Lemma 1. Lemma 4. Recall condition 2) in Causal de Finetti states that 8 i, 8 n 2 N: X The first equality holds by well-defindedness. The fourth equality follow from well-definedness.



A Graphical Terminology An arbitrary graph

Neural Information Processing Systems

We refer the readers to ( Peters et al., 2017) for more detailed graphical terminology. We base our proof mostly on ( Kirsch, 2019). The first statement follows directly from the first theorem in ( Haviland, 1936). Without loss of generality, we reorder the variables according to reversed topological ordering, i.e. a Follows directly from Lemma 1. Lemma 4. Recall condition 2) in Causal de Finetti states that 8 i, 8 n 2 N: X The first equality holds by well-defindedness. The fourth equality follow from well-definedness.


Curiosity-Driven LLM-as-a-judge for Personalized Creative Judgment

Kumar, Vanya Bannihatti, Goyal, Divyanshu, Eppa, Akhil, Bhandari, Neel

arXiv.org Artificial Intelligence

Creative Thinking(TTCW) benchmark introduced in Chakrabarty et al. (2024), Rigorous, standardized evaluation has repeatedly catalyzed progress in machine learning, from ImageNetRussakovsky et al. (2015) and GLUEWang et al. (2019), driving leaps in the fields of computer vision and Natural Language Processing, respectively. The same effect is evident in objective math reasoning, where benchmarks like GSM8KCobbe et al. (2021), together with RL-trained reasoning models such as OpenAI's o1OpenAI et al. (2024) and DeepSeek-R1DeepSeek-AI Models(LLM) as a judge prefer their own generations making them unreliable. As shown in Chakrabarty et al. (2024) and Table 12 and Table 2, even Specifically, when the model is "surprised" by an expert's explanation, it signals a mismatch between the LLM's prior belief and the expert's The intuition behind predicting the annotator is that the model can learn which annotator caused the belief shift, allowing it to calibrate the curiosity signal for each annotator individually, thereby improving personalization. In our experiments, we establish a baseline using an SFT model that predicts annotators' binary More details about the results can be found in Fig 4.Figure 1: Overview of Architecture during training for Curiosity Driven LLM-as-a-judgeFigure 2: Overview of Architecture during inference for Curiosity Driven LLM-as-a-judge 2 (a) Baseline without using explanations (b) Baseline using explanations TTCW dataset Chakrabarty et al. (2024), which is based on the Torrance Test of Creative Thinking Torrance (1966) but adapted for LLMs. All the distinct dimensions in the TTCW dataset are mentioned in Appendix A.1.


Intra-Cluster Mixup: An Effective Data Augmentation Technique for Complementary-Label Learning

Mai, Tan-Ha, Lin, Hsuan-Tien

arXiv.org Artificial Intelligence

In this paper, we investigate the challenges of complementary-label learning (CLL), a specialized form of weakly-supervised learning (WSL) where models are trained with labels indicating classes to which instances do not belong, rather than standard ordinary labels. This alternative supervision is appealing because collecting complementary labels is generally cheaper and less labor-intensive. Although most existing research in CLL emphasizes the development of novel loss functions, the potential of data augmentation in this domain remains largely underexplored. In this work, we uncover that the widely-used Mixup data augmentation technique is ineffective when directly applied to CLL. Through in-depth analysis, we identify that the complementary-label noise generated by Mixup negatively impacts the performance of CLL models. We then propose an improved technique called Intra-Cluster Mixup (ICM), which only synthesizes augmented data from nearby examples, to mitigate the noise effect. ICM carries the benefits of encouraging complementary label sharing of nearby examples, and leads to substantial performance improvements across synthetic and real-world labeled datasets. In particular, our wide spectrum of experimental results on both balanced and imbalanced CLL settings justifies the potential of ICM in allying with state-of-the-art CLL algorithms, achieving significant accuracy increases of 30% and 10% on MNIST and CIF AR datasets, respectively.


Empirical Validation of the Independent Chip Model

Kim, Juho

arXiv.org Artificial Intelligence

The independent chip model (ICM) forms a cornerstone of all modern poker tournament strategy. However, despite its prominence, the ICM's performance in the real world has not been sufficiently scrutinized, especially at a large scale. In this paper, we introduce our new dataset of poker tournaments, consisting of results of over ten thousand events. Then, using this dataset, we perform two experiments as part of a large-scale empirical validation of the ICM. First, we verify that the ICM performs more accurately than a baseline we propose. Second, we obtain empirical evidence of the ICM underestimating the performances of players with larger stacks while overestimating those who are short-stacked. Our contributions may be useful to future researchers developing new algorithms for estimating a player's value in poker tournaments.



Interpretable Learning Dynamics in Unsupervised Reinforcement Learning

Pandey, Shashwat

arXiv.org Machine Learning

We present an interpretability framework for unsupervised reinforcement learning (URL) agents, aimed at understanding how intrinsic motivation shapes attention, behavior, and representation learning. We analyze five agents DQN, RND, ICM, PPO, and a Transformer-RND variant trained on procedurally generated environments, using Grad-CAM, Layer-wise Relevance Propagation (LRP), exploration metrics, and latent space clustering. To capture how agents perceive and adapt over time, we introduce two metrics: attention diversity, which measures the spatial breadth of focus, and attention change rate, which quantifies temporal shifts in attention. Our findings show that curiosity-driven agents display broader, more dynamic attention and exploratory behavior than their extrinsically motivated counterparts. Among them, TransformerRND combines wide attention, high exploration coverage, and compact, structured latent representations. Our results highlight the influence of architectural inductive biases and training signals on internal agent dynamics. Beyond reward-centric evaluation, the proposed framework offers diagnostic tools to probe perception and abstraction in RL agents, enabling more interpretable and generalizable behavior.


On the Role of Priors in Bayesian Causal Learning

Geiger, Bernhard C., Kern, Roman

arXiv.org Machine Learning

--In this work, we investigate causal learning of independent causal mechanisms from a Bayesian perspective. Confirming previous claims from the literature, we show in a didactically accessible manner that unlabeled data (i.e., cause realizations) do not improve the estimation of the parameters defining the mechanism. Furthermore, we observe the importance of choosing an appropriate prior for the cause and mechanism parameters, respectively. Specifically, we show that a factorized prior results in a factorized posterior, which resonates with Janz-ing and Sch olkopf's definition of independent causal mechanisms via the Kolmogorov complexity of the involved distributions and with the concept of parameter independence of Heckerman et al. Impact Statement --Learning the effect from a given cause is an important problem in many engineering disciplines, specifically in the field of surrogate modeling, which aims to reduce the computational cost of numerical simulations. Causal learning, however, cannot make use of unlabeled data - i.e., cause realizations - if the mechanism that produces the effect is independent from the cause. In this work, we recover this well-known fact from a Bayesian perspective.


Inverse Flow and Consistency Models

Zhang, Yuchen, Zhou, Jian

arXiv.org Artificial Intelligence

Inverse generation problems, such as denoising without ground truth observations, is a critical challenge in many scientific inquiries and real-world applications. While recent advances in generative models like diffusion models, conditional flow matching, and consistency models achieved impressive results by casting generation as denoising problems, they cannot be directly used for inverse generation without access to clean data. Here we introduce Inverse Flow (IF), a novel framework that enables using these generative models for inverse generation problems including denoising without ground truth. Inverse Flow can be flexibly applied to nearly any continuous noise distribution and allows complex dependencies. We propose two algorithms for learning Inverse Flows, Inverse Flow Matching (IFM) and Inverse Consistency Model (ICM). Notably, to derive the computationally efficient, simulation-free inverse consistency model objective, we generalized consistency training to any forward diffusion processes or conditional flows, which have applications beyond denoising. We demonstrate the effectiveness of IF on synthetic and real datasets, outperforming prior approaches while enabling noise distributions that previous methods cannot support. Finally, we showcase applications of our techniques to fluorescence microscopy and single-cell genomics data, highlighting IF's utility in scientific problems. Overall, this work expands the applications of powerful generative models to inversion generation problems.