Goto

Collaborating Authors

 counterfactual example


COCO-Counterfactuals: Automatically Constructed Counterfactual Examples for Image-Text Pairs

Neural Information Processing Systems

Despite their demonstrated utility for NLP, multimodal counterfactual examples have been relatively unexplored due to the difficulty of creating paired image-text data with minimal counterfactual changes. To address this challenge, we introduce a scalable framework for automatic generation of counterfactual examples using text-to-image diffusion models.


COCO-Counterfactuals: Automatically Constructed Counterfactual Examples for Image-Text Pairs

Neural Information Processing Systems

Counterfactual examples have proven to be valuable in the field of natural language processing (NLP) for both evaluating and improving the robustness of language models to spurious correlations in datasets. Despite their demonstrated utility for NLP, multimodal counterfactual examples have been relatively unexplored due to the difficulty of creating paired image-text data with minimal counterfactual changes. To address this challenge, we introduce a scalable framework for automatic generation of counterfactual examples using text-to-image diffusion models. We use our framework to create COCO-Counterfactuals, a multimodal counterfactual dataset of paired image and text captions based on the MS-COCO dataset. We validate the quality of COCO-Counterfactuals through human evaluations and show that existing multimodal models are challenged by our counterfactual image-text pairs. Additionally, we demonstrate the usefulness of COCO-Counterfactuals for improving out-of-domain generalization of multimodal vision-language models via training data augmentation. We make our code and the COCO-Counterfactuals dataset publicly available.



LLM-Guided Synthetic Augmentation (LGSA) for Mitigating Bias in AI Systems

Karri, Sai Suhruth Reddy, Nallapuneni, Yashwanth Sai, Mallireddy, Laxmi Narasimha Reddy, G, Gopichand

arXiv.org Artificial Intelligence

This is the preprint version of the article "LLM - Guided Synthetic Augmentation (LGSA) for Mitigating Bias in AI Systems." This version is made available on arXiv for early dissemination. If accepted, the final authenticated version will be published in the respective venue. Dr. G opichand G School of Computer Science and Engineering Vellore Institute of Technology Vellore - 632014, TamilNadu, India gopichand.g@vit.ac.in Abstract -- Bias in Artificial Intelligence systems, especially those that rely on natural language data, brings up serious ethical and practical issues. When certain groups are underrepresented, it often leads to uneven performance across different demographics. Whil e traditional fairness methods like pre - processing, in - processing, and post - processing can be helpful, they usually depend on protected - attribute labels, create a trade - off between accuracy and fairness, and struggle to adapt across various datas ets. To tackle these challenges, this study presents LLM - Guided Synthetic Augmentation (LGSA), a process that leverages large language models to create counterfactual examples for underrepresented groups while keeping label integrity intact. We put LGSA to the test on a controlled dataset of short English sentences that included gendered pronouns, professions, and binary task labels. The process involved using structured prompts to a large language model to generate gender - swapped paraphrases, followed by a thorough quality control process. This included checking for semantic similarity, verifying attributes, screening for toxi city, and conducting human spot checks. The augmented dataset broadened training coverage and was utilized to train a classifier under consistent experimental conditions. The results showed that LGSA significantly lessens performance disparities without co mpromising accuracy. The baseline model achieved an impressive 96.7% accuracy but had a gender bias gap of 7.2%. A simple swap augmentation brought the gap down to 0.7% but also reduced accuracy to 95.6%. In contrast, LGSA achieved an overall accuracy of 9 9.1%, showing strong performance on female - labeled examples and a reduced gap of 1.9%. These results indicate that LGSA is a powerful and dependable strategy for mitigating bias. By generating diverse and semantically accurate counterfactuals, this method enhances the balance of subgroup performance, narrows bias gaps, and maintains high ove rall task accuracy and label fidelity, showcasing its potential as a practical framework for fairness - focused AI systems.


CaRT: Teaching LLM Agents to Know When They Know Enough

Liu, Grace, Qu, Yuxiao, Schneider, Jeff, Singh, Aarti, Kumar, Aviral

arXiv.org Artificial Intelligence

Many tasks require learned models to strategically gather relevant information over multiple rounds of interaction before actually acting on a task. Strategic information gathering requires models to know not only how to effectively acquire information, but also when to stop gathering information and make a decision, in order to avoid overthinking or getting derailed when acting. In this paper, we formalize this problem and introduce Counterfactuals and Reasoning for Termination (CaRT), an approach for teaching LLMs when to stop seeking information. To appropriately learn when to terminate, CaRT fine-tunes LLMs using counterfactual pairs of trajectories, one where termination is appropriate and a minimally modified version of the same trajectory where it is not. It trains the LLM to explain the rationale for the termination decision in either case via verbal reasoning, and imbues this capability into the base LLM via fine-tuning. We instantiate CaRT in two domains: interactive medical diagnosis and math problem solving. In both domains, we find that CaRT improves the efficiency of information gathering and task success rate compared to other fine-tuning methods.



Truth or Twist? Optimal Model Selection for Reliable Label Flipping Evaluation in LLM-based Counterfactuals

Wang, Qianli, Nguyen, Van Bach, Feldhus, Nils, Villa-Arenas, Luis Felipe, Seifert, Christin, Möller, Sebastian, Schmitt, Vera

arXiv.org Artificial Intelligence

Counterfactual examples are widely employed to enhance the performance and robustness of large language models (LLMs) through counterfactual data augmentation (CDA). However, the selection of the judge model used to evaluate label flipping, the primary metric for assessing the validity of generated counterfactuals for CDA, yields inconsistent results. To decipher this, we define four types of relationships between the counterfactual generator and judge models: being the same model, belonging to the same model family, being independent models, and having an distillation relationship. Through extensive experiments involving two state-of-the-art LLM-based methods, three datasets, four generator models, and 15 judge models, complemented by a user study (n = 90), we demonstrate that judge models with an independent, non-fine-tuned relationship to the generator model provide the most reliable label flipping evaluations. Relationships between the generator and judge models, which are closely aligned with the user study for CDA, result in better model performance and robustness. Nevertheless, we find that the gap between the most effective judge models and the results obtained from the user study remains considerably large. This suggests that a fully automated pipeline for CDA may be inadequate and requires human intervention.


Enhancing Chemical Explainability Through Counterfactual Masking

Janisiów, Łukasz, Kochańczyk, Marek, Zieliński, Bartosz, Danel, Tomasz

arXiv.org Artificial Intelligence

Molecular property prediction is a crucial task that guides the design of new compounds, including drugs and materials. While explainable artificial intelligence methods aim to scrutinize model predictions by identifying influential molecular substructures, many existing approaches rely on masking strategies that remove either atoms or atom-level features to assess importance via fidelity metrics. These methods, however, often fail to adhere to the underlying molecular distribution and thus yield unintuitive explanations. In this work, we propose counterfactual masking, a novel framework that replaces masked substructures with chemically reasonable fragments sampled from generative models trained to complete molecular graphs. Rather than evaluating masked predictions against implausible zeroed-out baselines, we assess them relative to counterfactual molecules drawn from the data distribution. Our method offers two key benefits: (1) molecular realism underpinning robust and distribution-consistent explanations, and (2) meaningful counterfactuals that directly indicate how structural modifications may affect predicted properties. We demonstrate that counterfactual masking is well-suited for benchmarking model explainers and yields more actionable insights across multiple datasets and property prediction tasks.


Explainable Counterfactual Reasoning in Depression Medication Selection at Multi-Levels (Personalized and Population)

Qin, Xinyu, Chignell, Mark H., Greifenberger, Alexandria, Lokuge, Sachinthya, Toumeh, Elssa, Sternat, Tia, Katzman, Martin, Wang, Lu

arXiv.org Artificial Intelligence

Background: This study investigates how variations in Major Depressive Disorder (MDD) symptoms, quantified by the Hamilton Rating Scale for Depression (HAM-D), causally influence the prescription of SSRIs versus SNRIs. Methods: We applied explainable counterfactual reasoning with counterfactual explanations (CFs) to assess the impact of specific symptom changes on antidepressant choice. Results: Among 17 binary classifiers, Random Forest achieved highest performance (accuracy, F1, precision, recall, ROC-AUC near 0.85). Sample-based CFs revealed both local and global feature importance of individual symptoms in medication selection. Conclusions: Counterfactual reasoning elucidates which MDD symptoms most strongly drive SSRI versus SNRI selection, enhancing interpretability of AI-based clinical decision support systems. Future work should validate these findings on more diverse cohorts and refine algorithms for clinical deployment.