Causal-HalBench: Uncovering LVLMs Object Hallucinations Through Causal Intervention

Xu, Zhe, Wang, Zhicai, Wu, Junkang, Lu, Jinda, Wang, Xiang

arXiv.org Artificial Intelligence 

Large Vision-Language Models (L VLMs) often suffer from object hallucination, making erroneous judgments about the presence of objects in images. We propose this primarily stems from spurious correlations arising when models strongly associate highly co-occurring objects during training, leading to hallucinated objects influenced by visual context. Current benchmarks mainly focus on hallucination detection but lack a formal characterization and quantitative evaluation of spurious correlations in L VLMs. To address this, we introduce causal analysis into the object recognition scenario of L VLMs, establishing a Structural Causal Model (SCM). Utilizing the language of causality, we formally define spurious correlations arising from co-occurrence bias. To quantify the influence induced by these spurious correlations, we develop Causal-HalBench, a benchmark specifically constructed with counterfactual samples and integrated with comprehensive causal metrics designed to assess model robustness against spurious correlations. Concurrently, we propose an extensible pipeline for the construction of these counterfactual samples, leveraging the capabilities of proprietary L VLMs and Text-to-Image (T2I) models for their generation. Our evaluations on mainstream L VLMs using Causal-HalBench demonstrate these models exhibit susceptibility to spurious correlations, albeit to varying extents.

Duplicate Docs Excel Report

Title
None found

Similar Docs  Excel Report  more

TitleSimilaritySource
None found