explanation 2
Explaining Multi-modal Large Language Models by Analyzing their Vision Perception
Giulivi, Loris, Boracchi, Giacomo
Multi-modal Large Language Models (MLLMs) have demonstrated remarkable capabilities in understanding and generating content across various modalities, such as images and text. However, their interpretability remains a challenge, hindering their adoption in critical applications. This research proposes a novel approach to enhance the interpretability of MLLMs by focusing on the image embedding component. We combine an open-world localization model with a MLLM, thus creating a new architecture able to simultaneously produce text and object localization outputs from the same vision embedding. The proposed architecture greatly promotes interpretability, enabling us to design a novel saliency map to explain any output token, to identify model hallucinations, and to assess model biases through semantic adversarial perturbations.
- Europe > Switzerland > Zürich > Zürich (0.14)
- North America > United States (0.04)
Verification and Refinement of Natural Language Explanations through LLM-Symbolic Theorem Proving
Quan, Xin, Valentino, Marco, Dennis, Louise A., Freitas, André
Natural language explanations have become a proxy for evaluating explainable and multi-step Natural Language Inference (NLI) models. However, assessing the validity of explanations for NLI is challenging as it typically involves the crowd-sourcing of apposite datasets, a process that is time-consuming and prone to logical errors. To address existing limitations, this paper investigates the verification and refinement of natural language explanations through the integration of Large Language Models (LLMs) and Theorem Provers (TPs). Specifically, we present a neuro-symbolic framework, named Explanation-Refiner, that augments a TP with LLMs to generate and formalise explanatory sentences and suggest potential inference strategies for NLI. In turn, the TP is employed to provide formal guarantees on the logical validity of the explanations and to generate feedback for subsequent improvements. We demonstrate how Explanation-Refiner can be jointly used to evaluate explanatory reasoning, autoformalisation, and error correction mechanisms of state-of-the-art LLMs as well as to automatically enhance the quality of human-annotated explanations of variable complexity in different domains.
- North America > Canada > Ontario > Toronto (0.04)
- Asia > Singapore (0.04)
- Europe > United Kingdom > England > Greater Manchester > Manchester (0.04)
- (7 more...)
- Health & Medicine (0.68)
- Leisure & Entertainment (0.67)