Ranking Counterfactual Explanations
Lim, Suryani, Prade, Henri, Richard, Gilles
–arXiv.org Artificial Intelligence
AI-driven outcomes can be challenging for end-users to understand. Explanations can address two key questions: "Why this outcome?" (factual) and "Why not another?" (counterfactual). While substantial efforts have been made to formalize factual explanations, a precise and comprehensive study of counterfactual explanations is still lacking. This paper proposes a formal definition of counterfactual explanations, proving some properties they satisfy, and examining the relationship with factual explanations. Given that multiple counterfactual explanations generally exist for a specific case, we also introduce a rigorous method to rank these counterfactual explanations, going beyond a simple minimality condition, and to identify the optimal ones. Our experiments with 12 real-world datasets highlight that, in most cases, a single optimal counterfactual explanation emerges. We also demonstrate, via three metrics, that the selected optimal explanation exhibits higher representativeness and can explain a broader range of elements than a random minimal counterfactual. This result highlights the effectiveness of our approach in identifying more robust and comprehensive counterfactual explanations.
arXiv.org Artificial Intelligence
Mar-19-2025
- Country:
- Europe
- France > Occitanie
- Haute-Garonne > Toulouse (0.04)
- Italy (0.04)
- Portugal (0.04)
- Spain > Galicia
- A Coruña Province > Santiago de Compostela (0.04)
- Sweden > Uppsala County
- Uppsala (0.04)
- France > Occitanie
- North America
- Canada > Alberta
- United States
- Hawaii > Honolulu County
- Honolulu (0.04)
- New York > New York County
- New York City (0.04)
- Hawaii > Honolulu County
- Oceania > Australia (0.04)
- South America > Chile
- Europe
- Genre:
- Research Report (0.64)
- Industry:
- Government (0.46)
- Information Technology > Security & Privacy (0.46)
- Technology: