RORA: Robust Free-Text Rationale Evaluation
Jiang, Zhengping, Lu, Yining, Chen, Hanjie, Khashabi, Daniel, Van Durme, Benjamin, Liu, Anqi
–arXiv.org Artificial Intelligence
Free-text rationales play a pivotal role in explainable NLP, bridging the knowledge and reasoning gaps behind a model's decision-making. However, due to the diversity of potential reasoning paths and a corresponding lack of definitive ground truth, their evaluation remains a challenge. Existing evaluation metrics rely on the degree to which a rationale supports a target label, but we find these fall short in evaluating rationales that inadvertently leak the labels. To address this problem, we propose RORA, a Robust free-text Rationale evaluation against label leakage. RORA quantifies the new information supplied by a rationale to justify the label. This is achieved by assessing the conditional V-information \citep{hewitt-etal-2021-conditional} with a predictive family robust against leaky features that can be exploited by a small model. RORA consistently outperforms existing approaches in evaluating human-written, synthetic, or model-generated rationales, particularly demonstrating robustness against label leakage. We also show that RORA aligns well with human judgment, providing a more reliable and accurate measurement across diverse free-text rationales.
arXiv.org Artificial Intelligence
Jun-14-2024
- Country:
- Asia
- China (0.05)
- Middle East > Jordan (0.04)
- Europe
- Germany (0.04)
- Italy > Marche
- Ancona Province > Ancona (0.04)
- United Kingdom > England
- Cambridgeshire > Cambridge (0.04)
- North America
- Dominican Republic (0.04)
- United States
- California > Los Angeles County
- Los Angeles (0.04)
- Illinois > Cook County
- Chicago (0.04)
- Minnesota > Hennepin County
- Minneapolis (0.14)
- New York > New York County
- New York City (0.04)
- California > Los Angeles County
- South America > Brazil
- Rio de Janeiro > Rio de Janeiro (0.05)
- Asia
- Genre:
- Research Report (1.00)
- Technology: