Evaluating Saliency Explanations in NLP by Crowdsourcing
Lu, Xiaotian, Li, Jiyi, Wan, Zhen, Lin, Xiaofeng, Takeuchi, Koh, Kashima, Hisashi
–arXiv.org Artificial Intelligence
Deep learning models have performed well on many NLP tasks. However, their internal mechanisms are typically difficult for humans to understand. The development of methods to explain models has become a key issue in the reliability of deep learning models in many important applications. Various saliency explanation methods, which give each feature of input a score proportional to the contribution of output, have been proposed to determine the part of the input which a model values most. Despite a considerable body of work on the evaluation of saliency methods, whether the results of various evaluation metrics agree with human cognition remains an open question. In this study, we propose a new human-based method to evaluate saliency methods in NLP by crowdsourcing. We recruited 800 crowd workers and empirically evaluated seven saliency methods on two datasets with the proposed method. We analyzed the performance of saliency methods, compared our results with existing automated evaluation methods, and identified notable differences between NLP and computer vision (CV) fields when using saliency methods. The instance-level data of our crowdsourced experiments and the code to reproduce the explanations are available at https://github.com/xtlu/lreccoling_evaluation.
arXiv.org Artificial Intelligence
May-17-2024
- Country:
- Asia
- China > Beijing
- Beijing (0.04)
- India (0.04)
- Japan > Honshū
- Kansai > Kyoto Prefecture > Kyoto (0.04)
- Middle East
- Palestine > Gaza Strip
- Gaza Governorate > Gaza (0.04)
- UAE > Abu Dhabi Emirate
- Abu Dhabi (0.04)
- Palestine > Gaza Strip
- Singapore (0.04)
- China > Beijing
- Europe > Italy
- Marche > Ancona Province > Ancona (0.04)
- North America
- Canada
- Guatemala (0.14)
- United States
- California > San Diego County
- San Diego (0.04)
- Oregon (0.04)
- California > San Diego County
- Asia
- Genre:
- Research Report > New Finding (1.00)
- Industry:
- Technology: