Crowdsourcing Evaluations of Classifier Interpretability
Hutton, Amanda (The University of Texas at Austin) | Liu, Alexander (The University of Texas at Austin) | Martin, Cheryl (The University of Texas at Austin)
This paper presents work using crowdsourcing to assess explanations for supervised text classification. In this paper, an explanation is defined to be a set of words from the input text that a classifier or human believes to be most useful for making a classification decision. We compared two types of explanations for classification decisions: human-generated and computer-generated. The comparison is based on whether the type of the explanation was identifiable and on which type of explanation was preferred. Crowdsourcing was used to collect two types of data for these experiments. First, human-generated explanations were collected by having users select an appropriate category for a piece of text and highlight words that best support this category. Second, users were asked to compare human- and computer-generated explanations and indicate which they preferred and why. The crowdsourced data used for this paper was collected primarily via Amazon’s Mechanical Turk, using several quality control methods. We found that in one test corpus, the two explanation types were virtually indistinguishable, and that participants did not have a significant preference for one type over another. For another corpus, the explanations were slightly more distinguishable, and participants preferred the computer-generated explanations at a small, but statistically significant, level. We conclude that computer-generated explanations for text classification can be comparable in quality to human-generated explanations.
Mar-25-2012
- Country:
- North America > United States > Texas > Travis County > Austin (0.14)
- Genre:
- Research Report > New Finding (0.68)
- Industry:
- Energy (0.33)
- Technology: