Human Evaluation of Interpretability: The Case of AI-Generated Music Knowledge
Yu, Haizi, Taube, Heinrich, Evans, James A., Varshney, Lav R.
–arXiv.org Artificial Intelligence
Interpretability of machine learning models has gained more and more attention among researchers in the artificial intelligence (AI) and human-computer interaction (HCI) communities. Most existing work focuses on decision making, whereas we consider knowledge discovery. In particular, we focus on evaluating AI-discovered knowledge/rules in the arts and humanities. From a specific scenario, we present an experimental procedure to collect and assess human-generated verbal interpretations of AI-generated music theory/rules rendered as sophisticated symbolic/numeric objects. Our goal is to reveal both the possibilities and the challenges in such a process of decoding expressive messages from AI sources. We treat this as a first step towards 1) better design of AI representations that are human interpretable and 2) a general methodology to evaluate interpretability of AI-discovered knowledge representations.
arXiv.org Artificial Intelligence
Apr-15-2020
- Country:
- North America > United States
- California > Santa Clara County
- Palo Alto (0.04)
- Hawaii > Honolulu County
- Honolulu (0.04)
- Illinois > Cook County
- Chicago (0.05)
- California > Santa Clara County
- North America > United States
- Genre:
- Research Report (0.50)
- Industry:
- Leisure & Entertainment (1.00)
- Media > Music (1.00)
- Technology: