Partially Interpretable Estimators (PIE): Black-Box-Refined Interpretable Machine Learning
Wang, Tong, Yang, Jingyi, Li, Yunyi, Wang, Boxiang
–arXiv.org Artificial Intelligence
We propose Partially Interpretable Estimators (PIE) which attribute a prediction to individual features via an interpretable model, while a (possibly) small part of the PIE prediction is attributed to the interaction of features via a black-box model, with the goal to boost the predictive performance while maintaining interpretability. As such, the interpretable model captures the main contributions of features, and the black-box model attempts to complement the interpretable piece by capturing the "nuances" of feature interactions as a refinement. We design an iterative training algorithm to jointly train the two types of models. Experimental results show that PIE is highly competitive to black-box models while outperforming interpretable baselines. In addition, the understandability of PIE is comparable to simple linear models as validated via a human evaluation.
arXiv.org Artificial Intelligence
May-5-2021
- Country:
- North America > United States > Texas (0.14)
- Genre:
- Research Report > New Finding (0.48)
- Industry:
- Health & Medicine (0.94)
- Transportation > Air (1.00)
- Technology: