NormLime: A New Feature Importance Metric for Explaining Deep Neural Networks
Ahern, Isaac, Noack, Adam, Guzman-Nateras, Luis, Dou, Dejing, Li, Boyang, Huan, Jun
The problem of explaining deep learning models, and model predictions generally, has attracted intensive interest recently. Many successful approaches forgo global approximations in order to provide more faithful local interpretations of the model's behavior. LIME develops multiple interpretable models, each approximating a large neural network on a small region of the data manifold and SP-LIME aggregates the local models to form a global interpretation. Extending this line of research, we propose a simple yet effective method, Norm-LIME for aggregating local models into global and class-specific interpretations. A human user study strongly favored class-specific interpretations created by NormLIME to other feature importance metrics. Numerical experiments confirm that NormLIME is effective at recognizing important features. Introduction As the applications of deep neural networks continue to expand, the intrinsic black-box nature of neural networks creates a potential trust issue. For application domains with high cost of prediction error, such as healthcare (Phan et al. 2017), it is necessary that human users can verify that a model learns reasonable representation of data and the rationale for its decisions are justifiable according to societal norms (Koh and Liang 2017; Fong and V edaldi 2018; Zhou et al. 2018; Lipton 2016; Langley 2019). An interpretable model, such as a linear sparse regression, lends itself readily to model explanation.
Sep-9-2019
- Country:
- Asia > Middle East
- Jordan (0.04)
- Europe > Italy
- Marche > Ancona Province > Ancona (0.04)
- North America > United States
- Oregon (0.04)
- Asia > Middle East
- Genre:
- Research Report > Experimental Study (0.68)
- Technology: