Representer Point Selection for Explaining Regularized High-dimensional Models
Tsai, Che-Ping, Zhang, Jiong, Chien, Eli, Yu, Hsiang-Fu, Hsieh, Cho-Jui, Ravikumar, Pradeep
–arXiv.org Artificial Intelligence
We introduce a novel class of sample-based explanations we term high-dimensional representers, that can be used to explain the predictions of a regularized high-dimensional model in terms of importance weights for each of the training samples. Our workhorse is a novel representer theorem for general regularized high-dimensional models, which decomposes the model prediction in terms of contributions from each of the training samples: with positive (negative) values corresponding to positive (negative) impact training samples to the model's prediction. We derive consequences for the canonical instances of $\ell_1$ regularized sparse models, and nuclear norm regularized low-rank models. As a case study, we further investigate the application of low-rank models in the context of collaborative filtering, where we instantiate high-dimensional representers for specific popular classes of models. Finally, we study the empirical performance of our proposed methods on three real-world binary classification datasets and two recommender system datasets. We also showcase the utility of high-dimensional representers in explaining model recommendations.
arXiv.org Artificial Intelligence
Jun-30-2023
- Country:
- Asia > Middle East
- Republic of Türkiye > Batman Province > Batman (0.04)
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.04)
- North America > United States
- California
- Los Angeles County > Los Angeles (0.28)
- Santa Clara County > Palo Alto (0.04)
- Illinois > Champaign County
- New York (0.04)
- Pennsylvania > Allegheny County
- Pittsburgh (0.14)
- California
- Asia > Middle East
- Genre:
- Research Report (1.00)
- Industry:
- Leisure & Entertainment (1.00)
- Media > Film (0.93)
- Technology: