Learning with Explanation Constraints
–Neural Information Processing Systems
As larger deep learning models are hard to interpret, there has been a recent focus on generating explanations of these black-box models. In contrast, we may have apriori explanations of how models should behave. In this paper, we formalize this notion as learning from explanation constraints and provide a learning theoretic framework to analyze how such explanations can improve the learning of our models. One may naturally ask, "When would these explanations be helpful?"
Neural Information Processing Systems
Mar-27-2025, 15:03:34 GMT