Evaluating the Robustness of Interpretability Methods through Explanation Invariance and Equivariance
–Neural Information Processing Systems
Interpretability methods are valuable only if their explanations faithfully describe the explained model. In this work, we consider neural networks whose predictions are invariant under a specific symmetry group. This includes popular architectures, ranging from convolutional to graph neural networks. Any explanation that faithfully explains this type of model needs to be in agreement with this invariance property.
Neural Information Processing Systems
May-25-2025, 14:31:16 GMT
- Country:
- Europe > United Kingdom
- England > Cambridgeshire > Cambridge (0.14)
- North America > United States
- New York > New York County > New York City (0.14)
- Europe > United Kingdom
- Industry:
- Health & Medicine (0.93)
- Technology: