On the Complexity-Faithfulness Trade-off of Gradient-Based Explanations
Mehrpanah, Amir, Gamba, Matteo, Smith, Kevin, Azizpour, Hossein
–arXiv.org Artificial Intelligence
ReLU networks, while prevalent for visual data, have sharp transitions, sometimes relying on individual pixels for predictions, making vanilla gradient-based explanations noisy and difficult to interpret. Existing methods, such as Grad-CAM, smooth these explanations by producing surrogate models at the cost of faithfulness. W e introduce a unifying spectral framework to systematically analyze and quantify smoothness, faithfulness, and their trade-off in explanations. Using this framework, we quantify and regularize the contribution of ReLU networks to high-frequency information, providing a principled approach to identifying this trade-off. Our analysis characterizes how surrogate-based smoothing distorts explanations, leading to an "explanation gap" that we formally define and measure for different post-hoc methods.
arXiv.org Artificial Intelligence
Aug-15-2025
- Country:
- Asia
- Japan > Honshū
- Kantō > Kanagawa Prefecture (0.04)
- Middle East > Jordan (0.04)
- Japan > Honshū
- Europe
- Sweden > Stockholm
- Stockholm (0.04)
- Switzerland (0.04)
- Sweden > Stockholm
- Asia
- Genre:
- Research Report > New Finding (1.00)
- Technology: