Tackling Shortcut Learning in Deep Neural Networks: An Iterative Approach with Interpretable Models
Ghosh, Shantanu, Yu, Ke, Arabshahi, Forough, Batmanghelich, Kayhan
–arXiv.org Artificial Intelligence
We use concept-based interpretable models to mitigate shortcut learning. Existing methods lack interpretability. Beginning with a Blackbox, we iteratively carve out a mixture of interpretable experts (MoIE) and a residual network. Each expert explains a subset of data using First Order Logic (FOL). While explaining a sample, the FOL from biased BB-derived MoIE detects the shortcut effectively. Finetuning the BB with Metadata Normalization (MDN) eliminates the shortcut. The FOLs from the finetuned-BB-derived MoIE verify the elimination of the shortcut. Our experiments show that MoIE does not hurt the accuracy of the original BB and eliminates shortcuts effectively.
arXiv.org Artificial Intelligence
Jul-7-2023
- Country:
- North America > United States > Hawaii (0.14)
- Genre:
- Research Report (0.64)
- Industry:
- Health & Medicine > Diagnostic Medicine > Imaging (0.46)
- Technology: