Tackling Shortcut Learning in Deep Neural Networks: An Iterative Approach with Interpretable Models
Ghosh, Shantanu, Yu, Ke, Arabshahi, Forough, Batmanghelich, Kayhan
–arXiv.org Artificial Intelligence
We use concept-based interpretable models to mitigate shortcut learning. Existing methods lack interpretability. Beginning with a Blackbox, we iteratively carve out a mixture of interpretable experts (MoIE) and a residual network. Each expert explains a subset of data using First Order Logic (FOL). While explaining a sample, the FOL from biased BB-derived MoIE detects the shortcut effectively. Finetuning the BB with Metadata Normalization (MDN) eliminates the shortcut. The FOLs from the finetuned-BB-derived MoIE verify the elimination of the shortcut. Our experiments show that MoIE does not hurt the accuracy of the original BB and eliminates shortcuts effectively.
arXiv.org Artificial Intelligence
Jul-7-2023
- Country:
- Asia > China
- Hong Kong (0.04)
- North America > United States
- Hawaii > Honolulu County
- Honolulu (0.04)
- Pennsylvania > Allegheny County
- Pittsburgh (0.04)
- Hawaii > Honolulu County
- Asia > China
- Genre:
- Research Report (0.64)
- Industry:
- Health & Medicine > Diagnostic Medicine > Imaging (0.46)
- Technology: