AnyCBMs: How to Turn Any Black Box into a Concept Bottleneck Model
Dominici, Gabriele, Barbiero, Pietro, Giannini, Francesco, Gjoreski, Martin, Langhenirich, Marc
–arXiv.org Artificial Intelligence
Interpretable deep learning aims at developing neural architectures whose decision-making processes could be understood by their users. Among these techniqes, Concept Bottleneck Models enhance the interpretability of neural networks by integrating a layer of human-understandable concepts. These models, however, necessitate training a new model from the beginning, consuming significant resources and failing to utilize already trained large models. To address this issue, we introduce "AnyCBM", a method that transforms any existing trained model into a Concept Bottleneck Model with minimal impact on computational resources. We provide both theoretical and experimental insights showing the effectiveness of AnyCBMs in terms of classification performances and effectivenss of concept-based interventions on downstream tasks.
arXiv.org Artificial Intelligence
May-26-2024
- Country:
- Europe
- Italy (0.14)
- Middle East > Malta (0.14)
- North America > United States (0.14)
- Europe
- Genre:
- Research Report > New Finding (0.69)
- Industry:
- Government (0.47)
- Transportation > Air (0.53)
- Technology: