Decomposing and Interpreting Image Representations via Text in ViTs Beyond CLIP
–Neural Information Processing Systems
Recent work has explored how individual components of the CLIP-ViT model contribute to the final representation by leveraging the shared image-text representation space of CLIP. These components, such as attention heads and MLPs, have been shown to capture distinct image features like shape, color or texture. However, understanding the role of these components in arbitrary vision transformers (ViTs) is challenging. To this end, we introduce a general framework which can identify the roles of various components in ViTs beyond CLIP. Specifically, we (a) automate the decomposition of the final representation into contributions from different model components, and (b) linearly map these contributions to CLIP space to interpret them via text. Additionally, we introduce a novel scoring function to rank components by their importance with respect to specific features.
Neural Information Processing Systems
May-31-2025, 11:48:25 GMT
- Country:
- North America > United States > Maryland (0.14)
- Genre:
- Research Report > Experimental Study (0.93)
- Technology: