Interpretable Visual Understanding with Cognitive Attention Network
Tang, Xuejiao, Zhang, Wenbin, Yu, Yi, Turner, Kea, Derr, Tyler, Wang, Mengyu, Ntoutsi, Eirini
–arXiv.org Artificial Intelligence
While image understanding on recognition-level has achieved remarkable advancements, reliable visual scene understanding requires comprehensive image understanding on recognition-level but also cognition-level, which calls for exploiting the multi-source information as well as learning different levels of understanding and extensive commonsense knowledge. In this paper, we propose a novel Cognitive Attention Network (CAN) for visual commonsense reasoning to achieve interpretable visual understanding. Specifically, we first introduce an image-text fusion module to fuse information from images and text collectively. Second, a novel inference module is designed to encode commonsense among image, query and response. Extensive experiments on large-scale Visual Commonsense Reasoning (VCR) benchmark dataset demonstrate the effectiveness of our approach. The implementation is publicly available at https://github.com/tanjatang/CAN
arXiv.org Artificial Intelligence
Aug-14-2021
- Country:
- Asia > Japan
- Honshū > Chūbu > Ishikawa Prefecture > Kanazawa (0.04)
- Europe > Germany
- Berlin (0.04)
- Lower Saxony > Hanover (0.04)
- North America > United States
- California > Yolo County
- Davis (0.04)
- Pennsylvania > Allegheny County
- Pittsburgh (0.04)
- California > Yolo County
- Asia > Japan
- Genre:
- Research Report (0.82)
- Industry:
- Health & Medicine (0.69)
- Technology: