Towards Interpretable Radiology Report Generation via Concept Bottlenecks using a Multi-Agentic RAG
Alam, Hasan Md Tusfiqur, Srivastav, Devansh, Kadir, Md Abdul, Sonntag, Daniel
–arXiv.org Artificial Intelligence
Deep learning has advanced medical image classification, but interpretability challenges hinder its clinical adoption. This study enhances interpretability in Chest X-ray (CXR) classification by using concept bottleneck models (CBMs) and a multi-agent Retrieval-Augmented Generation (RAG) system for report generation. By modeling relationships between visual features and clinical concepts, we create interpretable concept vectors that guide a multi-agent RAG system to generate radiology reports, enhancing clinical relevance, explainability, and transparency. Evaluation of the generated reports using an LLM-as-a-judge confirmed the interpretability and clinical utility of our model's outputs. On the COVID-QU dataset, our model achieved 81% classification accuracy and demonstrated robust report generation performance, with five key metrics ranging between 84% and 90%. This interpretable multi-agent framework bridges the gap between high-performance AI and the explainability required for reliable AI-driven CXR analysis in clinical settings.
arXiv.org Artificial Intelligence
Dec-20-2024
- Country:
- Asia > Middle East
- Qatar (0.04)
- Europe > Germany
- Lower Saxony > Oldenburg (0.04)
- Saarland > Saarbrücken (0.04)
- Asia > Middle East
- Genre:
- Research Report (0.40)
- Industry:
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
- Technology:
- Information Technology > Artificial Intelligence
- Machine Learning > Neural Networks
- Deep Learning (1.00)
- Natural Language
- Chatbot (1.00)
- Large Language Model (1.00)
- Representation & Reasoning > Agents (1.00)
- Vision (1.00)
- Machine Learning > Neural Networks
- Information Technology > Artificial Intelligence