A Lightweight Large Vision-language Model for Multimodal Medical Images
Alsinglawi, Belal, McCarthy, Chris, Webb, Sara, Fluke, Christopher, Saidy, Navid Toosy
–arXiv.org Artificial Intelligence
Medical Visual Question Answering (VQA) enhances clinical decision-making by enabling systems to interpret medical images and answer clinical queries. However, developing efficient, high-performance VQA models is challenging due to the complexity of medical imagery and diverse modalities. In this paper, we introduce a lightweight, multimodal VQA model integrating BiomedCLIP for image feature extraction and LLaMA-3 for text processing. Designed for medical VQA tasks, our model achieves state-of-the-art performance on the OmniMedVQA dataset. With approximately 8 billion parameters, it requires only two NVIDIA 40 GB A100 GPUs, demonstrating superior efficiency over larger models. Our results show 73.4% accuracy for open-end questions, surpassing existing models and validating its potential for real-world medical applications. Key contributions include a specialized multimodal VQA model, a resource-efficient architecture, and strong performance in answering open-ended clinical questions.
arXiv.org Artificial Intelligence
Apr-9-2025
- Country:
- North America > United States
- District of Columbia > Washington (0.04)
- Oceania > Australia
- Queensland > Brisbane (0.04)
- Victoria > Melbourne (0.04)
- North America > United States
- Genre:
- Research Report > New Finding (0.69)
- Industry:
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
- Technology: