Freeze the backbones: A Parameter-Efficient Contrastive Approach to Robust Medical Vision-Language Pre-training
Qin, Jiuming, Liu, Che, Cheng, Sibo, Guo, Yike, Arcucci, Rossella
–arXiv.org Artificial Intelligence
Modern healthcare often utilises radiographic images alongside textual reports for diagnostics, encouraging the use of Vision-Language Self-Supervised Learning (VL-SSL) with large pre-trained models to learn versatile medical vision representations. However, most existing VL-SSL frameworks are trained end-to-end, which is computation-heavy and can lose vital prior information embedded in pre-trained encoders. To address both issues, we introduce the backbone-agnostic Adaptor framework, which preserves medical knowledge in pre-trained image and text encoders by keeping them frozen, and employs a lightweight Adaptor module for cross-modal learning. Experiments on medical image classification and segmentation tasks across three datasets reveal that our framework delivers competitive performance while cutting trainable parameters by over 90% compared to current pre-training approaches. Notably, when fine-tuned with just 1% of data, Adaptor outperforms several Transformer-based methods trained on full datasets in medical image segmentation.
arXiv.org Artificial Intelligence
Jan-2-2024
- Country:
- Asia > China
- Hong Kong (0.04)
- Europe
- Slovenia > Drava
- Municipality of Benedikt > Benedikt (0.04)
- Switzerland (0.05)
- Slovenia > Drava
- North America > United States (0.04)
- Asia > China
- Genre:
- Industry:
- Health & Medicine > Diagnostic Medicine > Imaging (1.00)
- Technology: