MCE: Towards a General Framework for Handling Missing Modalities under Imbalanced Missing Rates
Zhao, Binyu, Zhang, Wei, Zou, Zhaonian
–arXiv.org Artificial Intelligence
Multi-modal learning has made significant advances across diverse pattern recognition applications. However, handling missing modalities, especially under imbalanced missing rates, remains a major challenge. This imbalance triggers a vicious cycle: modalities with higher missing rates receive fewer updates, leading to inconsistent learning progress and representational degradation that further diminishes their contribution. Existing methods typically focus on global dataset-level balancing, often overlooking critical sample-level variations in modality utility and the underlying issue of degraded feature quality. We propose Modality Capability Enhancement (MCE) to tackle these limitations. MCE includes two synergistic components: i) Learning Capability Enhancement (LCE), which introduces multi-level factors to dynamically balance modality-specific learning progress, and ii) Representation Capability Enhancement (RCE), which improves feature semantics and robustness through subset prediction and cross-modal completion tasks. Comprehensive evaluations on four multi-modal benchmarks show that MCE consistently outperforms state-of-the-art methods under various missing configurations. The final published version is now available at https://doi.org/10.1016/j.patcog.2025.112591. Our code is available at https://github.com/byzhaoAI/MCE.
arXiv.org Artificial Intelligence
Nov-11-2025
- Country:
- Asia > China
- Heilongjiang Province > Harbin (0.04)
- Europe > Germany
- Bavaria > Upper Bavaria > Munich (0.04)
- Asia > China
- Genre:
- Research Report > New Finding (0.93)
- Industry:
- Health & Medicine
- Diagnostic Medicine > Imaging (0.93)
- Therapeutic Area (1.00)
- Health & Medicine
- Technology: