MoMA: A Mixture-of-Multimodal-Agents Architecture for Enhancing Clinical Prediction Modelling
Gao, Jifan, Rahman, Mahmudur, Caskey, John, Oguss, Madeline, O'Rourke, Ann, Brown, Randy, Stey, Anne, Mayampurath, Anoop, Churpek, Matthew M., Chen, Guanhua, Afshar, Majid
–arXiv.org Artificial Intelligence
Multimodal electronic health record (EHR) data provide richer, complementary insights into patient health compared to single-modality data. However, effectively integrating diverse data modalities for clinical prediction modeling remains challenging due to the substantial data requirements. We introduce a novel architecture, Mixture-of-Multimodal-Agents (MoMA), designed to leverage multiple large language model (LLM) agents for clinical prediction tasks using multimodal EHR data. MoMA employs specialized LLM agents ("specialist agents") to convert non-textual modalities, such as medical images and laboratory results, into structured textual summaries. These summaries, together with clinical notes, are combined by another LLM ("aggregator agent") to generate a unified multimodal summary, which is then used by a third LLM ("predictor agent") to produce clinical predictions. Evaluating MoMA on three prediction tasks using real-world datasets with different modality combinations and prediction settings, MoMA outperforms current state-of-the-art methods, highlighting its enhanced accuracy and flexibility across various tasks.
arXiv.org Artificial Intelligence
Aug-8-2025
- Country:
- North America > United States
- Illinois > Cook County
- Chicago (0.04)
- Wisconsin > Dane County
- Madison (0.14)
- Illinois > Cook County
- North America > United States
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Research Report
- Industry:
- Health & Medicine
- Diagnostic Medicine > Imaging (1.00)
- Health Care Providers & Services (0.93)
- Health Care Technology > Medical Record (0.90)
- Therapeutic Area > Psychiatry/Psychology (0.97)
- Health & Medicine
- Technology: