Generative Medical Event Models Improve with Scale
Waxler, Shane, Blazek, Paul, White, Davis, Sneider, Daniel, Chung, Kevin, Nagarathnam, Mani, Williams, Patrick, Voeller, Hank, Wong, Karen, Swanhorst, Matthew, Zhang, Sheng, Usuyama, Naoto, Wong, Cliff, Naumann, Tristan, Poon, Hoifung, Loza, Andrew, Meeker, Daniella, Hain, Seth, Shah, Rahul
–arXiv.org Artificial Intelligence
Realizing personalized medicine at scale calls for methods that distill insights from longitudinal patient journeys, which can be viewed as a sequence of medical events. Foundation models pretrained on large-scale medical event data represent a promising direction for scaling real-world evidence generation and generalizing to diverse downstream tasks. Using Epic Cosmos, a dataset with medical events from de-identified longitudinal health records for 16.3 billion encounters over 300 million unique patient records from 310 health systems, we introduce the Curiosity models, a family of decoder-only transformer models pretrained on 118 million patients representing 115 billion discrete medical events (151 billion tokens). We present the largest scaling-law study of medical event data, establishing a methodology for pretraining and revealing power-law scaling relationships for compute, tokens, and model size. Consequently, we pretrained a series of compute-optimal models with up to 1 billion parameters. Conditioned on a patient's real-world history, Curiosity autoregressively predicts the next medical event to simulate patient health timelines. We studied 78 real-world tasks, including diagnosis prediction, disease prognosis, and healthcare operations. Remarkably for a foundation model with generic pretraining and simulation-based inference, Curiosity generally outperformed or matched task-specific supervised models on these tasks, without requiring task-specific fine-tuning or few-shot examples. Curiosity's predictive power consistently improves as the model and pretraining scale. Our results show that Curiosity, a generative medical event foundation model, can effectively capture complex clinical dynamics, providing an extensible and generalizable framework to support clinical decision-making, streamline healthcare operations, and improve patient outcomes.
arXiv.org Artificial Intelligence
Nov-11-2025
- Country:
- Asia > Middle East
- Jordan (0.04)
- Lebanon (0.04)
- Saudi Arabia (0.04)
- North America
- Canada (0.04)
- United States > Alaska (0.04)
- Asia > Middle East
- Genre:
- Research Report
- Experimental Study (1.00)
- New Finding (1.00)
- Research Report
- Industry:
- Government > Regional Government
- Health & Medicine
- Consumer Health (1.00)
- Government Relations & Public Policy (1.00)
- Health Care Providers & Services > Reimbursement (0.67)
- Health Care Technology > Medical Record (1.00)
- Pharmaceuticals & Biotechnology (1.00)
- Therapeutic Area
- Internal Medicine (0.92)
- Pulmonary/Respiratory Diseases (0.92)
- Nephrology (0.67)
- Neurology (1.00)
- Musculoskeletal (1.00)
- Immunology (1.00)
- Cardiology/Vascular Diseases (1.00)
- Hematology (0.67)
- Hepatology (0.93)
- Gastroenterology (1.00)
- Oncology (1.00)
- Rheumatology (1.00)
- Endocrinology > Diabetes (1.00)
- Infections and Infectious Diseases (1.00)
- Technology:
- Information Technology
- Artificial Intelligence
- Machine Learning
- Neural Networks > Deep Learning (0.93)
- Performance Analysis > Accuracy (1.00)
- Statistical Learning (1.00)
- Natural Language
- Chatbot (0.67)
- Large Language Model (1.00)
- Representation & Reasoning > Diagnosis (0.68)
- Machine Learning
- Data Science (1.00)
- Information Management (0.67)
- Artificial Intelligence
- Information Technology