Exploring Scaling Laws for EHR Foundation Models

📅 2025-05-29
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Prior work lacks systematic understanding of scaling laws for electronic health record (EHR) foundation models, despite their critical role in clinical AI. EHR data exhibit strong temporal dependencies and structural heterogeneity, posing unique challenges for scalable modeling. Method: We construct multi-scale Transformer-based pretraining models on MIMIC-IV, systematically varying compute (FLOPs), parameter count, and dataset size, and quantitatively evaluate their impact on clinical prediction performance. Contribution/Results: We discover that EHR foundation models follow LLM-like scaling behavior: performance peaks quadratically under fixed FLOPs, while clinical utility scales as a power law with model size. This work establishes the first predictive, resource-efficient scaling paradigm for EHR models—improving downstream task performance by up to 12.7%—and provides both theoretical foundations and empirical guidelines for designing, training, and deploying medical foundation models.

Technology Category

Application Category

📝 Abstract
The emergence of scaling laws has profoundly shaped the development of large language models (LLMs), enabling predictable performance gains through systematic increases in model size, dataset volume, and compute. Yet, these principles remain largely unexplored in the context of electronic health records (EHRs) -- a rich, sequential, and globally abundant data source that differs structurally from natural language. In this work, we present the first empirical investigation of scaling laws for EHR foundation models. By training transformer architectures on patient timeline data from the MIMIC-IV database across varying model sizes and compute budgets, we identify consistent scaling patterns, including parabolic IsoFLOPs curves and power-law relationships between compute, model parameters, data size, and clinical utility. These findings demonstrate that EHR models exhibit scaling behavior analogous to LLMs, offering predictive insights into resource-efficient training strategies. Our results lay the groundwork for developing powerful EHR foundation models capable of transforming clinical prediction tasks and advancing personalized healthcare.
Problem

Research questions and friction points this paper is trying to address.

Investigates scaling laws for EHR foundation models
Explores performance gains via model size and data volume
Identifies scaling patterns in EHR models akin to LLMs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Transformer architectures on EHR data
Scaling laws for EHR foundation models
Resource-efficient EHR training strategies
🔎 Similar Papers
No similar papers found.
S
Sheng Zhang
Microsoft Research
Q
Qin Liu
University of Southern California
N
N. Usuyama
Microsoft Research
C
Cliff Wong
Microsoft Research
Tristan Naumann
Tristan Naumann
Principal Researcher, Microsoft Research Health Futures
Artificial IntelligenceMachine LearningNatural Language ProcessingClinical Inference
H
H. Poon
Microsoft Research