← Back to Paper List

EveryQuery: Zero-Shot Clinical Prediction via Task-Conditioned Pretraining over Electronic Health Records

Payal Chandak, Gregory Kondas, Isaac Kohane, Matthew McDermott
Biomedical Informatics, Harvard Medical School
arXiv (2026)
Pretraining QA Benchmark

📝 Paper Summary

Clinical Foundation Models Zero-Shot Prediction Electronic Health Records (EHR)
EveryQuery enables zero-shot clinical prediction by pretraining a transformer to answer structured queries about future events directly from patient history, bypassing expensive trajectory sampling.
Core Problem
Existing autoregressive EHR foundation models are computationally expensive, statistically noisy, and lack native promptability because they require sampling many future trajectories to estimate event probabilities.
Why it matters:
  • Inference is inefficient: long-horizon predictions require many generation steps per trajectory, compounding cost.
  • Estimates are noisy: rare event probabilities are often quantized to zero if the event doesn't appear in the limited sample of trajectories.
  • Limited interactivity: users cannot directly condition predictions on specific clinical questions without designing custom aggregation pipelines.
Concrete Example: To predict a rare event with prevalence p < 0.05, an autoregressive model sampling 20 trajectories often yields a probability of 0, missing the risk entirely. EveryQuery outputs a direct probability estimate in a single pass.
Key Novelty
Task-Conditioned Pretraining for EHR
  • Treats clinical prediction as a query-answering task where the input is (patient history, query code) and output is the probability of that code occurring.
  • Pretrains on randomly sampled queries (Does code 'c' occur in next 30 days?) across patient histories, learning to condition representations on the specific task requested.
  • Replaces generative next-token prediction with a discriminative objective that directly answers the user's structured query.
Evaluation Highlights
  • Outperforms an autoregressive baseline on 82% of 39 randomly sampled tasks, with a mean AUC improvement of +0.16.
  • Inference is ~3,000x faster than the autoregressive baseline (single forward pass vs. 20 trajectory rollouts).
  • Generalizes effectively to rare events where trajectory-based methods fail (Spearman correlation ρ = -0.32 between event rarity and performance gain).
Breakthrough Assessment
7/10
Significant efficiency and usability gains for clinical models. Solves the 'rare event' problem of autoregressive sampling. Currently limited by a simple query language (fixed 30-day window).
×