Iterative Learning of Computable Phenotypes for Treatment Resistant Hypertension using Large Language Models

📅 2025-08-07
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses the challenge of automatically generating accurate, concise, and interpretable computable phenotypes (CPs) for clinical decision support in treatment-resistant hypertension using large language models (LLMs). We propose a “Generate–Execute–Debug–Instruct” iterative learning framework tailored to six key clinical phenotypes, integrating program synthesis, execution-based validation, and few-shot feedback to drastically reduce reliance on large-scale annotated data. Our method automatically translates domain-specific medical knowledge into executable, verifiable clinical logic programs. Empirically, it achieves accuracy comparable to state-of-the-art machine learning models while ensuring strong interpretability through transparent, rule-based reasoning. Experiments demonstrate that clinically deployable performance is attainable with only a minimal number of labeled examples—fewer than 10 per phenotype—thereby establishing a novel paradigm for CP development in low-resource settings.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) have demonstrated remarkable capabilities for medical question answering and programming, but their potential for generating interpretable computable phenotypes (CPs) is under-explored. In this work, we investigate whether LLMs can generate accurate and concise CPs for six clinical phenotypes of varying complexity, which could be leveraged to enable scalable clinical decision support to improve care for patients with hypertension. In addition to evaluating zero-short performance, we propose and test a synthesize, execute, debug, instruct strategy that uses LLMs to generate and iteratively refine CPs using data-driven feedback. Our results show that LLMs, coupled with iterative learning, can generate interpretable and reasonably accurate programs that approach the performance of state-of-the-art ML methods while requiring significantly fewer training examples.
Problem

Research questions and friction points this paper is trying to address.

Generate accurate computable phenotypes for hypertension
Improve clinical decision support using LLMs
Refine phenotypes iteratively with data-driven feedback
Innovation

Methods, ideas, or system contributions that make the work stand out.

LLMs generate computable phenotypes iteratively
Synthesize, execute, debug, instruct strategy used
Fewer training examples than ML methods needed
🔎 Similar Papers
No similar papers found.
G
Guilherme Seidyo Imai Aldeia
Federal University of ABC, Santo André, São Paulo, Brazil
D
Daniel S. Herman
Department of Pathology and Laboratory Medicine, University of Pennsylvania, Philadelphia, PA, USA
William G. La Cava
William G. La Cava
Harvard, Boston Children's Hospital
biomedical informaticsmachine learningfairnessinterpretabilitysymbolic regression