🤖 AI Summary
Frequent misdiagnoses escalate healthcare costs and compromise patient safety. To address this, we propose MedRAG—the first intelligent multimodal clinical decision support system integrating voice-based patient interviews, unstructured clinical notes, and structured electronic health records (EHRs) to generate diagnostic hypotheses, treatment plans, medication recommendations, and follow-up questions. Its core innovation is a knowledge graph–guided retrieval-augmented generation (KG-RAG) framework that enables medical-logic-driven, cross-modal reasoning over heterogeneous data sources, markedly improving diagnostic interpretability and clinical applicability. Evaluated on multiple public and proprietary medical datasets, MedRAG consistently outperforms state-of-the-art baselines in recommendation accuracy and clinical specificity. The source code and demonstration video are publicly available.
📝 Abstract
Misdiagnosis causes significant harm to healthcare systems worldwide, leading to increased costs and patient risks. MedRAG is a smart multimodal healthcare copilot equipped with powerful large language model (LLM) reasoning, designed to enhance medical decision-making. It supports multiple input modalities, including non-intrusive voice monitoring, general medical queries, and electronic health records. MedRAG provides recommendations on diagnosis, treatment, medication, and follow-up questioning. Leveraging retrieval-augmented generation enhanced by knowledge graph-elicited reasoning, MedRAG retrieves and integrates critical diagnostic insights, reducing the risk of misdiagnosis. It has been evaluated on both public and private datasets, outperforming existing models and offering more specific and accurate healthcare assistance. A demonstration video of MedRAG is available at: https://www.youtube.com/watch?v=PNIBDMYRfDM. The source code is available at: https://github.com/SNOWTEAM2023/MedRAG.