MKG-Rank: Enhancing Large Language Models with Knowledge Graph for Multilingual Medical Question Answering

📅 2025-03-20
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the low accuracy of medical question answering (QA) in low-resource languages, this paper proposes a word-level cross-lingual knowledge alignment method that lightweightly integrates English medical knowledge graphs (KGs) into large language models (LLMs) to enable precise cross-lingual reasoning. Our key contributions are threefold: (1) a novel translation-driven word-level KG alignment mechanism; (2) a lightweight retrieval framework integrating KG embeddings, retrieval-augmented generation (RAG), and multi-perspective semantic ranking; and (3) a cache-optimized strategy enabling millisecond-scale response times. Evaluated on medical QA benchmarks across Chinese, Japanese, Korean, and Swahili, our approach achieves up to a 33.89% absolute accuracy improvement over baselines, with an average retrieval latency of only 0.0009 seconds—significantly outperforming existing cross-lingual medical QA methods.

Technology Category

Application Category

📝 Abstract
Large Language Models (LLMs) have shown remarkable progress in medical question answering (QA), yet their effectiveness remains predominantly limited to English due to imbalanced multilingual training data and scarce medical resources for low-resource languages. To address this critical language gap in medical QA, we propose Multilingual Knowledge Graph-based Retrieval Ranking (MKG-Rank), a knowledge graph-enhanced framework that enables English-centric LLMs to perform multilingual medical QA. Through a word-level translation mechanism, our framework efficiently integrates comprehensive English-centric medical knowledge graphs into LLM reasoning at a low cost, mitigating cross-lingual semantic distortion and achieving precise medical QA across language barriers. To enhance efficiency, we introduce caching and multi-angle ranking strategies to optimize the retrieval process, significantly reducing response times and prioritizing relevant medical knowledge. Extensive evaluations on multilingual medical QA benchmarks across Chinese, Japanese, Korean, and Swahili demonstrate that MKG-Rank consistently outperforms zero-shot LLMs, achieving maximum 33.89% increase in accuracy, while maintaining an average retrieval time of only 0.0009 seconds.
Problem

Research questions and friction points this paper is trying to address.

Addresses multilingual medical QA limitations in LLMs
Integrates English-centric knowledge graphs for cross-lingual QA
Optimizes retrieval efficiency with caching and ranking strategies
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates English-centric medical knowledge graphs
Uses word-level translation for multilingual QA
Optimizes retrieval with caching and multi-angle ranking
🔎 Similar Papers
No similar papers found.
F
Feiyang Li
University of Tokyo
Y
Yingjian Chen
University of Tokyo
Haoran Liu
Haoran Liu
Ph.D. Student, Department of Computer Science & Engineering, Texas A&M University
LLMsGraph/Geometric LearningAI for ScienceGenerative Models
R
Rui Yang
Duke-NUS Medical School
H
Han Yuan
Duke-NUS Medical School
Y
Yuang Jiang
Smartor LLC, Japan
T
Tianxiao Li
NEC Laboratories America
E
Edison Marrese Taylor
University of Tokyo
H
H. Rouhizadeh
University of Geneva, Switzerland
Yusuke Iwasawa
Yusuke Iwasawa
The University of Tokyo
deep learningtransfer learningfoundation modelmeta learning
Douglas Teodoro
Douglas Teodoro
Professor, University of Geneva
biomedical NLPmachine learning for healthcaremedical informatics
Y
Yutaka Matsuo
University of Tokyo
Irene Li
Irene Li
Project Lecturer (特任講師) at University of Tokyo
Large Language ModelsGraph Neural NetworksBioNLPMedical NLPText Summarization