Hybrid Fact-Checking that Integrates Knowledge Graphs, Large Language Models, and Search-Based Retrieval Agents Improves Interpretable Claim Verification

📅 2025-11-05
🏛️ Proceedings of the 9th Widening NLP Workshop
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
To address the lack of verifiable evidence in large language model (LLM)-based fact-checking and the limited coverage and outdatedness of knowledge graphs (KGs), this paper proposes a hybrid fact-checking framework integrating DBpedia KG retrieval, prompt-driven LLM classification, rule-based logical reasoning, and dynamic invocation of web search agents. Its key contribution is a zero-shot fallback mechanism: when both KG and LLM fail to determine veracity, the system automatically triggers real-time web search to retrieve up-to-date evidence—significantly improving verification capability for “information-scarce” claims. Evaluated on the FEVER benchmark’s Supported/Refuted binary classification task, the system achieves an F1 score of 0.93. Further validation via re-annotation confirms its ability to uncover critical evidence missed in the original annotations. The framework thus delivers high accuracy, strong interpretability, and broad coverage without requiring model fine-tuning.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) excel in generating fluent utterances but can lack reliable grounding in verified information. At the same time, knowledge-graph-based fact-checkers deliver precise and interpretable evidence, yet suffer from limited coverage or latency. By integrating LLMs with knowledge graphs and real-time search agents, we introduce a hybrid fact-checking approach that leverages the individual strengths of each component. Our system comprises three autonomous steps: 1) a Knowledge Graph (KG) Retrieval for rapid one - hop lookups in DBpedia, 2) an LM-based classification guided by a task-specific labeling prompt, producing outputs with internal rule-based logic, and 3) a Web Search Agent invoked only when KG coverage is insufficient. Our pipeline achieves an F1 score of 0.93 on the FEVER benchmark on the Supported/Refuted split without task- specific fine - tuning. To address Not enough information cases, we conduct a targeted reannotation study showing that our approach frequently uncovers valid evidence for claims originally labeled as Not Enough Information (NEI), as confirmed by both expert annotators and LLM reviewers. With this paper, we present a modular, opensource fact-checking pipeline with fallback strategies and generalization across datasets.
Problem

Research questions and friction points this paper is trying to address.

LLMs lack reliable grounding in verified information sources
Knowledge-graph fact-checkers have limited coverage and latency issues
Existing methods struggle with Not Enough Information claim verification
Innovation

Methods, ideas, or system contributions that make the work stand out.

Hybrid fact-checking integrates knowledge graphs and LLMs
System uses KG retrieval, LM classification, and web search
Modular pipeline achieves high accuracy without fine-tuning
🔎 Similar Papers
No similar papers found.
S
Shaghayegh Kolli
Technical University Munich
R
Richard Rosenbaum
Technical University Munich
T
Timo Cavelius
Technical University Munich
L
Lasse Strothe
Technical University Munich
A
Andrii Lata
Technical University Munich
Jana Diesner
Jana Diesner
Professor, Technical University of Munich & Affiliate Associate Professor, UIUC
natural language processingcomputational social sciencenetwork analysisresponsible computing