On the Use of Large Language Models for Qualitative Synthesis

📅 2025-10-18
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
This study addresses methodological risks arising from the application of large language models (LLMs) in the qualitative synthesis (QS) phase of systematic reviews—particularly bias amplification and diminished result credibility. Adopting a novel collaborative reflexive research paradigm, it integrates LLM technical mechanism analysis, iterative empirical trials, and methodological critique frameworks to systematically evaluate LLMs’ text summarization capabilities and practical limitations. Findings reveal that inconsistent methodological reporting frequently leads to LLM misuse, while intrinsic structural properties—including training data biases and opaque reasoning processes—significantly undermine the reproducibility and interpretability of synthesized outputs. Based on these insights, the study proposes three foundational principles: human-led oversight, model-assisted execution, and process transparency—highlighting the critical role of embedded supervision and auditable operational logging in safeguarding QS rigor. This work delivers the first reflexively grounded risk identification framework and governance pathway for integrating LLMs into evidence-based research.

Technology Category

Application Category

📝 Abstract
Large language models (LLMs) show promise for supporting systematic reviews (SR), even complex tasks such as qualitative synthesis (QS). However, applying them to a stage that is unevenly reported and variably conducted carries important risks: misuse can amplify existing weaknesses and erode confidence in the SR findings. To examine the challenges of using LLMs for QS, we conducted a collaborative autoethnography involving two trials. We evaluated each trial for methodological rigor and practical usefulness, and interpreted the results through a technical lens informed by how LLMs are built and their current limitations.
Problem

Research questions and friction points this paper is trying to address.

Applying LLMs to qualitative synthesis in systematic reviews
Examining risks of misuse that amplify existing weaknesses
Evaluating methodological rigor and practical usefulness of LLMs
Innovation

Methods, ideas, or system contributions that make the work stand out.

Using collaborative autoethnography to evaluate LLMs
Assessing methodological rigor and practical usefulness
Interpreting results through technical lens of LLMs
🔎 Similar Papers
No similar papers found.
S
Sebastián Pizard
Facultad de Ingeniería, Universidad de la República, Montevideo, Uruguay
R
Ramiro Moreira
Facultad de Ingeniería, Universidad de la República, Montevideo, Uruguay
F
Federico Galiano
Facultad de Ingeniería, Universidad de la República, Montevideo, Uruguay
I
Ignacio Sastre
Facultad de Ingeniería, Universidad de la República, Montevideo, Uruguay
Lorena Etcheverry
Lorena Etcheverry
Associate Professor, Instituto de Computación, Facultad de Ingeniería, Universidad de la Republica
Data ManagementGraph DatabasesSemantic WebData Quality