Detecting LLM-Generated Spam Reviews by Integrating Language Model Embeddings and Graph Neural Network

📅 2025-10-02
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Existing detection methods struggle to identify highly realistic, adversarial spam reviews generated by large language models (LLMs). Method: We propose an end-to-end detection framework that jointly models textual semantics and user behavioral graph structure. Leveraging three LLM-synthesized spam review datasets, we extract text embeddings via a pretrained language model and employ a gated graph neural network to encode heterogeneous user interaction graphs—eliminating the need for manual feature engineering. Contribution/Results: To our knowledge, this is the first work to unify semantic representation learning with heterogeneous behavioral graph analysis within a lightweight hybrid architecture. On three LLM-generated benchmarks, our method outperforms state-of-the-art approaches by +44.22% in accuracy and +43.01% in recall. It also generalizes effectively to real human-written spam reviews, exhibits low annotation dependency, and demonstrates strong practical deployability.

Technology Category

Application Category

📝 Abstract
The rise of large language models (LLMs) has enabled the generation of highly persuasive spam reviews that closely mimic human writing. These reviews pose significant challenges for existing detection systems and threaten the credibility of online platforms. In this work, we first create three realistic LLM-generated spam review datasets using three distinct LLMs, each guided by product metadata and genuine reference reviews. Evaluations by GPT-4.1 confirm the high persuasion and deceptive potential of these reviews. To address this threat, we propose FraudSquad, a hybrid detection model that integrates text embeddings from a pre-trained language model with a gated graph transformer for spam node classification. FraudSquad captures both semantic and behavioral signals without relying on manual feature engineering or massive training resources. Experiments show that FraudSquad outperforms state-of-the-art baselines by up to 44.22% in precision and 43.01% in recall on three LLM-generated datasets, while also achieving promising results on two human-written spam datasets. Furthermore, FraudSquad maintains a modest model size and requires minimal labeled training data, making it a practical solution for real-world applications. Our contributions include new synthetic datasets, a practical detection framework, and empirical evidence highlighting the urgency of adapting spam detection to the LLM era. Our code and datasets are available at: https://anonymous.4open.science/r/FraudSquad-5389/.
Problem

Research questions and friction points this paper is trying to address.

Detecting persuasive spam reviews generated by large language models
Integrating semantic embeddings with graph networks for classification
Addressing threats to online platform credibility from deceptive content
Innovation

Methods, ideas, or system contributions that make the work stand out.

Integrates language model embeddings with graph neural network
Uses gated graph transformer for spam node classification
Captures semantic and behavioral signals without manual features
🔎 Similar Papers
No similar papers found.
X
Xin Liu
Tsinghua University, Beijing, China; SUPCON, Hangzhou, China
Rongwu Xu
Rongwu Xu
University of Washington
AIHuman-AINLPRL
X
Xinyi Jia
Tsinghua University, Beijing, China
J
Jason Liao
University of British Columbia, Vancouver, British Columbia, Canada
Jiao Sun
Jiao Sun
Google DeepMind
Natural Language Generation
L
Ling Huang
AHI Fintech, Beijing, China
W
Wei Xu
Tsinghua University, Beijing, China