Correcting False Alarms from Unseen: Adapting Graph Anomaly Detectors at Test Time

📅 2025-11-10
📈 Citations: 0
Influential: 0
📄 PDF
🤖 AI Summary
Graph anomaly detection (GAD) suffers significant performance degradation under train-test distribution shift, primarily due to unseen normal samples causing semantic confusion (misclassifying novel normals as anomalies) and aggregation contamination (distorting representations of known normal nodes). To address this, we propose TUNE, a lightweight test-time adaptation framework that enables online correction without retraining. TUNE introduces a graph aligner to minimize attribute representation shift and jointly optimizes semantic discrimination and neighborhood aggregation based on estimated aggregation contamination severity. Designed as a plug-and-play module, TUNE is compatible with diverse pre-trained GAD models. Extensive experiments across 10 real-world datasets demonstrate that TUNE substantially improves robustness against both synthetic and real-world unseen normal patterns, while effectively reducing false positive rates.

Technology Category

Application Category

📝 Abstract
Graph anomaly detection (GAD), which aims to detect outliers in graph-structured data, has received increasing research attention recently. However, existing GAD methods assume identical training and testing distributions, which is rarely valid in practice. In real-world scenarios, unseen but normal samples may emerge during deployment, leading to a normality shift that degrades the performance of GAD models trained on the original data. Through empirical analysis, we reveal that the degradation arises from (1) semantic confusion, where unseen normal samples are misinterpreted as anomalies due to their novel patterns, and (2) aggregation contamination, where the representations of seen normal nodes are distorted by unseen normals through message aggregation. While retraining or fine-tuning GAD models could be a potential solution to the above challenges, the high cost of model retraining and the difficulty of obtaining labeled data often render this approach impractical in real-world applications. To bridge the gap, we proposed a lightweight and plug-and-play Test-time adaptation framework for correcting Unseen Normal pattErns (TUNE) in GAD. To address semantic confusion, a graph aligner is employed to align the shifted data to the original one at the graph attribute level. Moreover, we utilize the minimization of representation-level shift as a supervision signal to train the aligner, which leverages the estimated aggregation contamination as a key indicator of normality shift. Extensive experiments on 10 real-world datasets demonstrate that TUNE significantly enhances the generalizability of pre-trained GAD models to both synthetic and real unseen normal patterns.
Problem

Research questions and friction points this paper is trying to address.

Addressing performance degradation in graph anomaly detection due to unseen normal patterns
Correcting semantic confusion where novel normal samples are misinterpreted as anomalies
Mitigating aggregation contamination that distorts representations through message passing
Innovation

Methods, ideas, or system contributions that make the work stand out.

Test-time adaptation framework for unseen normals
Graph aligner corrects attribute-level semantic confusion
Minimizes representation shift using aggregation contamination
Junjun Pan
Junjun Pan
School of Information and Communication Technology, Griffith University, Queensland, Australia
Y
Yixin Liu
School of Information and Communication Technology, Griffith University, Queensland, Australia
C
Chuan Zhou
Academy of Mathematics and Systems Science, Chinese Academy of Sciences, Beijing, China
F
Fei Xiong
School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, China
Alan Wee-Chung Liew
Alan Wee-Chung Liew
Professor, School of ICT, Griffith University
Machine learningmedical imagingcomputer visionensemble learningdata stream learning
Shirui Pan
Shirui Pan
Professor, ARC Future Fellow, FQA, Director of TrustAGI Lab, Griffith University
Data MiningMachine LearningGraph Neural NetworksTrustworthy AITime Series