Don't Lag, RAG: Training-Free Adversarial Detection Using RAG

๐Ÿ“… 2025-04-07
๐Ÿ“ˆ Citations: 0
โœจ Influential: 0
๐Ÿ“„ PDF
๐Ÿค– AI Summary
To address the real-time detection challenge of local adversarial patch attacks in vision systems, this paper proposes Vision Retrieval-Augmented Generation (VRAG), a training-free and fine-tuning-free framework. VRAG integrates multiple heterogeneous vision-language models (VLMs) and enables zero-shot generative discriminative reasoning for attack identification. It establishes the first training-free paradigm for adversarial patch detection. We construct a scalable visual attack database, enabling plug-and-play compatibility with both open- and closed-source large VLMsโ€”including Qwen-VL-Plus, UI-TARS-72B-DPO, and Gemini-2.0. Experiments demonstrate that UI-TARS-72B-DPO achieves 95% accuracy (state-of-the-art among open-source models), while Gemini-2.0 attains the highest accuracy of 98%. The approach significantly reduces reliance on manual annotation and exhibits strong generalization, effectively detecting diverse novel adversarial patches across unseen attack types and domains.

Technology Category

Application Category

๐Ÿ“ Abstract
Adversarial patch attacks pose a major threat to vision systems by embedding localized perturbations that mislead deep models. Traditional defense methods often require retraining or fine-tuning, making them impractical for real-world deployment. We propose a training-free Visual Retrieval-Augmented Generation (VRAG) framework that integrates Vision-Language Models (VLMs) for adversarial patch detection. By retrieving visually similar patches and images that resemble stored attacks in a continuously expanding database, VRAG performs generative reasoning to identify diverse attack types, all without additional training or fine-tuning. We extensively evaluate open-source large-scale VLMs, including Qwen-VL-Plus, Qwen2.5-VL-72B, and UI-TARS-72B-DPO, alongside Gemini-2.0, a closed-source model. Notably, the open-source UI-TARS-72B-DPO model achieves up to 95 percent classification accuracy, setting a new state-of-the-art for open-source adversarial patch detection. Gemini-2.0 attains the highest overall accuracy, 98 percent, but remains closed-source. Experimental results demonstrate VRAG's effectiveness in identifying a variety of adversarial patches with minimal human annotation, paving the way for robust, practical defenses against evolving adversarial patch attacks.
Problem

Research questions and friction points this paper is trying to address.

Detect adversarial patch attacks without retraining models
Use Vision-Language Models for training-free detection
Achieve high accuracy in identifying diverse attack types
Innovation

Methods, ideas, or system contributions that make the work stand out.

Training-free VRAG framework for adversarial detection
Uses Vision-Language Models for patch identification
Leverages generative reasoning with expanding database
๐Ÿ”Ž Similar Papers
No similar papers found.