Patch of Invisibility: Naturalistic Physical Black-Box Adversarial Attacks on Object Detectors

📅 2023-03-07
📈 Citations: 3
Influential: 0
📄 PDF
🤖 AI Summary
Addressing the challenge of adversarial attacks against object detectors in black-box, gradient-free, and physically deployable settings, this paper proposes the first fully black-box, gradient-agnostic, naturalistic physical patch attack method grounded in a pre-trained GAN’s image manifold. Our approach optimizes the patch representation within the GAN’s latent space via black-box query-driven search, jointly modeling digital robustness and physical-domain distortions—including lighting variations, viewpoint changes, and print degradation—to achieve model-agnostic attack generation. Evaluated on mainstream detectors (e.g., YOLOv5, Faster R-CNN), the digital attack reduces mAP by over 60%; physical experiments using printed patches also significantly degrade detection performance. Compared to four state-of-the-art black-box attack methods, ours achieves superior attack success rate, cross-model transferability, and physical feasibility—demonstrating both theoretical rigor and practical deployability.
📝 Abstract
Adversarial attacks on deep-learning models have been receiving increased attention in recent years. Work in this area has mostly focused on gradient-based techniques, so-called"white-box"attacks, wherein the attacker has access to the targeted model's internal parameters; such an assumption is usually unrealistic in the real world. Some attacks additionally use the entire pixel space to fool a given model, which is neither practical nor physical (i.e., real-world). On the contrary, we propose herein a direct, black-box, gradient-free method that uses the learned image manifold of a pretrained generative adversarial network (GAN) to generate naturalistic physical adversarial patches for object detectors. To our knowledge this is the first and only method that performs black-box physical attacks directly on object-detection models, which results with a model-agnostic attack. We show that our proposed method works both digitally and physically. We compared our approach against four different black-box attacks with different configurations. Our approach outperformed all other approaches that were tested in our experiments by a large margin.
Problem

Research questions and friction points this paper is trying to address.

Generating naturalistic adversarial patches for object detectors
Performing black-box attacks without model internal parameters
Using GAN-learned image manifold for gradient-free optimization
Innovation

Methods, ideas, or system contributions that make the work stand out.

Black-box attacks using GAN-generated naturalistic patches
Gradient-free method leveraging pretrained generative networks
Model-agnostic approach relying solely on model outputs
🔎 Similar Papers
No similar papers found.
R
Raz Lapid
Department of Computer Science, Ben-Gurion University of the Negev, Beer-Sheva, 8410501, Israel, DeepKeep, Tel-Aviv, Israel
E
Eylon Mizrahi
Department of Computer Science, Ben-Gurion University of the Negev, Beer-Sheva, 8410501, Israel
Moshe Sipper
Moshe Sipper
Ben-Gurion University
evolutionary machine learningevolutionary deep learningbio-inspired computingcellular computing