NaturalAE: Natural and Robust Physical Adversarial Examples for Object
Detectors
- URL: http://arxiv.org/abs/2011.13692v2
- Date: Wed, 17 Mar 2021 08:47:26 GMT
- Title: NaturalAE: Natural and Robust Physical Adversarial Examples for Object
Detectors
- Authors: Mingfu Xue, Chengxiang Yuan, Can He, Jian Wang, Weiqiang Liu
- Abstract summary: We propose a natural and robust physical adversarial example attack method targeting object detectors under real-world conditions.
The generated adversarial examples are robust to various physical constraints and visually look similar to the original images.
The proposed method ensures the naturalness of the generated adversarial example, and the size of added perturbations is much smaller than the perturbations in the existing works.
- Score: 6.4039013462213
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we propose a natural and robust physical adversarial example
attack method targeting object detectors under real-world conditions. The
generated adversarial examples are robust to various physical constraints and
visually look similar to the original images, thus these adversarial examples
are natural to humans and will not cause any suspicions. First, to ensure the
robustness of the adversarial examples in real-world conditions, the proposed
method exploits different image transformation functions, to simulate various
physical changes during the iterative optimization of the adversarial examples
generation. Second, to construct natural adversarial examples, the proposed
method uses an adaptive mask to constrain the area and intensities of the added
perturbations, and utilizes the real-world perturbation score (RPS) to make the
perturbations be similar to those real noises in physical world. Compared with
existing studies, our generated adversarial examples can achieve a high success
rate with less conspicuous perturbations. Experimental results demonstrate
that, the generated adversarial examples are robust under various indoor and
outdoor physical conditions, including different distances, angles,
illuminations, and photographing. Specifically, the attack success rate of
generated adversarial examples indoors and outdoors is high up to 73.33% and
82.22%, respectively. Meanwhile, the proposed method ensures the naturalness of
the generated adversarial example, and the size of added perturbations is much
smaller than the perturbations in the existing works. Further, the proposed
physical adversarial attack method can be transferred from the white-box models
to other object detection models.
Related papers
- Imperceptible Adversarial Examples in the Physical World [10.981325924844167]
We make adversarial examples imperceptible in the physical world using a straight-through estimator (STE, a.k.a. BPDA)
Our differentiable rendering extension to STE also enables imperceptible adversarial patches in the physical world.
To the best of our knowledge, this is the first work demonstrating imperceptible adversarial examples bounded by small norms in the physical world.
arXiv Detail & Related papers (2024-11-25T18:02:23Z) - Imperceptible Face Forgery Attack via Adversarial Semantic Mask [59.23247545399068]
We propose an Adversarial Semantic Mask Attack framework (ASMA) which can generate adversarial examples with good transferability and invisibility.
Specifically, we propose a novel adversarial semantic mask generative model, which can constrain generated perturbations in local semantic regions for good stealthiness.
arXiv Detail & Related papers (2024-06-16T10:38:11Z) - Extreme Miscalibration and the Illusion of Adversarial Robustness [66.29268991629085]
Adversarial Training is often used to increase model robustness.
We show that this observed gain in robustness is an illusion of robustness (IOR)
We urge the NLP community to incorporate test-time temperature scaling into their robustness evaluations.
arXiv Detail & Related papers (2024-02-27T13:49:12Z) - Diffusion to Confusion: Naturalistic Adversarial Patch Generation Based
on Diffusion Model for Object Detector [18.021582628066554]
We propose a novel naturalistic adversarial patch generation method based on the diffusion models (DM)
We are the first to propose DM-based naturalistic adversarial patch generation for object detectors.
arXiv Detail & Related papers (2023-07-16T15:22:30Z) - The Enemy of My Enemy is My Friend: Exploring Inverse Adversaries for
Improving Adversarial Training [72.39526433794707]
Adversarial training and its variants have been shown to be the most effective approaches to defend against adversarial examples.
We propose a novel adversarial training scheme that encourages the model to produce similar outputs for an adversarial example and its inverse adversarial'' counterpart.
Our training method achieves state-of-the-art robustness as well as natural accuracy.
arXiv Detail & Related papers (2022-11-01T15:24:26Z) - Shadows can be Dangerous: Stealthy and Effective Physical-world
Adversarial Attack by Natural Phenomenon [79.33449311057088]
We study a new type of optical adversarial examples, in which the perturbations are generated by a very common natural phenomenon, shadow.
We extensively evaluate the effectiveness of this new attack on both simulated and real-world environments.
arXiv Detail & Related papers (2022-03-08T02:40:18Z) - Improving White-box Robustness of Pre-processing Defenses via Joint Adversarial Training [106.34722726264522]
A range of adversarial defense techniques have been proposed to mitigate the interference of adversarial noise.
Pre-processing methods may suffer from the robustness degradation effect.
A potential cause of this negative effect is that adversarial training examples are static and independent to the pre-processing model.
We propose a method called Joint Adversarial Training based Pre-processing (JATP) defense.
arXiv Detail & Related papers (2021-06-10T01:45:32Z) - Adversarial Examples Detection beyond Image Space [88.7651422751216]
We find that there exists compliance between perturbations and prediction confidence, which guides us to detect few-perturbation attacks from the aspect of prediction confidence.
We propose a method beyond image space by a two-stream architecture, in which the image stream focuses on the pixel artifacts and the gradient stream copes with the confidence artifacts.
arXiv Detail & Related papers (2021-02-23T09:55:03Z) - On the Similarity of Deep Learning Representations Across Didactic and Adversarial Examples [0.0]
Adrial examples in the wild may inadvertently prove deleterious for accurate predictive modeling.
We show that representational similarity and performance vary according to the frequency of adversarial examples in the input space.
arXiv Detail & Related papers (2020-02-17T07:49:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.