Reflection Backdoor: A Natural Backdoor Attack on Deep Neural Networks
- URL: http://arxiv.org/abs/2007.02343v2
- Date: Mon, 13 Jul 2020 13:46:10 GMT
- Title: Reflection Backdoor: A Natural Backdoor Attack on Deep Neural Networks
- Authors: Yunfei Liu, Xingjun Ma, James Bailey, Feng Lu
- Abstract summary: A backdoor attack installs a backdoor into the victim model by injecting a backdoor pattern into a small proportion of the training data.
We propose reflection backdoor (Refool) to plant reflections as backdoor into a victim model.
We demonstrate on 3 computer vision tasks and 5 datasets that, Refool can attack state-of-the-art DNNs with high success rate.
- Score: 46.99548490594115
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies have shown that DNNs can be compromised by backdoor attacks
crafted at training time. A backdoor attack installs a backdoor into the victim
model by injecting a backdoor pattern into a small proportion of the training
data. At test time, the victim model behaves normally on clean test data, yet
consistently predicts a specific (likely incorrect) target class whenever the
backdoor pattern is present in a test example. While existing backdoor attacks
are effective, they are not stealthy. The modifications made on training data
or labels are often suspicious and can be easily detected by simple data
filtering or human inspection. In this paper, we present a new type of backdoor
attack inspired by an important natural phenomenon: reflection. Using
mathematical modeling of physical reflection models, we propose reflection
backdoor (Refool) to plant reflections as backdoor into a victim model. We
demonstrate on 3 computer vision tasks and 5 datasets that, Refool can attack
state-of-the-art DNNs with high success rate, and is resistant to
state-of-the-art backdoor defenses.
Related papers
- Expose Before You Defend: Unifying and Enhancing Backdoor Defenses via Exposed Models [68.40324627475499]
We introduce a novel two-step defense framework named Expose Before You Defend.
EBYD unifies existing backdoor defense methods into a comprehensive defense system with enhanced performance.
We conduct extensive experiments on 10 image attacks and 6 text attacks across 2 vision datasets and 4 language datasets.
arXiv Detail & Related papers (2024-10-25T09:36:04Z) - Mitigating Backdoor Attack by Injecting Proactive Defensive Backdoor [63.84477483795964]
Data-poisoning backdoor attacks are serious security threats to machine learning models.
In this paper, we focus on in-training backdoor defense, aiming to train a clean model even when the dataset may be potentially poisoned.
We propose a novel defense approach called PDB (Proactive Defensive Backdoor)
arXiv Detail & Related papers (2024-05-25T07:52:26Z) - Backdoor Defense via Deconfounded Representation Learning [17.28760299048368]
We propose a Causality-inspired Backdoor Defense (CBD) to learn deconfounded representations for reliable classification.
CBD is effective in reducing backdoor threats while maintaining high accuracy in predicting benign samples.
arXiv Detail & Related papers (2023-03-13T02:25:59Z) - Untargeted Backdoor Attack against Object Detection [69.63097724439886]
We design a poison-only backdoor attack in an untargeted manner, based on task characteristics.
We show that, once the backdoor is embedded into the target model by our attack, it can trick the model to lose detection of any object stamped with our trigger patterns.
arXiv Detail & Related papers (2022-11-02T17:05:45Z) - BATT: Backdoor Attack with Transformation-based Triggers [72.61840273364311]
Deep neural networks (DNNs) are vulnerable to backdoor attacks.
Backdoor adversaries inject hidden backdoors that can be activated by adversary-specified trigger patterns.
One recent research revealed that most of the existing attacks failed in the real physical world.
arXiv Detail & Related papers (2022-11-02T16:03:43Z) - Can Backdoor Attacks Survive Time-Varying Models? [35.836598031681426]
Backdoors are powerful attacks against deep neural networks (DNNs)
We study the impact of backdoor attacks on a more realistic scenario of time-varying DNN models.
Our results show that one-shot backdoor attacks do not survive past a few model updates.
arXiv Detail & Related papers (2022-06-08T01:32:49Z) - Test-Time Detection of Backdoor Triggers for Poisoned Deep Neural
Networks [24.532269628999025]
Backdoor (Trojan) attacks are emerging threats against deep neural networks (DNN)
In this paper, we propose an "in-flight" defense against backdoor attacks on image classification.
arXiv Detail & Related papers (2021-12-06T20:52:00Z) - Black-box Detection of Backdoor Attacks with Limited Information and
Data [56.0735480850555]
We propose a black-box backdoor detection (B3D) method to identify backdoor attacks with only query access to the model.
In addition to backdoor detection, we also propose a simple strategy for reliable predictions using the identified backdoored models.
arXiv Detail & Related papers (2021-03-24T12:06:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.