Backdoor Attacks and Countermeasures on Deep Learning: A Comprehensive
Review
- URL: http://arxiv.org/abs/2007.10760v3
- Date: Sun, 2 Aug 2020 08:38:25 GMT
- Title: Backdoor Attacks and Countermeasures on Deep Learning: A Comprehensive
Review
- Authors: Yansong Gao, Bao Gia Doan, Zhi Zhang, Siqi Ma, Jiliang Zhang, Anmin
Fu, Surya Nepal, and Hyoungshick Kim
- Abstract summary: This work provides the community with a timely comprehensive review of backdoor attacks and countermeasures on deep learning.
According to the attacker's capability and affected stage of the machine learning pipeline, the attack surfaces are recognized to be wide.
Countermeasures are categorized into four general classes: blind backdoor removal, offline backdoor inspection, online backdoor inspection, and post backdoor removal.
- Score: 40.36824357892676
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This work provides the community with a timely comprehensive review of
backdoor attacks and countermeasures on deep learning. According to the
attacker's capability and affected stage of the machine learning pipeline, the
attack surfaces are recognized to be wide and then formalized into six
categorizations: code poisoning, outsourcing, pretrained, data collection,
collaborative learning and post-deployment. Accordingly, attacks under each
categorization are combed. The countermeasures are categorized into four
general classes: blind backdoor removal, offline backdoor inspection, online
backdoor inspection, and post backdoor removal. Accordingly, we review
countermeasures, and compare and analyze their advantages and disadvantages. We
have also reviewed the flip side of backdoor attacks, which are explored for i)
protecting intellectual property of deep learning models, ii) acting as a
honeypot to catch adversarial example attacks, and iii) verifying data deletion
requested by the data contributor.Overall, the research on defense is far
behind the attack, and there is no single defense that can prevent all types of
backdoor attacks. In some cases, an attacker can intelligently bypass existing
defenses with an adaptive attack. Drawing the insights from the systematic
review, we also present key areas for future research on the backdoor, such as
empirical security evaluations from physical trigger attacks, and in
particular, more efficient and practical countermeasures are solicited.
Related papers
- Mitigating Backdoor Attack by Injecting Proactive Defensive Backdoor [63.84477483795964]
Data-poisoning backdoor attacks are serious security threats to machine learning models.
In this paper, we focus on in-training backdoor defense, aiming to train a clean model even when the dataset may be potentially poisoned.
We propose a novel defense approach called PDB (Proactive Defensive Backdoor)
arXiv Detail & Related papers (2024-05-25T07:52:26Z) - On the Difficulty of Defending Contrastive Learning against Backdoor
Attacks [58.824074124014224]
We show how contrastive backdoor attacks operate through distinctive mechanisms.
Our findings highlight the need for defenses tailored to the specificities of contrastive backdoor attacks.
arXiv Detail & Related papers (2023-12-14T15:54:52Z) - Backdoor Attacks and Countermeasures in Natural Language Processing Models: A Comprehensive Security Review [15.179940846141873]
Applicating third-party data and models has become a new paradigm for language modeling in NLP.
backdoor attacks can induce the model to exhibit expected behaviors through specific triggers.
There is still no systematic and comprehensive review to reflect the security challenges, attacker's capabilities, and purposes.
arXiv Detail & Related papers (2023-09-12T08:48:38Z) - Rethinking Backdoor Attacks [122.1008188058615]
In a backdoor attack, an adversary inserts maliciously constructed backdoor examples into a training set to make the resulting model vulnerable to manipulation.
Defending against such attacks typically involves viewing these inserted examples as outliers in the training set and using techniques from robust statistics to detect and remove them.
We show that without structural information about the training data distribution, backdoor attacks are indistinguishable from naturally-occurring features in the data.
arXiv Detail & Related papers (2023-07-19T17:44:54Z) - Detecting Backdoors in Deep Text Classifiers [43.36440869257781]
We present the first robust defence mechanism that generalizes to several backdoor attacks against text classification models.
Our technique is highly accurate at defending against state-of-the-art backdoor attacks, including data poisoning and weight poisoning.
arXiv Detail & Related papers (2022-10-11T07:48:03Z) - On the Effectiveness of Adversarial Training against Backdoor Attacks [111.8963365326168]
A backdoored model always predicts a target class in the presence of a predefined trigger pattern.
In general, adversarial training is believed to defend against backdoor attacks.
We propose a hybrid strategy which provides satisfactory robustness across different backdoor attacks.
arXiv Detail & Related papers (2022-02-22T02:24:46Z) - Backdoor Learning: A Survey [75.59571756777342]
Backdoor attack intends to embed hidden backdoor into deep neural networks (DNNs)
Backdoor learning is an emerging and rapidly growing research area.
This paper presents the first comprehensive survey of this realm.
arXiv Detail & Related papers (2020-07-17T04:09:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.