Friendly Noise against Adversarial Noise: A Powerful Defense against
Data Poisoning Attacks
- URL: http://arxiv.org/abs/2208.10224v4
- Date: Thu, 20 Jul 2023 05:42:46 GMT
- Title: Friendly Noise against Adversarial Noise: A Powerful Defense against
Data Poisoning Attacks
- Authors: Tian Yu Liu, Yu Yang, Baharan Mirzasoleiman
- Abstract summary: A powerful category of (invisible) data poisoning attacks modify a subset of training examples by small adversarial perturbations to change the prediction of certain test-time data.
Here, we propose a highly effective approach that unlike existing methods breaks various types of invisible poisoning attacks with the slightest drop in the generalization performance.
Our approach comprises two components: an optimized friendly noise that is generated to maximally perturb examples without degrading the performance, and a randomly varying noise component.
- Score: 15.761683760167777
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A powerful category of (invisible) data poisoning attacks modify a subset of
training examples by small adversarial perturbations to change the prediction
of certain test-time data. Existing defense mechanisms are not desirable to
deploy in practice, as they often either drastically harm the generalization
performance, or are attack-specific, and prohibitively slow to apply. Here, we
propose a simple but highly effective approach that unlike existing methods
breaks various types of invisible poisoning attacks with the slightest drop in
the generalization performance. We make the key observation that attacks
introduce local sharp regions of high training loss, which when minimized,
results in learning the adversarial perturbations and makes the attack
successful. To break poisoning attacks, our key idea is to alleviate the sharp
loss regions introduced by poisons. To do so, our approach comprises two
components: an optimized friendly noise that is generated to maximally perturb
examples without degrading the performance, and a randomly varying noise
component. The combination of both components builds a very light-weight but
extremely effective defense against the most powerful triggerless targeted and
hidden-trigger backdoor poisoning attacks, including Gradient Matching,
Bulls-eye Polytope, and Sleeper Agent. We show that our friendly noise is
transferable to other architectures, and adaptive attacks cannot break our
defense due to its random noise component. Our code is available at:
https://github.com/tianyu139/friendly-noise
Related papers
- ECLIPSE: Expunging Clean-label Indiscriminate Poisons via Sparse Diffusion Purification [29.28977815669999]
Clean-label indiscriminate poisoning attacks add invisible perturbations to correctly labeled training images.
We propose a more universally effective, practical, and robust defense scheme called ECLIPSE.
arXiv Detail & Related papers (2024-06-21T12:14:24Z) - FreqFed: A Frequency Analysis-Based Approach for Mitigating Poisoning
Attacks in Federated Learning [98.43475653490219]
Federated learning (FL) is susceptible to poisoning attacks.
FreqFed is a novel aggregation mechanism that transforms the model updates into the frequency domain.
We demonstrate that FreqFed can mitigate poisoning attacks effectively with a negligible impact on the utility of the aggregated model.
arXiv Detail & Related papers (2023-12-07T16:56:24Z) - HINT: Healthy Influential-Noise based Training to Defend against Data
Poisoning Attacks [12.929357709840975]
We propose an efficient and robust training approach to defend against data poisoning attacks based on influence functions.
Using influence functions, we craft healthy noise that helps to harden the classification model against poisoning attacks.
Our empirical results show that HINT can efficiently protect deep learning models against the effect of both untargeted and targeted poisoning attacks.
arXiv Detail & Related papers (2023-09-15T17:12:19Z) - Adversarial Robustness of Deep Reinforcement Learning based Dynamic
Recommender Systems [50.758281304737444]
We propose to explore adversarial examples and attack detection on reinforcement learning-based interactive recommendation systems.
We first craft different types of adversarial examples by adding perturbations to the input and intervening on the casual factors.
Then, we augment recommendation systems by detecting potential attacks with a deep learning-based classifier based on the crafted data.
arXiv Detail & Related papers (2021-12-02T04:12:24Z) - Sparse and Imperceptible Adversarial Attack via a Homotopy Algorithm [93.80082636284922]
Sparse adversarial attacks can fool deep networks (DNNs) by only perturbing a few pixels.
Recent efforts combine it with another l_infty perturbation on magnitudes.
We propose a homotopy algorithm to tackle the sparsity and neural perturbation framework.
arXiv Detail & Related papers (2021-06-10T20:11:36Z) - Towards Defending against Adversarial Examples via Attack-Invariant
Features [147.85346057241605]
Deep neural networks (DNNs) are vulnerable to adversarial noise.
adversarial robustness can be improved by exploiting adversarial examples.
Models trained on seen types of adversarial examples generally cannot generalize well to unseen types of adversarial examples.
arXiv Detail & Related papers (2021-06-09T12:49:54Z) - PDPGD: Primal-Dual Proximal Gradient Descent Adversarial Attack [92.94132883915876]
State-of-the-art deep neural networks are sensitive to small input perturbations.
Many defence methods have been proposed that attempt to improve robustness to adversarial noise.
evaluating adversarial robustness has proven to be extremely challenging.
arXiv Detail & Related papers (2021-06-03T01:45:48Z) - Removing Adversarial Noise in Class Activation Feature Space [160.78488162713498]
We propose to remove adversarial noise by implementing a self-supervised adversarial training mechanism in a class activation feature space.
We train a denoising model to minimize the distances between the adversarial examples and the natural examples in the class activation feature space.
Empirical evaluations demonstrate that our method could significantly enhance adversarial robustness in comparison to previous state-of-the-art approaches.
arXiv Detail & Related papers (2021-04-19T10:42:24Z) - Adversarial Feature Desensitization [12.401175943131268]
We propose a novel approach to adversarial robustness, which builds upon the insights from the domain adaptation field.
Our method, called Adversarial Feature Desensitization (AFD), aims at learning features that are invariant towards adversarial perturbations of the inputs.
arXiv Detail & Related papers (2020-06-08T14:20:02Z) - Ensemble Noise Simulation to Handle Uncertainty about Gradient-based
Adversarial Attacks [5.4572790062292125]
A gradient-based adversarial attack on neural networks can be crafted in a variety of ways by varying how the attack algorithm relies on the gradient.
Most recent work has focused on defending classifiers in a case where there is no uncertainty about the attacker's behavior.
We fill this gap by simulating the attacker's noisy perturbation using a variety of attack algorithms based on gradients of various classifiers.
We demonstrate significant improvements in post-attack accuracy, using our proposed ensemble-trained defense.
arXiv Detail & Related papers (2020-01-26T17:12:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.