Staircase Sign Method for Boosting Adversarial Attacks
- URL: http://arxiv.org/abs/2104.09722v1
- Date: Tue, 20 Apr 2021 02:31:55 GMT
- Title: Staircase Sign Method for Boosting Adversarial Attacks
- Authors: Lianli Gao, Qilong Zhang, Xiaosu Zhu, Jingkuan Song and Heng Tao Shen
- Abstract summary: Crafting adversarial examples for the transfer-based attack is challenging and remains a research hot spot.
We propose a novel Staircase Sign Method (S$2$M) to alleviate this issue, thus boosting transfer-based attacks.
Our method can be generally integrated into any transfer-based attacks, and the computational overhead is negligible.
- Score: 123.19227129979943
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Crafting adversarial examples for the transfer-based attack is challenging
and remains a research hot spot. Currently, such attack methods are based on
the hypothesis that the substitute model and the victim's model learn similar
decision boundaries, and they conventionally apply Sign Method (SM) to
manipulate the gradient as the resultant perturbation. Although SM is
efficient, it only extracts the sign of gradient units but ignores their value
difference, which inevitably leads to a serious deviation. Therefore, we
propose a novel Staircase Sign Method (S$^2$M) to alleviate this issue, thus
boosting transfer-based attacks. Technically, our method heuristically divides
the gradient sign into several segments according to the values of the gradient
units, and then assigns each segment with a staircase weight for better
crafting adversarial perturbation. As a result, our adversarial examples
perform better in both white-box and black-box manner without being more
visible. Since S$^2$M just manipulates the resultant gradient, our method can
be generally integrated into any transfer-based attacks, and the computational
overhead is negligible. Extensive experiments on the ImageNet dataset
demonstrate the effectiveness of our proposed methods, which significantly
improve the transferability (i.e., on average, \textbf{5.1\%} for normally
trained models and \textbf{11.2\%} for adversarially trained defenses). Our
code is available at:
\url{https://github.com/qilong-zhang/Staircase-sign-method}.
Related papers
- Hard No-Box Adversarial Attack on Skeleton-Based Human Action
Recognition with Skeleton-Motion-Informed Gradient [14.392853911242923]
Methods for skeleton-based human activity recognition have been shown to be vulnerable to adversarial attacks.
In this paper, we consider a new attack task: the attacker has no access to the victim model or the training data or labels.
Specifically, we define an adversarial loss to compute a new gradient for the attack, named skeleton-motion-informed (SMI) gradient.
arXiv Detail & Related papers (2023-08-10T16:34:20Z) - Sampling-based Fast Gradient Rescaling Method for Highly Transferable
Adversarial Attacks [18.05924632169541]
We propose a Sampling-based Fast Gradient Rescaling Method (S-FGRM)
Specifically, we use data rescaling to substitute the sign function without extra computational cost.
Our method could significantly boost the transferability of gradient-based attacks and outperform the state-of-the-art baselines.
arXiv Detail & Related papers (2023-07-06T07:52:42Z) - Boosting Adversarial Attacks by Leveraging Decision Boundary Information [68.07365511533675]
gradients of different models are more similar on the decision boundary than in the original position.
We propose a Boundary Fitting Attack to improve transferability.
Our method obtains an average attack success rate of 58.2%, which is 10.8% higher than other state-of-the-art transfer-based attacks.
arXiv Detail & Related papers (2023-03-10T05:54:11Z) - Sampling-based Fast Gradient Rescaling Method for Highly Transferable
Adversarial Attacks [19.917677500613788]
gradient-based approaches generally use the $sign$ function to generate perturbations at the end of the process.
We propose a Sampling-based Fast Gradient Rescaling Method (S-FGRM) to improve the transferability of crafted adversarial examples.
arXiv Detail & Related papers (2022-04-06T15:12:20Z) - Query-Efficient Black-box Adversarial Attacks Guided by a Transfer-based
Prior [50.393092185611536]
We consider the black-box adversarial setting, where the adversary needs to craft adversarial examples without access to the gradients of a target model.
Previous methods attempted to approximate the true gradient either by using the transfer gradient of a surrogate white-box model or based on the feedback of model queries.
We propose two prior-guided random gradient-free (PRGF) algorithms based on biased sampling and gradient averaging.
arXiv Detail & Related papers (2022-03-13T04:06:27Z) - Adversarially Robust Classification by Conditional Generative Model
Inversion [4.913248451323163]
We propose a classification model that does not obfuscate gradients and is robust by construction without assuming prior knowledge about the attack.
Our method casts classification as an optimization problem where we "invert" a conditional generator trained on unperturbed, natural images.
We demonstrate that our model is extremely robust against black-box attacks and has improved robustness against white-box attacks.
arXiv Detail & Related papers (2022-01-12T23:11:16Z) - Adaptive Perturbation for Adversarial Attack [50.77612889697216]
We propose a new gradient-based attack method for adversarial examples.
We use the exact gradient direction with a scaling factor for generating adversarial perturbations.
Our method exhibits higher transferability and outperforms the state-of-the-art methods.
arXiv Detail & Related papers (2021-11-27T07:57:41Z) - Boosting Transferability of Targeted Adversarial Examples via
Hierarchical Generative Networks [56.96241557830253]
Transfer-based adversarial attacks can effectively evaluate model robustness in the black-box setting.
We propose a conditional generative attacking model, which can generate the adversarial examples targeted at different classes.
Our method improves the success rates of targeted black-box attacks by a significant margin over the existing methods.
arXiv Detail & Related papers (2021-07-05T06:17:47Z) - Improving the Transferability of Adversarial Examples with New Iteration
Framework and Input Dropout [8.24029748310858]
We propose a new gradient iteration framework, which redefines the relationship between the iteration step size, the number of perturbations, and the maximum iterations.
Under this framework, we easily improve the attack success rate of DI-TI-MIM.
In addition, we propose a gradient iterative attack method based on input dropout, which can be well combined with our framework.
arXiv Detail & Related papers (2021-06-03T06:36:38Z) - Transferable Sparse Adversarial Attack [62.134905824604104]
We introduce a generator architecture to alleviate the overfitting issue and thus efficiently craft transferable sparse adversarial examples.
Our method achieves superior inference speed, 700$times$ faster than other optimization-based methods.
arXiv Detail & Related papers (2021-05-31T06:44:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.