A Multiclass Boosting Framework for Achieving Fast and Provable
Adversarial Robustness
- URL: http://arxiv.org/abs/2103.01276v2
- Date: Wed, 3 Mar 2021 17:17:28 GMT
- Title: A Multiclass Boosting Framework for Achieving Fast and Provable
Adversarial Robustness
- Authors: Jacob Abernethy, Pranjal Awasthi, Satyen Kale
- Abstract summary: deep neural networks can be corrupted in order to modify output predictions.
This apparent lack of robustness has led researchers to propose methods that can help to prevent an adversary from having such capabilities.
We propose a multiclass boosting framework to ensure adversarial robustness.
- Score: 32.90358643120235
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Alongside the well-publicized accomplishments of deep neural networks there
has emerged an apparent bug in their success on tasks such as object
recognition: with deep models trained using vanilla methods, input images can
be slightly corrupted in order to modify output predictions, even when these
corruptions are practically invisible. This apparent lack of robustness has led
researchers to propose methods that can help to prevent an adversary from
having such capabilities. The state-of-the-art approaches have incorporated the
robustness requirement into the loss function, and the training process
involves taking stochastic gradient descent steps not using original inputs but
on adversarially-corrupted ones. In this paper we propose a multiclass boosting
framework to ensure adversarial robustness. Boosting algorithms are generally
well-suited for adversarial scenarios, as they were classically designed to
satisfy a minimax guarantee. We provide a theoretical foundation for this
methodology and describe conditions under which robustness can be achieved
given a weak training oracle. We show empirically that adversarially-robust
multiclass boosting not only outperforms the state-of-the-art methods, it does
so at a fraction of the training time.
Related papers
- Adversarial Training Can Provably Improve Robustness: Theoretical Analysis of Feature Learning Process Under Structured Data [38.44734564565478]
We provide a theoretical understanding of adversarial examples and adversarial training algorithms from the perspective of feature learning theory.
We show that the adversarial training method can provably strengthen the robust feature learning and suppress the non-robust feature learning.
arXiv Detail & Related papers (2024-10-11T03:59:49Z) - Adversarial Robustification via Text-to-Image Diffusion Models [56.37291240867549]
Adrial robustness has been conventionally believed as a challenging property to encode for neural networks.
We develop a scalable and model-agnostic solution to achieve adversarial robustness without using any data.
arXiv Detail & Related papers (2024-07-26T10:49:14Z) - Unlearning Backdoor Threats: Enhancing Backdoor Defense in Multimodal Contrastive Learning via Local Token Unlearning [49.242828934501986]
Multimodal contrastive learning has emerged as a powerful paradigm for building high-quality features.
backdoor attacks subtly embed malicious behaviors within the model during training.
We introduce an innovative token-based localized forgetting training regime.
arXiv Detail & Related papers (2024-03-24T18:33:15Z) - Doubly Robust Instance-Reweighted Adversarial Training [107.40683655362285]
We propose a novel doubly-robust instance reweighted adversarial framework.
Our importance weights are obtained by optimizing the KL-divergence regularized loss function.
Our proposed approach outperforms related state-of-the-art baseline methods in terms of average robust performance.
arXiv Detail & Related papers (2023-08-01T06:16:18Z) - Adversarial Training Should Be Cast as a Non-Zero-Sum Game [121.95628660889628]
Two-player zero-sum paradigm of adversarial training has not engendered sufficient levels of robustness.
We show that the commonly used surrogate-based relaxation used in adversarial training algorithms voids all guarantees on robustness.
A novel non-zero-sum bilevel formulation of adversarial training yields a framework that matches and in some cases outperforms state-of-the-art attacks.
arXiv Detail & Related papers (2023-06-19T16:00:48Z) - Confidence-aware Training of Smoothed Classifiers for Certified
Robustness [75.95332266383417]
We use "accuracy under Gaussian noise" as an easy-to-compute proxy of adversarial robustness for an input.
Our experiments show that the proposed method consistently exhibits improved certified robustness upon state-of-the-art training methods.
arXiv Detail & Related papers (2022-12-18T03:57:12Z) - Certifiably-Robust Federated Adversarial Learning via Randomized
Smoothing [16.528628447356496]
In this paper, we incorporate smoothing techniques into federated adversarial training to enable data-private distributed learning.
Our experiments show that such an advanced federated adversarial learning framework can deliver models as robust as those trained by the centralized training.
arXiv Detail & Related papers (2021-03-30T02:19:45Z) - Towards Achieving Adversarial Robustness by Enforcing Feature
Consistency Across Bit Planes [51.31334977346847]
We train networks to form coarse impressions based on the information in higher bit planes, and use the lower bit planes only to refine their prediction.
We demonstrate that, by imposing consistency on the representations learned across differently quantized images, the adversarial robustness of networks improves significantly.
arXiv Detail & Related papers (2020-04-01T09:31:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.