Rethinking the Effect of Data Augmentation in Adversarial Contrastive
Learning
- URL: http://arxiv.org/abs/2303.01289v2
- Date: Fri, 3 Mar 2023 02:21:48 GMT
- Title: Rethinking the Effect of Data Augmentation in Adversarial Contrastive
Learning
- Authors: Rundong Luo, Yifei Wang, Yisen Wang
- Abstract summary: We show that DYNACL can improve state-of-the-art self-AT robustness by 8.84% under Auto-Attack on the CIFAR-10 dataset.
We also show that DYNACL can even outperform vanilla supervised adversarial training for the first time.
- Score: 15.259867823352012
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent works have shown that self-supervised learning can achieve remarkable
robustness when integrated with adversarial training (AT). However, the
robustness gap between supervised AT (sup-AT) and self-supervised AT (self-AT)
remains significant. Motivated by this observation, we revisit existing self-AT
methods and discover an inherent dilemma that affects self-AT robustness:
either strong or weak data augmentations are harmful to self-AT, and a medium
strength is insufficient to bridge the gap. To resolve this dilemma, we propose
a simple remedy named DYNACL (Dynamic Adversarial Contrastive Learning). In
particular, we propose an augmentation schedule that gradually anneals from a
strong augmentation to a weak one to benefit from both extreme cases. Besides,
we adopt a fast post-processing stage for adapting it to downstream tasks.
Through extensive experiments, we show that DYNACL can improve state-of-the-art
self-AT robustness by 8.84% under Auto-Attack on the CIFAR-10 dataset, and can
even outperform vanilla supervised adversarial training for the first time. Our
code is available at \url{https://github.com/PKU-ML/DYNACL}.
Related papers
- Can Large Reasoning Models Self-Train? [58.953117118687096]
Scaling the performance of large language models increasingly depends on methods that reduce reliance on human supervision.<n>We propose an online self-training reinforcement learning algorithm that leverages the model's self-consistency to infer correctness signals and train without any ground-truth supervision.
arXiv Detail & Related papers (2025-05-27T17:16:00Z) - The Effectiveness of Random Forgetting for Robust Generalization [21.163070161951868]
We introduce a novel learning paradigm called "Forget to Mitigate Overfitting" (FOMO)
FOMO alternates between the forgetting phase, which randomly forgets a subset of weights, and the relearning phase, which emphasizes learning generalizable features.
Our experiments show that FOMO alleviates robust overfitting by significantly reducing the gap between the best and last robust test accuracy.
arXiv Detail & Related papers (2024-02-18T23:14:40Z) - SEDA: Self-Ensembling ViT with Defensive Distillation and Adversarial
Training for robust Chest X-rays Classification [0.8812173669205372]
Vision Transformer (ViT) to adversarial, privacy, and confidentiality attacks raise serious concerns about their reliability in medical settings.
We propose Self-Ensembling ViT with defensive Distillation and Adversarial training (SEDA)
SEDA utilizes efficient CNN blocks to learn spatial features with various levels of abstraction from feature representations extracted from intermediate ViT blocks.
arXiv Detail & Related papers (2023-08-15T16:40:46Z) - RelaxLoss: Defending Membership Inference Attacks without Losing Utility [68.48117818874155]
We propose a novel training framework based on a relaxed loss with a more achievable learning target.
RelaxLoss is applicable to any classification model with added benefits of easy implementation and negligible overhead.
Our approach consistently outperforms state-of-the-art defense mechanisms in terms of resilience against MIAs.
arXiv Detail & Related papers (2022-07-12T19:34:47Z) - When Does Contrastive Learning Preserve Adversarial Robustness from
Pretraining to Finetuning? [99.4914671654374]
We propose AdvCL, a novel adversarial contrastive pretraining framework.
We show that AdvCL is able to enhance cross-task robustness transferability without loss of model accuracy and finetuning efficiency.
arXiv Detail & Related papers (2021-11-01T17:59:43Z) - Adversarial Visual Robustness by Causal Intervention [56.766342028800445]
Adversarial training is the de facto most promising defense against adversarial examples.
Yet, its passive nature inevitably prevents it from being immune to unknown attackers.
We provide a causal viewpoint of adversarial vulnerability: the cause is the confounder ubiquitously existing in learning.
arXiv Detail & Related papers (2021-06-17T14:23:54Z) - Consistency Regularization for Adversarial Robustness [88.65786118562005]
Adversarial training is one of the most successful methods to obtain the adversarial robustness of deep neural networks.
However, a significant generalization gap in the robustness obtained from AT has been problematic.
In this paper, we investigate data augmentation techniques to address the issue.
arXiv Detail & Related papers (2021-03-08T09:21:41Z) - Adversarial Momentum-Contrastive Pre-Training [20.336258934272102]
Adversarial self-supervised pre-training is helpful to extract the invariant representations under both data augmentations and adversarial perturbations.
This paper proposes a novel adversarial momentum-contrastive (AMOC) pre-training approach.
Compared with the existing self-supervised pre-training approaches, AMOC can use a smaller batch size and fewer training epochs but learn more robust features.
arXiv Detail & Related papers (2020-12-24T07:49:10Z) - Robust Pre-Training by Adversarial Contrastive Learning [120.33706897927391]
Recent work has shown that, when integrated with adversarial training, self-supervised pre-training can lead to state-of-the-art robustness.
We improve robustness-aware self-supervised pre-training by learning representations consistent under both data augmentations and adversarial perturbations.
arXiv Detail & Related papers (2020-10-26T04:44:43Z) - Boosting Adversarial Training with Hypersphere Embedding [53.75693100495097]
Adversarial training is one of the most effective defenses against adversarial attacks for deep learning models.
In this work, we advocate incorporating the hypersphere embedding mechanism into the AT procedure.
We validate our methods under a wide range of adversarial attacks on the CIFAR-10 and ImageNet datasets.
arXiv Detail & Related papers (2020-02-20T08:42:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.