On the Loss Landscape of Adversarial Training: Identifying Challenges
and How to Overcome Them
- URL: http://arxiv.org/abs/2006.08403v2
- Date: Mon, 2 Nov 2020 22:43:42 GMT
- Title: On the Loss Landscape of Adversarial Training: Identifying Challenges
and How to Overcome Them
- Authors: Chen Liu, Mathieu Salzmann, Tao Lin, Ryota Tomioka, Sabine S\"usstrunk
- Abstract summary: We analyze the influence of adversarial training on the loss landscape of machine learning models.
We show that the adversarial loss landscape is less favorable to optimization, due to increased curvature and more scattered gradients.
- Score: 57.957466608543676
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We analyze the influence of adversarial training on the loss landscape of
machine learning models. To this end, we first provide analytical studies of
the properties of adversarial loss functions under different adversarial
budgets. We then demonstrate that the adversarial loss landscape is less
favorable to optimization, due to increased curvature and more scattered
gradients. Our conclusions are validated by numerical analyses, which show that
training under large adversarial budgets impede the escape from suboptimal
random initialization, cause non-vanishing gradients and make the model find
sharper minima. Based on these observations, we show that a periodic
adversarial scheduling (PAS) strategy can effectively overcome these
challenges, yielding better results than vanilla adversarial training while
being much less sensitive to the choice of learning rate.
Related papers
- Robust Losses for Decision-Focused Learning [2.9652474178611405]
Decision-focused learning aims at training the predictive model to minimize regret by making a suboptimal decision.
empirical regret can be an ineffective surrogate because empirical optimal decisions can vary substantially from expected optimal decisions.
We propose three novel loss functions that approximate expected regret more robustly.
arXiv Detail & Related papers (2023-10-06T15:45:10Z) - Gradient constrained sharpness-aware prompt learning for vision-language
models [99.74832984957025]
This paper targets a novel trade-off problem in generalizable prompt learning for vision-language models (VLM)
By analyzing the loss landscapes of the state-of-the-art method and vanilla Sharpness-aware Minimization (SAM) based method, we conclude that the trade-off performance correlates to both loss value and loss sharpness.
We propose a novel SAM-based method for prompt learning, denoted as Gradient Constrained Sharpness-aware Context Optimization (GCSCoOp)
arXiv Detail & Related papers (2023-09-14T17:13:54Z) - Enhancing Adversarial Training with Feature Separability [52.39305978984573]
We introduce a new concept of adversarial training graph (ATG) with which the proposed adversarial training with feature separability (ATFS) enables to boost the intra-class feature similarity and increase inter-class feature variance.
Through comprehensive experiments, we demonstrate that the proposed ATFS framework significantly improves both clean and robust performance.
arXiv Detail & Related papers (2022-05-02T04:04:23Z) - Imitating, Fast and Slow: Robust learning from demonstrations via
decision-time planning [96.72185761508668]
Planning at Test-time (IMPLANT) is a new meta-algorithm for imitation learning.
We demonstrate that IMPLANT significantly outperforms benchmark imitation learning approaches on standard control environments.
arXiv Detail & Related papers (2022-04-07T17:16:52Z) - A Loss Curvature Perspective on Training Instability in Deep Learning [28.70491071044542]
We study the evolution of the loss Hessian across many classification tasks in order to understand the effect curvature of the loss has on the training dynamics.
Inspired by the conditioning perspective, we show that learning rate warmup can improve training stability just as much as batch normalization.
arXiv Detail & Related papers (2021-10-08T20:25:48Z) - Unleashing the Power of Contrastive Self-Supervised Visual Models via
Contrast-Regularized Fine-Tuning [94.35586521144117]
We investigate whether applying contrastive learning to fine-tuning would bring further benefits.
We propose Contrast-regularized tuning (Core-tuning), a novel approach for fine-tuning contrastive self-supervised visual models.
arXiv Detail & Related papers (2021-02-12T16:31:24Z) - Adversarially Robust Learning via Entropic Regularization [31.6158163883893]
We propose a new family of algorithms, ATENT, for training adversarially robust deep neural networks.
Our approach achieves competitive (or better) performance in terms of robust classification accuracy.
arXiv Detail & Related papers (2020-08-27T18:54:43Z) - On the Generalization Properties of Adversarial Training [21.79888306754263]
This paper studies the generalization performance of a generic adversarial training algorithm.
A series of numerical studies are conducted to demonstrate how the smoothness and L1 penalization help improve the adversarial robustness of models.
arXiv Detail & Related papers (2020-08-15T02:32:09Z) - Reparameterized Variational Divergence Minimization for Stable Imitation [57.06909373038396]
We study the extent to which variations in the choice of probabilistic divergence may yield more performant ILO algorithms.
We contribute a re parameterization trick for adversarial imitation learning to alleviate the challenges of the promising $f$-divergence minimization framework.
Empirically, we demonstrate that our design choices allow for ILO algorithms that outperform baseline approaches and more closely match expert performance in low-dimensional continuous-control tasks.
arXiv Detail & Related papers (2020-06-18T19:04:09Z) - Perturbation Analysis of Gradient-based Adversarial Attacks [2.3016608994689274]
We investigate the objective functions of three popular methods for adversarial example generation: the L-BFGS attack, the Iterative Fast Gradient Sign attack, and Carlini & Wagner's attack (CW)
Specifically, we perform a comparative and formal analysis of the loss functions underlying the aforementioned attacks while laying out large-scale experimental results on ImageNet dataset.
Our experiments reveal that the Iterative Fast Gradient Sign attack, which is thought to be fast for generating adversarial examples, is the worst attack in terms of the number of iterations required to create adversarial examples.
arXiv Detail & Related papers (2020-06-02T08:51:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.