The shape and simplicity biases of adversarially robust ImageNet-trained
CNNs
- URL: http://arxiv.org/abs/2006.09373v6
- Date: Mon, 12 Sep 2022 13:13:07 GMT
- Title: The shape and simplicity biases of adversarially robust ImageNet-trained
CNNs
- Authors: Peijie Chen, Chirag Agarwal, Anh Nguyen
- Abstract summary: We study the shape bias and internal mechanisms that enable the generalizability of AlexNet, GoogLeNet, and ResNet-50 models trained via adversarial training.
Remarkably, adversarial training induces three simplicity biases into hidden neurons in the process of "robustifying" CNNs.
- Score: 9.707679445925516
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Increasingly more similarities between human vision and convolutional neural
networks (CNNs) have been revealed in the past few years. Yet, vanilla CNNs
often fall short in generalizing to adversarial or out-of-distribution (OOD)
examples which humans demonstrate superior performance. Adversarial training is
a leading learning algorithm for improving the robustness of CNNs on
adversarial and OOD data; however, little is known about the properties,
specifically the shape bias and internal features learned inside
adversarially-robust CNNs. In this paper, we perform a thorough, systematic
study to understand the shape bias and some internal mechanisms that enable the
generalizability of AlexNet, GoogLeNet, and ResNet-50 models trained via
adversarial training. We find that while standard ImageNet classifiers have a
strong texture bias, their R counterparts rely heavily on shapes. Remarkably,
adversarial training induces three simplicity biases into hidden neurons in the
process of "robustifying" CNNs. That is, each convolutional neuron in R
networks often changes to detecting (1) pixel-wise smoother patterns, i.e., a
mechanism that blocks high-frequency noise from passing through the network;
(2) more lower-level features i.e. textures and colors (instead of objects);and
(3) fewer types of inputs. Our findings reveal the interesting mechanisms that
made networks more adversarially robust and also explain some recent findings
e.g., why R networks benefit from a much larger capacity (Xie et al. 2020) and
can act as a strong image prior in image synthesis (Santurkar et al. 2019).
Related papers
- Using Human-like Mechanism to Weaken Effect of Pre-training Weight Bias
in Face-Recognition Convolutional Neural Network [6.0950431324191845]
We focus on 4 extensively studied CNNs (AlexNet, VGG11, VGG13, and VGG16) which has been analyzed as human-like models by neuroscientists.
We trained these CNNs to emotion valence classification task by transfer learning.
We then update the object-based AlexNet using self-attention mechanism based on neuroscience and behavioral data.
arXiv Detail & Related papers (2023-10-20T17:22:57Z) - Improving the Accuracy and Robustness of CNNs Using a Deep CCA Neural
Data Regularizer [2.026424957803652]
As convolutional neural networks (CNNs) become more accurate at object recognition, their representations become more similar to the primate visual system.
Previous attempts to address this question showed very modest gains in accuracy, owing in part to limitations of the regularization method.
We develop a new neural data regularizer for CNNs that uses Deep Correlation Analysis (DCCA) to optimize the resemblance of the CNN's image representations to that of the monkey visual cortex.
arXiv Detail & Related papers (2022-09-06T15:40:39Z) - Exploring Adversarial Examples and Adversarial Robustness of
Convolutional Neural Networks by Mutual Information [44.841339443764696]
This work investigates similarities and differences between two types of convolutional neural networks (CNNs) in information extraction.
The reason why adversarial examples mislead CNNs may be that they contain more texture-based information about other categories.
Normally trained CNNs tend to extract texture-based information from the inputs, while adversarially trained models prefer to shape-based information.
arXiv Detail & Related papers (2022-07-12T13:25:42Z) - Neural Architecture Dilation for Adversarial Robustness [56.18555072877193]
A shortcoming of convolutional neural networks is that they are vulnerable to adversarial attacks.
This paper aims to improve the adversarial robustness of the backbone CNNs that have a satisfactory accuracy.
Under a minimal computational overhead, a dilation architecture is expected to be friendly with the standard performance of the backbone CNN.
arXiv Detail & Related papers (2021-08-16T03:58:00Z) - BreakingBED -- Breaking Binary and Efficient Deep Neural Networks by
Adversarial Attacks [65.2021953284622]
We study robustness of CNNs against white-box and black-box adversarial attacks.
Results are shown for distilled CNNs, agent-based state-of-the-art pruned models, and binarized neural networks.
arXiv Detail & Related papers (2021-03-14T20:43:19Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Extreme Value Preserving Networks [65.2037926048262]
Recent evidence shows that convolutional neural networks (CNNs) are biased towards textures so that CNNs are non-robust to adversarial perturbations over textures.
This paper aims to leverage good properties of SIFT to renovate CNN architectures towards better accuracy and robustness.
arXiv Detail & Related papers (2020-11-17T02:06:52Z) - Informative Dropout for Robust Representation Learning: A Shape-bias
Perspective [84.30946377024297]
We propose a light-weight model-agnostic method, namely Informative Dropout (InfoDrop), to improve interpretability and reduce texture bias.
Specifically, we discriminate texture from shape based on local self-information in an image, and adopt a Dropout-like algorithm to decorrelate the model output from the local texture.
arXiv Detail & Related papers (2020-08-10T16:52:24Z) - Teaching CNNs to mimic Human Visual Cognitive Process & regularise
Texture-Shape bias [18.003188982585737]
Recent experiments in computer vision demonstrate texture bias as the primary reason for supreme results in models employing Convolutional Neural Networks (CNNs)
It is believed that the cost function forces the CNN to take a greedy approach and develop a proclivity for local information like texture to increase accuracy, thus failing to explore any global statistics.
We propose CognitiveCNN, a new intuitive architecture, inspired from feature integration theory in psychology to utilise human interpretable feature like shape, texture, edges etc. to reconstruct, and classify the image.
arXiv Detail & Related papers (2020-06-25T22:32:54Z) - Deep Polynomial Neural Networks [77.70761658507507]
$Pi$Nets are a new class of function approximators based on expansions.
$Pi$Nets produce state-the-art results in three challenging tasks, i.e. image generation, face verification and 3D mesh representation learning.
arXiv Detail & Related papers (2020-06-20T16:23:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.