Permuted AdaIN: Reducing the Bias Towards Global Statistics in Image
Classification
- URL: http://arxiv.org/abs/2010.05785v3
- Date: Wed, 23 Jun 2021 17:04:49 GMT
- Title: Permuted AdaIN: Reducing the Bias Towards Global Statistics in Image
Classification
- Authors: Oren Nuriel, Sagie Benaim, Lior Wolf
- Abstract summary: Recent work has shown that convolutional neural network classifiers overly rely on texture at the expense of shape cues.
We make a similar but different distinction between shape and local image cues, on the one hand, and global image statistics, on the other.
Our method, called Permuted Adaptive Instance Normalization (pAdaIN), reduces the representation of global statistics in the hidden layers of image classifiers.
- Score: 97.81205777897043
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent work has shown that convolutional neural network classifiers overly
rely on texture at the expense of shape cues. We make a similar but different
distinction between shape and local image cues, on the one hand, and global
image statistics, on the other. Our method, called Permuted Adaptive Instance
Normalization (pAdaIN), reduces the representation of global statistics in the
hidden layers of image classifiers. pAdaIN samples a random permutation $\pi$
that rearranges the samples in a given batch. Adaptive Instance Normalization
(AdaIN) is then applied between the activations of each (non-permuted) sample
$i$ and the corresponding activations of the sample $\pi(i)$, thus swapping
statistics between the samples of the batch. Since the global image statistics
are distorted, this swapping procedure causes the network to rely on cues, such
as shape or texture. By choosing the random permutation with probability $p$
and the identity permutation otherwise, one can control the effect's strength.
With the correct choice of $p$, fixed apriori for all experiments and
selected without considering test data, our method consistently outperforms
baselines in multiple settings. In image classification, our method improves on
both CIFAR100 and ImageNet using multiple architectures. In the setting of
robustness, our method improves on both ImageNet-C and Cifar-100-C for multiple
architectures. In the setting of domain adaptation and domain generalization,
our method achieves state of the art results on the transfer learning task from
GTAV to Cityscapes and on the PACS benchmark.
Related papers
- MB-RACS: Measurement-Bounds-based Rate-Adaptive Image Compressed Sensing Network [65.1004435124796]
We propose a Measurement-Bounds-based Rate-Adaptive Image Compressed Sensing Network (MB-RACS) framework.
Our experiments demonstrate that the proposed MB-RACS method surpasses current leading methods.
arXiv Detail & Related papers (2024-01-19T04:40:20Z) - High-fidelity Pseudo-labels for Boosting Weakly-Supervised Segmentation [17.804090651425955]
Image-level weakly-supervised segmentation (WSSS) reduces the usually vast data annotation cost by surrogate segmentation masks during training.
Our work is based on two techniques for improving CAMs; importance sampling, which is a substitute for GAP, and the feature similarity loss.
We reformulate both techniques based on binomial posteriors of multiple independent binary problems.
This has two benefits; their performance is improved and they become more general, resulting in an add-on method that can boost virtually any WSSS method.
arXiv Detail & Related papers (2023-04-05T17:43:57Z) - Generalized Differentiable RANSAC [95.95627475224231]
$nabla$-RANSAC is a differentiable RANSAC that allows learning the entire randomized robust estimation pipeline.
$nabla$-RANSAC is superior to the state-of-the-art in terms of accuracy while running at a similar speed to its less accurate alternatives.
arXiv Detail & Related papers (2022-12-26T15:13:13Z) - Test-time Batch Statistics Calibration for Covariate Shift [66.7044675981449]
We propose to adapt the deep models to the novel environment during inference.
We present a general formulation $alpha$-BN to calibrate the batch statistics.
We also present a novel loss function to form a unified test time adaptation framework Core.
arXiv Detail & Related papers (2021-10-06T08:45:03Z) - A Hierarchical Transformation-Discriminating Generative Model for Few
Shot Anomaly Detection [93.38607559281601]
We devise a hierarchical generative model that captures the multi-scale patch distribution of each training image.
The anomaly score is obtained by aggregating the patch-based votes of the correct transformation across scales and image regions.
arXiv Detail & Related papers (2021-04-29T17:49:48Z) - Exploiting Invariance in Training Deep Neural Networks [4.169130102668252]
Inspired by two basic mechanisms in animal visual systems, we introduce a feature transform technique that imposes invariance properties in the training of deep neural networks.
The resulting algorithm requires less parameter tuning, trains well with an initial learning rate 1.0, and easily generalizes to different tasks.
Tested on ImageNet, MS COCO, and Cityscapes datasets, our proposed technique requires fewer iterations to train, surpasses all baselines by a large margin, seamlessly works on both small and large batch size training, and applies to different computer vision tasks of image classification, object detection, and semantic segmentation.
arXiv Detail & Related papers (2021-03-30T19:18:31Z) - Center-wise Local Image Mixture For Contrastive Representation Learning [37.806687971373954]
Contrastive learning based on instance discrimination trains model to discriminate different transformations of the anchor sample from other samples.
This paper proposes a new kind of contrastive learning method, named CLIM, which uses positives from other samples in the dataset.
We reach 75.5% top-1 accuracy with linear evaluation over ResNet-50, and 59.3% top-1 accuracy when fine-tuned with only 1% labels.
arXiv Detail & Related papers (2020-11-05T08:20:31Z) - Tent: Fully Test-time Adaptation by Entropy Minimization [77.85911673550851]
A model must adapt itself to generalize to new and different data during testing.
In this setting of fully test-time adaptation the model has only the test data and its own parameters.
We propose to adapt by test entropy minimization (tent): we optimize the model for confidence as measured by the entropy of its predictions.
arXiv Detail & Related papers (2020-06-18T17:55:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.