Overcoming Simplicity Bias in Deep Networks using a Feature Sieve
- URL: http://arxiv.org/abs/2301.13293v3
- Date: Tue, 6 Jun 2023 16:45:31 GMT
- Title: Overcoming Simplicity Bias in Deep Networks using a Feature Sieve
- Authors: Rishabh Tiwari, Pradeep Shenoy
- Abstract summary: We propose a direct, interventional method for addressing simplicity bias in deep networks.
We aim to automatically identify and suppress easily-computable spurious features in lower layers of the network.
We report substantial gains on many real-world debiasing benchmarks.
- Score: 5.33024001730262
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Simplicity bias is the concerning tendency of deep networks to over-depend on
simple, weakly predictive features, to the exclusion of stronger, more complex
features. This is exacerbated in real-world applications by limited training
data and spurious feature-label correlations, leading to biased, incorrect
predictions. We propose a direct, interventional method for addressing
simplicity bias in DNNs, which we call the feature sieve. We aim to
automatically identify and suppress easily-computable spurious features in
lower layers of the network, thereby allowing the higher network levels to
extract and utilize richer, more meaningful representations. We provide
concrete evidence of this differential suppression & enhancement of relevant
features on both controlled datasets and real-world images, and report
substantial gains on many real-world debiasing benchmarks (11.4% relative gain
on Imagenet-A; 3.2% on BAR, etc). Crucially, we do not depend on prior
knowledge of spurious attributes or features, and in fact outperform many
baselines that explicitly incorporate such information. We believe that our
feature sieve work opens up exciting new research directions in automated
adversarial feature extraction and representation learning for deep networks.
Related papers
- Improving Network Interpretability via Explanation Consistency Evaluation [56.14036428778861]
We propose a framework that acquires more explainable activation heatmaps and simultaneously increase the model performance.
Specifically, our framework introduces a new metric, i.e., explanation consistency, to reweight the training samples adaptively in model learning.
Our framework then promotes the model learning by paying closer attention to those training samples with a high difference in explanations.
arXiv Detail & Related papers (2024-08-08T17:20:08Z) - Simplicity Bias of Two-Layer Networks beyond Linearly Separable Data [4.14360329494344]
We characterize simplicity bias for general datasets in the context of two-layer neural networks with small weights and trained with gradient flow.
For datasets with an XOR-like pattern, we precisely identify the learned features and demonstrate that simplicity bias intensifies during later training stages.
These results indicate that features learned in the middle stages of training may be more useful for OOD transfer.
arXiv Detail & Related papers (2024-05-27T16:00:45Z) - Hierarchical Simplicity Bias of Neural Networks [0.0]
We introduce a novel method called imbalanced label coupling to explore and extend this simplicity bias across hierarchical levels.
Our approach demonstrates that trained networks sequentially consider features of increasing complexity based on their correlation with labels in the training set.
arXiv Detail & Related papers (2023-11-05T11:27:03Z) - Implicit Visual Bias Mitigation by Posterior Estimate Sharpening of a
Bayesian Neural Network [7.488317734152586]
We propose a novel implicit mitigation method using a Bayesian neural network.
Our proposed posterior estimate sharpening procedure encourages the network to focus on core features that do not contribute to high uncertainties.
arXiv Detail & Related papers (2023-03-29T09:47:35Z) - Look Beyond Bias with Entropic Adversarial Data Augmentation [4.893694715581673]
Deep neural networks do not discriminate between spurious and causal patterns, and will only learn the most predictive ones while ignoring the others.
Debiasing methods were developed to make networks robust to such spurious biases but require to know in advance if a dataset is biased.
In this paper, we argue that such samples should not be necessarily needed because the ''hidden'' causal information is often also contained in biased images.
arXiv Detail & Related papers (2023-01-10T08:25:24Z) - Neural networks trained with SGD learn distributions of increasing
complexity [78.30235086565388]
We show that neural networks trained using gradient descent initially classify their inputs using lower-order input statistics.
We then exploit higher-order statistics only later during training.
We discuss the relation of DSB to other simplicity biases and consider its implications for the principle of universality in learning.
arXiv Detail & Related papers (2022-11-21T15:27:22Z) - Learning an Invertible Output Mapping Can Mitigate Simplicity Bias in
Neural Networks [66.76034024335833]
We investigate why diverse/ complex features are learned by the backbone, and their brittleness is due to the linear classification head relying primarily on the simplest features.
We propose Feature Reconstruction Regularizer (FRR) to ensure that the learned features can be reconstructed back from the logits.
We demonstrate up to 15% gains in OOD accuracy on the recently introduced semi-synthetic datasets with extreme distribution shifts.
arXiv Detail & Related papers (2022-10-04T04:01:15Z) - Correlation-Aware Deep Tracking [83.51092789908677]
We propose a novel target-dependent feature network inspired by the self-/cross-attention scheme.
Our network deeply embeds cross-image feature correlation in multiple layers of the feature network.
Our model can be flexibly pre-trained on abundant unpaired images, leading to notably faster convergence than the existing methods.
arXiv Detail & Related papers (2022-03-03T11:53:54Z) - Learning Deep Interleaved Networks with Asymmetric Co-Attention for
Image Restoration [65.11022516031463]
We present a deep interleaved network (DIN) that learns how information at different states should be combined for high-quality (HQ) images reconstruction.
In this paper, we propose asymmetric co-attention (AsyCA) which is attached at each interleaved node to model the feature dependencies.
Our presented DIN can be trained end-to-end and applied to various image restoration tasks.
arXiv Detail & Related papers (2020-10-29T15:32:00Z) - Resolution Adaptive Networks for Efficient Inference [53.04907454606711]
We propose a novel Resolution Adaptive Network (RANet), which is inspired by the intuition that low-resolution representations are sufficient for classifying "easy" inputs.
In RANet, the input images are first routed to a lightweight sub-network that efficiently extracts low-resolution representations.
High-resolution paths in the network maintain the capability to recognize the "hard" samples.
arXiv Detail & Related papers (2020-03-16T16:54:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.