BiasBed -- Rigorous Texture Bias Evaluation
- URL: http://arxiv.org/abs/2211.13190v3
- Date: Fri, 24 Mar 2023 20:21:52 GMT
- Title: BiasBed -- Rigorous Texture Bias Evaluation
- Authors: Nikolai Kalischek, Rodrigo C. Daudt, Torben Peters, Reinhard Furrer,
Jan D. Wegner, Konrad Schindler
- Abstract summary: We introduce BiasBed, a testbed for texture- and style-biased training.
It comes with rigorous hypothesis testing to gauge the significance of the results.
E.g., we find that some algorithms proposed in the literature do not significantly mitigate the impact of style bias at all.
- Score: 21.55506905780658
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The well-documented presence of texture bias in modern convolutional neural
networks has led to a plethora of algorithms that promote an emphasis on shape
cues, often to support generalization to new domains. Yet, common datasets,
benchmarks and general model selection strategies are missing, and there is no
agreed, rigorous evaluation protocol. In this paper, we investigate
difficulties and limitations when training networks with reduced texture bias.
In particular, we also show that proper evaluation and meaningful comparisons
between methods are not trivial. We introduce BiasBed, a testbed for texture-
and style-biased training, including multiple datasets and a range of existing
algorithms. It comes with an extensive evaluation protocol that includes
rigorous hypothesis testing to gauge the significance of the results, despite
the considerable training instability of some style bias methods. Our extensive
experiments, shed new light on the need for careful, statistically founded
evaluation protocols for style bias (and beyond). E.g., we find that some
algorithms proposed in the literature do not significantly mitigate the impact
of style bias at all. With the release of BiasBed, we hope to foster a common
understanding of consistent and meaningful comparisons, and consequently faster
progress towards learning methods free of texture bias. Code is available at
https://github.com/D1noFuzi/BiasBed
Related papers
- CosFairNet:A Parameter-Space based Approach for Bias Free Learning [1.9116784879310025]
Deep neural networks trained on biased data often inadvertently learn unintended inference rules.
We introduce a novel approach to address bias directly in the model's parameter space, preventing its propagation across layers.
We show enhanced classification accuracy and debiasing effectiveness across various synthetic and real-world datasets.
arXiv Detail & Related papers (2024-10-19T13:06:40Z) - Language-guided Detection and Mitigation of Unknown Dataset Bias [23.299264313976213]
We propose a framework to identify potential biases as keywords without prior knowledge based on the partial occurrence in the captions.
Our framework not only outperforms existing methods without prior knowledge, but also is even comparable with a method that assumes prior knowledge.
arXiv Detail & Related papers (2024-06-05T03:11:33Z) - Towards Real World Debiasing: A Fine-grained Analysis On Spurious Correlation [17.080528126651977]
We revisit biased distributions in existing benchmarks and real-world datasets, and propose a fine-grained framework for analyzing dataset bias.
Results show that existing methods are incapable of handling real-world biases.
We propose a simple yet effective approach that can be easily applied to existing debias methods, named Debias in Destruction (DiD)
arXiv Detail & Related papers (2024-05-24T06:06:41Z) - Evaluating Graph Neural Networks for Link Prediction: Current Pitfalls
and New Benchmarking [66.83273589348758]
Link prediction attempts to predict whether an unseen edge exists based on only a portion of edges of a graph.
A flurry of methods have been introduced in recent years that attempt to make use of graph neural networks (GNNs) for this task.
New and diverse datasets have also been created to better evaluate the effectiveness of these new models.
arXiv Detail & Related papers (2023-06-18T01:58:59Z) - Spuriosity Rankings: Sorting Data to Measure and Mitigate Biases [62.54519787811138]
We present a simple but effective method to measure and mitigate model biases caused by reliance on spurious cues.
We rank images within their classes based on spuriosity, proxied via deep neural features of an interpretable network.
Our results suggest that model bias due to spurious feature reliance is influenced far more by what the model is trained on than how it is trained.
arXiv Detail & Related papers (2022-12-05T23:15:43Z) - Prototype-Anchored Learning for Learning with Imperfect Annotations [83.7763875464011]
It is challenging to learn unbiased classification models from imperfectly annotated datasets.
We propose a prototype-anchored learning (PAL) method, which can be easily incorporated into various learning-based classification schemes.
We verify the effectiveness of PAL on class-imbalanced learning and noise-tolerant learning by extensive experiments on synthetic and real-world datasets.
arXiv Detail & Related papers (2022-06-23T10:25:37Z) - Pseudo Bias-Balanced Learning for Debiased Chest X-ray Classification [57.53567756716656]
We study the problem of developing debiased chest X-ray diagnosis models without knowing exactly the bias labels.
We propose a novel algorithm, pseudo bias-balanced learning, which first captures and predicts per-sample bias labels.
Our proposed method achieved consistent improvements over other state-of-the-art approaches.
arXiv Detail & Related papers (2022-03-18T11:02:18Z) - Evading the Simplicity Bias: Training a Diverse Set of Models Discovers
Solutions with Superior OOD Generalization [93.8373619657239]
Neural networks trained with SGD were recently shown to rely preferentially on linearly-predictive features.
This simplicity bias can explain their lack of robustness out of distribution (OOD)
We demonstrate that the simplicity bias can be mitigated and OOD generalization improved.
arXiv Detail & Related papers (2021-05-12T12:12:24Z) - Are Bias Mitigation Techniques for Deep Learning Effective? [24.84797949716142]
We introduce an improved evaluation protocol, sensible metrics, and a new dataset.
We evaluate seven state-of-the-art algorithms using the same network architecture.
We find that algorithms exploit hidden biases, are unable to scale to multiple forms of bias, and are highly sensitive to the choice of tuning set.
arXiv Detail & Related papers (2021-04-01T00:14:45Z) - Debiasing Pre-trained Contextualised Embeddings [28.378270372391498]
We propose a fine-tuning method that can be applied at token- or sentence-levels to debias pre-trained contextualised embeddings.
Using gender bias as an illustrative example, we then conduct a systematic study using several state-of-the-art (SoTA) contextualised representations.
We find that applying token-level debiasing for all tokens and across all layers of a contextualised embedding model produces the best performance.
arXiv Detail & Related papers (2021-01-23T15:28:48Z) - Learning from Failure: Training Debiased Classifier from Biased
Classifier [76.52804102765931]
We show that neural networks learn to rely on spurious correlation only when it is "easier" to learn than the desired knowledge.
We propose a failure-based debiasing scheme by training a pair of neural networks simultaneously.
Our method significantly improves the training of the network against various types of biases in both synthetic and real-world datasets.
arXiv Detail & Related papers (2020-07-06T07:20:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.