The World of an Octopus: How Reporting Bias Influences a Language
Model's Perception of Color
- URL: http://arxiv.org/abs/2110.08182v1
- Date: Fri, 15 Oct 2021 16:28:17 GMT
- Title: The World of an Octopus: How Reporting Bias Influences a Language
Model's Perception of Color
- Authors: Cory Paik, St\'ephane Aroca-Ouellette, Alessandro Roncone and
Katharina Kann
- Abstract summary: We show that reporting bias negatively impacts and inherently limits text-only training.
We then demonstrate that multimodal models can leverage their visual training to mitigate these effects.
- Score: 73.70233477125781
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent work has raised concerns about the inherent limitations of text-only
pretraining. In this paper, we first demonstrate that reporting bias, the
tendency of people to not state the obvious, is one of the causes of this
limitation, and then investigate to what extent multimodal training can
mitigate this issue. To accomplish this, we 1) generate the Color Dataset
(CoDa), a dataset of human-perceived color distributions for 521 common
objects; 2) use CoDa to analyze and compare the color distribution found in
text, the distribution captured by language models, and a human's perception of
color; and 3) investigate the performance differences between text-only and
multimodal models on CoDa. Our results show that the distribution of colors
that a language model recovers correlates more strongly with the inaccurate
distribution found in text than with the ground-truth, supporting the claim
that reporting bias negatively impacts and inherently limits text-only
training. We then demonstrate that multimodal models can leverage their visual
training to mitigate these effects, providing a promising avenue for future
research.
Related papers
- Multilingual Text-to-Image Generation Magnifies Gender Stereotypes and Prompt Engineering May Not Help You [64.74707085021858]
We show that multilingual models suffer from significant gender biases just as monolingual models do.
We propose a novel benchmark, MAGBIG, intended to foster research on gender bias in multilingual models.
Our results show that not only do models exhibit strong gender biases but they also behave differently across languages.
arXiv Detail & Related papers (2024-01-29T12:02:28Z) - All Should Be Equal in the Eyes of Language Models: Counterfactually
Aware Fair Text Generation [16.016546693767403]
We propose a framework that dynamically compares the model understanding of diverse demographics to generate more equitable sentences.
CAFIE produces fairer text and strikes the best balance between fairness and language modeling capability.
arXiv Detail & Related papers (2023-11-09T15:39:40Z) - Exposing Bias in Online Communities through Large-Scale Language Models [3.04585143845864]
This work uses the flaw of bias in language models to explore the biases of six different online communities.
The bias of the resulting models is evaluated by prompting the models with different demographics and comparing the sentiment and toxicity values of these generations.
This work not only affirms how easily bias is absorbed from training data but also presents a scalable method to identify and compare the bias of different datasets or communities.
arXiv Detail & Related papers (2023-06-04T08:09:26Z) - Stubborn Lexical Bias in Data and Models [50.79738900885665]
We use a new statistical method to examine whether spurious patterns in data appear in models trained on the data.
We apply an optimization approach to *reweight* the training data, reducing thousands of spurious correlations.
Surprisingly, though this method can successfully reduce lexical biases in the training data, we still find strong evidence of corresponding bias in the trained models.
arXiv Detail & Related papers (2023-06-03T20:12:27Z) - Debiasing Vision-Language Models via Biased Prompts [79.04467131711775]
We propose a general approach for debiasing vision-language foundation models by projecting out biased directions in the text embedding.
We show that debiasing only the text embedding with a calibrated projection matrix suffices to yield robust classifiers and fair generative models.
arXiv Detail & Related papers (2023-01-31T20:09:33Z) - Debiasing Stance Detection Models with Counterfactual Reasoning and
Adversarial Bias Learning [15.68462203989933]
Stance detection models tend to rely on dataset bias in the text part as a shortcut.
We propose an adversarial bias learning module to model the bias more accurately.
arXiv Detail & Related papers (2022-12-20T16:20:56Z) - Naturalistic Causal Probing for Morpho-Syntax [76.83735391276547]
We suggest a naturalistic strategy for input-level intervention on real world data in Spanish.
Using our approach, we isolate morpho-syntactic features from counfounders in sentences.
We apply this methodology to analyze causal effects of gender and number on contextualized representations extracted from pre-trained models.
arXiv Detail & Related papers (2022-05-14T11:47:58Z) - Visual Commonsense in Pretrained Unimodal and Multimodal Models [29.462625570767123]
We investigate what degree unimodal (language-only) and multimodal (image and language) models capture a broad range of visually salient attributes.
We create the Visual Commonsense Tests dataset covering 5 property types (color, shape, material, size, and visual co-occurrence) for over 5000 subjects.
We then use our dataset to evaluate pretrained unimodal models and multimodal models.
arXiv Detail & Related papers (2022-05-04T02:07:55Z) - Balancing out Bias: Achieving Fairness Through Training Reweighting [58.201275105195485]
Bias in natural language processing arises from models learning characteristics of the author such as gender and race.
Existing methods for mitigating and measuring bias do not directly account for correlations between author demographics and linguistic variables.
This paper introduces a very simple but highly effective method for countering bias using instance reweighting.
arXiv Detail & Related papers (2021-09-16T23:40:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.