FineDeb: A Debiasing Framework for Language Models
- URL: http://arxiv.org/abs/2302.02453v1
- Date: Sun, 5 Feb 2023 18:35:21 GMT
- Title: FineDeb: A Debiasing Framework for Language Models
- Authors: Akash Saravanan, Dhruv Mullick, Habibur Rahman, Nidhi Hegde
- Abstract summary: We propose FineDeb, a two-phase debiasing framework for language models.
Our results show that FineDeb offers stronger debiasing in comparison to other methods.
Our framework is generalizable for demographics with multiple classes.
- Score: 3.7698299781999376
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As language models are increasingly included in human-facing machine learning
tools, bias against demographic subgroups has gained attention. We propose
FineDeb, a two-phase debiasing framework for language models that starts with
contextual debiasing of embeddings learned by pretrained language models. The
model is then fine-tuned on a language modeling objective. Our results show
that FineDeb offers stronger debiasing in comparison to other methods which
often result in models as biased as the original language model. Our framework
is generalizable for demographics with multiple classes, and we demonstrate its
effectiveness through extensive experiments and comparisons with state of the
art techniques. We release our code and data on GitHub.
Related papers
- Collapsed Language Models Promote Fairness [88.48232731113306]
We find that debiased language models exhibit collapsed alignment between token representations and word embeddings.
We design a principled fine-tuning method that can effectively improve fairness in a wide range of debiasing methods.
arXiv Detail & Related papers (2024-10-06T13:09:48Z) - Language Models for Text Classification: Is In-Context Learning Enough? [54.869097980761595]
Recent foundational language models have shown state-of-the-art performance in many NLP tasks in zero- and few-shot settings.
An advantage of these models over more standard approaches is the ability to understand instructions written in natural language (prompts)
This makes them suitable for addressing text classification problems for domains with limited amounts of annotated instances.
arXiv Detail & Related papers (2024-03-26T12:47:39Z) - Revisiting Topic-Guided Language Models [20.21486464604549]
We study four topic-guided language models and two baselines, evaluating the held-out predictive performance of each model on four corpora.
We find that none of these methods outperform a standard LSTM language model baseline, and most fail to learn good topics.
arXiv Detail & Related papers (2023-12-04T20:33:24Z) - FairPy: A Toolkit for Evaluation of Social Biases and their Mitigation
in Large Language Models [7.250074804839615]
Studies have shown that large pretrained language models exhibit biases against social groups based on race, gender etc.
Various researchers have proposed mathematical tools for quantifying and identifying these biases.
We present a comprehensive quantitative evaluation of different kinds of biases such as race, gender, ethnicity, age etc.
arXiv Detail & Related papers (2023-02-10T20:54:10Z) - Debiasing Vision-Language Models via Biased Prompts [79.04467131711775]
We propose a general approach for debiasing vision-language foundation models by projecting out biased directions in the text embedding.
We show that debiasing only the text embedding with a calibrated projection matrix suffices to yield robust classifiers and fair generative models.
arXiv Detail & Related papers (2023-01-31T20:09:33Z) - Robust Preference Learning for Storytelling via Contrastive
Reinforcement Learning [53.92465205531759]
Controlled automated story generation seeks to generate natural language stories satisfying constraints from natural language critiques or preferences.
We train a contrastive bi-encoder model to align stories with human critiques, building a general purpose preference model.
We further fine-tune the contrastive reward model using a prompt-learning technique to increase story generation robustness.
arXiv Detail & Related papers (2022-10-14T13:21:33Z) - Read Like Humans: Autonomous, Bidirectional and Iterative Language
Modeling for Scene Text Recognition [80.446770909975]
Linguistic knowledge is of great benefit to scene text recognition.
How to effectively model linguistic rules in end-to-end deep networks remains a research challenge.
We propose an autonomous, bidirectional and iterative ABINet for scene text recognition.
arXiv Detail & Related papers (2021-03-11T06:47:45Z) - Comparison of Interactive Knowledge Base Spelling Correction Models for
Low-Resource Languages [81.90356787324481]
Spelling normalization for low resource languages is a challenging task because the patterns are hard to predict.
This work shows a comparison of a neural model and character language models with varying amounts on target language data.
Our usage scenario is interactive correction with nearly zero amounts of training examples, improving models as more data is collected.
arXiv Detail & Related papers (2020-10-20T17:31:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.