Reinforcement Guided Multi-Task Learning Framework for Low-Resource
Stereotype Detection
- URL: http://arxiv.org/abs/2203.14349v1
- Date: Sun, 27 Mar 2022 17:16:11 GMT
- Title: Reinforcement Guided Multi-Task Learning Framework for Low-Resource
Stereotype Detection
- Authors: Rajkumar Pujari, Erik Oveson, Priyanka Kulkarni, Elnaz Nouri
- Abstract summary: "Stereotype Detection" datasets mainly adopt a diagnostic approach toward large Pre-trained Language Models.
Annotating a reliable dataset requires a precise understanding of the subtle nuances of how stereotypes manifest in text.
We present a multi-task model that leverages the abundance of data-rich neighboring tasks to improve the empirical performance on "Stereotype Detection"
- Score: 3.7223111129285096
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As large Pre-trained Language Models (PLMs) trained on large amounts of data
in an unsupervised manner become more ubiquitous, identifying various types of
bias in the text has come into sharp focus. Existing "Stereotype Detection"
datasets mainly adopt a diagnostic approach toward large PLMs. Blodgett et. al
(2021a) show that there are significant reliability issues with the existing
benchmark datasets. Annotating a reliable dataset requires a precise
understanding of the subtle nuances of how stereotypes manifest in text. In
this paper, we annotate a focused evaluation set for "Stereotype Detection"
that addresses those pitfalls by de-constructing various ways in which
stereotypes manifest in text. Further, we present a multi-task model that
leverages the abundance of data-rich neighboring tasks such as hate speech
detection, offensive language detection, misogyny detection, etc., to improve
the empirical performance on "Stereotype Detection". We then propose a
reinforcement-learning agent that guides the multi-task learning model by
learning to identify the training examples from the neighboring tasks that help
the target task the most. We show that the proposed models achieve significant
empirical gains over existing baselines on all the tasks.
Related papers
- Unsupervised Model Diagnosis [49.36194740479798]
This paper proposes Unsupervised Model Diagnosis (UMO) to produce semantic counterfactual explanations without any user guidance.
Our approach identifies and visualizes changes in semantics, and then matches these changes to attributes from wide-ranging text sources.
arXiv Detail & Related papers (2024-10-08T17:59:03Z) - Training-Free Deepfake Voice Recognition by Leveraging Large-Scale Pre-Trained Models [52.04189118767758]
Generalization is a main issue for current audio deepfake detectors.
In this paper we study the potential of large-scale pre-trained models for audio deepfake detection.
arXiv Detail & Related papers (2024-05-03T15:27:11Z) - Auditing Large Language Models for Enhanced Text-Based Stereotype Detection and Probing-Based Bias Evaluation [4.908389661988191]
This work introduces the Multi-Grain Stereotype dataset, encompassing 51,867 instances across gender, race, profession, religion, and stereotypical text.
We explore different machine learning approaches aimed at establishing baselines for stereotype detection.
We develop a series of stereotype elicitation prompts and evaluate the presence of stereotypes in text generation tasks with popular Large Language Models.
arXiv Detail & Related papers (2024-04-02T09:31:32Z) - Unified Demonstration Retriever for In-Context Learning [56.06473069923567]
Unified Demonstration Retriever (textbfUDR) is a single model to retrieve demonstrations for a wide range of tasks.
We propose a multi-task list-wise ranking training framework, with an iterative mining strategy to find high-quality candidates.
Experiments on 30+ tasks across 13 task families and multiple data domains show that UDR significantly outperforms baselines.
arXiv Detail & Related papers (2023-05-07T16:07:11Z) - OmDet: Large-scale vision-language multi-dataset pre-training with
multimodal detection network [17.980765138522322]
This work introduces OmDet, a novel language-aware object detection architecture.
Leveraging natural language as a universal knowledge representation, OmDet accumulates a "visual vocabulary" from diverse datasets.
We demonstrate superior performance of OmDet over strong baselines in object detection in the wild, open-vocabulary detection, and phrase grounding.
arXiv Detail & Related papers (2022-09-10T14:25:14Z) - Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of
Language Models [86.02610674750345]
Adversarial GLUE (AdvGLUE) is a new multi-task benchmark to explore and evaluate the vulnerabilities of modern large-scale language models under various types of adversarial attacks.
We apply 14 adversarial attack methods to GLUE tasks to construct AdvGLUE, which is further validated by humans for reliable annotations.
All the language models and robust training methods we tested perform poorly on AdvGLUE, with scores lagging far behind the benign accuracy.
arXiv Detail & Related papers (2021-11-04T12:59:55Z) - Learning What Makes a Difference from Counterfactual Examples and
Gradient Supervision [57.14468881854616]
We propose an auxiliary training objective that improves the generalization capabilities of neural networks.
We use pairs of minimally-different examples with different labels, a.k.a counterfactual or contrasting examples, which provide a signal indicative of the underlying causal structure of the task.
Models trained with this technique demonstrate improved performance on out-of-distribution test sets.
arXiv Detail & Related papers (2020-04-20T02:47:49Z) - Stance Detection Benchmark: How Robust Is Your Stance Detection? [65.91772010586605]
Stance Detection (StD) aims to detect an author's stance towards a certain topic or claim.
We introduce a StD benchmark that learns from ten StD datasets of various domains in a multi-dataset learning setting.
Within this benchmark setup, we are able to present new state-of-the-art results on five of the datasets.
arXiv Detail & Related papers (2020-01-06T13:37:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.