SeedBERT: Recovering Annotator Rating Distributions from an Aggregated
Label
- URL: http://arxiv.org/abs/2211.13196v1
- Date: Wed, 23 Nov 2022 18:35:15 GMT
- Title: SeedBERT: Recovering Annotator Rating Distributions from an Aggregated
Label
- Authors: Aneesha Sampath, Victoria Lin, Louis-Philippe Morency
- Abstract summary: We propose SeedBERT, a method for recovering annotator rating distributions from a single label.
Our human evaluations indicate that SeedBERT's attention mechanism is consistent with human sources of annotator disagreement.
- Score: 43.23903984174963
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Many machine learning tasks -- particularly those in affective computing --
are inherently subjective. When asked to classify facial expressions or to rate
an individual's attractiveness, humans may disagree with one another, and no
single answer may be objectively correct. However, machine learning datasets
commonly have just one "ground truth" label for each sample, so models trained
on these labels may not perform well on tasks that are subjective in nature.
Though allowing models to learn from the individual annotators' ratings may
help, most datasets do not provide annotator-specific labels for each sample.
To address this issue, we propose SeedBERT, a method for recovering annotator
rating distributions from a single label by inducing pre-trained models to
attend to different portions of the input. Our human evaluations indicate that
SeedBERT's attention mechanism is consistent with human sources of annotator
disagreement. Moreover, in our empirical evaluations using large language
models, SeedBERT demonstrates substantial gains in performance on downstream
subjective tasks compared both to standard deep learning models and to other
current models that account explicitly for annotator disagreement.
Related papers
- Capturing Perspectives of Crowdsourced Annotators in Subjective Learning Tasks [9.110872603799839]
Supervised classification heavily depends on datasets annotated by humans.
In subjective tasks such as toxicity classification, these annotations often exhibit low agreement among raters.
In this work, we propose textbfAnnotator Awares for Texts (AART) for subjective classification tasks.
arXiv Detail & Related papers (2023-11-16T10:18:32Z) - ACTOR: Active Learning with Annotator-specific Classification Heads to
Embrace Human Label Variation [35.10805667891489]
Active learning, as an annotation cost-saving strategy, has not been fully explored in the context of learning from disagreement.
We show that in the active learning setting, a multi-head model performs significantly better than a single-head model in terms of uncertainty estimation.
arXiv Detail & Related papers (2023-10-23T14:26:43Z) - Learn What Is Possible, Then Choose What Is Best: Disentangling
One-To-Many Relations in Language Through Text-based Games [3.615981646205045]
We present an approach to train language models that can emulate the desirable behaviours, but not the undesirable ones.
Using text-based games as a testbed, our approach, PASA, uses discrete latent variables to capture the range of different behaviours.
Results show up to 49% empirical improvement over the previous state-of-the-art model.
arXiv Detail & Related papers (2023-04-14T17:11:26Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Debiased Pseudo Labeling in Self-Training [77.83549261035277]
Deep neural networks achieve remarkable performances on a wide range of tasks with the aid of large-scale labeled datasets.
To mitigate the requirement for labeled data, self-training is widely used in both academia and industry by pseudo labeling on readily-available unlabeled data.
We propose Debiased, in which the generation and utilization of pseudo labels are decoupled by two independent heads.
arXiv Detail & Related papers (2022-02-15T02:14:33Z) - Few-shot Instruction Prompts for Pretrained Language Models to Detect
Social Biases [55.45617404586874]
We propose a few-shot instruction-based method for prompting pre-trained language models (LMs)
We show that large LMs can detect different types of fine-grained biases with similar and sometimes superior accuracy to fine-tuned models.
arXiv Detail & Related papers (2021-12-15T04:19:52Z) - End-to-End Annotator Bias Approximation on Crowdsourced Single-Label
Sentiment Analysis [0.4925222726301579]
Sentiment analysis is often a crowdsourcing task prone to subjective labels given by many annotators.
It is not yet fully understood how the annotation bias of each annotator can be modeled correctly with state-of-the-art methods.
Our contribution is an explanation and improvement for precise neural end-to-end bias modeling and ground truth estimation.
arXiv Detail & Related papers (2021-11-03T16:20:16Z) - Towards Good Practices for Efficiently Annotating Large-Scale Image
Classification Datasets [90.61266099147053]
We investigate efficient annotation strategies for collecting multi-class classification labels for a large collection of images.
We propose modifications and best practices aimed at minimizing human labeling effort.
Simulated experiments on a 125k image subset of the ImageNet100 show that it can be annotated to 80% top-1 accuracy with 0.35 annotations per image on average.
arXiv Detail & Related papers (2021-04-26T16:29:32Z) - One-bit Supervision for Image Classification [121.87598671087494]
One-bit supervision is a novel setting of learning from incomplete annotations.
We propose a multi-stage training paradigm which incorporates negative label suppression into an off-the-shelf semi-supervised learning algorithm.
arXiv Detail & Related papers (2020-09-14T03:06:23Z) - GPM: A Generic Probabilistic Model to Recover Annotator's Behavior and
Ground Truth Labeling [34.48095564497967]
We propose a probabilistic graphical annotation model to infer the underlying ground truth and annotator's behavior.
The proposed model is able to identify whether an annotator has worked diligently towards the task during the labeling procedure.
arXiv Detail & Related papers (2020-03-01T12:14:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.