Neural Contrastive Clustering: Fully Unsupervised Bias Reduction for
Sentiment Classification
- URL: http://arxiv.org/abs/2204.10467v1
- Date: Fri, 22 Apr 2022 02:34:41 GMT
- Title: Neural Contrastive Clustering: Fully Unsupervised Bias Reduction for
Sentiment Classification
- Authors: Jared Mowery
- Abstract summary: Correlation bias in sentiment classification often arises in conversations about controversial topics.
This study uses adversarial learning to contrast clusters based on sentiment classification labels, with clusters produced by unsupervised topic modeling.
This discourages the neural network from learning topic-related features that produce biased classification results.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Background: Neural networks produce biased classification results due to
correlation bias (they learn correlations between their inputs and outputs to
classify samples, even when those correlations do not represent
cause-and-effect relationships).
Objective: This study introduces a fully unsupervised method of mitigating
correlation bias, demonstrated with sentiment classification on COVID-19 social
media data.
Methods: Correlation bias in sentiment classification often arises in
conversations about controversial topics. Therefore, this study uses
adversarial learning to contrast clusters based on sentiment classification
labels, with clusters produced by unsupervised topic modeling. This discourages
the neural network from learning topic-related features that produce biased
classification results.
Results: Compared to a baseline classifier, neural contrastive clustering
approximately doubles accuracy on bias-prone sentences for human-labeled
COVID-19 social media data, without adversely affecting the classifier's
overall F1 score. Despite being a fully unsupervised approach, neural
contrastive clustering achieves a larger improvement in accuracy on bias-prone
sentences than a supervised masking approach.
Conclusions: Neural contrastive clustering reduces correlation bias in
sentiment text classification. Further research is needed to explore
generalizing this technique to other neural network architectures and
application domains.
Related papers
- Causality and Independence Enhancement for Biased Node Classification [56.38828085943763]
We propose a novel Causality and Independence Enhancement (CIE) framework, applicable to various graph neural networks (GNNs)
Our approach estimates causal and spurious features at the node representation level and mitigates the influence of spurious correlations.
Our approach CIE not only significantly enhances the performance of GNNs but outperforms state-of-the-art debiased node classification methods.
arXiv Detail & Related papers (2023-10-14T13:56:24Z) - Debiasing Counterfactuals In the Presence of Spurious Correlations [0.98342301244574]
We introduce the first end-to-end training framework that integrates both (i) popular debiasing classifiers and (ii) counterfactual image generation.
We demonstrate that the debiasing method: learns generalizable markers across the population, and (ii) successfully ignores spurious correlations and focuses on the underlying disease pathology.
arXiv Detail & Related papers (2023-08-21T19:01:45Z) - How does Contrastive Learning Organize Images? [8.077578967149561]
Contrastive learning, a dominant self-supervised technique, emphasizes similarity in representations between augmentations of the same input and dissimilarity for different ones.
Recent studies challenge this direct relationship, spotlighting the crucial role of inductive biases.
We introduce the "RLD (Relative Local Density)" metric to capture this discrepancy.
arXiv Detail & Related papers (2023-05-17T14:10:54Z) - Self-supervised debiasing using low rank regularization [59.84695042540525]
Spurious correlations can cause strong biases in deep neural networks, impairing generalization ability.
We propose a self-supervised debiasing framework potentially compatible with unlabeled samples.
Remarkably, the proposed debiasing framework significantly improves the generalization performance of self-supervised learning baselines.
arXiv Detail & Related papers (2022-10-11T08:26:19Z) - Contrastive Clustering: Toward Unsupervised Bias Reduction for Emotion
and Sentiment Classification [0.0]
This study assesses the impact of bias on COVID-19 topics.
It demonstrates an automatic algorithm for reducing bias when applied to COVID-19 social media texts.
arXiv Detail & Related papers (2021-11-14T20:58:04Z) - LOGAN: Local Group Bias Detection by Clustering [86.38331353310114]
We argue that evaluating bias at the corpus level is not enough for understanding how biases are embedded in a model.
We propose LOGAN, a new bias detection technique based on clustering.
Experiments on toxicity classification and object classification tasks show that LOGAN identifies bias in a local region.
arXiv Detail & Related papers (2020-10-06T16:42:51Z) - Identifying Spurious Correlations for Robust Text Classification [9.457737910527829]
We propose a method to distinguish spurious and genuine correlations in text classification.
We use features derived from treatment effect estimators to distinguish spurious correlations from "genuine" ones.
Experiments on four datasets suggest that using this approach to inform feature selection also leads to more robust classification.
arXiv Detail & Related papers (2020-10-06T03:49:22Z) - Learning from Failure: Training Debiased Classifier from Biased
Classifier [76.52804102765931]
We show that neural networks learn to rely on spurious correlation only when it is "easier" to learn than the desired knowledge.
We propose a failure-based debiasing scheme by training a pair of neural networks simultaneously.
Our method significantly improves the training of the network against various types of biases in both synthetic and real-world datasets.
arXiv Detail & Related papers (2020-07-06T07:20:29Z) - Decorrelated Clustering with Data Selection Bias [55.91842043124102]
We propose a novel Decorrelation regularized K-Means algorithm (DCKM) for clustering with data selection bias.
Our DCKM algorithm achieves significant performance gains, indicating the necessity of removing unexpected feature correlations induced by selection bias.
arXiv Detail & Related papers (2020-06-29T08:55:50Z) - Learning from Aggregate Observations [82.44304647051243]
We study the problem of learning from aggregate observations where supervision signals are given to sets of instances.
We present a general probabilistic framework that accommodates a variety of aggregate observations.
Simple maximum likelihood solutions can be applied to various differentiable models.
arXiv Detail & Related papers (2020-04-14T06:18:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.