Robustifying Sentiment Classification by Maximally Exploiting Few
Counterfactuals
- URL: http://arxiv.org/abs/2210.11805v1
- Date: Fri, 21 Oct 2022 08:30:09 GMT
- Title: Robustifying Sentiment Classification by Maximally Exploiting Few
Counterfactuals
- Authors: Maarten De Raedt, Fr\'ederic Godin, Chris Develder, Thomas Demeester
- Abstract summary: We propose a novel solution that only requires annotation of a small fraction of the original training data.
We achieve noticeable accuracy improvements by adding only 1% manual counterfactuals.
- Score: 16.731183915325584
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: For text classification tasks, finetuned language models perform remarkably
well. Yet, they tend to rely on spurious patterns in training data, thus
limiting their performance on out-of-distribution (OOD) test data. Among recent
models aiming to avoid this spurious pattern problem, adding extra
counterfactual samples to the training data has proven to be very effective.
Yet, counterfactual data generation is costly since it relies on human
annotation. Thus, we propose a novel solution that only requires annotation of
a small fraction (e.g., 1%) of the original training data, and uses automatic
generation of extra counterfactuals in an encoding vector space. We demonstrate
the effectiveness of our approach in sentiment classification, using IMDb data
for training and other sets for OOD tests (i.e., Amazon, SemEval and Yelp). We
achieve noticeable accuracy improvements by adding only 1% manual
counterfactuals: +3% compared to adding +100% in-distribution training samples,
+1.3% compared to alternate counterfactual approaches.
Related papers
- Adaptive Pre-training Data Detection for Large Language Models via Surprising Tokens [1.2549198550400134]
Large language models (LLMs) are extensively used, but there are concerns regarding privacy, security, and copyright due to their opaque training data.
Current solutions to this problem leverage techniques explored in machine learning privacy such as Membership Inference Attacks (MIAs)
We propose an adaptive pre-training data detection method which alleviates this reliance and effectively amplify the identification.
arXiv Detail & Related papers (2024-07-30T23:43:59Z) - WC-SBERT: Zero-Shot Text Classification via SBERT with Self-Training for
Wikipedia Categories [5.652290685410878]
Our research focuses on solving the zero-shot text classification problem in NLP.
We propose a novel self-training strategy that uses labels rather than text for training.
Our method achieves state-of-the-art results on both the Yahoo Topic and AG News datasets.
arXiv Detail & Related papers (2023-07-28T04:17:41Z) - Quantifying Overfitting: Evaluating Neural Network Performance through
Analysis of Null Space [10.698553177585973]
We analyze the null space in the last layer of neural networks to quantify overfitting without access to training data or knowledge of the accuracy of those data.
Our work represents the first attempt to quantify overfitting without access to training data or knowing any knowledge about the training samples.
arXiv Detail & Related papers (2023-05-30T21:31:24Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Effective Robustness against Natural Distribution Shifts for Models with
Different Training Data [113.21868839569]
"Effective robustness" measures the extra out-of-distribution robustness beyond what can be predicted from the in-distribution (ID) performance.
We propose a new evaluation metric to evaluate and compare the effective robustness of models trained on different data.
arXiv Detail & Related papers (2023-02-02T19:28:41Z) - A Data-Centric Approach for Improving Adversarial Training Through the
Lens of Out-of-Distribution Detection [0.4893345190925178]
We propose detecting and removing hard samples directly from the training procedure rather than applying complicated algorithms to mitigate their effects.
Our results on SVHN and CIFAR-10 datasets show the effectiveness of this method in improving the adversarial training without adding too much computational cost.
arXiv Detail & Related papers (2023-01-25T08:13:50Z) - CellTranspose: Few-shot Domain Adaptation for Cellular Instance
Segmentation [4.38301148531795]
We address the problem of designing an approach that requires minimal amounts of new annotated data as well as training time.
We do so by designing specialized contrastive losses that leverage the few annotated samples very efficiently.
arXiv Detail & Related papers (2022-12-28T23:00:50Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Agree to Disagree: Diversity through Disagreement for Better
Transferability [54.308327969778155]
We propose D-BAT (Diversity-By-disAgreement Training), which enforces agreement among the models on the training data.
We show how D-BAT naturally emerges from the notion of generalized discrepancy.
arXiv Detail & Related papers (2022-02-09T12:03:02Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Uncertainty-aware Self-training for Text Classification with Few Labels [54.13279574908808]
We study self-training as one of the earliest semi-supervised learning approaches to reduce the annotation bottleneck.
We propose an approach to improve self-training by incorporating uncertainty estimates of the underlying neural network.
We show our methods leveraging only 20-30 labeled samples per class for each task for training and for validation can perform within 3% of fully supervised pre-trained language models.
arXiv Detail & Related papers (2020-06-27T08:13:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.