Semi-Supervised Models via Data Augmentationfor Classifying Interactive
Affective Responses
- URL: http://arxiv.org/abs/2004.10972v1
- Date: Thu, 23 Apr 2020 05:02:31 GMT
- Title: Semi-Supervised Models via Data Augmentationfor Classifying Interactive
Affective Responses
- Authors: Jiaao Chen, Yuwei Wu, Diyi Yang
- Abstract summary: We present semi-supervised models with data augmentation (SMDA), a semi-supervised text classification system to classify interactive affective responses.
For labeled sentences, we performed data augmentations to uniform the label distributions and computed supervised loss during training process.
For unlabeled sentences, we explored self-training by regarding low-entropy predictions over unlabeled sentences as pseudo labels.
- Score: 85.04362095899656
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present semi-supervised models with data augmentation (SMDA), a
semi-supervised text classification system to classify interactive affective
responses. SMDA utilizes recent transformer-based models to encode each
sentence and employs back translation techniques to paraphrase given sentences
as augmented data. For labeled sentences, we performed data augmentations to
uniform the label distributions and computed supervised loss during training
process. For unlabeled sentences, we explored self-training by regarding
low-entropy predictions over unlabeled sentences as pseudo labels, assuming
high-confidence predictions as labeled data for training. We further introduced
consistency regularization as unsupervised loss after data augmentations on
unlabeled data, based on the assumption that the model should predict similar
class distributions with original unlabeled sentences as input and augmented
sentences as input. Via a set of experiments, we demonstrated that our system
outperformed baseline models in terms of F1-score and accuracy.
Related papers
- SenTest: Evaluating Robustness of Sentence Encoders [0.4194295877935868]
This work focuses on evaluating the robustness of the sentence encoders.
We employ several adversarial attacks to evaluate its robustness.
The results of the experiments strongly undermine the robustness of sentence encoders.
arXiv Detail & Related papers (2023-11-29T15:21:35Z) - Measuring and Improving Attentiveness to Partial Inputs with Counterfactuals [91.59906995214209]
We propose a new evaluation method, Counterfactual Attentiveness Test (CAT)
CAT uses counterfactuals by replacing part of the input with its counterpart from a different example, expecting an attentive model to change its prediction.
We show that GPT3 becomes less attentive with an increased number of demonstrations, while its accuracy on the test data improves.
arXiv Detail & Related papers (2023-11-16T06:27:35Z) - Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot
Text Classification Tasks [75.42002070547267]
We propose a self evolution learning (SE) based mixup approach for data augmentation in text classification.
We introduce a novel instance specific label smoothing approach, which linearly interpolates the model's output and one hot labels of the original samples to generate new soft for label mixing up.
arXiv Detail & Related papers (2023-05-22T23:43:23Z) - CAFA: Class-Aware Feature Alignment for Test-Time Adaptation [50.26963784271912]
Test-time adaptation (TTA) aims to address this challenge by adapting a model to unlabeled data at test time.
We propose a simple yet effective feature alignment loss, termed as Class-Aware Feature Alignment (CAFA), which simultaneously encourages a model to learn target representations in a class-discriminative manner.
arXiv Detail & Related papers (2022-06-01T03:02:07Z) - Discriminative and Generative Transformer-based Models For Situation
Entity Classification [8.029049649310211]
We re-examine the situation entity (SE) classification task with varying amounts of available training data.
We exploit a Transformer-based variational autoencoder to encode sentences into a lower dimensional latent space.
arXiv Detail & Related papers (2021-09-15T17:07:07Z) - Unsupervised neural adaptation model based on optimal transport for
spoken language identification [54.96267179988487]
Due to the mismatch of statistical distributions of acoustic speech between training and testing sets, the performance of spoken language identification (SLID) could be drastically degraded.
We propose an unsupervised neural adaptation model to deal with the distribution mismatch problem for SLID.
arXiv Detail & Related papers (2020-12-24T07:37:19Z) - Robustness to Spurious Correlations in Text Classification via
Automatically Generated Counterfactuals [8.827892752465958]
We propose to train a robust text classifier by augmenting the training data with automatically generated counterfactual data.
We show that the robust classifier makes meaningful and trustworthy predictions by emphasizing causal features and de-emphasizing non-causal features.
arXiv Detail & Related papers (2020-12-18T03:57:32Z) - Unsupervised Data Augmentation with Naive Augmentation and without
Unlabeled Data [40.82826366059613]
Unsupervised Data Augmentation (UDA) is a semi-supervised technique that applies a consistency loss to penalize differences between a model's predictions.
In this paper, we re-examine UDA and demonstrate its efficacy on several sequential tasks.
We find that applying its consistency loss affords meaningful gains without any unlabeled data at all.
arXiv Detail & Related papers (2020-10-22T18:01:51Z) - Self-training Improves Pre-training for Natural Language Understanding [63.78927366363178]
We study self-training as another way to leverage unlabeled data through semi-supervised learning.
We introduce SentAugment, a data augmentation method which computes task-specific query embeddings from labeled data.
Our approach leads to scalable and effective self-training with improvements of up to 2.6% on standard text classification benchmarks.
arXiv Detail & Related papers (2020-10-05T17:52:25Z) - Learning the Prediction Distribution for Semi-Supervised Learning with
Normalising Flows [6.789370732159177]
Impressive results have been achieved in semi-supervised learning (SSL) for image classification, nearing fully supervised performance.
We propose a probabilistically principled general approach to SSL that considers the distribution over label predictions.
We demonstrate the general applicability of this approach on a range of computer vision tasks with varying output complexity.
arXiv Detail & Related papers (2020-07-06T13:36:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.