Self-supervised Regularization for Text Classification
- URL: http://arxiv.org/abs/2103.05231v1
- Date: Tue, 9 Mar 2021 05:35:52 GMT
- Title: Self-supervised Regularization for Text Classification
- Authors: Meng Zhou, Zechen Li, Pengtao Xie
- Abstract summary: In many real-world problems, the number of texts for training classification models is limited, which renders these models prone to overfitting.
We propose SSL-Reg, a data-dependent regularization approach based on self-supervised learning (SSL)
SSL is an unsupervised learning approach which defines auxiliary tasks on input data without using any human-provided labels.
- Score: 14.824073299035675
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Text classification is a widely studied problem and has broad applications.
In many real-world problems, the number of texts for training classification
models is limited, which renders these models prone to overfitting. To address
this problem, we propose SSL-Reg, a data-dependent regularization approach
based on self-supervised learning (SSL). SSL is an unsupervised learning
approach which defines auxiliary tasks on input data without using any
human-provided labels and learns data representations by solving these
auxiliary tasks. In SSL-Reg, a supervised classification task and an
unsupervised SSL task are performed simultaneously. The SSL task is
unsupervised, which is defined purely on input texts without using any
human-provided labels. Training a model using an SSL task can prevent the model
from being overfitted to a limited number of class labels in the classification
task. Experiments on 17 text classification datasets demonstrate the
effectiveness of our proposed method.
Related papers
- OwMatch: Conditional Self-Labeling with Consistency for Open-World Semi-Supervised Learning [4.462726364160216]
Semi-supervised learning (SSL) offers a robust framework for harnessing the potential of unannotated data.
The emergence of open-world SSL (OwSSL) introduces a more practical challenge, wherein unlabeled data may encompass samples from unseen classes.
We propose an effective framework called OwMatch, combining conditional self-labeling and open-world hierarchical thresholding.
arXiv Detail & Related papers (2024-11-04T06:07:43Z) - Semi-Supervised Sparse Gaussian Classification: Provable Benefits of Unlabeled Data [6.812609988733991]
We study SSL for high dimensional Gaussian classification.
We analyze information theoretic lower bounds for accurate feature selection.
We present simulations that complement our theoretical analysis.
arXiv Detail & Related papers (2024-09-05T08:21:05Z) - A Closer Look at Benchmarking Self-Supervised Pre-training with Image Classification [51.35500308126506]
Self-supervised learning (SSL) is a machine learning approach where the data itself provides supervision, eliminating the need for external labels.
We study how classification-based evaluation protocols for SSL correlate and how well they predict downstream performance on different dataset types.
arXiv Detail & Related papers (2024-07-16T23:17:36Z) - Generalized Semi-Supervised Learning via Self-Supervised Feature Adaptation [87.17768598044427]
Traditional semi-supervised learning assumes that the feature distributions of labeled and unlabeled data are consistent.
We propose Self-Supervised Feature Adaptation (SSFA), a generic framework for improving SSL performance when labeled and unlabeled data come from different distributions.
Our proposed SSFA is applicable to various pseudo-label-based SSL learners and significantly improves performance in labeled, unlabeled, and even unseen distributions.
arXiv Detail & Related papers (2024-05-31T03:13:45Z) - Rethinking Semi-supervised Learning with Language Models [33.70349754359132]
Semi-supervised learning (SSL) is a popular setting aiming to effectively utilize unlabelled data to improve model performance.
There are two popular approaches to make use of unlabelled data: Self-training (ST) and Task-adaptive pre-training (TAPT)
arXiv Detail & Related papers (2023-05-22T13:07:35Z) - Complementing Semi-Supervised Learning with Uncertainty Quantification [6.612035830987296]
We propose a novel unsupervised uncertainty-aware objective that relies on aleatoric and epistemic uncertainty quantification.
Our results outperform the state-of-the-art results on complex datasets such as CIFAR-100 and Mini-ImageNet.
arXiv Detail & Related papers (2022-07-22T00:15:02Z) - OpenLDN: Learning to Discover Novel Classes for Open-World
Semi-Supervised Learning [110.40285771431687]
Semi-supervised learning (SSL) is one of the dominant approaches to address the annotation bottleneck of supervised learning.
Recent SSL methods can effectively leverage a large repository of unlabeled data to improve performance while relying on a small set of labeled data.
This work introduces OpenLDN that utilizes a pairwise similarity loss to discover novel classes.
arXiv Detail & Related papers (2022-07-05T18:51:05Z) - DATA: Domain-Aware and Task-Aware Pre-training [94.62676913928831]
We present DATA, a simple yet effective NAS approach specialized for self-supervised learning (SSL)
Our method achieves promising results across a wide range of computation costs on downstream tasks, including image classification, object detection and semantic segmentation.
arXiv Detail & Related papers (2022-03-17T02:38:49Z) - Robust Deep Semi-Supervised Learning: A Brief Introduction [63.09703308309176]
Semi-supervised learning (SSL) aims to improve learning performance by leveraging unlabeled data when labels are insufficient.
SSL with deep models has proven to be successful on standard benchmark tasks.
However, they are still vulnerable to various robustness threats in real-world applications.
arXiv Detail & Related papers (2022-02-12T04:16:41Z) - Self-supervised Learning is More Robust to Dataset Imbalance [65.84339596595383]
We investigate self-supervised learning under dataset imbalance.
Off-the-shelf self-supervised representations are already more robust to class imbalance than supervised representations.
We devise a re-weighted regularization technique that consistently improves the SSL representation quality on imbalanced datasets.
arXiv Detail & Related papers (2021-10-11T06:29:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.