Semi-Supervised Learning with Data Augmentation for End-to-End ASR
- URL: http://arxiv.org/abs/2007.13876v1
- Date: Mon, 27 Jul 2020 21:24:52 GMT
- Title: Semi-Supervised Learning with Data Augmentation for End-to-End ASR
- Authors: Felix Weninger, Franco Mana, Roberto Gemello, Jes\'us Andr\'es-Ferrer,
Puming Zhan
- Abstract summary: We focus on the consistency regularization principle, which has been successfully applied to image classification tasks.
We present sequence-to-sequence (seq2seq) versions of the FixMatch and Noisy Student algorithms.
- Score: 4.878819328459915
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we apply Semi-Supervised Learning (SSL) along with Data
Augmentation (DA) for improving the accuracy of End-to-End ASR. We focus on the
consistency regularization principle, which has been successfully applied to
image classification tasks, and present sequence-to-sequence (seq2seq) versions
of the FixMatch and Noisy Student algorithms. Specifically, we generate the
pseudo labels for the unlabeled data on-the-fly with a seq2seq model after
perturbing the input features with DA. We also propose soft label variants of
both algorithms to cope with pseudo label errors, showing further performance
improvements. We conduct SSL experiments on a conversational speech data set
with 1.9kh manually transcribed training data, using only 25% of the original
labels (475h labeled data). In the result, the Noisy Student algorithm with
soft labels and consistency regularization achieves 10.4% word error rate (WER)
reduction when adding 475h of unlabeled data, corresponding to a recovery rate
of 92%. Furthermore, when iteratively adding 950h more unlabeled data, our best
SSL performance is within 5% WER increase compared to using the full labeled
training set (recovery rate: 78%).
Related papers
- Learning from Spatio-temporal Correlation for Semi-Supervised LiDAR Semantic Segmentation [17.151511119485246]
Two main issues in low-budget SSLS are the poor-quality pseudo-labels for unlabeled data, and the performance drops.
We propose a proximity-based label estimation, which generates highly accurate pseudo-labels for unlabeled data.
Experimental results demonstrate remarkable performance in low-budget settings.
arXiv Detail & Related papers (2024-10-09T13:57:39Z) - Addressing Long-Tail Noisy Label Learning Problems: a Two-Stage Solution
with Label Refurbishment Considering Label Rarity [13.490974408726323]
We introduce an effective two-stage approach by combining soft-label refurbishing with multi-expert ensemble learning.
In the first stage of robust soft label refurbishing, we acquire unbiased features through contrastive learning.
In the second stage, our label refurbishment method is applied to obtain soft labels for multi-expert ensemble learning.
arXiv Detail & Related papers (2024-03-04T08:06:57Z) - FlatMatch: Bridging Labeled Data and Unlabeled Data with Cross-Sharpness
for Semi-Supervised Learning [73.13448439554497]
Semi-Supervised Learning (SSL) has been an effective way to leverage abundant unlabeled data with extremely scarce labeled data.
Most SSL methods are commonly based on instance-wise consistency between different data transformations.
We propose FlatMatch which minimizes a cross-sharpness measure to ensure consistent learning performance between the two datasets.
arXiv Detail & Related papers (2023-10-25T06:57:59Z) - MarginMatch: Improving Semi-Supervised Learning with Pseudo-Margins [73.17295479535161]
MarginMatch is a new SSL approach combining consistency regularization and pseudo-labeling.
We analyze the behavior of the model on the pseudo-labeled examples as the training progresses to ensure low quality predictions are masked out.
We obtain an improvement in error rate over the state-of-the-art of 3.25% on CIFAR-100 with only 25 labels per class and of 3.78% on STL-10 using as few as 4 labels per class.
arXiv Detail & Related papers (2023-08-17T15:19:04Z) - SoftMatch: Addressing the Quantity-Quality Trade-off in Semi-supervised
Learning [101.86916775218403]
This paper revisits the popular pseudo-labeling methods via a unified sample weighting formulation.
We propose SoftMatch to overcome the trade-off by maintaining both high quantity and high quality of pseudo-labels during training.
In experiments, SoftMatch shows substantial improvements across a wide variety of benchmarks, including image, text, and imbalanced classification.
arXiv Detail & Related papers (2023-01-26T03:53:25Z) - Dense FixMatch: a simple semi-supervised learning method for pixel-wise
prediction tasks [68.36996813591425]
We propose Dense FixMatch, a simple method for online semi-supervised learning of dense and structured prediction tasks.
We enable the application of FixMatch in semi-supervised learning problems beyond image classification by adding a matching operation on the pseudo-labels.
Dense FixMatch significantly improves results compared to supervised learning using only labeled data, approaching its performance with 1/4 of the labeled samples.
arXiv Detail & Related papers (2022-10-18T15:02:51Z) - Class-Aware Contrastive Semi-Supervised Learning [51.205844705156046]
We propose a general method named Class-aware Contrastive Semi-Supervised Learning (CCSSL) to improve pseudo-label quality and enhance the model's robustness in the real-world setting.
Our proposed CCSSL has significant performance improvements over the state-of-the-art SSL methods on the standard datasets CIFAR100 and STL10.
arXiv Detail & Related papers (2022-03-04T12:18:23Z) - AlphaMatch: Improving Consistency for Semi-supervised Learning with
Alpha-divergence [44.88886269629515]
Semi-supervised learning (SSL) is a key approach toward more data-efficient machine learning by jointly leverage both labeled and unlabeled data.
We propose AlphaMatch, an efficient SSL method that leverages data augmentations, by efficiently enforcing the label consistency between the data points and the augmented data derived from them.
arXiv Detail & Related papers (2020-11-23T22:43:45Z) - FixMatch: Simplifying Semi-Supervised Learning with Consistency and
Confidence [93.91751021370638]
Semi-supervised learning (SSL) provides an effective means of leveraging unlabeled data to improve a model's performance.
In this paper, we demonstrate the power of a simple combination of two common SSL methods: consistency regularization and pseudo-labeling.
Our algorithm, FixMatch, first generates pseudo-labels using the model's predictions on weakly-augmented unlabeled images.
arXiv Detail & Related papers (2020-01-21T18:32:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.