ApproBiVT: Lead ASR Models to Generalize Better Using Approximated
Bias-Variance Tradeoff Guided Early Stopping and Checkpoint Averaging
- URL: http://arxiv.org/abs/2308.02870v1
- Date: Sat, 5 Aug 2023 12:50:54 GMT
- Title: ApproBiVT: Lead ASR Models to Generalize Better Using Approximated
Bias-Variance Tradeoff Guided Early Stopping and Checkpoint Averaging
- Authors: Fangyuan Wang, Ming Hao, Yuhai Shi, Bo Xu
- Abstract summary: We take the training loss and validation loss as proxies of bias and variance and guide the early stopping and checkpoint averaging.
When evaluating with advanced ASR models, our recipe provides 2.5%-3.7% and 3.1%-4.6% CER reduction.
- Score: 7.0626076422397475
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The conventional recipe for Automatic Speech Recognition (ASR) models is to
1) train multiple checkpoints on a training set while relying on a validation
set to prevent overfitting using early stopping and 2) average several last
checkpoints or that of the lowest validation losses to obtain the final model.
In this paper, we rethink and update the early stopping and checkpoint
averaging from the perspective of the bias-variance tradeoff. Theoretically,
the bias and variance represent the fitness and variability of a model and the
tradeoff of them determines the overall generalization error. But, it's
impractical to evaluate them precisely. As an alternative, we take the training
loss and validation loss as proxies of bias and variance and guide the early
stopping and checkpoint averaging using their tradeoff, namely an Approximated
Bias-Variance Tradeoff (ApproBiVT). When evaluating with advanced ASR models,
our recipe provides 2.5%-3.7% and 3.1%-4.6% CER reduction on the AISHELL-1 and
AISHELL-2, respectively.
Related papers
- Uncertainty-Calibrated Test-Time Model Adaptation without Forgetting [55.17761802332469]
Test-time adaptation (TTA) seeks to tackle potential distribution shifts between training and test data by adapting a given model w.r.t. any test sample.
Prior methods perform backpropagation for each test sample, resulting in unbearable optimization costs to many applications.
We propose an Efficient Anti-Forgetting Test-Time Adaptation (EATA) method which develops an active sample selection criterion to identify reliable and non-redundant samples.
arXiv Detail & Related papers (2024-03-18T05:49:45Z) - Selective Learning: Towards Robust Calibration with Dynamic Regularization [79.92633587914659]
Miscalibration in deep learning refers to there is a discrepancy between the predicted confidence and performance.
We introduce Dynamic Regularization (DReg) which aims to learn what should be learned during training thereby circumventing the confidence adjusting trade-off.
arXiv Detail & Related papers (2024-02-13T11:25:20Z) - Generalized Logit Adjustment: Calibrating Fine-tuned Models by Removing Label Bias in Foundation Models [75.9543301303586]
Foundation models like CLIP allow zero-shot transfer on various tasks without additional training data.
Fine-tuning and ensembling are also commonly adopted to better fit the downstream tasks.
However, we argue that prior work has overlooked the inherent biases in foundation models.
arXiv Detail & Related papers (2023-10-12T08:01:11Z) - Beyond calibration: estimating the grouping loss of modern neural
networks [68.8204255655161]
Proper scoring rule theory shows that given the calibration loss, the missing piece to characterize individual errors is the grouping loss.
We show that modern neural network architectures in vision and NLP exhibit grouping loss, notably in distribution shifts settings.
arXiv Detail & Related papers (2022-10-28T07:04:20Z) - Recycling Scraps: Improving Private Learning by Leveraging Intermediate
Checkpoints [17.654346227497403]
This work explores various methods that aggregate intermediate checkpoints to improve the utility of DP training.
We show that checkpoint aggregations provide significant gains in the prediction accuracy over the existing SOTA for CIFAR10 and StackOverflow datasets.
Finally, we show that the sample variance from last few checkpoints provides a good approximation of the variance of the final model of a DP run.
arXiv Detail & Related papers (2022-10-04T19:21:00Z) - Ensembling over Classifiers: a Bias-Variance Perspective [13.006468721874372]
We build upon the extension to the bias-variance decomposition by Pfau (2013) in order to gain crucial insights into the behavior of ensembles of classifiers.
We show that conditional estimates necessarily incur an irreducible error.
Empirically, standard ensembling reducesthe bias, leading us to hypothesize that ensembles of classifiers may perform well in part because of this unexpected reduction.
arXiv Detail & Related papers (2022-06-21T17:46:35Z) - Sample-Efficient Optimisation with Probabilistic Transformer Surrogates [66.98962321504085]
This paper investigates the feasibility of employing state-of-the-art probabilistic transformers in Bayesian optimisation.
We observe two drawbacks stemming from their training procedure and loss definition, hindering their direct deployment as proxies in black-box optimisation.
We introduce two components: 1) a BO-tailored training prior supporting non-uniformly distributed points, and 2) a novel approximate posterior regulariser trading-off accuracy and input sensitivity to filter favourable stationary points for improved predictive performance.
arXiv Detail & Related papers (2022-05-27T11:13:17Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.