An evaluation of word-level confidence estimation for end-to-end
automatic speech recognition
- URL: http://arxiv.org/abs/2101.05525v1
- Date: Thu, 14 Jan 2021 09:51:59 GMT
- Title: An evaluation of word-level confidence estimation for end-to-end
automatic speech recognition
- Authors: Dan Oneata, Alexandru Caranica, Adriana Stan, Horia Cucu
- Abstract summary: We investigate confidence estimation for end-to-end automatic speech recognition (ASR)
We provide an extensive benchmark of popular confidence methods on four well-known speech datasets.
Our results suggest a strong baseline can be obtained by scaling the logits by a learnt temperature.
- Score: 70.61280174637913
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Quantifying the confidence (or conversely the uncertainty) of a prediction is
a highly desirable trait of an automatic system, as it improves the robustness
and usefulness in downstream tasks. In this paper we investigate confidence
estimation for end-to-end automatic speech recognition (ASR). Previous work has
addressed confidence measures for lattice-based ASR, while current machine
learning research mostly focuses on confidence measures for unstructured deep
learning. However, as the ASR systems are increasingly being built upon deep
end-to-end methods, there is little work that tries to develop confidence
measures in this context. We fill this gap by providing an extensive benchmark
of popular confidence methods on four well-known speech datasets. There are two
challenges we overcome in adapting existing methods: working on structured data
(sequences) and obtaining confidences at a coarser level than the predictions
(words instead of tokens). Our results suggest that a strong baseline can be
obtained by scaling the logits by a learnt temperature, followed by estimating
the confidence as the negative entropy of the predictive distribution and,
finally, sum pooling to aggregate at word level.
Related papers
- Confidence Intervals and Simultaneous Confidence Bands Based on Deep Learning [0.36832029288386137]
We provide a valid non-parametric bootstrap method that correctly disentangles data uncertainty from the noise inherent in the adopted optimization algorithm.
The proposed ad-hoc method can be easily integrated into any deep neural network without interfering with the training process.
arXiv Detail & Related papers (2024-06-20T05:51:37Z) - Revisiting Confidence Estimation: Towards Reliable Failure Prediction [53.79160907725975]
We find a general, widely existing but actually-neglected phenomenon that most confidence estimation methods are harmful for detecting misclassification errors.
We propose to enlarge the confidence gap by finding flat minima, which yields state-of-the-art failure prediction performance.
arXiv Detail & Related papers (2024-03-05T11:44:14Z) - Confidence Estimation Using Unlabeled Data [12.512654188295764]
We propose the first confidence estimation method for a semi-supervised setting, when most training labels are unavailable.
We use training consistency as a surrogate function and propose a consistency ranking loss for confidence estimation.
On both image classification and segmentation tasks, our method achieves state-of-the-art performances in confidence estimation.
arXiv Detail & Related papers (2023-07-19T20:11:30Z) - Fast Entropy-Based Methods of Word-Level Confidence Estimation for
End-To-End Automatic Speech Recognition [86.21889574126878]
We show how per-frame entropy values can be normalized and aggregated to obtain a confidence measure per unit and per word.
We evaluate the proposed confidence measures on LibriSpeech test sets, and show that they are up to 2 and 4 times better than confidence estimation based on the maximum per-frame probability.
arXiv Detail & Related papers (2022-12-16T20:27:40Z) - Learning Uncertainty For Safety-Oriented Semantic Segmentation In
Autonomous Driving [77.39239190539871]
We show how uncertainty estimation can be leveraged to enable safety critical image segmentation in autonomous driving.
We introduce a new uncertainty measure based on disagreeing predictions as measured by a dissimilarity function.
We show experimentally that our proposed approach is much less computationally intensive at inference time than competing methods.
arXiv Detail & Related papers (2021-05-28T09:23:05Z) - Multi-Task Learning for End-to-End ASR Word and Utterance Confidence
with Deletion Prediction [20.00640459241358]
Confidence scores are very useful for downstream applications of automatic speech recognition (ASR) systems.
Recent works have proposed using neural networks to learn word or utterance confidence scores for end-to-end ASR.
This paper proposes to jointly learn word confidence, word deletion, and utterance confidence.
arXiv Detail & Related papers (2021-04-26T20:38:42Z) - CoinDICE: Off-Policy Confidence Interval Estimation [107.86876722777535]
We study high-confidence behavior-agnostic off-policy evaluation in reinforcement learning.
We show in a variety of benchmarks that the confidence interval estimates are tighter and more accurate than existing methods.
arXiv Detail & Related papers (2020-10-22T12:39:11Z) - Confidence Estimation for Attention-based Sequence-to-sequence Models
for Speech Recognition [31.25931550876392]
Confidence scores from a speech recogniser are a useful measure to assess the quality of transcriptions.
We propose a lightweight and effective approach named confidence estimation module (CEM) on top of an existing end-to-end ASR model.
arXiv Detail & Related papers (2020-10-22T04:02:27Z) - Binary Classification from Positive Data with Skewed Confidence [85.18941440826309]
Positive-confidence (Pconf) classification is a promising weakly-supervised learning method.
In practice, the confidence may be skewed by bias arising in an annotation process.
We introduce the parameterized model of the skewed confidence, and propose the method for selecting the hyper parameter.
arXiv Detail & Related papers (2020-01-29T00:04:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.