Integrating Statistical Uncertainty into Neural Network-Based Speech
Enhancement
- URL: http://arxiv.org/abs/2203.02288v1
- Date: Fri, 4 Mar 2022 12:55:46 GMT
- Title: Integrating Statistical Uncertainty into Neural Network-Based Speech
Enhancement
- Authors: Huajian Fang, Tal Peer, Stefan Wermter, Timo Gerkmann
- Abstract summary: We study the benefits of modeling uncertainty in neural network-based speech enhancement.
By estimating the distribution instead of the point estimate, one can model the uncertainty associated with each estimate.
- Score: 27.868722093985006
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Speech enhancement in the time-frequency domain is often performed by
estimating a multiplicative mask to extract clean speech. However, most neural
network-based methods perform point estimation, i.e., their output consists of
a single mask. In this paper, we study the benefits of modeling uncertainty in
neural network-based speech enhancement. For this, our neural network is
trained to map a noisy spectrogram to the Wiener filter and its associated
variance, which quantifies uncertainty, based on the maximum a posteriori (MAP)
inference of spectral coefficients. By estimating the distribution instead of
the point estimate, one can model the uncertainty associated with each
estimate. We further propose to use the estimated Wiener filter and its
uncertainty to build an approximate MAP (A-MAP) estimator of spectral
magnitudes, which in turn is combined with the MAP inference of spectral
coefficients to form a hybrid loss function to jointly reinforce the
estimation. Experimental results on different datasets show that the proposed
method can not only capture the uncertainty associated with the estimated
filters, but also yield a higher enhancement performance over comparable models
that do not take uncertainty into account.
Related papers
- Robust Non-parametric Knowledge-based Diffusion Least Mean Squares over
Adaptive Networks [12.266804067030455]
The proposed algorithm leads to a robust estimation of an unknown parameter vector in a group of cooperative estimators.
Results show the robustness of the proposed algorithm in the presence of different noise types.
arXiv Detail & Related papers (2023-12-03T06:18:59Z) - Integrating Uncertainty into Neural Network-based Speech Enhancement [27.868722093985006]
Supervised masking approaches in the time-frequency domain aim to employ deep neural networks to estimate a multiplicative mask to extract clean speech.
This leads to a single estimate for each input without any guarantees or measures of reliability.
We study the benefits of modeling uncertainty in clean speech estimation.
arXiv Detail & Related papers (2023-05-15T15:55:12Z) - Uncertainty Estimation in Deep Speech Enhancement Using Complex Gaussian
Mixture Models [19.442685015494316]
Single-channel deep speech enhancement approaches often estimate a single multiplicative mask to extract clean speech without a measure of its accuracy.
We propose to quantify the uncertainty associated with clean speech estimates in neural network-based speech enhancement.
arXiv Detail & Related papers (2022-12-09T13:03:09Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Dense Uncertainty Estimation [62.23555922631451]
In this paper, we investigate neural networks and uncertainty estimation techniques to achieve both accurate deterministic prediction and reliable uncertainty estimation.
We work on two types of uncertainty estimations solutions, namely ensemble based methods and generative model based methods, and explain their pros and cons while using them in fully/semi/weakly-supervised framework.
arXiv Detail & Related papers (2021-10-13T01:23:48Z) - Sampling-free Variational Inference for Neural Networks with
Multiplicative Activation Noise [51.080620762639434]
We propose a more efficient parameterization of the posterior approximation for sampling-free variational inference.
Our approach yields competitive results for standard regression problems and scales well to large-scale image classification tasks.
arXiv Detail & Related papers (2021-03-15T16:16:18Z) - Uncertainty Intervals for Graph-based Spatio-Temporal Traffic Prediction [0.0]
We propose a Spatio-Temporal neural network that is trained to estimate a density given the measurements of previous timesteps, conditioned on a quantile.
Our method of density estimation is fully parameterised by our neural network and does not use a likelihood approximation internally.
This approach produces uncertainty estimates without the need to sample during inference, such as in Monte Carlo Dropout.
arXiv Detail & Related papers (2020-12-09T18:02:26Z) - Unlabelled Data Improves Bayesian Uncertainty Calibration under
Covariate Shift [100.52588638477862]
We develop an approximate Bayesian inference scheme based on posterior regularisation.
We demonstrate the utility of our method in the context of transferring prognostic models of prostate cancer across globally diverse populations.
arXiv Detail & Related papers (2020-06-26T13:50:19Z) - Nonparametric Score Estimators [49.42469547970041]
Estimating the score from a set of samples generated by an unknown distribution is a fundamental task in inference and learning of probabilistic models.
We provide a unifying view of these estimators under the framework of regularized nonparametric regression.
We propose score estimators based on iterative regularization that enjoy computational benefits from curl-free kernels and fast convergence.
arXiv Detail & Related papers (2020-05-20T15:01:03Z) - Uncertainty Estimation Using a Single Deep Deterministic Neural Network [66.26231423824089]
We propose a method for training a deterministic deep model that can find and reject out of distribution data points at test time with a single forward pass.
We scale training in these with a novel loss function and centroid updating scheme and match the accuracy of softmax models.
arXiv Detail & Related papers (2020-03-04T12:27:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.