SERIL: Noise Adaptive Speech Enhancement using Regularization-based
Incremental Learning
- URL: http://arxiv.org/abs/2005.11760v2
- Date: Thu, 17 Sep 2020 20:08:40 GMT
- Title: SERIL: Noise Adaptive Speech Enhancement using Regularization-based
Incremental Learning
- Authors: Chi-Chang Lee, Yu-Chen Lin, Hsuan-Tien Lin, Hsin-Min Wang, Yu Tsao
- Abstract summary: Adaptation to a new environment may lead to catastrophic forgetting of the previously learned environments.
In this paper, we propose a regularization-based incremental learning SE (SERIL) strategy.
With a regularization constraint, the parameters are updated to the new noise environment while retaining the knowledge of the previous noise environments.
- Score: 36.24803486242198
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Numerous noise adaptation techniques have been proposed to fine-tune
deep-learning models in speech enhancement (SE) for mismatched noise
environments. Nevertheless, adaptation to a new environment may lead to
catastrophic forgetting of the previously learned environments. The
catastrophic forgetting issue degrades the performance of SE in real-world
embedded devices, which often revisit previous noise environments. The nature
of embedded devices does not allow solving the issue with additional storage of
all pre-trained models or earlier training data. In this paper, we propose a
regularization-based incremental learning SE (SERIL) strategy, complementing
existing noise adaptation strategies without using additional storage. With a
regularization constraint, the parameters are updated to the new noise
environment while retaining the knowledge of the previous noise environments.
The experimental results show that, when faced with a new noise domain, the
SERIL model outperforms the unadapted SE model. Meanwhile, compared with the
current adaptive technique based on fine-tuning, the SERIL model can reduce the
forgetting of previous noise environments by 52%. The results verify that the
SERIL model can effectively adjust itself to new noise environments while
overcoming the catastrophic forgetting issue. The results make SERIL a
favorable choice for real-world SE applications, where the noise environment
changes frequently.
Related papers
- Beyond Image Prior: Embedding Noise Prior into Conditional Denoising Transformer [17.430622649002427]
Existing learning-based denoising methods typically train models to generalize the image prior from large-scale datasets.
We propose a new perspective on the denoising challenge by highlighting the distinct separation between noise and image priors.
We introduce a Locally Noise Prior Estimation algorithm, which accurately estimates the noise prior directly from a single raw noisy image.
arXiv Detail & Related papers (2024-07-12T08:43:11Z) - Learning with Noisy Foundation Models [95.50968225050012]
This paper is the first work to comprehensively understand and analyze the nature of noise in pre-training datasets.
We propose a tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise and improve generalization.
arXiv Detail & Related papers (2024-03-11T16:22:41Z) - Understanding and Mitigating the Label Noise in Pre-training on
Downstream Tasks [91.15120211190519]
This paper aims to understand the nature of noise in pre-training datasets and to mitigate its impact on downstream tasks.
We propose a light-weight black-box tuning method (NMTune) to affine the feature space to mitigate the malignant effect of noise.
arXiv Detail & Related papers (2023-09-29T06:18:15Z) - Latent Class-Conditional Noise Model [54.56899309997246]
We introduce a Latent Class-Conditional Noise model (LCCN) to parameterize the noise transition under a Bayesian framework.
We then deduce a dynamic label regression method for LCCN, whose Gibbs sampler allows us efficiently infer the latent true labels.
Our approach safeguards the stable update of the noise transition, which avoids previous arbitrarily tuning from a mini-batch of samples.
arXiv Detail & Related papers (2023-02-19T15:24:37Z) - NLIP: Noise-robust Language-Image Pre-training [95.13287735264937]
We propose a principled Noise-robust Language-Image Pre-training framework (NLIP) to stabilize pre-training via two schemes: noise-harmonization and noise-completion.
Our NLIP can alleviate the common noise effects during image-text pre-training in a more efficient way.
arXiv Detail & Related papers (2022-12-14T08:19:30Z) - Identifying Hard Noise in Long-Tailed Sample Distribution [76.16113794808001]
We introduce Noisy Long-Tailed Classification (NLT)
Most de-noising methods fail to identify the hard noises.
We design an iterative noisy learning framework called Hard-to-Easy (H2E)
arXiv Detail & Related papers (2022-07-27T09:03:03Z) - NASTAR: Noise Adaptive Speech Enhancement with Target-Conditional
Resampling [34.565077865854484]
We propose noise adaptive speech enhancement with target-conditional resampling (NASTAR)
NASTAR uses a feedback mechanism to simulate adaptive training data via a noise extractor and a retrieval model.
Experimental results show that NASTAR can effectively use one noisy speech sample to adapt an SE model to a target condition.
arXiv Detail & Related papers (2022-06-18T00:15:48Z) - Variational Autoencoder for Speech Enhancement with a Noise-Aware
Encoder [30.318947721658862]
We propose to include noise information in the training phase by using a noise-aware encoder trained on noisy-clean speech pairs.
We show that our proposed noise-aware VAE outperforms the standard VAE in terms of overall distortion without increasing the number of model parameters.
arXiv Detail & Related papers (2021-02-17T11:40:42Z) - Dynamic Layer Customization for Noise Robust Speech Emotion Recognition
in Heterogeneous Condition Training [16.807298318504156]
We show that we can improve performance by dynamically routing samples to specialized feature encoders for each noise condition.
We extend these improvements to the multimodal setting by dynamically routing samples to maintain temporal ordering.
arXiv Detail & Related papers (2020-10-21T18:07:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.