Dynamic Layer Customization for Noise Robust Speech Emotion Recognition
in Heterogeneous Condition Training
- URL: http://arxiv.org/abs/2010.11226v1
- Date: Wed, 21 Oct 2020 18:07:32 GMT
- Title: Dynamic Layer Customization for Noise Robust Speech Emotion Recognition
in Heterogeneous Condition Training
- Authors: Alex Wilf, Emily Mower Provost
- Abstract summary: We show that we can improve performance by dynamically routing samples to specialized feature encoders for each noise condition.
We extend these improvements to the multimodal setting by dynamically routing samples to maintain temporal ordering.
- Score: 16.807298318504156
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Robustness to environmental noise is important to creating automatic speech
emotion recognition systems that are deployable in the real world. Prior work
on noise robustness has assumed that systems would not make use of
sample-by-sample training noise conditions, or that they would have access to
unlabelled testing data to generalize across noise conditions. We avoid these
assumptions and introduce the resulting task as heterogeneous condition
training. We show that with full knowledge of the test noise conditions, we can
improve performance by dynamically routing samples to specialized feature
encoders for each noise condition, and with partial knowledge, we can use known
noise conditions and domain adaptation algorithms to train systems that
generalize well to unseen noise conditions. We then extend these improvements
to the multimodal setting by dynamically routing samples to maintain temporal
ordering, resulting in significant improvements over approaches that do not
specialize or generalize based on noise type.
Related papers
- Effective Noise-aware Data Simulation for Domain-adaptive Speech Enhancement Leveraging Dynamic Stochastic Perturbation [25.410770364140856]
Cross-domain speech enhancement (SE) is often faced with severe challenges due to the scarcity of noise and background information in an unseen target domain.
This study puts forward a novel data simulation method to address this issue, leveraging noise-extractive techniques and generative adversarial networks (GANs)
We introduce the notion of dynamic perturbation, which can inject controlled perturbations into the noise embeddings during inference.
arXiv Detail & Related papers (2024-09-03T02:29:01Z) - Noise-BERT: A Unified Perturbation-Robust Framework with Noise Alignment
Pre-training for Noisy Slot Filling Task [14.707646721729228]
In a realistic dialogue system, the input information from users is often subject to various types of input perturbations.
We propose Noise-BERT, a unified Perturbation-Robust Framework with Noise Alignment Pre-training.
Our framework incorporates two Noise Alignment Pre-training tasks: Slot Masked Prediction and Sentence Noisiness Discrimination.
arXiv Detail & Related papers (2024-02-22T12:39:50Z) - Understanding the Effect of Noise in LLM Training Data with Algorithmic
Chains of Thought [0.0]
We study how noise in chain of thought impacts task performance in highly-controlled setting.
We define two types of noise: textitstatic noise, a local form of noise which is applied after the CoT trace is computed, and textitdynamic noise, a global form of noise which propagates errors in the trace as it is computed.
We find fine-tuned models are extremely robust to high levels of static noise but struggle significantly more with lower levels of dynamic noise.
arXiv Detail & Related papers (2024-02-06T13:59:56Z) - Robustness Enhancement in Neural Networks with Alpha-Stable Training
Noise [0.0]
We explore the possibility of stronger robustness for non-Gaussian impulsive noise, specifically alpha-stable noise.
By comparing the testing accuracy of models trained with Gaussian noise and alpha-stable noise on data corrupted by different noise, we find that training with alpha-stable noise is more effective than Gaussian noise.
We propose a novel data augmentation method that replaces Gaussian noise, which is typically added to the training data, with alpha-stable noise.
arXiv Detail & Related papers (2023-11-17T10:00:47Z) - DiffSED: Sound Event Detection with Denoising Diffusion [70.18051526555512]
We reformulate the SED problem by taking a generative learning perspective.
Specifically, we aim to generate sound temporal boundaries from noisy proposals in a denoising diffusion process.
During training, our model learns to reverse the noising process by converting noisy latent queries to the groundtruth versions.
arXiv Detail & Related papers (2023-08-14T17:29:41Z) - Explainable Artificial Intelligence driven mask design for
self-supervised seismic denoising [0.0]
Self-supervised coherent noise suppression methods require extensive knowledge of the noise statistics.
We propose the use of explainable artificial intelligence approaches to see inside the black box that is the denoising network.
We show that a simple averaging of the Jacobian contributions over a number of randomly selected input pixels, provides an indication of the most effective mask.
arXiv Detail & Related papers (2023-07-13T11:02:55Z) - Realistic Noise Synthesis with Diffusion Models [68.48859665320828]
Deep image denoising models often rely on large amount of training data for the high quality performance.
We propose a novel method that synthesizes realistic noise using diffusion models, namely Realistic Noise Synthesize Diffusor (RNSD)
RNSD can incorporate guided multiscale content, such as more realistic noise with spatial correlations can be generated at multiple frequencies.
arXiv Detail & Related papers (2023-05-23T12:56:01Z) - Improve Noise Tolerance of Robust Loss via Noise-Awareness [60.34670515595074]
We propose a meta-learning method which is capable of adaptively learning a hyper parameter prediction function, called Noise-Aware-Robust-Loss-Adjuster (NARL-Adjuster for brevity)
Four SOTA robust loss functions are attempted to be integrated with our algorithm, and comprehensive experiments substantiate the general availability and effectiveness of the proposed method in both its noise tolerance and performance.
arXiv Detail & Related papers (2023-01-18T04:54:58Z) - Sources of Noise in Dialogue and How to Deal with Them [63.02707014103651]
Training dialogue systems often entails dealing with noisy training examples and unexpected user inputs.
Despite their prevalence, there currently lacks an accurate survey of dialogue noise.
This paper addresses this gap by first constructing a taxonomy of noise encountered by dialogue systems.
arXiv Detail & Related papers (2022-12-06T04:36:32Z) - C2N: Practical Generative Noise Modeling for Real-World Denoising [53.96391787869974]
We introduce a Clean-to-Noisy image generation framework, namely C2N, to imitate complex real-world noise without using paired examples.
We construct the noise generator in C2N accordingly with each component of real-world noise characteristics to express a wide range of noise accurately.
arXiv Detail & Related papers (2022-02-19T05:53:46Z) - Removing Noise from Extracellular Neural Recordings Using Fully
Convolutional Denoising Autoencoders [62.997667081978825]
We propose a Fully Convolutional Denoising Autoencoder, which learns to produce a clean neuronal activity signal from a noisy multichannel input.
The experimental results on simulated data show that our proposed method can improve significantly the quality of noise-corrupted neural signals.
arXiv Detail & Related papers (2021-09-18T14:51:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.