Residual acoustic echo suppression based on efficient multi-task
convolutional neural network
- URL: http://arxiv.org/abs/2009.13931v2
- Date: Fri, 6 Nov 2020 03:33:36 GMT
- Title: Residual acoustic echo suppression based on efficient multi-task
convolutional neural network
- Authors: Xinquan Zhou, Yanhong Leng
- Abstract summary: We propose a real-time residual acoustic echo suppression (RAES) method using an efficient convolutional neural network.
The training criterion is based on a novel loss function, which we call as the suppression loss, to balance the suppression of residual echo and the distortion of near-end signals.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Acoustic echo degrades the user experience in voice communication systems
thus needs to be suppressed completely. We propose a real-time residual
acoustic echo suppression (RAES) method using an efficient convolutional neural
network. The double talk detector is used as an auxiliary task to improve the
performance of RAES in the context of multi-task learning. The training
criterion is based on a novel loss function, which we call as the suppression
loss, to balance the suppression of residual echo and the distortion of
near-end signals. The experimental results show that the proposed method can
efficiently suppress the residual echo under different circumstances.
Related papers
- End-to-End Binaural Speech Synthesis [71.1869877389535]
We present an end-to-end speech synthesis system that combines a low-bitrate audio system with a powerful decoder.
We demonstrate the capability of the adversarial loss in capturing environment effects needed to create an authentic auditory scene.
arXiv Detail & Related papers (2022-07-08T05:18:36Z) - Audio-visual multi-channel speech separation, dereverberation and
recognition [70.34433820322323]
This paper proposes an audio-visual multi-channel speech separation, dereverberation and recognition approach.
The advantage of the additional visual modality over using audio only is demonstrated on two neural dereverberation approaches.
Experiments conducted on the LRS2 dataset suggest that the proposed audio-visual multi-channel speech separation, dereverberation and recognition system outperforms the baseline.
arXiv Detail & Related papers (2022-04-05T04:16:03Z) - Improving Noise Robustness of Contrastive Speech Representation Learning
with Speech Reconstruction [109.44933866397123]
Noise robustness is essential for deploying automatic speech recognition systems in real-world environments.
We employ a noise-robust representation learned by a refined self-supervised framework for noisy speech recognition.
We achieve comparable performance to the best supervised approach reported with only 16% of labeled data.
arXiv Detail & Related papers (2021-10-28T20:39:02Z) - Removing Noise from Extracellular Neural Recordings Using Fully
Convolutional Denoising Autoencoders [62.997667081978825]
We propose a Fully Convolutional Denoising Autoencoder, which learns to produce a clean neuronal activity signal from a noisy multichannel input.
The experimental results on simulated data show that our proposed method can improve significantly the quality of noise-corrupted neural signals.
arXiv Detail & Related papers (2021-09-18T14:51:24Z) - Deep Residual Echo Suppression with A Tunable Tradeoff Between Signal
Distortion and Echo Suppression [13.558688470594676]
A UNet neural network maps the outputs of a linear acoustic echo canceler to the desired signal in the spectral domain.
The system employs 136 thousand parameters, and requires 1.6 Giga floating-point operations per second and 10 Mega-bytes of memory.
arXiv Detail & Related papers (2021-06-25T09:49:18Z) - On End-to-end Multi-channel Time Domain Speech Separation in Reverberant
Environments [33.79711018198589]
This paper introduces a new method for multi-channel time domain speech separation in reverberant environments.
A fully-convolutional neural network structure has been used to directly separate speech from multiple microphone recordings.
To reduce the influence of reverberation on spatial feature extraction, a dereverberation pre-processing method has been applied.
arXiv Detail & Related papers (2020-11-11T18:25:07Z) - Acoustic Echo Cancellation by Combining Adaptive Digital Filter and
Recurrent Neural Network [11.335343110341354]
A fusion scheme by combining adaptive filter and neural network is proposed for Acoustic Echo Cancellation.
The echo could be reduced in a large scale by adaptive filtering, resulting in little residual echo.
The neural network is elaborately designed and trained for suppressing such residual echo.
arXiv Detail & Related papers (2020-05-19T06:25:52Z) - Nonlinear Residual Echo Suppression Based on Multi-stream Conv-TasNet [22.56178941790508]
We propose a residual echo suppression method based on the modification of fully convolutional time-domain audio separation network (Conv-TasNet)
Both the residual signal of the linear acoustic echo cancellation system, and the output of the adaptive filter are adopted to form multiple streams for the Conv-TasNet.
arXiv Detail & Related papers (2020-05-15T16:41:16Z) - Simultaneous Denoising and Dereverberation Using Deep Embedding Features [64.58693911070228]
We propose a joint training method for simultaneous speech denoising and dereverberation using deep embedding features.
At the denoising stage, the DC network is leveraged to extract noise-free deep embedding features.
At the dereverberation stage, instead of using the unsupervised K-means clustering algorithm, another neural network is utilized to estimate the anechoic speech.
arXiv Detail & Related papers (2020-04-06T06:34:01Z) - ADRN: Attention-based Deep Residual Network for Hyperspectral Image
Denoising [52.01041506447195]
We propose an attention-based deep residual network to learn a mapping from noisy HSI to the clean one.
Experimental results demonstrate that our proposed ADRN scheme outperforms the state-of-the-art methods both in quantitative and visual evaluations.
arXiv Detail & Related papers (2020-03-04T08:36:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.