A DNN based Normalized Time-frequency Weighted Criterion for Robust
Wideband DoA Estimation
- URL: http://arxiv.org/abs/2302.10147v1
- Date: Mon, 20 Feb 2023 18:26:52 GMT
- Title: A DNN based Normalized Time-frequency Weighted Criterion for Robust
Wideband DoA Estimation
- Authors: Kuan-Lin Chen and Ching-Hua Lee and Bhaskar D. Rao and Harinath
Garudadri
- Abstract summary: We propose a normalized time-frequency weighted criterion which minimizes the distance between the candidate steering vectors and the filtered snapshots in the T-F domain.
Our method requires no eigendecomposition and uses a simple normalization to prevent the optimization objective from being misled by noisy snapshots.
Experiments show that the proposed method outperforms popular DNN based DoA estimation methods including widely used subspace methods in noisy and reverberant environments.
- Score: 24.175086158375464
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep neural networks (DNNs) have greatly benefited direction of arrival (DoA)
estimation methods for speech source localization in noisy environments.
However, their localization accuracy is still far from satisfactory due to the
vulnerability to nonspeech interference. To improve the robustness against
interference, we propose a DNN based normalized time-frequency (T-F) weighted
criterion which minimizes the distance between the candidate steering vectors
and the filtered snapshots in the T-F domain. Our method requires no
eigendecomposition and uses a simple normalization to prevent the optimization
objective from being misled by noisy filtered snapshots. We also study
different designs of T-F weights guided by a DNN. We find that duplicating the
Hadamard product of speech ratio masks is highly effective and better than
other techniques such as direct masking and taking the mean in the proposed
approach. However, the best-performing design of T-F weights is
criterion-dependent in general. Experiments show that the proposed method
outperforms popular DNN based DoA estimation methods including widely used
subspace methods in noisy and reverberant environments.
Related papers
- Run-Time Adaptation of Neural Beamforming for Robust Speech Dereverberation and Denoising [15.152748065111194]
This paper describes speech enhancement for realtime automatic speech recognition in real environments.
It estimates the masks of clean dry speech from a noisy echoic mixture spectrogram with a deep neural network (DNN) and then computes a enhancement filter used for beamforming.
The performance of such a supervised approach, however, is drastically degraded under mismatched conditions.
arXiv Detail & Related papers (2024-10-30T08:32:47Z) - Stable Neighbor Denoising for Source-free Domain Adaptive Segmentation [91.83820250747935]
Pseudo-label noise is mainly contained in unstable samples in which predictions of most pixels undergo significant variations during self-training.
We introduce the Stable Neighbor Denoising (SND) approach, which effectively discovers highly correlated stable and unstable samples.
SND consistently outperforms state-of-the-art methods in various SFUDA semantic segmentation settings.
arXiv Detail & Related papers (2024-06-10T21:44:52Z) - Domain Generalization Guided by Gradient Signal to Noise Ratio of
Parameters [69.24377241408851]
Overfitting to the source domain is a common issue in gradient-based training of deep neural networks.
We propose to base the selection on gradient-signal-to-noise ratio (GSNR) of network's parameters.
arXiv Detail & Related papers (2023-10-11T10:21:34Z) - One-Dimensional Deep Image Prior for Curve Fitting of S-Parameters from
Electromagnetic Solvers [57.441926088870325]
Deep Image Prior (DIP) is a technique that optimized the weights of a randomly-d convolutional neural network to fit a signal from noisy or under-determined measurements.
Relative to publicly available implementations of Vector Fitting (VF), our method shows superior performance on nearly all test examples.
arXiv Detail & Related papers (2023-06-06T20:28:37Z) - Noise Injection Node Regularization for Robust Learning [0.0]
Noise Injection Node Regularization (NINR) is a method of injecting structured noise into Deep Neural Networks (DNN) during the training stage, resulting in an emergent regularizing effect.
We present theoretical and empirical evidence for substantial improvement in robustness against various test data perturbations for feed-forward DNNs when trained under NINR.
arXiv Detail & Related papers (2022-10-27T20:51:15Z) - Mixture GAN For Modulation Classification Resiliency Against Adversarial
Attacks [55.92475932732775]
We propose a novel generative adversarial network (GAN)-based countermeasure approach.
GAN-based aims to eliminate the adversarial attack examples before feeding to the DNN-based classifier.
Simulation results show the effectiveness of our proposed defense GAN so that it could enhance the accuracy of the DNN-based AMC under adversarial attacks to 81%, approximately.
arXiv Detail & Related papers (2022-05-29T22:30:32Z) - Neural Network-augmented Kalman Filtering for Robust Online Speech
Dereverberation in Noisy Reverberant Environments [13.49645012479288]
A neural network-augmented algorithm for noise-robust online dereverberation is proposed.
The presented framework allows for robust dereverberation on a single-channel noisy reverberant dataset.
arXiv Detail & Related papers (2022-04-06T11:38:04Z) - Time-Domain Mapping Based Single-Channel Speech Separation With
Hierarchical Constraint Training [10.883458728718047]
Single-channel speech separation is required for multi-speaker speech recognition.
Recent deep learning-based approaches focused on time-domain audio separation net (TasNet)
We introduce attention augmented DPRNN (AttnAugDPRNN) which directly approximates the clean sources from the mixture for speech separation.
arXiv Detail & Related papers (2021-10-20T14:42:50Z) - Using Kalman Filter The Right Way: Noise Estimation Is Not Optimal [46.556605821252276]
We show that even a seemingly small violation of KF assumptions can significantly modify the effective noise.
We suggest a method to apply gradient-based optimization efficiently to the symmetric and positive-definite (SPD) parameters of KF.
arXiv Detail & Related papers (2021-04-06T08:59:15Z) - Positive-Negative Momentum: Manipulating Stochastic Gradient Noise to
Improve Generalization [89.7882166459412]
gradient noise (SGN) acts as implicit regularization for deep learning.
Some works attempted to artificially simulate SGN by injecting random noise to improve deep learning.
For simulating SGN at low computational costs and without changing the learning rate or batch size, we propose the Positive-Negative Momentum (PNM) approach.
arXiv Detail & Related papers (2021-03-31T16:08:06Z) - Deep Networks for Direction-of-Arrival Estimation in Low SNR [89.45026632977456]
We introduce a Convolutional Neural Network (CNN) that is trained from mutli-channel data of the true array manifold matrix.
We train a CNN in the low-SNR regime to predict DoAs across all SNRs.
Our robust solution can be applied in several fields, ranging from wireless array sensors to acoustic microphones or sonars.
arXiv Detail & Related papers (2020-11-17T12:52:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.