Rethinking complex-valued deep neural networks for monaural speech
enhancement
- URL: http://arxiv.org/abs/2301.04320v1
- Date: Wed, 11 Jan 2023 05:59:50 GMT
- Title: Rethinking complex-valued deep neural networks for monaural speech
enhancement
- Authors: Haibin Wu, Ke Tan, Buye Xu, Anurag Kumar, Daniel Wong
- Abstract summary: We show that complex-valued deep neural networks (DNNs) do not provide a performance gain over their real-valued counterparts for monaural speech enhancement.
We also find that the use of complex-valued operations hinders the model capacity when the model size is small.
- Score: 22.033822936410246
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Despite multiple efforts made towards adopting complex-valued deep neural
networks (DNNs), it remains an open question whether complex-valued DNNs are
generally more effective than real-valued DNNs for monaural speech enhancement.
This work is devoted to presenting a critical assessment by systematically
examining complex-valued DNNs against their real-valued counterparts.
Specifically, we investigate complex-valued DNN atomic units, including linear
layers, convolutional layers, long short-term memory (LSTM), and gated linear
units. By comparing complex- and real-valued versions of fundamental building
blocks in the recently developed gated convolutional recurrent network (GCRN),
we show how different mechanisms for basic blocks affect the performance. We
also find that the use of complex-valued operations hinders the model capacity
when the model size is small. In addition, we examine two recent complex-valued
DNNs, i.e. deep complex convolutional recurrent network (DCCRN) and deep
complex U-Net (DCUNET). Evaluation results show that both DNNs produce
identical performance to their real-valued counterparts while requiring much
more computation. Based on these comprehensive comparisons, we conclude that
complex-valued DNNs do not provide a performance gain over their real-valued
counterparts for monaural speech enhancement, and thus are less desirable due
to their higher computational costs.
Related papers
- Hybrid Real- and Complex-valued Neural Network Architecture [2.6739705603496327]
We propose a emphhybrid real- and complex-valued emphneural network (HNN) architecture, designed to combine the computational efficiency of real-valued processing with the ability to handle complex-valued data.
Experiments with the AudioMNIST dataset demonstrate that the HNN reduces cross-entropy loss and consumes less parameters compared to an RVNN for all considered cases.
arXiv Detail & Related papers (2025-04-04T14:52:44Z) - Deep-Unrolling Multidimensional Harmonic Retrieval Algorithms on Neuromorphic Hardware [78.17783007774295]
This paper explores the potential of conversion-based neuromorphic algorithms for highly accurate and energy-efficient single-snapshot multidimensional harmonic retrieval.
A novel method for converting the complex-valued convolutional layers and activations into spiking neural networks (SNNs) is developed.
The converted SNNs achieve almost five-fold power efficiency at moderate performance loss compared to the original CNNs.
arXiv Detail & Related papers (2024-12-05T09:41:33Z) - Harnessing Neuron Stability to Improve DNN Verification [42.65507402735545]
We present VeriStable, a novel extension of recently proposed DPLL-based constraint DNN verification approach.
We evaluate the effectiveness of VeriStable across a range of challenging benchmarks including fully-connected feed networks (FNNs), convolutional neural networks (CNNs) and residual networks (ResNets)
Preliminary results show that VeriStable is competitive and outperforms state-of-the-art verification tools, including $alpha$-$beta$-CROWN and MN-BaB, the first and second performers of the VNN-COMP, respectively.
arXiv Detail & Related papers (2024-01-19T23:48:04Z) - An Automata-Theoretic Approach to Synthesizing Binarized Neural Networks [13.271286153792058]
Quantized neural networks (QNNs) have been developed, with binarized neural networks (BNNs) restricted to binary values as a special case.
This paper presents an automata-theoretic approach to synthesizing BNNs that meet designated properties.
arXiv Detail & Related papers (2023-07-29T06:27:28Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - On the Intrinsic Structures of Spiking Neural Networks [66.57589494713515]
Recent years have emerged a surge of interest in SNNs owing to their remarkable potential to handle time-dependent and event-driven data.
There has been a dearth of comprehensive studies examining the impact of intrinsic structures within spiking computations.
This work delves deep into the intrinsic structures of SNNs, by elucidating their influence on the expressivity of SNNs.
arXiv Detail & Related papers (2022-06-21T09:42:30Z) - Ultra-compact Binary Neural Networks for Human Activity Recognition on
RISC-V Processors [10.195581493173643]
Human Activity Recognition (HAR) is a relevant inference task in many mobile applications.
We propose a novel implementation of HAR based on deep neural networks, and precisely on Binary Neural Networks (BNNs)
BNNs yield very small memory footprints and low inference complexity, thanks to the replacement of arithmetic operations with bit-wise ones.
arXiv Detail & Related papers (2022-05-25T13:52:35Z) - Deep Architecture Connectivity Matters for Its Convergence: A
Fine-Grained Analysis [94.64007376939735]
We theoretically characterize the impact of connectivity patterns on the convergence of deep neural networks (DNNs) under gradient descent training.
We show that by a simple filtration on "unpromising" connectivity patterns, we can trim down the number of models to evaluate.
arXiv Detail & Related papers (2022-05-11T17:43:54Z) - Topological Measurement of Deep Neural Networks Using Persistent
Homology [0.7919213739992464]
The inner representation of deep neural networks (DNNs) is indecipherable.
Persistent homology (PH) was employed for investigating the complexities of trained DNNs.
arXiv Detail & Related papers (2021-06-06T03:06:15Z) - Deep Time Delay Neural Network for Speech Enhancement with Full Data
Learning [60.20150317299749]
This paper proposes a deep time delay neural network (TDNN) for speech enhancement with full data learning.
To make full use of the training data, we propose a full data learning method for speech enhancement.
arXiv Detail & Related papers (2020-11-11T06:32:37Z) - Spiking Neural Networks with Single-Spike Temporal-Coded Neurons for
Network Intrusion Detection [6.980076213134383]
Spiking neural network (SNN) is interesting due to its strong bio-plausibility and high energy efficiency.
However, its performance is falling far behind conventional deep neural networks (DNNs)
arXiv Detail & Related papers (2020-10-15T14:46:18Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Neural Additive Models: Interpretable Machine Learning with Neural Nets [77.66871378302774]
Deep neural networks (DNNs) are powerful black-box predictors that have achieved impressive performance on a wide variety of tasks.
We propose Neural Additive Models (NAMs) which combine some of the expressivity of DNNs with the inherent intelligibility of generalized additive models.
NAMs learn a linear combination of neural networks that each attend to a single input feature.
arXiv Detail & Related papers (2020-04-29T01:28:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.