Bayesian Sparsification Methods for Deep Complex-valued Networks
- URL: http://arxiv.org/abs/2003.11413v2
- Date: Sun, 28 Jun 2020 03:53:14 GMT
- Title: Bayesian Sparsification Methods for Deep Complex-valued Networks
- Authors: Ivan Nazarov and Evgeny Burnaev
- Abstract summary: We extend Sparse Variational Dropout to complex-valued neural networks.
We conduct a large numerical study of the performance-compression trade-off of C-valued networks on two tasks.
We replicate the state-of-the-art result by Trabelsi et al. [ 2018] on MusicNet with a complex-valued network compressed by 50-100x at a small performance penalty.
- Score: 18.00411355850543
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: With continual miniaturization ever more applications of deep learning can be
found in embedded systems, where it is common to encounter data with natural
complex domain representation. To this end we extend Sparse Variational Dropout
to complex-valued neural networks and verify the proposed Bayesian technique by
conducting a large numerical study of the performance-compression trade-off of
C-valued networks on two tasks: image recognition on MNIST-like and CIFAR10
datasets and music transcription on MusicNet. We replicate the state-of-the-art
result by Trabelsi et al. [2018] on MusicNet with a complex-valued network
compressed by 50-100x at a small performance penalty.
Related papers
- i-SpaSP: Structured Neural Pruning via Sparse Signal Recovery [11.119895959906085]
We propose a novel, structured pruning algorithm for neural networks -- the iterative, Sparse Structured Pruning, dubbed as i-SpaSP.
i-SpaSP operates by identifying a larger set of important parameter groups within a network that contribute most to the residual between pruned and dense network output.
It is shown to discover high-performing sub-networks and improve upon the pruning efficiency of provable baseline methodologies by several orders of magnitude.
arXiv Detail & Related papers (2021-12-07T05:26:45Z) - Self-supervised Neural Networks for Spectral Snapshot Compressive
Imaging [15.616674529295366]
We consider using untrained neural networks to solve the reconstruction problem of snapshot compressive imaging (SCI)
In this paper, inspired by the untrained neural networks such as deep image priors (DIP) and deep decoders, we develop a framework by integrating DIP into the plug-and-play regime, leading to a self-supervised network for spectral SCI reconstruction.
arXiv Detail & Related papers (2021-08-28T14:17:38Z) - Semi-supervised Network Embedding with Differentiable Deep Quantisation [81.49184987430333]
We develop d-SNEQ, a differentiable quantisation method for network embedding.
d-SNEQ incorporates a rank loss to equip the learned quantisation codes with rich high-order information.
It is able to substantially compress the size of trained embeddings, thus reducing storage footprint and accelerating retrieval speed.
arXiv Detail & Related papers (2021-08-20T11:53:05Z) - Over-and-Under Complete Convolutional RNN for MRI Reconstruction [57.95363471940937]
Recent deep learning-based methods for MR image reconstruction usually leverage a generic auto-encoder architecture.
We propose an Over-and-Under Complete Convolu?tional Recurrent Neural Network (OUCR), which consists of an overcomplete and an undercomplete Convolutional Recurrent Neural Network(CRNN)
The proposed method achieves significant improvements over the compressed sensing and popular deep learning-based methods with less number of trainable parameters.
arXiv Detail & Related papers (2021-06-16T15:56:34Z) - Deep Neural Networks and End-to-End Learning for Audio Compression [2.084078990567849]
We present an end-to-end deep learning approach that combines recurrent neural networks (RNNs) within the training strategy of variational autoencoders (VAEs) with a binary representation of the latent space.
This is the first end-to-end learning for a single audio compression model with RNNs, and our model achieves a Signal to Distortion Ratio (SDR) of 20.54.
arXiv Detail & Related papers (2021-05-25T05:36:30Z) - Learning Frequency-aware Dynamic Network for Efficient Super-Resolution [56.98668484450857]
This paper explores a novel frequency-aware dynamic network for dividing the input into multiple parts according to its coefficients in the discrete cosine transform (DCT) domain.
In practice, the high-frequency part will be processed using expensive operations and the lower-frequency part is assigned with cheap operations to relieve the computation burden.
Experiments conducted on benchmark SISR models and datasets show that the frequency-aware dynamic network can be employed for various SISR neural architectures.
arXiv Detail & Related papers (2021-03-15T12:54:26Z) - Anomaly Detection on Attributed Networks via Contrastive Self-Supervised
Learning [50.24174211654775]
We present a novel contrastive self-supervised learning framework for anomaly detection on attributed networks.
Our framework fully exploits the local information from network data by sampling a novel type of contrastive instance pair.
A graph neural network-based contrastive learning model is proposed to learn informative embedding from high-dimensional attributes and local structure.
arXiv Detail & Related papers (2021-02-27T03:17:20Z) - Learning low-rank latent mesoscale structures in networks [1.1470070927586016]
We present a new approach for describing low-rank mesoscale structures in networks.
We use several synthetic network models and empirical friendship, collaboration, and protein--protein interaction (PPI) networks.
We show how to denoise a corrupted network by using only the latent motifs that one learns directly from the corrupted network.
arXiv Detail & Related papers (2021-02-13T18:54:49Z) - ESPN: Extremely Sparse Pruned Networks [50.436905934791035]
We show that a simple iterative mask discovery method can achieve state-of-the-art compression of very deep networks.
Our algorithm represents a hybrid approach between single shot network pruning methods and Lottery-Ticket type approaches.
arXiv Detail & Related papers (2020-06-28T23:09:27Z) - Complexity Analysis of an Edge Preserving CNN SAR Despeckling Algorithm [1.933681537640272]
We exploit the effect of the complexity of the convolutional neural network for SAR despeckling.
Deeper networks better generalize on both simulated and real images.
arXiv Detail & Related papers (2020-04-17T17:02:01Z) - Large-Scale Gradient-Free Deep Learning with Recursive Local
Representation Alignment [84.57874289554839]
Training deep neural networks on large-scale datasets requires significant hardware resources.
Backpropagation, the workhorse for training these networks, is an inherently sequential process that is difficult to parallelize.
We propose a neuro-biologically-plausible alternative to backprop that can be used to train deep networks.
arXiv Detail & Related papers (2020-02-10T16:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.