Deep Transform and Metric Learning Network: Wedding Deep Dictionary
Learning and Neural Networks
- URL: http://arxiv.org/abs/2002.07898v2
- Date: Wed, 21 Oct 2020 01:57:18 GMT
- Title: Deep Transform and Metric Learning Network: Wedding Deep Dictionary
Learning and Neural Networks
- Authors: Wen Tang, Emilie Chouzenoux, Jean-Christophe Pesquet, and Hamid Krim
- Abstract summary: We propose a novel DDL approach where each DL layer can be formulated as a combination of one linear layer and a Recurrent Neural Network (RNN)
Our proposed work unveils new insights into Neural Networks and DDL and provides a new, efficient and competitive approach to jointly learn a deep transform and a metric for inference applications.
- Score: 34.49034775978504
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: On account of its many successes in inference tasks and denoising
applications, Dictionary Learning (DL) and its related sparse optimization
problems have garnered a lot of research interest. While most solutions have
focused on single layer dictionaries, the improved recently proposed Deep DL
(DDL) methods have also fallen short on a number of issues. We propose herein,
a novel DDL approach where each DL layer can be formulated as a combination of
one linear layer and a Recurrent Neural Network (RNN). The RNN is shown to
flexibly account for the layer-associated and learned metric. Our proposed work
unveils new insights into Neural Networks and DDL and provides a new, efficient
and competitive approach to jointly learn a deep transform and a metric for
inference applications. Extensive experiments are carried out to demonstrate
that the proposed method can not only outperform existing DDL but also
state-of-the-art generic CNNs.
Related papers
- Piecewise Linear Neural Networks and Deep Learning [27.02556725989978]
PieceWise Linear Neural Networks (PWLNNs) have proven successful in various fields, most recently in deep learning.
In 1977, the canonical representation pioneered the works of shallow PWLNNs learned by incremental designs.
In 2010, the Rectified Linear Unit (ReLU) advocated the prevalence of PWLNNs in deep learning.
arXiv Detail & Related papers (2022-06-18T08:41:42Z) - Low-bit Quantization of Recurrent Neural Network Language Models Using
Alternating Direction Methods of Multipliers [67.688697838109]
This paper presents a novel method to train quantized RNNLMs from scratch using alternating direction methods of multipliers (ADMM)
Experiments on two tasks suggest the proposed ADMM quantization achieved a model size compression factor of up to 31 times over the full precision baseline RNNLMs.
arXiv Detail & Related papers (2021-11-29T09:30:06Z) - FFNB: Forgetting-Free Neural Blocks for Deep Continual Visual Learning [14.924672048447338]
We devise a dynamic network architecture for continual learning based on a novel forgetting-free neural block (FFNB)
Training FFNB features on new tasks is achieved using a novel procedure that constrains the underlying parameters in the null-space of the previous tasks.
arXiv Detail & Related papers (2021-11-22T17:23:34Z) - Deep Transform and Metric Learning Networks [30.753361147185995]
We propose a novel Deep DL approach where each DL layer can be formulated and solved as a combination of one linear layer and a Recurrent Neural Network.
Our proposed work unveils new insights between the Neural Networks and Deep DL, and provides a novel, efficient and competitive approach to jointly learn the deep transforms and metrics.
arXiv Detail & Related papers (2021-04-21T03:10:15Z) - Reinforcement Learning with External Knowledge by using Logical Neural
Networks [67.46162586940905]
A recent neuro-symbolic framework called the Logical Neural Networks (LNNs) can simultaneously provide key-properties of both neural networks and symbolic logic.
We propose an integrated method that enables model-free reinforcement learning from external knowledge sources.
arXiv Detail & Related papers (2021-03-03T12:34:59Z) - Local Critic Training for Model-Parallel Learning of Deep Neural
Networks [94.69202357137452]
We propose a novel model-parallel learning method, called local critic training.
We show that the proposed approach successfully decouples the update process of the layer groups for both convolutional neural networks (CNNs) and recurrent neural networks (RNNs)
We also show that trained networks by the proposed method can be used for structural optimization.
arXiv Detail & Related papers (2021-02-03T09:30:45Z) - Solving Sparse Linear Inverse Problems in Communication Systems: A Deep
Learning Approach With Adaptive Depth [51.40441097625201]
We propose an end-to-end trainable deep learning architecture for sparse signal recovery problems.
The proposed method learns how many layers to execute to emit an output, and the network depth is dynamically adjusted for each task in the inference phase.
arXiv Detail & Related papers (2020-10-29T06:32:53Z) - Improved Neural Language Model Fusion for Streaming Recurrent Neural
Network Transducer [28.697119605752643]
Recurrent Neural Network Transducer (RNN-T) has an implicit neural network language model (NNLM) and cannot easily leverage unpaired text data during training.
Previous work has proposed various fusion methods to incorporate external NNLMs into end-to-end ASR to address this weakness.
We propose extensions to these techniques that allow RNN-T to exploit external NNLMs during both training and inference time.
arXiv Detail & Related papers (2020-10-26T20:10:12Z) - Progressive Tandem Learning for Pattern Recognition with Deep Spiking
Neural Networks [80.15411508088522]
Spiking neural networks (SNNs) have shown advantages over traditional artificial neural networks (ANNs) for low latency and high computational efficiency.
We propose a novel ANN-to-SNN conversion and layer-wise learning framework for rapid and efficient pattern recognition.
arXiv Detail & Related papers (2020-07-02T15:38:44Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.