When Dictionary Learning Meets Deep Learning: Deep Dictionary Learning
and Coding Network for Image Recognition with Limited Data
- URL: http://arxiv.org/abs/2005.10940v1
- Date: Thu, 21 May 2020 23:12:10 GMT
- Title: When Dictionary Learning Meets Deep Learning: Deep Dictionary Learning
and Coding Network for Image Recognition with Limited Data
- Authors: Hao Tang, Hong Liu, Wei Xiao, Nicu Sebe
- Abstract summary: We present a new Deep Dictionary Learning and Coding Network (DDLCN) for image recognition tasks with limited data.
We empirically compare DDLCN with several leading dictionary learning methods and deep learning models.
Experimental results on five popular datasets show that DDLCN achieves competitive results compared with state-of-the-art methods when the training data is limited.
- Score: 74.75557280245643
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a new Deep Dictionary Learning and Coding Network (DDLCN) for
image recognition tasks with limited data. The proposed DDLCN has most of the
standard deep learning layers (e.g., input/output, pooling, fully connected,
etc.), but the fundamental convolutional layers are replaced by our proposed
compound dictionary learning and coding layers. The dictionary learning learns
an over-complete dictionary for input training data. At the deep coding layer,
a locality constraint is added to guarantee that the activated dictionary bases
are close to each other. Then the activated dictionary atoms are assembled and
passed to the compound dictionary learning and coding layers. In this way, the
activated atoms in the first layer can be represented by the deeper atoms in
the second dictionary. Intuitively, the second dictionary is designed to learn
the fine-grained components shared among the input dictionary atoms, thus a
more informative and discriminative low-level representation of the dictionary
atoms can be obtained. We empirically compare DDLCN with several leading
dictionary learning methods and deep learning models. Experimental results on
five popular datasets show that DDLCN achieves competitive results compared
with state-of-the-art methods when the training data is limited. Code is
available at https://github.com/Ha0Tang/DDLCN.
Related papers
- Low Rank Multi-Dictionary Selection at Scale [5.827700856320355]
We propose a multi-dictionary atom selection technique for low-rank sparse coding named LRMDS.
We demonstrate the scalability quality of LRMDS in both synthetic and real-world datasets.
arXiv Detail & Related papers (2024-06-11T05:40:45Z) - Dictionary Learning Improves Patch-Free Circuit Discovery in Mechanistic
Interpretability: A Case Study on Othello-GPT [59.245414547751636]
We propose a circuit discovery framework alternative to activation patching.
Our framework suffers less from out-of-distribution and proves to be more efficient in terms of complexity.
We dig in a small transformer trained on a synthetic task named Othello and find a number of human-understandable fine-grained circuits inside of it.
arXiv Detail & Related papers (2024-02-19T15:04:53Z) - Incorporating Dictionaries into a Neural Network Architecture to Extract
COVID-19 Medical Concepts From Social Media [0.2302001830524133]
We investigate the potential benefit of incorporating dictionary information into a neural network architecture for natural language processing.
In particular, we make use of this architecture to extract several concepts related to COVID-19 from an on-line medical forum.
Our results show that incorporating small domain dictionaries to deep learning models can improve concept extraction tasks.
arXiv Detail & Related papers (2023-09-05T12:47:44Z) - Hiding Data Helps: On the Benefits of Masking for Sparse Coding [22.712098918769243]
We show that in the presence of noise, minimizing the standard dictionary learning objective can fail to recover the elements of the ground-truth dictionary in the over-realized regime.
We propose a novel masking objective for which recovering the ground-truth dictionary is in fact optimal as the signal increases for a large class of data-generating processes.
arXiv Detail & Related papers (2023-02-24T16:16:19Z) - Efficient CNN with uncorrelated Bag of Features pooling [98.78384185493624]
Bag of Features (BoF) has been recently proposed to reduce the complexity of convolution layers.
We propose an approach that builds on top of BoF pooling to boost its efficiency by ensuring that the items of the learned dictionary are non-redundant.
The proposed strategy yields an efficient variant of BoF and further boosts its performance, without any additional parameters.
arXiv Detail & Related papers (2022-09-22T09:00:30Z) - DetCLIP: Dictionary-Enriched Visual-Concept Paralleled Pre-training for
Open-world Detection [118.36746273425354]
This paper presents a paralleled visual-concept pre-training method for open-world detection by resorting to knowledge enrichment from a designed concept dictionary.
By enriching the concepts with their descriptions, we explicitly build the relationships among various concepts to facilitate the open-domain learning.
The proposed framework demonstrates strong zero-shot detection performances, e.g., on the LVIS dataset, our DetCLIP-T outperforms GLIP-T by 9.9% mAP and obtains a 13.5% improvement on rare categories.
arXiv Detail & Related papers (2022-09-20T02:01:01Z) - Neural Implicit Dictionary via Mixture-of-Expert Training [111.08941206369508]
We present a generic INR framework that achieves both data and training efficiency by learning a Neural Implicit Dictionary (NID)
Our NID assembles a group of coordinate-based Impworks which are tuned to span the desired function space.
Our experiments show that, NID can improve reconstruction of 2D images or 3D scenes by 2 orders of magnitude faster with up to 98% less input data.
arXiv Detail & Related papers (2022-07-08T05:07:19Z) - CodeRetriever: Unimodal and Bimodal Contrastive Learning [128.06072658302165]
We propose the CodeRetriever model, which combines the unimodal and bimodal contrastive learning to train function-level code semantic representations.
For unimodal contrastive learning, we design a semantic-guided method to build positive code pairs based on the documentation and function name.
For bimodal contrastive learning, we leverage the documentation and in-line comments of code to build text-code pairs.
arXiv Detail & Related papers (2022-01-26T10:54:30Z) - PUDLE: Implicit Acceleration of Dictionary Learning by Backpropagation [4.081440927534577]
This paper offers the first theoretical proof for empirical results through PUDLE, a Provable Unfolded Dictionary LEarning method.
We highlight the minimization impact of loss, unfolding, and backpropagation on convergence.
We complement our findings through synthetic and image denoising experiments.
arXiv Detail & Related papers (2021-05-31T18:49:58Z) - Learning Deep Analysis Dictionaries -- Part II: Convolutional
Dictionaries [38.7315182732103]
We introduce a Deep Convolutional Analysis Dictionary Model (DeepCAM) by learning convolutional dictionaries instead of unstructured dictionaries.
A L-layer DeepCAM consists of L layers of convolutional analysis dictionary and element-wise soft-thresholding pairs.
We demonstrate that DeepCAM is an effective multilayer convolutional model and, on single image super-resolution, achieves performance comparable with other methods.
arXiv Detail & Related papers (2020-01-31T19:02:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.