Multi-Scale Label Relation Learning for Multi-Label Classification Using
1-Dimensional Convolutional Neural Networks
- URL: http://arxiv.org/abs/2107.05941v1
- Date: Tue, 13 Jul 2021 09:26:34 GMT
- Title: Multi-Scale Label Relation Learning for Multi-Label Classification Using
1-Dimensional Convolutional Neural Networks
- Authors: Junhyung Kim, Byungyoon Park, Charmgil Hong
- Abstract summary: We present Multi-Scale Label Dependence Relation Networks (MSDN), a novel approach to multi-label classification (MLC)
MSDN uses 1-dimensional convolution kernels to learn label dependencies at multi-scale.
We demonstrate that our model can achieve better accuracies with much smaller number of model parameters compared to RNN-based MLC models.
- Score: 0.5801044612920815
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present Multi-Scale Label Dependence Relation Networks (MSDN), a novel
approach to multi-label classification (MLC) using 1-dimensional convolution
kernels to learn label dependencies at multi-scale. Modern multi-label
classifiers have been adopting recurrent neural networks (RNNs) as a memory
structure to capture and exploit label dependency relations. The RNN-based MLC
models however tend to introduce a very large number of parameters that may
cause under-/over-fitting problems. The proposed method uses the 1-dimensional
convolutional neural network (1D-CNN) to serve the same purpose in a more
efficient manner. By training a model with multiple kernel sizes, the method is
able to learn the dependency relations among labels at multiple scales, while
it uses a drastically smaller number of parameters. With public benchmark
datasets, we demonstrate that our model can achieve better accuracies with much
smaller number of model parameters compared to RNN-based MLC models.
Related papers
- Heterogenous Memory Augmented Neural Networks [84.29338268789684]
We introduce a novel heterogeneous memory augmentation approach for neural networks.
By introducing learnable memory tokens with attention mechanism, we can effectively boost performance without huge computational overhead.
We show our approach on various image and graph-based tasks under both in-distribution (ID) and out-of-distribution (OOD) conditions.
arXiv Detail & Related papers (2023-10-17T01:05:28Z) - Deep Dependency Networks for Multi-Label Classification [24.24496964886951]
We show that the performance of previous approaches that combine Markov Random Fields with neural networks can be modestly improved.
We propose a new modeling framework called deep dependency networks, which augments a dependency network.
Despite its simplicity, jointly learning this new architecture yields significant improvements in performance.
arXiv Detail & Related papers (2023-02-01T17:52:40Z) - Granger Causality using Neural Networks [7.62566998854384]
We propose novel classes of models that can handle underlying non-linearity in a computationally efficient manner.
We show one can directly decouple lags and individual time series importance via decoupled penalties.
This is important as we want to select the lag order during the process of GC estimation.
arXiv Detail & Related papers (2022-08-07T12:02:48Z) - Dynamically-Scaled Deep Canonical Correlation Analysis [77.34726150561087]
Canonical Correlation Analysis (CCA) is a method for feature extraction of two views by finding maximally correlated linear projections of them.
We introduce a novel dynamic scaling method for training an input-dependent canonical correlation model.
arXiv Detail & Related papers (2022-03-23T12:52:49Z) - Multi network InfoMax: A pre-training method involving graph
convolutional networks [0.0]
This paper presents a pre-training method involving graph convolutional/neural networks (GCNs/GNNs)
The learned high-level graph latent representations help increase performance for downstream graph classification tasks.
We apply our method to a neuroimaging dataset for classifying subjects into healthy control (HC) and schizophrenia (SZ) groups.
arXiv Detail & Related papers (2021-11-01T21:53:20Z) - Gated recurrent units and temporal convolutional network for multilabel
classification [122.84638446560663]
This work proposes a new ensemble method for managing multilabel classification.
The core of the proposed approach combines a set of gated recurrent units and temporal convolutional neural networks trained with variants of the Adam gradients optimization approach.
arXiv Detail & Related papers (2021-10-09T00:00:16Z) - ACDC: Weight Sharing in Atom-Coefficient Decomposed Convolution [57.635467829558664]
We introduce a structural regularization across convolutional kernels in a CNN.
We show that CNNs now maintain performance with dramatic reduction in parameters and computations.
arXiv Detail & Related papers (2020-09-04T20:41:47Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z) - Model Fusion via Optimal Transport [64.13185244219353]
We present a layer-wise model fusion algorithm for neural networks.
We show that this can successfully yield "one-shot" knowledge transfer between neural networks trained on heterogeneous non-i.i.d. data.
arXiv Detail & Related papers (2019-10-12T22:07:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.