Multi-Subspace Neural Network for Image Recognition
- URL: http://arxiv.org/abs/2006.09618v1
- Date: Wed, 17 Jun 2020 02:55:34 GMT
- Title: Multi-Subspace Neural Network for Image Recognition
- Authors: Chieh-Ning Fang, Chin-Teng Lin
- Abstract summary: In image classification task, feature extraction is always a big issue. Intra-class variability increases the difficulty in designing the extractors.
Recently, deep learning has drawn lots of attention on automatically learning features from data.
In this study, we proposed multi-subspace neural network (MSNN) which integrates key components of the convolutional neural network (CNN), receptive field, with subspace concept.
- Score: 33.61205842747625
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In image classification task, feature extraction is always a big issue.
Intra-class variability increases the difficulty in designing the extractors.
Furthermore, hand-crafted feature extractor cannot simply adapt new situation.
Recently, deep learning has drawn lots of attention on automatically learning
features from data. In this study, we proposed multi-subspace neural network
(MSNN) which integrates key components of the convolutional neural network
(CNN), receptive field, with subspace concept. Associating subspace with the
deep network is a novel designing, providing various viewpoints of data. Basis
vectors, trained by adaptive subspace self-organization map (ASSOM) span the
subspace, serve as a transfer function to access axial components and define
the receptive field to extract basic patterns of data without distorting the
topology in the visual task. Moreover, the multiple-subspace strategy is
implemented as parallel blocks to adapt real-world data and contribute various
interpretations of data hoping to be more robust dealing with intra-class
variability issues. To this end, handwritten digit and object image datasets
(i.e., MNIST and COIL-20) for classification are employed to validate the
proposed MSNN architecture. Experimental results show MSNN is competitive to
other state-of-the-art approaches.
Related papers
- On Characterizing the Evolution of Embedding Space of Neural Networks
using Algebraic Topology [9.537910170141467]
We study how the topology of feature embedding space changes as it passes through the layers of a well-trained deep neural network (DNN) through Betti numbers.
We demonstrate that as depth increases, a topologically complicated dataset is transformed into a simple one, resulting in Betti numbers attaining their lowest possible value.
arXiv Detail & Related papers (2023-11-08T10:45:12Z) - Variation Spaces for Multi-Output Neural Networks: Insights on Multi-Task Learning and Network Compression [28.851519959657466]
This paper introduces a novel theoretical framework for the analysis of vector-valued neural networks.
A key contribution of this work is the development of a representer theorem for the vector-valued variation spaces.
This observation reveals that the norm associated with these vector-valued variation spaces encourages the learning of features that are useful for multiple tasks.
arXiv Detail & Related papers (2023-05-25T23:32:10Z) - SDA-$x$Net: Selective Depth Attention Networks for Adaptive Multi-scale
Feature Representation [14.7929472540577]
Existing multi-scale solutions lead to a risk of just increasing the receptive field sizes while neglecting small receptive fields.
We introduce a new attention dimension, i.e. depth, in addition to existing attention dimensions such as channel, spatial, and branch.
We present a novel selective depth attention network to symmetrically handle multi-scale objects in various vision tasks.
arXiv Detail & Related papers (2022-09-21T12:49:55Z) - SAR Despeckling Using Overcomplete Convolutional Networks [53.99620005035804]
despeckling is an important problem in remote sensing as speckle degrades SAR images.
Recent studies show that convolutional neural networks(CNNs) outperform classical despeckling methods.
This study employs an overcomplete CNN architecture to focus on learning low-level features by restricting the receptive field.
We show that the proposed network improves despeckling performance compared to recent despeckling methods on synthetic and real SAR images.
arXiv Detail & Related papers (2022-05-31T15:55:37Z) - Network Comparison Study of Deep Activation Feature Discriminability
with Novel Objects [0.5076419064097732]
State-of-the-art computer visions algorithms have incorporated Deep Neural Networks (DNN) in feature extracting roles, creating Deep Convolutional Activation Features (DeCAF)
This study analyzes the general discriminability of novel object visual appearances encoded into the DeCAF space of six of the leading visual recognition DNN architectures.
arXiv Detail & Related papers (2022-02-08T07:40:53Z) - Exploring the Common Principal Subspace of Deep Features in Neural
Networks [50.37178960258464]
We find that different Deep Neural Networks (DNNs) trained with the same dataset share a common principal subspace in latent spaces.
Specifically, we design a new metric $mathcalP$-vector to represent the principal subspace of deep features learned in a DNN.
Small angles (with cosine close to $1.0$) have been found in the comparisons between any two DNNs trained with different algorithms/architectures.
arXiv Detail & Related papers (2021-10-06T15:48:32Z) - Spatial Dependency Networks: Neural Layers for Improved Generative Image
Modeling [79.15521784128102]
We introduce a novel neural network for building image generators (decoders) and apply it to variational autoencoders (VAEs)
In our spatial dependency networks (SDNs), feature maps at each level of a deep neural net are computed in a spatially coherent way.
We show that augmenting the decoder of a hierarchical VAE by spatial dependency layers considerably improves density estimation.
arXiv Detail & Related papers (2021-03-16T07:01:08Z) - Separable Batch Normalization for Robust Facial Landmark Localization
with Cross-protocol Network Training [41.82379935715916]
A big, diverse and balanced training data is the key to the success of deep neural network training.
A small dataset without diverse and balanced training samples cannot support the training of a deep network effectively.
This paper presents a novel Separable Batch Normalization (SepBN) module with a Cross-protocol Network Training (CNT) strategy for robust facial landmark localization.
arXiv Detail & Related papers (2021-01-17T13:04:06Z) - Spatio-Temporal Inception Graph Convolutional Networks for
Skeleton-Based Action Recognition [126.51241919472356]
We design a simple and highly modularized graph convolutional network architecture for skeleton-based action recognition.
Our network is constructed by repeating a building block that aggregates multi-granularity information from both the spatial and temporal paths.
arXiv Detail & Related papers (2020-11-26T14:43:04Z) - Learning Deep Interleaved Networks with Asymmetric Co-Attention for
Image Restoration [65.11022516031463]
We present a deep interleaved network (DIN) that learns how information at different states should be combined for high-quality (HQ) images reconstruction.
In this paper, we propose asymmetric co-attention (AsyCA) which is attached at each interleaved node to model the feature dependencies.
Our presented DIN can be trained end-to-end and applied to various image restoration tasks.
arXiv Detail & Related papers (2020-10-29T15:32:00Z) - SPIN: Structure-Preserving Inner Offset Network for Scene Text
Recognition [48.676064155070556]
Arbitrary text appearance poses a great challenge in scene text recognition tasks.
We introduce a new learnable geometric-unrelated module, the Structure-Preserving Inner Offset Network (SPIN)
SPIN allows the color manipulation of source data within the network.
arXiv Detail & Related papers (2020-05-27T01:47:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.