A Structurally Regularized Convolutional Neural Network for Image
Classification using Wavelet-based SubBand Decomposition
- URL: http://arxiv.org/abs/2103.01823v1
- Date: Tue, 2 Mar 2021 16:01:22 GMT
- Title: A Structurally Regularized Convolutional Neural Network for Image
Classification using Wavelet-based SubBand Decomposition
- Authors: Pavel Sinha, Ioannis Psaromiligkos, Zeljko Zilic
- Abstract summary: We propose a convolutional neural network (CNN) architecture for image classification based on subband decomposition of the image using wavelets.
The proposed architecture decomposes the input image spectra into multiple critically sampled subbands, extracts features using a single CNN per subband, and finally, performs classification by combining the extracted features using a fully connected layer.
We show the proposed architecture is more robust than the regular full-band CNN to noise caused by weight-and-bias quantization and input quantization.
- Score: 2.127049691404299
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a convolutional neural network (CNN) architecture for image
classification based on subband decomposition of the image using wavelets. The
proposed architecture decomposes the input image spectra into multiple
critically sampled subbands, extracts features using a single CNN per subband,
and finally, performs classification by combining the extracted features using
a fully connected layer. Processing each of the subbands by an individual CNN,
thereby limiting the learning scope of each CNN to a single subband, imposes a
form of structural regularization. This provides better generalization
capability as seen by the presented results. The proposed architecture achieves
best-in-class performance in terms of total multiply-add-accumulator operations
and nearly best-in-class performance in terms of total parameters required, yet
it maintains competitive classification performance. We also show the proposed
architecture is more robust than the regular full-band CNN to noise caused by
weight-and-bias quantization and input quantization.
Related papers
- SENetV2: Aggregated dense layer for channelwise and global
representations [0.0]
We introduce a novel aggregated multilayer perceptron, a multi-branch dense layer, within the Squeeze residual module.
This fusion enhances the network's ability to capture channel-wise patterns and have global knowledge.
We conduct extensive experiments on benchmark datasets to validate the model and compare them with established architectures.
arXiv Detail & Related papers (2023-11-17T14:10:57Z) - Pushing the Efficiency Limit Using Structured Sparse Convolutions [82.31130122200578]
We propose Structured Sparse Convolution (SSC), which leverages the inherent structure in images to reduce the parameters in the convolutional filter.
We show that SSC is a generalization of commonly used layers (depthwise, groupwise and pointwise convolution) in efficient architectures''
Architectures based on SSC achieve state-of-the-art performance compared to baselines on CIFAR-10, CIFAR-100, Tiny-ImageNet, and ImageNet classification benchmarks.
arXiv Detail & Related papers (2022-10-23T18:37:22Z) - A heterogeneous group CNN for image super-resolution [127.2132400582117]
Convolutional neural networks (CNNs) have obtained remarkable performance via deep architectures.
We present a heterogeneous group SR CNN (HGSRCNN) via leveraging structure information of different types to obtain a high-quality image.
arXiv Detail & Related papers (2022-09-26T04:14:59Z) - SpectralFormer: Rethinking Hyperspectral Image Classification with
Transformers [91.09957836250209]
Hyperspectral (HS) images are characterized by approximately contiguous spectral information.
CNNs have been proven to be a powerful feature extractor in HS image classification.
We propose a novel backbone network called ulSpectralFormer for HS image classification.
arXiv Detail & Related papers (2021-07-07T02:59:21Z) - The Mind's Eye: Visualizing Class-Agnostic Features of CNNs [92.39082696657874]
We propose an approach to visually interpret CNN features given a set of images by creating corresponding images that depict the most informative features of a specific layer.
Our method uses a dual-objective activation and distance loss, without requiring a generator network nor modifications to the original model.
arXiv Detail & Related papers (2021-01-29T07:46:39Z) - Convolutional Neural Networks from Image Markers [62.997667081978825]
Feature Learning from Image Markers (FLIM) was recently proposed to estimate convolutional filters, with no backpropagation, from strokes drawn by a user on very few images.
This paper extends FLIM for fully connected layers and demonstrates it on different image classification problems.
The results show that FLIM-based convolutional neural networks can outperform the same architecture trained from scratch by backpropagation.
arXiv Detail & Related papers (2020-12-15T22:58:23Z) - ACDC: Weight Sharing in Atom-Coefficient Decomposed Convolution [57.635467829558664]
We introduce a structural regularization across convolutional kernels in a CNN.
We show that CNNs now maintain performance with dramatic reduction in parameters and computations.
arXiv Detail & Related papers (2020-09-04T20:41:47Z) - Ensemble learning in CNN augmented with fully connected subnetworks [0.0]
We propose a new model called EnsNet which is composed of one base CNN and multiple Fully Connected SubNetworks (FCSNs)
An EnsNet achieves a state-of-the-art error rate of 0.16% on MNIST.
arXiv Detail & Related papers (2020-03-19T04:02:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.