Exploring Novel Pooling Strategies for Edge Preserved Feature Maps in
Convolutional Neural Networks
- URL: http://arxiv.org/abs/2110.08842v1
- Date: Sun, 17 Oct 2021 15:11:51 GMT
- Title: Exploring Novel Pooling Strategies for Edge Preserved Feature Maps in
Convolutional Neural Networks
- Authors: Adithya Sineesh and Mahesh Raveendranatha Panicker
- Abstract summary: Anti-aliased convolutional neural networks (CNNs) have led to some resurgence in relooking the way pooling is done in CNNs.
Two novel pooling approaches are presented such as Laplacian-Gaussian Concatenation with Attention (LGCA) pooling and Wavelet based approximate-detailed concatenation with attention (WADCA) pooling.
Results suggest that the proposed pooling approaches outperform the conventional pooling as well as blur pooling for classification, segmentation and autoencoders.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: With the introduction of anti-aliased convolutional neural networks (CNN),
there has been some resurgence in relooking the way pooling is done in CNNs.
The fundamental building block of the anti-aliased CNN has been the application
of Gaussian smoothing before the pooling operation to reduce the distortion due
to aliasing thereby making CNNs shift invariant. Wavelet based approaches have
also been proposed as a possibility of additional noise removal capability and
gave interesting results for even segmentation tasks. However, all the
approaches proposed completely remove the high frequency components under the
assumption that they are noise. However, by removing high frequency components,
the edges in the feature maps are also smoothed. In this work, an exhaustive
analysis of the edge preserving pooling options for classification,
segmentation and autoencoders are presented. Two novel pooling approaches are
presented such as Laplacian-Gaussian Concatenation with Attention (LGCA)
pooling and Wavelet based approximate-detailed coefficient concatenation with
attention (WADCA) pooling. The results suggest that the proposed pooling
approaches outperform the conventional pooling as well as blur pooling for
classification, segmentation and autoencoders.
Related papers
- Simple Pooling Front-ends For Efficient Audio Classification [56.59107110017436]
We show that eliminating the temporal redundancy in the input audio features could be an effective approach for efficient audio classification.
We propose a family of simple pooling front-ends (SimPFs) which use simple non-parametric pooling operations to reduce the redundant information.
SimPFs can achieve a reduction in more than half of the number of floating point operations for off-the-shelf audio neural networks.
arXiv Detail & Related papers (2022-10-03T14:00:41Z) - Hierarchical Spherical CNNs with Lifting-based Adaptive Wavelets for
Pooling and Unpooling [101.72318949104627]
We propose a novel framework of hierarchical convolutional neural networks (HS-CNNs) with a lifting structure to learn adaptive spherical wavelets for pooling and unpooling.
LiftHS-CNN ensures a more efficient hierarchical feature learning for both image- and pixel-level tasks.
arXiv Detail & Related papers (2022-05-31T07:23:42Z) - A Passive Similarity based CNN Filter Pruning for Efficient Acoustic
Scene Classification [23.661189257759535]
We present a method to develop low-complexity convolutional neural networks (CNNs) for acoustic scene classification (ASC)
We propose a passive filter pruning framework, where a few convolutional filters from the CNNs are eliminated to yield compressed CNNs.
The proposed method is simple, reduces computations per inference by 27%, with 25% fewer parameters, with less than 1% drop in accuracy.
arXiv Detail & Related papers (2022-03-29T17:00:06Z) - Fuzzy Pooling [7.6146285961466]
Convolutional Neural Networks (CNNs) are artificial learning systems typically based on two operations: convolution and pooling.
We present a novel pooling operation based on (type-1) fuzzy sets to cope with the local imprecision of the feature maps.
Experiments using publicly available datasets show that the proposed approach can enhance the classification performance of a CNN.
arXiv Detail & Related papers (2022-02-12T11:18:32Z) - AdaPool: Exponential Adaptive Pooling for Information-Retaining
Downsampling [82.08631594071656]
Pooling layers are essential building blocks of Convolutional Neural Networks (CNNs)
We propose an adaptive and exponentially weighted pooling method named adaPool.
We demonstrate how adaPool improves the preservation of detail through a range of tasks including image and video classification and object detection.
arXiv Detail & Related papers (2021-11-01T08:50:37Z) - TSG: Target-Selective Gradient Backprop for Probing CNN Visual Saliency [72.9106103283475]
We study the visual saliency, a.k.a. visual explanation, to interpret convolutional neural networks.
Inspired by those observations, we propose a novel visual saliency framework, termed Target-Selective Gradient (TSG) backprop.
The proposed TSG consists of two components, namely, TSG-Conv and TSG-FC, which rectify the gradients for convolutional layers and fully-connected layers, respectively.
arXiv Detail & Related papers (2021-10-11T12:00:20Z) - Frequency Pooling: Shift-Equivalent and Anti-Aliasing Downsampling [9.249235534786072]
We show that frequency pooling is shift-equivalent and anti-aliasing based on the property of Fourier transform and Nyquist frequency.
Experiments on image classification show that frequency pooling improves accuracy and robustness with respect to the shifts of CNNs.
arXiv Detail & Related papers (2021-09-24T09:32:10Z) - Adversarial Feature Augmentation and Normalization for Visual
Recognition [109.6834687220478]
Recent advances in computer vision take advantage of adversarial data augmentation to ameliorate the generalization ability of classification models.
Here, we present an effective and efficient alternative that advocates adversarial augmentation on intermediate feature embeddings.
We validate the proposed approach across diverse visual recognition tasks with representative backbone networks.
arXiv Detail & Related papers (2021-03-22T20:36:34Z) - Refining activation downsampling with SoftPool [74.1840492087968]
Convolutional Neural Networks (CNNs) use pooling to decrease the size of activation maps.
We propose SoftPool: a fast and efficient method for exponentially weighted activation downsampling.
We show that SoftPool can retain more information in the reduced activation maps.
arXiv Detail & Related papers (2021-01-02T12:09:49Z) - Acoustic Scene Classification with Squeeze-Excitation Residual Networks [4.591851728010269]
We propose two novel squeeze-excitation blocks to improve the accuracy of a CNN-based ASC framework based on residual learning.
The behavior of the block that implements such operators and, therefore, the entire neural network, can be modified depending on the input to the block.
arXiv Detail & Related papers (2020-03-20T14:07:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.