RIC-CNN: Rotation-Invariant Coordinate Convolutional Neural Network
- URL: http://arxiv.org/abs/2211.11812v1
- Date: Mon, 21 Nov 2022 19:27:02 GMT
- Title: RIC-CNN: Rotation-Invariant Coordinate Convolutional Neural Network
- Authors: Hanlin Mo and Guoying Zhao
- Abstract summary: We propose a new convolutional operation, called Rotation-Invariant Coordinate Convolution (RIC-C)
By replacing all standard convolutional layers in a CNN with the corresponding RIC-C, a RIC-CNN can be derived.
It can be observed that RIC-CNN achieves the state-of-the-art classification on the rotated test dataset of MNIST.
- Score: 56.42518353373004
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In recent years, convolutional neural network has shown good performance in
many image processing and computer vision tasks. However, a standard CNN model
is not invariant to image rotations. In fact, even slight rotation of an input
image will seriously degrade its performance. This shortcoming precludes the
use of CNN in some practical scenarios. Thus, in this paper, we focus on
designing convolutional layer with good rotation invariance. Specifically,
based on a simple rotation-invariant coordinate system, we propose a new
convolutional operation, called Rotation-Invariant Coordinate Convolution
(RIC-C). Without additional trainable parameters and data augmentation, RIC-C
is naturally invariant to arbitrary rotations around the input center.
Furthermore, we find the connection between RIC-C and deformable convolution,
and propose a simple but efficient approach to implement RIC-C using Pytorch.
By replacing all standard convolutional layers in a CNN with the corresponding
RIC-C, a RIC-CNN can be derived. Using MNIST dataset, we first evaluate the
rotation invariance of RIC-CNN and compare its performance with most of
existing rotation-invariant CNN models. It can be observed that RIC-CNN
achieves the state-of-the-art classification on the rotated test dataset of
MNIST. Then, we deploy RIC-C to VGG, ResNet and DenseNet, and conduct the
classification experiments on two real image datasets. Also, a shallow CNN and
the corresponding RIC-CNN are trained to extract image patch descriptors, and
we compare their performance in patch verification. These experimental results
again show that RIC-C can be easily used as drop in replacement for standard
convolutions, and greatly enhances the rotation invariance of CNN models
designed for different applications.
Related papers
- Achieving Rotation Invariance in Convolution Operations: Shifting from Data-Driven to Mechanism-Assured [18.910817148765176]
This paper designs a set of new convolution operations that are natually invariant to arbitrary rotations.
We compare their performance with previous rotation-invariant convolutional neural networks (RI-CNNs)
The results show that RIConvs significantly improve the accuracy of these CNN backbones, especially when the training data is limited.
arXiv Detail & Related papers (2024-04-17T12:21:57Z) - Revisiting Data Augmentation for Rotational Invariance in Convolutional
Neural Networks [0.29127054707887967]
We investigate how best to include rotational invariance in a CNN for image classification.
Our experiments show that networks trained with data augmentation alone can classify rotated images nearly as well as in the normal unrotated case.
arXiv Detail & Related papers (2023-10-12T15:53:24Z) - Sorted Convolutional Network for Achieving Continuous Rotational
Invariance [56.42518353373004]
We propose a Sorting Convolution (SC) inspired by some hand-crafted features of texture images.
SC achieves continuous rotational invariance without requiring additional learnable parameters or data augmentation.
Our results demonstrate that SC achieves the best performance in the aforementioned tasks.
arXiv Detail & Related papers (2023-05-23T18:37:07Z) - What Does CNN Shift Invariance Look Like? A Visualization Study [87.79405274610681]
Feature extraction with convolutional neural networks (CNNs) is a popular method to represent images for machine learning tasks.
We focus on measuring and visualizing the shift invariance of extracted features from popular off-the-shelf CNN models.
We conclude that features extracted from popular networks are not globally invariant, and that biases and artifacts exist within this variance.
arXiv Detail & Related papers (2020-11-09T01:16:30Z) - CyCNN: A Rotation Invariant CNN using Polar Mapping and Cylindrical
Convolution Layers [2.4316550366482357]
This paper proposes a deep CNN model, called CyCNN, which exploits polar mapping of input images to convert rotation to translation.
A CyConv layer exploits the cylindrically sliding windows (CSW) mechanism that vertically extends the input-image receptive fields of boundary units in a convolutional layer.
We show that if there is no data augmentation during training, CyCNN significantly improves classification accuracies when compared to conventional CNN models.
arXiv Detail & Related papers (2020-07-21T04:05:35Z) - Dense Steerable Filter CNNs for Exploiting Rotational Symmetry in
Histology Images [3.053417311299492]
Histology images are inherently symmetric under rotation, where each orientation is equally as likely to appear.
Dense Steerable Filter CNNs (DSF-CNNs) use group convolutions with multiple rotated copies of each filter in a densely connected framework.
We show that DSF-CNNs achieve state-of-the-art performance, with significantly fewer parameters, when applied to three different tasks in the area of pathology computational.
arXiv Detail & Related papers (2020-04-06T23:12:31Z) - Visual Commonsense R-CNN [102.5061122013483]
We present a novel unsupervised feature representation learning method, Visual Commonsense Region-based Convolutional Neural Network (VC R-CNN)
VC R-CNN serves as an improved visual region encoder for high-level tasks such as captioning and VQA.
We extensively apply VC R-CNN features in prevailing models of three popular tasks: Image Captioning, VQA, and VCR, and observe consistent performance boosts across them.
arXiv Detail & Related papers (2020-02-27T15:51:19Z) - Computational optimization of convolutional neural networks using
separated filters architecture [69.73393478582027]
We consider a convolutional neural network transformation that reduces computation complexity and thus speedups neural network processing.
Use of convolutional neural networks (CNN) is the standard approach to image recognition despite the fact they can be too computationally demanding.
arXiv Detail & Related papers (2020-02-18T17:42:13Z) - Approximation and Non-parametric Estimation of ResNet-type Convolutional
Neural Networks [52.972605601174955]
We show a ResNet-type CNN can attain the minimax optimal error rates in important function classes.
We derive approximation and estimation error rates of the aformentioned type of CNNs for the Barron and H"older classes.
arXiv Detail & Related papers (2019-03-24T19:42:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.