ReF -- Rotation Equivariant Features for Local Feature Matching
- URL: http://arxiv.org/abs/2203.05206v1
- Date: Thu, 10 Mar 2022 07:36:09 GMT
- Title: ReF -- Rotation Equivariant Features for Local Feature Matching
- Authors: Abhishek Peri, Kinal Mehta, Avneesh Mishra, Michael Milford, Sourav
Garg, K. Madhava Krishna
- Abstract summary: We propose an alternative, complementary approach that centers on inducing bias in the model architecture itself to generate rotation-specific' features.
We demonstrate that this high performance, rotation-specific coverage from the steerable CNNs can be expanded to all rotation angles.
We present a detailed analysis of the performance effects of ensembling, robust estimation, network architecture variations, and the use of rotation priors.
- Score: 30.459559206664427
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Sparse local feature matching is pivotal for many computer vision and
robotics tasks. To improve their invariance to challenging appearance
conditions and viewing angles, and hence their usefulness, existing
learning-based methods have primarily focused on data augmentation-based
training. In this work, we propose an alternative, complementary approach that
centers on inducing bias in the model architecture itself to generate
`rotation-specific' features using Steerable E2-CNNs, that are then
group-pooled to achieve rotation-invariant local features. We demonstrate that
this high performance, rotation-specific coverage from the steerable CNNs can
be expanded to all rotation angles by combining it with augmentation-trained
standard CNNs which have broader coverage but are often inaccurate, thus
creating a state-of-the-art rotation-robust local feature matcher. We benchmark
our proposed methods against existing techniques on HPatches and a newly
proposed UrbanScenes3D-Air dataset for visual place recognition. Furthermore,
we present a detailed analysis of the performance effects of ensembling, robust
estimation, network architecture variations, and the use of rotation priors.
Related papers
- Boosting Cross-Domain Point Classification via Distilling Relational Priors from 2D Transformers [59.0181939916084]
Traditional 3D networks mainly focus on local geometric details and ignore the topological structure between local geometries.
We propose a novel Priors Distillation (RPD) method to extract priors from the well-trained transformers on massive images.
Experiments on the PointDA-10 and the Sim-to-Real datasets verify that the proposed method consistently achieves the state-of-the-art performance of UDA for point cloud classification.
arXiv Detail & Related papers (2024-07-26T06:29:09Z) - Achieving Rotation Invariance in Convolution Operations: Shifting from Data-Driven to Mechanism-Assured [18.910817148765176]
This paper designs a set of new convolution operations that are natually invariant to arbitrary rotations.
We compare their performance with previous rotation-invariant convolutional neural networks (RI-CNNs)
The results show that RIConvs significantly improve the accuracy of these CNN backbones, especially when the training data is limited.
arXiv Detail & Related papers (2024-04-17T12:21:57Z) - PNeRFLoc: Visual Localization with Point-based Neural Radiance Fields [54.8553158441296]
We propose a novel visual localization framework, ie, PNeRFLoc, based on a unified point-based representation.
On the one hand, PNeRFLoc supports the initial pose estimation by matching 2D and 3D feature points.
On the other hand, it also enables pose refinement with novel view synthesis using rendering-based optimization.
arXiv Detail & Related papers (2023-12-17T08:30:00Z) - ART-Point: Improving Rotation Robustness of Point Cloud Classifiers via
Adversarial Rotation [89.47574181669903]
In this study, we show that the rotation robustness of point cloud classifiers can also be acquired via adversarial training.
Specifically, our proposed framework named ART-Point regards the rotation of the point cloud as an attack.
We propose a fast one-step optimization to efficiently reach the final robust model.
arXiv Detail & Related papers (2022-03-08T07:20:16Z) - Attentive Rotation Invariant Convolution for Point Cloud-based Large
Scale Place Recognition [11.433270318356675]
We propose an Attentive Rotation Invariant Convolution (ARIConv) in this paper.
We experimentally demonstrate that our model can achieve state-of-the-art performance on large scale place recognition task when the point cloud scans are rotated.
arXiv Detail & Related papers (2021-08-29T09:10:56Z) - Adversarial Feature Augmentation and Normalization for Visual
Recognition [109.6834687220478]
Recent advances in computer vision take advantage of adversarial data augmentation to ameliorate the generalization ability of classification models.
Here, we present an effective and efficient alternative that advocates adversarial augmentation on intermediate feature embeddings.
We validate the proposed approach across diverse visual recognition tasks with representative backbone networks.
arXiv Detail & Related papers (2021-03-22T20:36:34Z) - RoRD: Rotation-Robust Descriptors and Orthographic Views for Local
Feature Matching [32.10261486751993]
We present a novel framework that combines learning of invariant descriptors through data augmentation and viewpoint projection.
We evaluate the effectiveness of the proposed approach on key tasks including pose estimation and visual place recognition.
arXiv Detail & Related papers (2021-03-15T17:40:25Z) - Adjoint Rigid Transform Network: Task-conditioned Alignment of 3D Shapes [86.2129580231191]
Adjoint Rigid Transform (ART) Network is a neural module which can be integrated with a variety of 3D networks.
ART learns to rotate input shapes to a learned canonical orientation, which is crucial for a lot of tasks.
We will release our code and pre-trained models for further research.
arXiv Detail & Related papers (2021-02-01T20:58:45Z) - Learning Rotation-Invariant Representations of Point Clouds Using
Aligned Edge Convolutional Neural Networks [29.3830445533532]
Point cloud analysis is an area of increasing interest due to the development of 3D sensors that are able to rapidly measure the depth of scenes accurately.
Applying deep learning techniques to perform point cloud analysis is non-trivial due to the inability of these methods to generalize to unseen rotations.
To address this limitation, one usually has to augment the training data, which can lead to extra computation and require larger model complexity.
This paper proposes a new neural network called the Aligned Edge Convolutional Neural Network (AECNN) that learns a feature representation of point clouds relative to Local Reference Frames (LRFs)
arXiv Detail & Related papers (2021-01-02T17:36:00Z) - A Rotation-Invariant Framework for Deep Point Cloud Analysis [132.91915346157018]
We introduce a new low-level purely rotation-invariant representation to replace common 3D Cartesian coordinates as the network inputs.
Also, we present a network architecture to embed these representations into features, encoding local relations between points and their neighbors, and the global shape structure.
We evaluate our method on multiple point cloud analysis tasks, including shape classification, part segmentation, and shape retrieval.
arXiv Detail & Related papers (2020-03-16T14:04:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.