Rotation-Invariant Autoencoders for Signals on Spheres
- URL: http://arxiv.org/abs/2012.04474v1
- Date: Tue, 8 Dec 2020 15:15:03 GMT
- Title: Rotation-Invariant Autoencoders for Signals on Spheres
- Authors: Suhas Lohit, Shubhendu Trivedi
- Abstract summary: We study the problem of unsupervised learning of rotation-invariant representations for spherical images.
In particular, we design an autoencoder architecture consisting of $S2$ and $SO(3)$ convolutional layers.
Experiments on multiple datasets demonstrate the usefulness of the learned representations on clustering, retrieval and classification applications.
- Score: 10.406659081400354
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Omnidirectional images and spherical representations of $3D$ shapes cannot be
processed with conventional 2D convolutional neural networks (CNNs) as the
unwrapping leads to large distortion. Using fast implementations of spherical
and $SO(3)$ convolutions, researchers have recently developed deep learning
methods better suited for classifying spherical images. These newly proposed
convolutional layers naturally extend the notion of convolution to functions on
the unit sphere $S^2$ and the group of rotations $SO(3)$ and these layers are
equivariant to 3D rotations. In this paper, we consider the problem of
unsupervised learning of rotation-invariant representations for spherical
images. In particular, we carefully design an autoencoder architecture
consisting of $S^2$ and $SO(3)$ convolutional layers. As 3D rotations are often
a nuisance factor, the latent space is constrained to be exactly invariant to
these input transformations. As the rotation information is discarded in the
latent space, we craft a novel rotation-invariant loss function for training
the network. Extensive experiments on multiple datasets demonstrate the
usefulness of the learned representations on clustering, retrieval and
classification applications.
Related papers
- Category-Level 6D Object Pose Estimation with Flexible Vector-Based
Rotation Representation [51.67545893892129]
We propose a novel 3D graph convolution based pipeline for category-level 6D pose and size estimation from monocular RGB-D images.
We first design an orientation-aware autoencoder with 3D graph convolution for latent feature learning.
Then, to efficiently decode the rotation information from the latent feature, we design a novel flexible vector-based decomposable rotation representation.
arXiv Detail & Related papers (2022-12-09T02:13:43Z) - TetraSphere: A Neural Descriptor for O(3)-Invariant Point Cloud Analysis [19.322295753674844]
We present a learnable descriptor invariant under 3D rotations and reflections, i.e., the O(3) actions.
We propose an embedding of the 3D spherical neurons into 4D vector neurons, which leverages end-to-end training of the model.
Our results reveal the practical value of steerable 3D spherical neurons for learning in 3D Euclidean space.
arXiv Detail & Related papers (2022-11-26T02:15:35Z) - Approximate Equivariance SO(3) Needlet Convolution [3.1050801937854504]
This paper develops a rotation-invariant needlet convolution for rotation group SO(3) to distill multiscale information of spherical signals.
The network establishes a powerful tool to extract geometric-invariant features of spherical signals.
The NES achieves state-of-the-art performance for quantum chemistry regression and Cosmic Microwave Background (CMB) delensing reconstruction.
arXiv Detail & Related papers (2022-06-17T17:21:56Z) - VNT-Net: Rotational Invariant Vector Neuron Transformers [3.04585143845864]
We introduce a rotational invariant neural network by combining recently introduced vector neurons with self-attention layers.
Experiments demonstrate that our network efficiently handles 3D point cloud objects in arbitrary poses.
arXiv Detail & Related papers (2022-05-19T16:51:56Z) - 3D Equivariant Graph Implicit Functions [51.5559264447605]
We introduce a novel family of graph implicit functions with equivariant layers that facilitates modeling fine local details.
Our method improves over the existing rotation-equivariant implicit function from 0.69 to 0.89 on the ShapeNet reconstruction task.
arXiv Detail & Related papers (2022-03-31T16:51:25Z) - Rotation Equivariant 3D Hand Mesh Generation from a Single RGB Image [1.8692254863855962]
We develop a rotation equivariant model for generating 3D hand meshes from 2D RGB images.
This guarantees that as the input image of a hand is rotated the generated mesh undergoes a corresponding rotation.
arXiv Detail & Related papers (2021-11-25T11:07:27Z) - Concentric Spherical GNN for 3D Representation Learning [53.45704095146161]
We propose a novel multi-resolution convolutional architecture for learning over concentric spherical feature maps.
Our hierarchical architecture is based on alternatively learning to incorporate both intra-sphere and inter-sphere information.
We demonstrate the effectiveness of our approach in improving state-of-the-art performance on 3D classification tasks with rotated data.
arXiv Detail & Related papers (2021-03-18T19:05:04Z) - Adjoint Rigid Transform Network: Task-conditioned Alignment of 3D Shapes [86.2129580231191]
Adjoint Rigid Transform (ART) Network is a neural module which can be integrated with a variety of 3D networks.
ART learns to rotate input shapes to a learned canonical orientation, which is crucial for a lot of tasks.
We will release our code and pre-trained models for further research.
arXiv Detail & Related papers (2021-02-01T20:58:45Z) - Rotation-Invariant Point Convolution With Multiple Equivariant
Alignments [1.0152838128195467]
We show that using rotation-equivariant alignments, it is possible to make any convolutional layer rotation-invariant.
With this core layer, we design rotation-invariant architectures which improve state-of-the-art results in both object classification and semantic segmentation.
arXiv Detail & Related papers (2020-12-07T20:47:46Z) - Spin-Weighted Spherical CNNs [58.013031812072356]
We present a new type of spherical CNN that allows anisotropic filters in an efficient way, without ever leaving the sphere domain.
The key idea is to consider spin-weighted spherical functions, which were introduced in physics in the study of gravitational waves.
Our method outperforms previous methods on tasks like classification of spherical images, classification of 3D shapes and semantic segmentation of spherical panoramas.
arXiv Detail & Related papers (2020-06-18T17:57:21Z) - A Rotation-Invariant Framework for Deep Point Cloud Analysis [132.91915346157018]
We introduce a new low-level purely rotation-invariant representation to replace common 3D Cartesian coordinates as the network inputs.
Also, we present a network architecture to embed these representations into features, encoding local relations between points and their neighbors, and the global shape structure.
We evaluate our method on multiple point cloud analysis tasks, including shape classification, part segmentation, and shape retrieval.
arXiv Detail & Related papers (2020-03-16T14:04:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.