Group Equivariant Conditional Neural Processes
- URL: http://arxiv.org/abs/2102.08759v1
- Date: Wed, 17 Feb 2021 13:50:07 GMT
- Title: Group Equivariant Conditional Neural Processes
- Authors: Makoto Kawano, Wataru Kumagai, Akiyoshi Sannai, Yusuke Iwasawa and
Yutaka Matsuo
- Abstract summary: We present the group equivariant conditional neural process (EquivCNP)
We show that EquivCNP achieves comparable performance to conventional conditional neural processes in a 1D regression task.
- Score: 30.134634059773703
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present the group equivariant conditional neural process (EquivCNP), a
meta-learning method with permutation invariance in a data set as in
conventional conditional neural processes (CNPs), and it also has
transformation equivariance in data space. Incorporating group equivariance,
such as rotation and scaling equivariance, provides a way to consider the
symmetry of real-world data. We give a decomposition theorem for
permutation-invariant and group-equivariant maps, which leads us to construct
EquivCNPs with an infinite-dimensional latent space to handle group symmetries.
In this paper, we build architecture using Lie group convolutional layers for
practical implementation. We show that EquivCNP with translation equivariance
achieves comparable performance to conventional CNPs in a 1D regression task.
Moreover, we demonstrate that incorporating an appropriate Lie group
equivariance, EquivCNP is capable of zero-shot generalization for an
image-completion task by selecting an appropriate Lie group equivariance.
Related papers
- Variational Partial Group Convolutions for Input-Aware Partial Equivariance of Rotations and Color-Shifts [21.397064770689795]
Group Equivariant CNNs (G-CNNs) have shown promising efficacy in various tasks, owing to their ability to capture hierarchical features in an equivariant manner.
We propose a novel approach, Variational Partial G-CNN (VP G-CNN), to capture varying levels of partial equivariance specific to each data instance.
We demonstrate the effectiveness of VP G-CNN on both toy and real-world datasets, including M67-180, CIFAR10, ColorMNIST, and Flowers102.
arXiv Detail & Related papers (2024-07-05T05:52:51Z) - Equivariant Representation Learning in the Presence of Stabilizers [13.11108596589607]
EquIN is suitable for group actions that are not free, i.e., that stabilize data via nontrivial symmetries.
EquIN is theoretically grounded in the orbit-stabilizer theorem from group theory.
arXiv Detail & Related papers (2023-01-12T11:43:26Z) - Architectural Optimization over Subgroups for Equivariant Neural
Networks [0.0]
We propose equivariance relaxation morphism and $[G]$-mixed equivariant layer to operate with equivariance constraints on a subgroup.
We present evolutionary and differentiable neural architecture search (NAS) algorithms that utilize these mechanisms respectively for equivariance-aware architectural optimization.
arXiv Detail & Related papers (2022-10-11T14:37:29Z) - Equivariant Disentangled Transformation for Domain Generalization under
Combination Shift [91.38796390449504]
Combinations of domains and labels are not observed during training but appear in the test environment.
We provide a unique formulation of the combination shift problem based on the concepts of homomorphism, equivariance, and a refined definition of disentanglement.
arXiv Detail & Related papers (2022-08-03T12:31:31Z) - Equivariance Discovery by Learned Parameter-Sharing [153.41877129746223]
We study how to discover interpretable equivariances from data.
Specifically, we formulate this discovery process as an optimization problem over a model's parameter-sharing schemes.
Also, we theoretically analyze the method for Gaussian data and provide a bound on the mean squared gap between the studied discovery scheme and the oracle scheme.
arXiv Detail & Related papers (2022-04-07T17:59:19Z) - Learning Equivariances and Partial Equivariances from Data [11.548853370822345]
Group equivariant Convolutional Neural Networks (G-CNNs) constrain features to respect the chosen symmetries, and lead to better generalization when these symmetries appear in the data.
We introduce Partial G-CNNs: a family of equivariant networks able to learn partial and full equivariances from data at every layer end-to-end.
arXiv Detail & Related papers (2021-10-19T19:17:32Z) - Group Equivariant Subsampling [60.53371517247382]
Subsampling is used in convolutional neural networks (CNNs) in the form of pooling or strided convolutions.
We first introduce translation equivariant subsampling/upsampling layers that can be used to construct exact translation equivariant CNNs.
We then generalise these layers beyond translations to general groups, thus proposing group equivariant subsampling/upsampling.
arXiv Detail & Related papers (2021-06-10T16:14:00Z) - Commutative Lie Group VAE for Disentanglement Learning [96.32813624341833]
We view disentanglement learning as discovering an underlying structure that equivariantly reflects the factorized variations shown in data.
A simple model named Commutative Lie Group VAE is introduced to realize the group-based disentanglement learning.
Experiments show that our model can effectively learn disentangled representations without supervision, and can achieve state-of-the-art performance without extra constraints.
arXiv Detail & Related papers (2021-06-07T07:03:14Z) - LieTransformer: Equivariant self-attention for Lie Groups [49.9625160479096]
Group equivariant neural networks are used as building blocks of group invariant neural networks.
We extend the scope of the literature to self-attention, that is emerging as a prominent building block of deep learning models.
We propose the LieTransformer, an architecture composed of LieSelfAttention layers that are equivariant to arbitrary Lie groups and their discrete subgroups.
arXiv Detail & Related papers (2020-12-20T11:02:49Z) - Generalizing Convolutional Neural Networks for Equivariance to Lie
Groups on Arbitrary Continuous Data [52.78581260260455]
We propose a general method to construct a convolutional layer that is equivariant to transformations from any specified Lie group.
We apply the same model architecture to images, ball-and-stick molecular data, and Hamiltonian dynamical systems.
arXiv Detail & Related papers (2020-02-25T17:40:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.