On the geometric and Riemannian structure of the spaces of group
equivariant non-expansive operators
- URL: http://arxiv.org/abs/2103.02543v2
- Date: Sun, 31 Dec 2023 08:55:23 GMT
- Title: On the geometric and Riemannian structure of the spaces of group
equivariant non-expansive operators
- Authors: Pasquale Cascarano, Patrizio Frosini, Nicola Quercioli and Amir Saki
- Abstract summary: Group equivariant non-expansive operators have been recently proposed as basic components in topological data analysis and deep learning.
We show how a space $mathcalF$ of group equivariant non-expansive operators can be endowed with the structure of a Riemannian manifold.
We also describe a procedure to select a finite set of representative group equivariant non-expansive operators in the considered manifold.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Group equivariant non-expansive operators have been recently proposed as
basic components in topological data analysis and deep learning. In this paper
we study some geometric properties of the spaces of group equivariant operators
and show how a space $\mathcal{F}$ of group equivariant non-expansive operators
can be endowed with the structure of a Riemannian manifold, so making available
the use of gradient descent methods for the minimization of cost functions on
$\mathcal{F}$. As an application of this approach, we also describe a procedure
to select a finite set of representative group equivariant non-expansive
operators in the considered manifold.
Related papers
- Equivariant Manifold Neural ODEs and Differential Invariants [1.6073704837297416]
We develop a manifestly geometric framework for equivariant manifold neural ordinary differential equations (NODEs)
We use it to analyse their modelling capabilities for symmetric data.
arXiv Detail & Related papers (2024-01-25T12:23:22Z) - Lie Group Decompositions for Equivariant Neural Networks [12.139222986297261]
We show how convolution kernels can be parametrized to build models equivariant with respect to affine transformations.
We evaluate the robustness and out-of-distribution generalisation capability of our model on the benchmark affine-invariant classification task.
arXiv Detail & Related papers (2023-10-17T16:04:33Z) - Geometry of Linear Neural Networks: Equivariance and Invariance under
Permutation Groups [0.0]
We investigate the subvariety of functions that are equivariant or invariant under the action of a permutation group.
We draw conclusions for the parameterization and the design of equivariant and invariant linear networks.
arXiv Detail & Related papers (2023-09-24T19:40:15Z) - Deep Learning Symmetries and Their Lie Groups, Algebras, and Subalgebras
from First Principles [55.41644538483948]
We design a deep-learning algorithm for the discovery and identification of the continuous group of symmetries present in a labeled dataset.
We use fully connected neural networks to model the transformations symmetry and the corresponding generators.
Our study also opens the door for using a machine learning approach in the mathematical study of Lie groups and their properties.
arXiv Detail & Related papers (2023-01-13T16:25:25Z) - Equivariant Transduction through Invariant Alignment [71.45263447328374]
We introduce a novel group-equivariant architecture that incorporates a group-in hard alignment mechanism.
We find that our network's structure allows it to develop stronger equivariant properties than existing group-equivariant approaches.
We additionally find that it outperforms previous group-equivariant networks empirically on the SCAN task.
arXiv Detail & Related papers (2022-09-22T11:19:45Z) - Capacity of Group-invariant Linear Readouts from Equivariant
Representations: How Many Objects can be Linearly Classified Under All
Possible Views? [21.06669693699965]
We find that the fraction of separable dichotomies is determined by the dimension of the space that is fixed by the group action.
We show how this relation extends to operations such as convolutions, element-wise nonlinearities, and global and local pooling.
arXiv Detail & Related papers (2021-10-14T15:46:53Z) - Group Equivariant Subsampling [60.53371517247382]
Subsampling is used in convolutional neural networks (CNNs) in the form of pooling or strided convolutions.
We first introduce translation equivariant subsampling/upsampling layers that can be used to construct exact translation equivariant CNNs.
We then generalise these layers beyond translations to general groups, thus proposing group equivariant subsampling/upsampling.
arXiv Detail & Related papers (2021-06-10T16:14:00Z) - A Practical Method for Constructing Equivariant Multilayer Perceptrons
for Arbitrary Matrix Groups [115.58550697886987]
We provide a completely general algorithm for solving for the equivariant layers of matrix groups.
In addition to recovering solutions from other works as special cases, we construct multilayer perceptrons equivariant to multiple groups that have never been tackled before.
Our approach outperforms non-equivariant baselines, with applications to particle physics and dynamical systems.
arXiv Detail & Related papers (2021-04-19T17:21:54Z) - LieTransformer: Equivariant self-attention for Lie Groups [49.9625160479096]
Group equivariant neural networks are used as building blocks of group invariant neural networks.
We extend the scope of the literature to self-attention, that is emerging as a prominent building block of deep learning models.
We propose the LieTransformer, an architecture composed of LieSelfAttention layers that are equivariant to arbitrary Lie groups and their discrete subgroups.
arXiv Detail & Related papers (2020-12-20T11:02:49Z) - On the finite representation of group equivariant operators via
permutant measures [0.0]
We show that each linear $G$-equivariant operator can be produced by a suitable permutant measure.
This result makes available a new method to build linear $G$-equivariant operators in the finite setting.
arXiv Detail & Related papers (2020-08-07T14:25:04Z) - Invariant Feature Coding using Tensor Product Representation [75.62232699377877]
We prove that the group-invariant feature vector contains sufficient discriminative information when learning a linear classifier.
A novel feature model that explicitly consider group action is proposed for principal component analysis and k-means clustering.
arXiv Detail & Related papers (2019-06-05T07:15:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.