Equivariant bifurcation, quadratic equivariants, and symmetry breaking
for the standard representation of $S_n$
- URL: http://arxiv.org/abs/2107.02422v1
- Date: Tue, 6 Jul 2021 06:43:06 GMT
- Title: Equivariant bifurcation, quadratic equivariants, and symmetry breaking
for the standard representation of $S_n$
- Authors: Yossi Arjevani and Michael Field
- Abstract summary: Motivated by questions originating from the study of a class of shallow student-teacher neural networks, methods are developed for the analysis of spurious minima in classes of equivariant dynamics related to neural nets.
It is shown that spurious minima do not arise from spontaneous symmetry breaking but rather through a complex deformation of the landscape geometry that can be encoded by a generic $S_n$-equivariant bifurcation.
Results on generic bifurcation when there are quadratic equivariants are also proved; this work extends and clarifies results of Ihrig & Golubitsky and Chossat, Lauterback &
- Score: 15.711517003382484
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motivated by questions originating from the study of a class of shallow
student-teacher neural networks, methods are developed for the analysis of
spurious minima in classes of gradient equivariant dynamics related to neural
nets. In the symmetric case, methods depend on the generic equivariant
bifurcation theory of irreducible representations of the symmetric group on $n$
symbols, $S_n$; in particular, the standard representation of $S_n$. It is
shown that spurious minima do not arise from spontaneous symmetry breaking but
rather through a complex deformation of the landscape geometry that can be
encoded by a generic $S_n$-equivariant bifurcation. We describe minimal models
for forced symmetry breaking that give a lower bound on the dynamic complexity
involved in the creation of spurious minima when there is no symmetry. Results
on generic bifurcation when there are quadratic equivariants are also proved;
this work extends and clarifies results of Ihrig & Golubitsky and Chossat,
Lauterback & Melbourne on the instability of solutions when there are quadratic
equivariants.
Related papers
- Equivariant score-based generative models provably learn distributions with symmetries efficiently [7.90752151686317]
Empirical studies have demonstrated that incorporating symmetries into generative models can provide better generalization and sampling efficiency.
We provide the first theoretical analysis and guarantees of score-based generative models (SGMs) for learning distributions that are invariant with respect to some group symmetry.
arXiv Detail & Related papers (2024-10-02T05:14:28Z) - Relative Representations: Topological and Geometric Perspectives [53.88896255693922]
Relative representations are an established approach to zero-shot model stitching.
We introduce a normalization procedure in the relative transformation, resulting in invariance to non-isotropic rescalings and permutations.
Second, we propose to deploy topological densification when fine-tuning relative representations, a topological regularization loss encouraging clustering within classes.
arXiv Detail & Related papers (2024-09-17T08:09:22Z) - Equivariant Manifold Neural ODEs and Differential Invariants [1.6073704837297416]
We develop a manifestly geometric framework for equivariant manifold neural ordinary differential equations (NODEs)
We use it to analyse their modelling capabilities for symmetric data.
arXiv Detail & Related papers (2024-01-25T12:23:22Z) - Symmetry Breaking and Equivariant Neural Networks [17.740760773905986]
We introduce a novel notion of'relaxed equiinjection'
We show how to incorporate this relaxation into equivariant multilayer perceptronrons (E-MLPs)
The relevance of symmetry breaking is then discussed in various application domains.
arXiv Detail & Related papers (2023-12-14T15:06:48Z) - Evaluating the Robustness of Interpretability Methods through
Explanation Invariance and Equivariance [72.50214227616728]
Interpretability methods are valuable only if their explanations faithfully describe the explained model.
We consider neural networks whose predictions are invariant under a specific symmetry group.
arXiv Detail & Related papers (2023-04-13T17:59:03Z) - Deep Learning Symmetries and Their Lie Groups, Algebras, and Subalgebras
from First Principles [55.41644538483948]
We design a deep-learning algorithm for the discovery and identification of the continuous group of symmetries present in a labeled dataset.
We use fully connected neural networks to model the transformations symmetry and the corresponding generators.
Our study also opens the door for using a machine learning approach in the mathematical study of Lie groups and their properties.
arXiv Detail & Related papers (2023-01-13T16:25:25Z) - Equivariant Disentangled Transformation for Domain Generalization under
Combination Shift [91.38796390449504]
Combinations of domains and labels are not observed during training but appear in the test environment.
We provide a unique formulation of the combination shift problem based on the concepts of homomorphism, equivariance, and a refined definition of disentanglement.
arXiv Detail & Related papers (2022-08-03T12:31:31Z) - Non-local order parameters for fermion chains via the partial transpose [0.0]
This paper takes up proposals for non-local order parameters defined through anti-unitary symmetries.
For matrix product states, an interpretation of these invariants is provided.
arXiv Detail & Related papers (2022-06-07T13:13:59Z) - Learning Equivariant Energy Based Models with Equivariant Stein
Variational Gradient Descent [80.73580820014242]
We focus on the problem of efficient sampling and learning of probability densities by incorporating symmetries in probabilistic models.
We first introduce Equivariant Stein Variational Gradient Descent algorithm -- an equivariant sampling method based on Stein's identity for sampling from densities with symmetries.
We propose new ways of improving and scaling up training of energy based models.
arXiv Detail & Related papers (2021-06-15T01:35:17Z) - Geometric Deep Learning and Equivariant Neural Networks [0.9381376621526817]
We survey the mathematical foundations of geometric deep learning, focusing on group equivariant and gauge equivariant neural networks.
We develop gauge equivariant convolutional neural networks on arbitrary manifold $mathcalM$ using principal bundles with structure group $K$ and equivariant maps between sections of associated vector bundles.
We analyze several applications of this formalism, including semantic segmentation and object detection networks.
arXiv Detail & Related papers (2021-05-28T15:41:52Z) - Generalized string-nets for unitary fusion categories without
tetrahedral symmetry [77.34726150561087]
We present a general construction of the Levin-Wen model for arbitrary multiplicity-free unitary fusion categories.
We explicitly calculate the matrix elements of the Hamiltonian and, furthermore, show that it has the same properties as the original one.
arXiv Detail & Related papers (2020-04-15T12:21:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.