Hierarchical Learning in Euclidean Neural Networks
- URL: http://arxiv.org/abs/2210.04766v1
- Date: Mon, 10 Oct 2022 15:26:00 GMT
- Title: Hierarchical Learning in Euclidean Neural Networks
- Authors: Joshua A. Rackers and Pranav Rao
- Abstract summary: We study the role of higher order (non-scalar) features in Euclidean Neural Networks (texttte3nn)
We find a natural hierarchy of features by $l$, reminiscent of a multipole expansion.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Equivariant machine learning methods have shown wide success at 3D learning
applications in recent years. These models explicitly build in the reflection,
translation and rotation symmetries of Euclidean space and have facilitated
large advances in accuracy and data efficiency for a range of applications in
the physical sciences. An outstanding question for equivariant models is why
they achieve such larger-than-expected advances in these applications. To probe
this question, we examine the role of higher order (non-scalar) features in
Euclidean Neural Networks (\texttt{e3nn}). We focus on the previously studied
application of \texttt{e3nn} to the problem of electron density prediction,
which allows for a variety of non-scalar outputs, and examine whether the
nature of the output (scalar $l=0$, vector $l=1$, or higher order $l>1$) is
relevant to the effectiveness of non-scalar hidden features in the network.
Further, we examine the behavior of non-scalar features throughout training,
finding a natural hierarchy of features by $l$, reminiscent of a multipole
expansion. We aim for our work to ultimately inform design principles and
choices of domain applications for {\tt e3nn} networks.
Related papers
- Global Convergence and Rich Feature Learning in $L$-Layer Infinite-Width Neural Networks under $μ$P Parametrization [66.03821840425539]
In this paper, we investigate the training dynamics of $L$-layer neural networks using the tensor gradient program (SGD) framework.
We show that SGD enables these networks to learn linearly independent features that substantially deviate from their initial values.
This rich feature space captures relevant data information and ensures that any convergent point of the training process is a global minimum.
arXiv Detail & Related papers (2025-03-12T17:33:13Z) - Deep Neural Networks are Adaptive to Function Regularity and Data Distribution in Approximation and Estimation [8.284464143581546]
We study how deep neural networks can adapt to different regularity in functions across different locations and scales.
Our results show that deep neural networks are adaptive to different regularity of functions and nonuniform data distributions.
arXiv Detail & Related papers (2024-06-08T02:01:50Z) - A singular Riemannian Geometry Approach to Deep Neural Networks III. Piecewise Differentiable Layers and Random Walks on $n$-dimensional Classes [49.32130498861987]
We study the case of non-differentiable activation functions, such as ReLU.
Two recent works introduced a geometric framework to study neural networks.
We illustrate our findings with some numerical experiments on classification of images and thermodynamic problems.
arXiv Detail & Related papers (2024-04-09T08:11:46Z) - Asymptotics of Learning with Deep Structured (Random) Features [9.366617422860543]
For a large class of feature maps we provide a tight characterisation of the test error associated with learning the readout layer.
In some cases our results can capture feature maps learned by deep, finite-width neural networks trained under gradient descent.
arXiv Detail & Related papers (2024-02-21T18:35:27Z) - Neural approximation of Wasserstein distance via a universal
architecture for symmetric and factorwise group invariant functions [6.994580267603235]
We first present a general neural network architecture for approximating SFGI functions.
The main contribution of this paper combines this general neural network with a sketching idea to develop a specific and efficient neural network.
Our work provides an interesting integration of sketching ideas for geometric problems with universal approximation of symmetric functions.
arXiv Detail & Related papers (2023-08-01T04:11:19Z) - A Unified Algebraic Perspective on Lipschitz Neural Networks [88.14073994459586]
This paper introduces a novel perspective unifying various types of 1-Lipschitz neural networks.
We show that many existing techniques can be derived and generalized via finding analytical solutions of a common semidefinite programming (SDP) condition.
Our approach, called SDP-based Lipschitz Layers (SLL), allows us to design non-trivial yet efficient generalization of convex potential layers.
arXiv Detail & Related papers (2023-03-06T14:31:09Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Spelunking the Deep: Guaranteed Queries for General Neural Implicit
Surfaces [35.438964954948574]
This work presents a new approach to perform queries directly on general neural implicit functions for a wide range of existing architectures.
Our key tool is the application of range analysis to neural networks, using automatic arithmetic rules to bound the output of a network over a region.
We use the resulting bounds to develop queries including ray casting, intersection testing, constructing spatial hierarchies, fast mesh extraction, closest-point evaluation.
arXiv Detail & Related papers (2022-02-05T00:37:08Z) - A singular Riemannian geometry approach to Deep Neural Networks II.
Reconstruction of 1-D equivalence classes [78.120734120667]
We build the preimage of a point in the output manifold in the input space.
We focus for simplicity on the case of neural networks maps from n-dimensional real spaces to (n - 1)-dimensional real spaces.
arXiv Detail & Related papers (2021-12-17T11:47:45Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - Deep Polynomial Neural Networks [77.70761658507507]
$Pi$Nets are a new class of function approximators based on expansions.
$Pi$Nets produce state-the-art results in three challenging tasks, i.e. image generation, face verification and 3D mesh representation learning.
arXiv Detail & Related papers (2020-06-20T16:23:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.