Dynamically Stable Poincar\'e Embeddings for Neural Manifolds
- URL: http://arxiv.org/abs/2112.11172v1
- Date: Tue, 21 Dec 2021 13:09:08 GMT
- Title: Dynamically Stable Poincar\'e Embeddings for Neural Manifolds
- Authors: Jun Chen, Yuang Liu, Xiangrui Zhao, Yong Liu
- Abstract summary: We prove that if initial metrics have an $L2$-norm perturbation which deviates from the Hyperbolic metric on the Poincar'e ball, the scaled Ricci-DeTurck flow of such metrics smoothly and exponentially converges to the Hyperbolic metric.
Specifically, the role of the Ricci flow is to serve as naturally evolving to the stable Poincar'e ball that will then be mapped back to the Euclidean space.
- Score: 10.76554740227876
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In a Riemannian manifold, the Ricci flow is a partial differential equation
for evolving the metric to become more regular. We hope that topological
structures from such metrics may be used to assist in the tasks of machine
learning. However, this part of the work is still missing. In this paper, we
bridge this gap between the Ricci flow and deep neural networks by dynamically
stable Poincar\'e embeddings for neural manifolds. As a result, we prove that,
if initial metrics have an $L^2$-norm perturbation which deviates from the
Hyperbolic metric on the Poincar\'e ball, the scaled Ricci-DeTurck flow of such
metrics smoothly and exponentially converges to the Hyperbolic metric.
Specifically, the role of the Ricci flow is to serve as naturally evolving to
the stable Poincar\'e ball that will then be mapped back to the Euclidean
space. For such dynamically stable neural manifolds under the Ricci flow, the
convergence of neural networks embedded with such manifolds is not susceptible
to perturbations. And we show that such Ricci flow assisted neural networks
outperform with their all Euclidean versions on image classification tasks
(CIFAR datasets).
Related papers
- von Mises Quasi-Processes for Bayesian Circular Regression [57.88921637944379]
We explore a family of expressive and interpretable distributions over circle-valued random functions.
The resulting probability model has connections with continuous spin models in statistical physics.
For posterior inference, we introduce a new Stratonovich-like augmentation that lends itself to fast Markov Chain Monte Carlo sampling.
arXiv Detail & Related papers (2024-06-19T01:57:21Z) - Enhancing lattice kinetic schemes for fluid dynamics with Lattice-Equivariant Neural Networks [79.16635054977068]
We present a new class of equivariant neural networks, dubbed Lattice-Equivariant Neural Networks (LENNs)
Our approach develops within a recently introduced framework aimed at learning neural network-based surrogate models Lattice Boltzmann collision operators.
Our work opens towards practical utilization of machine learning-augmented Lattice Boltzmann CFD in real-world simulations.
arXiv Detail & Related papers (2024-05-22T17:23:15Z) - Metric Flows with Neural Networks [0.0]
We develop a theory of flows in the space of Riemannian metrics induced by neural network gradient descent.
This is motivated in part by advances in approximating Calabi-Yau metrics with neural networks.
We demonstrate that well-learned numerical metrics at finite-width exhibit an evolving metric-NTK, associated with feature learning.
arXiv Detail & Related papers (2023-10-30T18:00:01Z) - Machine learning in and out of equilibrium [58.88325379746631]
Our study uses a Fokker-Planck approach, adapted from statistical physics, to explore these parallels.
We focus in particular on the stationary state of the system in the long-time limit, which in conventional SGD is out of equilibrium.
We propose a new variation of Langevin dynamics (SGLD) that harnesses without replacement minibatching.
arXiv Detail & Related papers (2023-06-06T09:12:49Z) - Scalable Stochastic Gradient Riemannian Langevin Dynamics in Non-Diagonal Metrics [3.8811062755861956]
We propose two non-diagonal metrics that can be used in-gradient samplers to improve convergence and exploration.
We show that for fully connected neural networks (NNs) with sparsity-inducing priors and convolutional NNs with correlated priors, using these metrics can provide improvements.
arXiv Detail & Related papers (2023-03-09T08:20:28Z) - Learning Discretized Neural Networks under Ricci Flow [51.36292559262042]
We study Discretized Neural Networks (DNNs) composed of low-precision weights and activations.
DNNs suffer from either infinite or zero gradients due to the non-differentiable discrete function during training.
arXiv Detail & Related papers (2023-02-07T10:51:53Z) - A physics-informed search for metric solutions to Ricci flow, their
embeddings, and visualisation [0.0]
Neural networks with PDEs embedded in their loss functions are employed as a function approximators.
A general method is developed and applied to the real torus.
The validity of the solution is verified by comparing the time evolution of scalar curvature with that found using a standard PDE solver.
arXiv Detail & Related papers (2022-11-30T08:17:06Z) - Thoughts on the Consistency between Ricci Flow and Neural Network
Behavior [11.912554495037362]
In this paper, we propose the linearly nearly Euclidean metric to assist manifold micro-surgery.
We prove the dynamical stability and convergence of the metrics close to the linearly nearly Euclidean metric under the Ricci-DeTurck flow.
arXiv Detail & Related papers (2021-11-16T12:23:09Z) - A Unifying and Canonical Description of Measure-Preserving Diffusions [60.59592461429012]
A complete recipe of measure-preserving diffusions in Euclidean space was recently derived unifying several MCMC algorithms into a single framework.
We develop a geometric theory that improves and generalises this construction to any manifold.
arXiv Detail & Related papers (2021-05-06T17:36:55Z) - Hyperbolic Neural Networks++ [66.16106727715061]
We generalize the fundamental components of neural networks in a single hyperbolic geometry model, namely, the Poincar'e ball model.
Experiments show the superior parameter efficiency of our methods compared to conventional hyperbolic components, and stability and outperformance over their Euclidean counterparts.
arXiv Detail & Related papers (2020-06-15T08:23:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.