The Landscape of Multi-Layer Linear Neural Network From the Perspective
of Algebraic Geometry
- URL: http://arxiv.org/abs/2102.04338v1
- Date: Sat, 30 Jan 2021 04:50:45 GMT
- Title: The Landscape of Multi-Layer Linear Neural Network From the Perspective
of Algebraic Geometry
- Authors: Xiuyi Yang
- Abstract summary: The clear understanding of the non-dual landscape of neural network is a complex incomplete problem.
By treating the gradient equations as equations, we use algebraic geometry tools to solve it.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The clear understanding of the non-convex landscape of neural network is a
complex incomplete problem. This paper studies the landscape of linear
(residual) network, the simplified version of the nonlinear network. By
treating the gradient equations as polynomial equations, we use algebraic
geometry tools to solve it over the complex number field, the attained solution
can be decomposed into different irreducible complex geometry objects. Then
three hypotheses are proposed, involving how to calculate the loss on each
irreducible geometry object, the losses of critical points have a certain range
and the relationship between the dimension of each irreducible geometry object
and strict saddle condition. Finally, numerical algebraic geometry is applied
to verify the rationality of these three hypotheses which further clarify the
landscape of linear network and the role of residual connection.
Related papers
- Fuse, Reason and Verify: Geometry Problem Solving with Parsed Clauses from Diagram [78.79651421493058]
We propose a neural-symbolic model for plane geometry problem solving (PGPS) with three key steps: modal fusion, reasoning process and knowledge verification.
For reasoning, we design an explicable solution program to describe the geometric reasoning process, and employ a self-limited decoder to generate solution program autoregressively.
We also construct a large-scale geometry problem dataset called PGPS9K, containing fine-grained annotations of textual clauses, solution program and involved knowledge solvers.
arXiv Detail & Related papers (2024-07-10T02:45:22Z) - Geometry-Informed Neural Networks [15.27249535281444]
We introduce geometry-informed neural networks (GINNs)
GINNs are a framework for training shape-generative neural fields without data.
We apply GINNs to several validation problems and a realistic 3D engineering design problem.
arXiv Detail & Related papers (2024-02-21T18:50:12Z) - Adaptive Surface Normal Constraint for Geometric Estimation from Monocular Images [56.86175251327466]
We introduce a novel approach to learn geometries such as depth and surface normal from images while incorporating geometric context.
Our approach extracts geometric context that encodes the geometric variations present in the input image and correlates depth estimation with geometric constraints.
Our method unifies depth and surface normal estimations within a cohesive framework, which enables the generation of high-quality 3D geometry from images.
arXiv Detail & Related papers (2024-02-08T17:57:59Z) - Algebraic Complexity and Neurovariety of Linear Convolutional Networks [0.0]
We study linear convolutional networks with one-dimensional and arbitrary strides.
We generate equations whose common zero locus corresponds to the Zariski closure of the corresponding neuromanifold.
Our findings reveal that the number of all complex critical points in the optimization of such a network is equal to the generic Euclidean distance of a Segre variety.
arXiv Detail & Related papers (2024-01-29T23:00:15Z) - Function Space and Critical Points of Linear Convolutional Networks [4.483341215742946]
We study the geometry of linear networks with one-dimensional convolutional layers.
We analyze the impact of the network's architecture on the function space's dimension, boundary, and singular points.
arXiv Detail & Related papers (2023-04-12T10:15:17Z) - Curved Geometric Networks for Visual Anomaly Recognition [39.91252195360767]
Learning a latent embedding to understand the underlying nature of data distribution is often formulated in Euclidean spaces with zero curvature.
In this work, we investigate benefits of the curved space for analyzing anomalies or out-of-distribution objects in data.
arXiv Detail & Related papers (2022-08-02T01:15:39Z) - Differential Geometry in Neural Implicits [0.6198237241838558]
We introduce a neural implicit framework that bridges discrete differential geometry of triangle meshes and continuous differential geometry of neural implicit surfaces.
It exploits the differentiable properties of neural networks and the discrete geometry of triangle meshes to approximate them as the zero-level sets of neural implicit functions.
arXiv Detail & Related papers (2022-01-23T13:40:45Z) - Multi-view 3D Reconstruction of a Texture-less Smooth Surface of Unknown
Generic Reflectance [86.05191217004415]
Multi-view reconstruction of texture-less objects with unknown surface reflectance is a challenging task.
This paper proposes a simple and robust solution to this problem based on a co-light scanner.
arXiv Detail & Related papers (2021-05-25T01:28:54Z) - ResNet-LDDMM: Advancing the LDDMM Framework Using Deep Residual Networks [86.37110868126548]
In this work, we make use of deep residual neural networks to solve the non-stationary ODE (flow equation) based on a Euler's discretization scheme.
We illustrate these ideas on diverse registration problems of 3D shapes under complex topology-preserving transformations.
arXiv Detail & Related papers (2021-02-16T04:07:13Z) - Eigendecomposition-Free Training of Deep Networks for Linear
Least-Square Problems [107.3868459697569]
We introduce an eigendecomposition-free approach to training a deep network.
We show that our approach is much more robust than explicit differentiation of the eigendecomposition.
Our method has better convergence properties and yields state-of-the-art results.
arXiv Detail & Related papers (2020-04-15T04:29:34Z) - On the Decision Boundaries of Neural Networks: A Tropical Geometry
Perspective [54.1171355815052]
This work tackles the problem of characterizing and understanding the decision boundaries of neural networks with piecewise linear non-linearity activations.
We use tropical geometry, a new development in the area of algebraic geometry, to characterize the decision boundaries of a simple network of the form (Affine, ReLU, Affine)
arXiv Detail & Related papers (2020-02-20T16:22:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.