Embedding Principle in Depth for the Loss Landscape Analysis of Deep
Neural Networks
- URL: http://arxiv.org/abs/2205.13283v1
- Date: Thu, 26 May 2022 11:42:44 GMT
- Title: Embedding Principle in Depth for the Loss Landscape Analysis of Deep
Neural Networks
- Authors: Zhiwei Bai, Tao Luo, Zhi-Qin John Xu, Yaoyu Zhang
- Abstract summary: We prove an embedding principle in depth that loss landscape of an NN "contains" all critical points of the loss landscapes for shallower NNs.
We empirically demonstrate that, through suppressing layer linearization, batch normalization helps avoid the lifted critical manifold.
- Score: 3.5208869573271446
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unraveling the general structure underlying the loss landscapes of deep
neural networks (DNNs) is important for the theoretical study of deep learning.
Inspired by the embedding principle of DNN loss landscape, we prove in this
work an embedding principle in depth that loss landscape of an NN "contains"
all critical points of the loss landscapes for shallower NNs. Specifically, we
propose a critical lifting operator that any critical point of a shallower
network can be lifted to a critical manifold of the target network while
preserving the outputs. Through lifting, local minimum of an NN can become a
strict saddle point of a deeper NN, which can be easily escaped by first-order
methods. The embedding principle in depth reveals a large family of critical
points in which layer linearization happens, i.e., computation of certain
layers is effectively linear for the training inputs. We empirically
demonstrate that, through suppressing layer linearization, batch normalization
helps avoid the lifted critical manifolds, resulting in a faster decay of loss.
We also demonstrate that increasing training data reduces the lifted critical
manifold thus could accelerate the training. Overall, the embedding principle
in depth well complements the embedding principle (in width), resulting in a
complete characterization of the hierarchical structure of critical
points/manifolds of a DNN loss landscape.
Related papers
- Component-based Sketching for Deep ReLU Nets [55.404661149594375]
We develop a sketching scheme based on deep net components for various tasks.
We transform deep net training into a linear empirical risk minimization problem.
We show that the proposed component-based sketching provides almost optimal rates in approximating saturated functions.
arXiv Detail & Related papers (2024-09-21T15:30:43Z) - Peeking Behind the Curtains of Residual Learning [10.915277646160707]
"The Plain Neural Net Hypothesis" (PNNH) identifies the internal path across non-linear layers as the most critical part in residual learning.
We thoroughly evaluate PNNH-enabled CNN architectures and Transformers on popular vision benchmarks, showing on-par accuracy, up to 0.3% higher training throughput, and 2x better parameter efficiency compared to ResNets and vision Transformers.
arXiv Detail & Related papers (2024-02-13T18:24:10Z) - Addressing caveats of neural persistence with deep graph persistence [54.424983583720675]
We find that the variance of network weights and spatial concentration of large weights are the main factors that impact neural persistence.
We propose an extension of the filtration underlying neural persistence to the whole neural network instead of single layers.
This yields our deep graph persistence measure, which implicitly incorporates persistent paths through the network and alleviates variance-related issues.
arXiv Detail & Related papers (2023-07-20T13:34:11Z) - Rank Diminishing in Deep Neural Networks [71.03777954670323]
Rank of neural networks measures information flowing across layers.
It is an instance of a key structural condition that applies across broad domains of machine learning.
For neural networks, however, the intrinsic mechanism that yields low-rank structures remains vague and unclear.
arXiv Detail & Related papers (2022-06-13T12:03:32Z) - Exact Solutions of a Deep Linear Network [2.2344764434954256]
This work finds the analytical expression of the global minima of a deep linear network with weight decay and neurons.
We show that weight decay strongly interacts with the model architecture and can create bad minima at zero in a network with more than $1$ hidden layer.
arXiv Detail & Related papers (2022-02-10T00:13:34Z) - Embedding Principle: a hierarchical structure of loss landscape of deep
neural networks [3.0871079010101963]
We prove a general Embedding Principle of loss landscape of deep neural networks (NNs)
We provide a gross estimate of the dimension of critical submanifolds embedded from critical points of narrower NNs.
arXiv Detail & Related papers (2021-11-30T16:15:50Z) - Embedding Principle of Loss Landscape of Deep Neural Networks [1.1958610985612828]
We show that the loss landscape of a deep neural network (DNN) "contains" all the critical principle of all DNNs.
We find that a wide DNN is often embedded by highlydegenerate critical points that are embedded from narrow DNNs.
arXiv Detail & Related papers (2021-05-30T15:32:32Z) - Statistical Mechanics of Deep Linear Neural Networks: The
Back-Propagating Renormalization Group [4.56877715768796]
We study the statistical mechanics of learning in Deep Linear Neural Networks (DLNNs) in which the input-output function of an individual unit is linear.
We solve exactly the network properties following supervised learning using an equilibrium Gibbs distribution in the weight space.
Our numerical simulations reveal that despite the nonlinearity, the predictions of our theory are largely shared by ReLU networks with modest depth.
arXiv Detail & Related papers (2020-12-07T20:08:31Z) - Learning Connectivity of Neural Networks from a Topological Perspective [80.35103711638548]
We propose a topological perspective to represent a network into a complete graph for analysis.
By assigning learnable parameters to the edges which reflect the magnitude of connections, the learning process can be performed in a differentiable manner.
This learning process is compatible with existing networks and owns adaptability to larger search spaces and different tasks.
arXiv Detail & Related papers (2020-08-19T04:53:31Z) - Revisiting Initialization of Neural Networks [72.24615341588846]
We propose a rigorous estimation of the global curvature of weights across layers by approximating and controlling the norm of their Hessian matrix.
Our experiments on Word2Vec and the MNIST/CIFAR image classification tasks confirm that tracking the Hessian norm is a useful diagnostic tool.
arXiv Detail & Related papers (2020-04-20T18:12:56Z) - Robust Pruning at Initialization [61.30574156442608]
A growing need for smaller, energy-efficient, neural networks to be able to use machine learning applications on devices with limited computational resources.
For Deep NNs, such procedures remain unsatisfactory as the resulting pruned networks can be difficult to train and, for instance, they do not prevent one layer from being fully pruned.
arXiv Detail & Related papers (2020-02-19T17:09:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.