On the Implicit Biases of Architecture & Gradient Descent
- URL: http://arxiv.org/abs/2110.04274v1
- Date: Fri, 8 Oct 2021 17:36:37 GMT
- Title: On the Implicit Biases of Architecture & Gradient Descent
- Authors: Jeremy Bernstein and Yisong Yue
- Abstract summary: This paper finds that while typical networks that fit the training data already generalise fairly well, gradient descent can further improve generalisation by selecting networks with a large margin.
New technical tools suggest a nuanced portrait of generalisation involving both the implicit biases of architecture and gradient descent.
- Score: 46.34988166338264
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Do neural networks generalise because of bias in the functions returned by
gradient descent, or bias already present in the network architecture? Por
qu\'e no los dos?
This paper finds that while typical networks that fit the training data
already generalise fairly well, gradient descent can further improve
generalisation by selecting networks with a large margin. This conclusion is
based on a careful study of the behaviour of infinite width networks trained by
Bayesian inference and finite width networks trained by gradient descent. To
measure the implicit bias of architecture, new technical tools are developed to
both analytically bound and consistently estimate the average test error of the
neural network--Gaussian process (NNGP) posterior. This error is found to be
already better than chance, corroborating the findings of Valle-P\'erez et al.
(2019) and underscoring the importance of architecture. Going beyond this
result, this paper finds that test performance can be substantially improved by
selecting a function with much larger margin than is typical under the NNGP
posterior. This highlights a curious fact: minimum a posteriori functions can
generalise best, and gradient descent can select for those functions. In
summary, new technical tools suggest a nuanced portrait of generalisation
involving both the implicit biases of architecture and gradient descent.
Code for this paper is available at: https://github.com/jxbz/implicit-bias/.
Related papers
- Bias of Stochastic Gradient Descent or the Architecture: Disentangling the Effects of Overparameterization of Neural Networks [37.02386277426315]
This paper aims to disentangle the factors that influence generalization by studying random and SGD-optimized networks that achieve zero training error.
We experimentally show, in the low sample regime, that over parameterization in terms of increasing width is beneficial for generalization.
For increasing depth, over parameterization is detrimental for generalization, but random and SGD-optimized networks behave similarly, so this can be attributed to an architectural bias.
arXiv Detail & Related papers (2024-07-04T11:29:50Z) - How to guess a gradient [68.98681202222664]
We show that gradients are more structured than previously thought.
Exploiting this structure can significantly improve gradient-free optimization schemes.
We highlight new challenges in overcoming the large gap between optimizing with exact gradients and guessing the gradients.
arXiv Detail & Related papers (2023-12-07T21:40:44Z) - Implicit Bias of Gradient Descent for Two-layer ReLU and Leaky ReLU
Networks on Nearly-orthogonal Data [66.1211659120882]
The implicit bias towards solutions with favorable properties is believed to be a key reason why neural networks trained by gradient-based optimization can generalize well.
While the implicit bias of gradient flow has been widely studied for homogeneous neural networks (including ReLU and leaky ReLU networks), the implicit bias of gradient descent is currently only understood for smooth neural networks.
arXiv Detail & Related papers (2023-10-29T08:47:48Z) - Automatic Gradient Descent: Deep Learning without Hyperparameters [35.350274248478804]
The architecture of a deep neural network is defined explicitly in terms of the number of layers, the width of each layer and the general network topology.
Paper builds a new framework for deriving objective functions: gradient idea is to transform a Bregman divergence to account for the non gradient structure of neural architecture.
arXiv Detail & Related papers (2023-04-11T12:45:52Z) - Theoretical Characterization of How Neural Network Pruning Affects its
Generalization [131.1347309639727]
This work makes the first attempt to study how different pruning fractions affect the model's gradient descent dynamics and generalization.
It is shown that as long as the pruning fraction is below a certain threshold, gradient descent can drive the training loss toward zero.
More surprisingly, the generalization bound gets better as the pruning fraction gets larger.
arXiv Detail & Related papers (2023-01-01T03:10:45Z) - Implicit Bias in Leaky ReLU Networks Trained on High-Dimensional Data [63.34506218832164]
In this work, we investigate the implicit bias of gradient flow and gradient descent in two-layer fully-connected neural networks with ReLU activations.
For gradient flow, we leverage recent work on the implicit bias for homogeneous neural networks to show that leakyally, gradient flow produces a neural network with rank at most two.
For gradient descent, provided the random variance is small enough, we show that a single step of gradient descent suffices to drastically reduce the rank of the network, and that the rank remains small throughout training.
arXiv Detail & Related papers (2022-10-13T15:09:54Z) - Optimization-Based Separations for Neural Networks [57.875347246373956]
We show that gradient descent can efficiently learn ball indicator functions using a depth 2 neural network with two layers of sigmoidal activations.
This is the first optimization-based separation result where the approximation benefits of the stronger architecture provably manifest in practice.
arXiv Detail & Related papers (2021-12-04T18:07:47Z) - Training Two-Layer ReLU Networks with Gradient Descent is Inconsistent [2.7793394375935088]
We prove that two-layer (Leaky)ReLU networks by e.g. the widely used method proposed by He et al. are not consistent.
arXiv Detail & Related papers (2020-02-12T09:22:45Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.