NeurIPS 2020 Competition: Predicting Generalization in Deep Learning
- URL: http://arxiv.org/abs/2012.07976v1
- Date: Mon, 14 Dec 2020 22:21:37 GMT
- Title: NeurIPS 2020 Competition: Predicting Generalization in Deep Learning
- Authors: Yiding Jiang (1), Pierre Foret (1), Scott Yak (1), Daniel M. Roy (2),
Hossein Mobahi (1), Gintare Karolina Dziugaite (3), Samy Bengio (1), Suriya
Gunasekar (4), Isabelle Guyon (5), Behnam Neyshabur (1) ((1) Google Research,
(2) University of Toronto, (3) Element AI, (4) Microsoft Research, (5)
University Paris-Saclay and ChaLearn)
- Abstract summary: Understanding generalization in deep learning is arguably one of the most important questions in deep learning.
We invite the community to propose complexity measures that can accurately predict generalization of models.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding generalization in deep learning is arguably one of the most
important questions in deep learning. Deep learning has been successfully
adopted to a large number of problems ranging from pattern recognition to
complex decision making, but many recent researchers have raised many concerns
about deep learning, among which the most important is generalization. Despite
numerous attempts, conventional statistical learning approaches have yet been
able to provide a satisfactory explanation on why deep learning works. A recent
line of works aims to address the problem by trying to predict the
generalization performance through complexity measures. In this competition, we
invite the community to propose complexity measures that can accurately predict
generalization of models. A robust and general complexity measure would
potentially lead to a better understanding of deep learning's underlying
mechanism and behavior of deep models on unseen data, or shed light on better
generalization bounds. All these outcomes will be important for making deep
learning more robust and reliable.
Related papers
- A Survey Analyzing Generalization in Deep Reinforcement Learning [14.141453107129403]
We will formalize and analyze generalization in deep reinforcement learning.
We will explain the fundamental reasons why deep reinforcement learning policies encounter overfitting problems that limit their generalization capabilities.
arXiv Detail & Related papers (2024-01-04T16:45:01Z) - A Comprehensive Survey of Continual Learning: Theory, Method and
Application [64.23253420555989]
We present a comprehensive survey of continual learning, seeking to bridge the basic settings, theoretical foundations, representative methods, and practical applications.
We summarize the general objectives of continual learning as ensuring a proper stability-plasticity trade-off and an adequate intra/inter-task generalizability in the context of resource efficiency.
arXiv Detail & Related papers (2023-01-31T11:34:56Z) - Generalization Through the Lens of Learning Dynamics [11.009483845261958]
A machine learning (ML) system must learn to generalize to novel situations in order to yield accurate predictions at deployment.
The impressive generalization performance of deep neural networks has stymied theoreticians.
This thesis will study the learning dynamics of deep neural networks in both supervised and reinforcement learning tasks.
arXiv Detail & Related papers (2022-12-11T00:07:24Z) - PAC-Bayes Compression Bounds So Tight That They Can Explain
Generalization [48.26492774959634]
We develop a compression approach based on quantizing neural network parameters in a linear subspace.
We find large models can be compressed to a much greater extent than previously known, encapsulating Occam's razor.
arXiv Detail & Related papers (2022-11-24T13:50:16Z) - Transferability in Deep Learning: A Survey [80.67296873915176]
The ability to acquire and reuse knowledge is known as transferability in deep learning.
We present this survey to connect different isolated areas in deep learning with their relation to transferability.
We implement a benchmark and an open-source library, enabling a fair evaluation of deep learning methods in terms of transferability.
arXiv Detail & Related papers (2022-01-15T15:03:17Z) - Deep Long-Tailed Learning: A Survey [163.16874896812885]
Deep long-tailed learning aims to train well-performing deep models from a large number of images that follow a long-tailed class distribution.
Long-tailed class imbalance is a common problem in practical visual recognition tasks.
This paper provides a comprehensive survey on recent advances in deep long-tailed learning.
arXiv Detail & Related papers (2021-10-09T15:25:22Z) - Discussion of Ensemble Learning under the Era of Deep Learning [4.061135251278187]
Ensemble deep learning has shown significant performances in improving the generalization of learning system.
Time and space overheads for training multiple base deep learners and testing with the ensemble deep learner are far greater than that of traditional ensemble learning.
An urgent problem needs to be solved is how to take the significant advantages of ensemble deep learning while reduce the required time and space overheads.
arXiv Detail & Related papers (2021-01-21T01:33:23Z) - In Search of Robust Measures of Generalization [79.75709926309703]
We develop bounds on generalization error, optimization error, and excess risk.
When evaluated empirically, most of these bounds are numerically vacuous.
We argue that generalization measures should instead be evaluated within the framework of distributional robustness.
arXiv Detail & Related papers (2020-10-22T17:54:25Z) - Structure preserving deep learning [1.2263454117570958]
deep learning has risen to the foreground as a topic of massive interest.
There are multiple challenging mathematical problems involved in applying deep learning.
A growing effort to mathematically understand the structure in existing deep learning methods.
arXiv Detail & Related papers (2020-06-05T10:59:09Z) - Generalization in Deep Learning [103.91623583928852]
This paper provides theoretical insights into why and how deep learning can generalize well, despite its large capacity, complexity, possible algorithmic instability, nonrobustness, and sharp minima.
We also discuss approaches to provide non-vacuous generalization guarantees for deep learning.
arXiv Detail & Related papers (2017-10-16T02:21:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.