Do Deep Neural Networks Always Perform Better When Eating More Data?
- URL: http://arxiv.org/abs/2205.15187v1
- Date: Mon, 30 May 2022 15:40:33 GMT
- Title: Do Deep Neural Networks Always Perform Better When Eating More Data?
- Authors: Jiachen Yang, Zhuo Zhang, Yicheng Gong, Shukun Ma, Xiaolan Guo, Yue
Yang, Shuai Xiao, Jiabao Wen, Yang Li, Xinbo Gao, Wen Lu and Qinggang Meng
- Abstract summary: We design experiments from Identically Independent Distribution(IID) and Out of Distribution(OOD)
Under IID condition, the amount of information determines the effectivity of each sample, the contribution of samples and difference between classes determine the amount of class information.
Under OOD condition, the cross-domain degree of samples determine the contributions, and the bias-fitting caused by irrelevant elements is a significant factor of cross-domain.
- Score: 82.6459747000664
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Data has now become a shortcoming of deep learning. Researchers in their own
fields share the thinking that "deep neural networks might not always perform
better when they eat more data," which still lacks experimental validation and
a convincing guiding theory. Here to fill this lack, we design experiments from
Identically Independent Distribution(IID) and Out of Distribution(OOD), which
give powerful answers. For the purpose of guidance, based on the discussion of
results, two theories are proposed: under IID condition, the amount of
information determines the effectivity of each sample, the contribution of
samples and difference between classes determine the amount of sample
information and the amount of class information; under OOD condition, the
cross-domain degree of samples determine the contributions, and the
bias-fitting caused by irrelevant elements is a significant factor of
cross-domain. The above theories provide guidance from the perspective of data,
which can promote a wide range of practical applications of artificial
intelligence.
Related papers
- Dimensionality-induced information loss of outliers in deep neural networks [29.15751143793406]
Out-of-distribution (OOD) detection is a critical issue for systems using a deep neural network (DNN)
We experimentally clarify this issue by investigating the layer dependence of feature representations from multiple perspectives.
We propose a dimensionality-aware OOD detection method based on alignment of features and weights.
arXiv Detail & Related papers (2024-10-29T01:52:46Z) - Hypothesis-Driven Deep Learning for Out of Distribution Detection [0.8191518216608217]
We propose a hypothesis-driven approach to quantify whether a new sample is InD or OoD.
We adapt our method to detect an unseen sample of bacteria to a trained deep learning model, and show that it reveals interpretable differences between InD and OoD latent responses.
arXiv Detail & Related papers (2024-03-21T01:06:47Z) - EAT: Towards Long-Tailed Out-of-Distribution Detection [55.380390767978554]
This paper addresses the challenging task of long-tailed OOD detection.
The main difficulty lies in distinguishing OOD data from samples belonging to the tail classes.
We propose two simple ideas: (1) Expanding the in-distribution class space by introducing multiple abstention classes, and (2) Augmenting the context-limited tail classes by overlaying images onto the context-rich OOD data.
arXiv Detail & Related papers (2023-12-14T13:47:13Z) - LINe: Out-of-Distribution Detection by Leveraging Important Neurons [15.797257361788812]
We introduce a new aspect for analyzing the difference in model outputs between in-distribution data and OOD data.
We propose a novel method, Leveraging Important Neurons (LINe), for post-hoc Out of distribution detection.
arXiv Detail & Related papers (2023-03-24T13:49:05Z) - Energy-based Out-of-Distribution Detection for Graph Neural Networks [76.0242218180483]
We propose a simple, powerful and efficient OOD detection model for GNN-based learning on graphs, which we call GNNSafe.
GNNSafe achieves up to $17.0%$ AUROC improvement over state-of-the-arts and it could serve as simple yet strong baselines in such an under-developed area.
arXiv Detail & Related papers (2023-02-06T16:38:43Z) - A Survey of Learning on Small Data: Generalization, Optimization, and
Challenge [101.27154181792567]
Learning on small data that approximates the generalization ability of big data is one of the ultimate purposes of AI.
This survey follows the active sampling theory under a PAC framework to analyze the generalization error and label complexity of learning on small data.
Multiple data applications that may benefit from efficient small data representation are surveyed.
arXiv Detail & Related papers (2022-07-29T02:34:19Z) - Improving the Level of Autism Discrimination through GraphRNN Link
Prediction [8.103074928419527]
This paper is based on the latter technique, which learns the edge distribution of real brain network through GraphRNN.
The experimental results show that the combination of original and synthetic data greatly improves the discrimination of the neural network.
arXiv Detail & Related papers (2022-02-19T06:50:32Z) - Negative Data Augmentation [127.28042046152954]
We show that negative data augmentation samples provide information on the support of the data distribution.
We introduce a new GAN training objective where we use NDA as an additional source of synthetic data for the discriminator.
Empirically, models trained with our method achieve improved conditional/unconditional image generation along with improved anomaly detection capabilities.
arXiv Detail & Related papers (2021-02-09T20:28:35Z) - Deep Transparent Prediction through Latent Representation Analysis [0.0]
The paper presents a novel deep learning approach, which extracts latent information from trained Deep Neural Networks (DNNs) and derives concise representations that are analyzed in an effective, unified way for prediction purposes.
Transparency combined with high prediction accuracy are the targeted goals of the proposed approach.
arXiv Detail & Related papers (2020-09-13T19:21:40Z) - Provably Efficient Causal Reinforcement Learning with Confounded
Observational Data [135.64775986546505]
We study how to incorporate the dataset (observational data) collected offline, which is often abundantly available in practice, to improve the sample efficiency in the online setting.
We propose the deconfounded optimistic value iteration (DOVI) algorithm, which incorporates the confounded observational data in a provably efficient manner.
arXiv Detail & Related papers (2020-06-22T14:49:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.