Confidence Dimension for Deep Learning based on Hoeffding Inequality and
Relative Evaluation
- URL: http://arxiv.org/abs/2203.09082v1
- Date: Thu, 17 Mar 2022 04:43:43 GMT
- Title: Confidence Dimension for Deep Learning based on Hoeffding Inequality and
Relative Evaluation
- Authors: Runqi Wang, Linlin Yang, Baochang Zhang, Wentao Zhu, David Doermann,
Guodong Guo
- Abstract summary: We propose to use multiple factors to measure and rank the relative generalization of deep neural networks (DNNs) based on a new concept of confidence dimension (CD)
Our CD yields a consistent and reliable measure and ranking for both full-precision DNNs and binary neural networks (BNNs) on all the tasks.
- Score: 44.393256948610016
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Research on the generalization ability of deep neural networks (DNNs) has
recently attracted a great deal of attention. However, due to their complex
architectures and large numbers of parameters, measuring the generalization
ability of specific DNN models remains an open challenge. In this paper, we
propose to use multiple factors to measure and rank the relative generalization
of DNNs based on a new concept of confidence dimension (CD). Furthermore, we
provide a feasible framework in our CD to theoretically calculate the upper
bound of generalization based on the conventional Vapnik-Chervonenk dimension
(VC-dimension) and Hoeffding's inequality. Experimental results on image
classification and object detection demonstrate that our CD can reflect the
relative generalization ability for different DNNs. In addition to
full-precision DNNs, we also analyze the generalization ability of binary
neural networks (BNNs), whose generalization ability remains an unsolved
problem. Our CD yields a consistent and reliable measure and ranking for both
full-precision DNNs and BNNs on all the tasks.
Related papers
- Generalization of Graph Neural Networks is Robust to Model Mismatch [84.01980526069075]
Graph neural networks (GNNs) have demonstrated their effectiveness in various tasks supported by their generalization capabilities.
In this paper, we examine GNNs that operate on geometric graphs generated from manifold models.
Our analysis reveals the robustness of the GNN generalization in the presence of such model mismatch.
arXiv Detail & Related papers (2024-08-25T16:00:44Z) - Unveiling and Mitigating Generalized Biases of DNNs through the Intrinsic Dimensions of Perceptual Manifolds [46.47992213722412]
Building fair deep neural networks (DNNs) is a crucial step towards achieving trustworthy artificial intelligence.
We propose Intrinsic Dimension Regularization (IDR), which enhances the fairness and performance of models.
In various image recognition benchmark tests, IDR significantly mitigates model bias while improving its performance.
arXiv Detail & Related papers (2024-04-22T04:16:40Z) - On Generalization Bounds for Deep Compound Gaussian Neural Networks [1.4425878137951238]
Unrolled deep neural networks (DNNs) provide better interpretability and superior empirical performance than standard DNNs.
We develop novel generalization error bounds for a class of unrolled DNNs informed by a compound Gaussian prior.
Under realistic conditions, we show that, at worst, the generalization error scales $mathcalO(nsqrt(n))$ in the signal dimension and $mathcalO(($Network Size$)3/2)$ in network size.
arXiv Detail & Related papers (2024-02-20T16:01:39Z) - Benign Overfitting in Deep Neural Networks under Lazy Training [72.28294823115502]
We show that when the data distribution is well-separated, DNNs can achieve Bayes-optimal test error for classification.
Our results indicate that interpolating with smoother functions leads to better generalization.
arXiv Detail & Related papers (2023-05-30T19:37:44Z) - Transferability of coVariance Neural Networks and Application to
Interpretable Brain Age Prediction using Anatomical Features [119.45320143101381]
Graph convolutional networks (GCN) leverage topology-driven graph convolutional operations to combine information across the graph for inference tasks.
We have studied GCNs with covariance matrices as graphs in the form of coVariance neural networks (VNNs)
VNNs inherit the scale-free data processing architecture from GCNs and here, we show that VNNs exhibit transferability of performance over datasets whose covariance matrices converge to a limit object.
arXiv Detail & Related papers (2023-05-02T22:15:54Z) - Graph Neural Networks are Inherently Good Generalizers: Insights by
Bridging GNNs and MLPs [71.93227401463199]
This paper pinpoints the major source of GNNs' performance gain to their intrinsic capability, by introducing an intermediate model class dubbed as P(ropagational)MLP.
We observe that PMLPs consistently perform on par with (or even exceed) their GNN counterparts, while being much more efficient in training.
arXiv Detail & Related papers (2022-12-18T08:17:32Z) - Estimating the Generalization in Deep Neural Networks via Sparsity [15.986873241115651]
We propose a novel method for estimating the generalization gap based on network sparsity.
By training DNNs with a wide range of generalization gap on popular datasets, we show that our key quantities and linear model could be efficient tools for estimating the generalization gap of DNNs.
arXiv Detail & Related papers (2021-04-02T02:10:32Z) - Neuron Coverage-Guided Domain Generalization [37.77033512313927]
This paper focuses on the domain generalization task where domain knowledge is unavailable, and even worse, only samples from a single domain can be utilized during training.
Our motivation originates from the recent progresses in deep neural network (DNN) testing, which has shown that maximizing neuron coverage of DNN can help to explore possible defects of DNN.
arXiv Detail & Related papers (2021-02-27T14:26:53Z) - Role Taxonomy of Units in Deep Neural Networks [15.067182415076148]
We identify the role of network units in deep neural networks (DNNs) via the retrieval-of-function test.
We show that ratios of the four categories are highly associated with the generalization ability of DNNs from two distinct perspectives.
arXiv Detail & Related papers (2020-11-02T07:37:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.