Computing the Testing Error without a Testing Set
- URL: http://arxiv.org/abs/2005.00450v1
- Date: Fri, 1 May 2020 15:35:50 GMT
- Title: Computing the Testing Error without a Testing Set
- Authors: Ciprian Corneanu, Meysam Madadi, Sergio Escalera, Aleix Martinez
- Abstract summary: We derive an algorithm to estimate the performance gap between training and testing that does not require any testing dataset.
This allows us to compute the DNN's testing error on unseen samples, even when we do not have access to them.
- Score: 33.068870286618655
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep Neural Networks (DNNs) have revolutionized computer vision. We now have
DNNs that achieve top (performance) results in many problems, including object
recognition, facial expression analysis, and semantic segmentation, to name but
a few. The design of the DNNs that achieve top results is, however, non-trivial
and mostly done by trail-and-error. That is, typically, researchers will derive
many DNN architectures (i.e., topologies) and then test them on multiple
datasets. However, there are no guarantees that the selected DNN will perform
well in the real world. One can use a testing set to estimate the performance
gap between the training and testing sets, but avoiding
overfitting-to-the-testing-data is almost impossible. Using a sequestered
testing dataset may address this problem, but this requires a constant update
of the dataset, a very expensive venture. Here, we derive an algorithm to
estimate the performance gap between training and testing that does not require
any testing dataset. Specifically, we derive a number of persistent topology
measures that identify when a DNN is learning to generalize to unseen samples.
This allows us to compute the DNN's testing error on unseen samples, even when
we do not have access to them. We provide extensive experimental validation on
multiple networks and datasets to demonstrate the feasibility of the proposed
approach.
Related papers
- DeepSample: DNN sampling-based testing for operational accuracy assessment [12.029919627622954]
Deep Neural Networks (DNN) are core components for classification and regression tasks of many software systems.
The challenge is to select a representative set of test inputs as small as possible to reduce the labelling cost.
This study presents DeepSample, a family of DNN testing techniques for cost-effective accuracy assessment.
arXiv Detail & Related papers (2024-03-28T09:56:26Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - DeepGD: A Multi-Objective Black-Box Test Selection Approach for Deep
Neural Networks [0.6249768559720121]
DeepGD is a black-box multi-objective test selection approach for Deep neural networks (DNNs)
It reduces the cost of labeling by prioritizing the selection of test inputs with high fault revealing power from large unlabeled datasets.
arXiv Detail & Related papers (2023-03-08T20:33:09Z) - The #DNN-Verification Problem: Counting Unsafe Inputs for Deep Neural
Networks [94.63547069706459]
#DNN-Verification problem involves counting the number of input configurations of a DNN that result in a violation of a safety property.
We propose a novel approach that returns the exact count of violations.
We present experimental results on a set of safety-critical benchmarks.
arXiv Detail & Related papers (2023-01-17T18:32:01Z) - Boosted Dynamic Neural Networks [53.559833501288146]
A typical EDNN has multiple prediction heads at different layers of the network backbone.
To optimize the model, these prediction heads together with the network backbone are trained on every batch of training data.
Treating training and testing inputs differently at the two phases will cause the mismatch between training and testing data distributions.
We formulate an EDNN as an additive model inspired by gradient boosting, and propose multiple training techniques to optimize the model effectively.
arXiv Detail & Related papers (2022-11-30T04:23:12Z) - Meta Input: How to Leverage Off-the-Shelf Deep Neural Networks [29.975937981538664]
We introduce a novel approach that allows end-users to exploit pretrained DNN models in their own testing environment without modifying the models.
We present a textitmeta input which is an additional input transforming the distribution of testing data to be aligned with that of training data.
As a result, end-users can exploit well-trained models in their own testing environment which can differ from the training environment.
arXiv Detail & Related papers (2022-10-21T02:11:38Z) - Reducing Flipping Errors in Deep Neural Networks [39.24451665215755]
Deep neural networks (DNNs) have been widely applied in various domains in artificial intelligence.
In this paper, we study how many test (unseen) samples that a DNN misclassifies in the last epoch were ever correctly classified.
We propose to restrict the behavior changes of a DNN on the correctly-classified samples so that the correct local boundaries can be maintained.
arXiv Detail & Related papers (2022-03-16T04:38:06Z) - Shift-Robust GNNs: Overcoming the Limitations of Localized Graph
Training data [52.771780951404565]
Shift-Robust GNN (SR-GNN) is designed to account for distributional differences between biased training data and the graph's true inference distribution.
We show that SR-GNN outperforms other GNN baselines by accuracy, eliminating at least (40%) of the negative effects introduced by biased training data.
arXiv Detail & Related papers (2021-08-02T18:00:38Z) - Self-Competitive Neural Networks [0.0]
Deep Neural Networks (DNNs) have improved the accuracy of classification problems in lots of applications.
One of the challenges in training a DNN is its need to be fed by an enriched dataset to increase its accuracy and avoid it suffering from overfitting.
Recently, researchers have worked extensively to propose methods for data augmentation.
In this paper, we generate adversarial samples to refine the Domains of Attraction (DoAs) of each class. In this approach, at each stage, we use the model learned by the primary and generated adversarial data (up to that stage) to manipulate the primary data in a way that look complicated to
arXiv Detail & Related papers (2020-08-22T12:28:35Z) - Boosting Deep Neural Networks with Geometrical Prior Knowledge: A Survey [77.99182201815763]
Deep Neural Networks (DNNs) achieve state-of-the-art results in many different problem settings.
DNNs are often treated as black box systems, which complicates their evaluation and validation.
One promising field, inspired by the success of convolutional neural networks (CNNs) in computer vision tasks, is to incorporate knowledge about symmetric geometrical transformations.
arXiv Detail & Related papers (2020-06-30T14:56:05Z) - Contextual-Bandit Anomaly Detection for IoT Data in Distributed
Hierarchical Edge Computing [65.78881372074983]
IoT devices can hardly afford complex deep neural networks (DNN) models, and offloading anomaly detection tasks to the cloud incurs long delay.
We propose and build a demo for an adaptive anomaly detection approach for distributed hierarchical edge computing (HEC) systems.
We show that our proposed approach significantly reduces detection delay without sacrificing accuracy, as compared to offloading detection tasks to the cloud.
arXiv Detail & Related papers (2020-04-15T06:13:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.