An Effective Data-Driven Approach for Localizing Deep Learning Faults
- URL: http://arxiv.org/abs/2307.08947v1
- Date: Tue, 18 Jul 2023 03:28:39 GMT
- Title: An Effective Data-Driven Approach for Localizing Deep Learning Faults
- Authors: Mohammad Wardat, Breno Dantas Cruz, Wei Le, Hridesh Rajan
- Abstract summary: We propose a novel data-driven approach that leverages model features to learn problem patterns.
Our methodology automatically links bug symptoms to their root causes, without the need for manually crafted mappings.
Our results demonstrate that our technique can effectively detect and diagnose different bug types.
- Score: 20.33411443073181
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deep Learning (DL) applications are being used to solve problems in critical
domains (e.g., autonomous driving or medical diagnosis systems). Thus,
developers need to debug their systems to ensure that the expected behavior is
delivered. However, it is hard and expensive to debug DNNs. When the failure
symptoms or unsatisfied accuracies are reported after training, we lose the
traceability as to which part of the DNN program is responsible for the
failure. Even worse, sometimes, a deep learning program has different types of
bugs. To address the challenges of debugging DNN models, we propose a novel
data-driven approach that leverages model features to learn problem patterns.
Our approach extracts these features, which represent semantic information of
faults during DNN training. Our technique uses these features as a training
dataset to learn and infer DNN fault patterns. Also, our methodology
automatically links bug symptoms to their root causes, without the need for
manually crafted mappings, so that developers can take the necessary steps to
fix faults. We evaluate our approach using real-world and mutated models. Our
results demonstrate that our technique can effectively detect and diagnose
different bug types. Finally, our technique achieved better accuracy,
precision, and recall than prior work for mutated models. Also, our approach
achieved comparable results for real-world models in terms of accuracy and
performance to the state-of-the-art.
Related papers
- Corrective Machine Unlearning [22.342035149807923]
We formalize Corrective Machine Unlearning as the problem of mitigating the impact of data affected by unknown manipulations on a trained model.
We find most existing unlearning methods, including retraining-from-scratch without the deletion set, require most of the manipulated data to be identified for effective corrective unlearning.
One approach, Selective Synaptic Dampening, achieves limited success, unlearning adverse effects with just a small portion of the manipulated samples in our setting.
arXiv Detail & Related papers (2024-02-21T18:54:37Z) - Learn to Unlearn for Deep Neural Networks: Minimizing Unlearning
Interference with Gradient Projection [56.292071534857946]
Recent data-privacy laws have sparked interest in machine unlearning.
Challenge is to discard information about the forget'' data without altering knowledge about remaining dataset.
We adopt a projected-gradient based learning method, named as Projected-Gradient Unlearning (PGU)
We provide empirically evidence to demonstrate that our unlearning method can produce models that behave similar to models retrained from scratch across various metrics even when the training dataset is no longer accessible.
arXiv Detail & Related papers (2023-12-07T07:17:24Z) - Bridging Precision and Confidence: A Train-Time Loss for Calibrating
Object Detection [58.789823426981044]
We propose a novel auxiliary loss formulation that aims to align the class confidence of bounding boxes with the accurateness of predictions.
Our results reveal that our train-time loss surpasses strong calibration baselines in reducing calibration error for both in and out-domain scenarios.
arXiv Detail & Related papers (2023-03-25T08:56:21Z) - Adversarial Learning Networks: Source-free Unsupervised Domain
Incremental Learning [0.0]
In a non-stationary environment, updating a DNN model requires parameter re-training or model fine-tuning.
We propose an unsupervised source-free method to update DNN classification models.
Unlike existing methods, our approach can update a DNN model incrementally for non-stationary source and target tasks without storing past training data.
arXiv Detail & Related papers (2023-01-28T02:16:13Z) - Testing Feedforward Neural Networks Training Programs [13.249453757295083]
Multiple testing techniques are proposed to generate test cases that can expose inconsistencies in the behavior of Deep Neural Networks.
These techniques assume implicitly that the training program is bug-free and appropriately configured.
We propose TheDeepChecker, an end-to-end property-based debug approach for DNN training programs.
arXiv Detail & Related papers (2022-04-01T20:49:14Z) - DapStep: Deep Assignee Prediction for Stack Trace Error rePresentation [61.99379022383108]
We propose new deep learning models to solve the bug triage problem.
The models are based on a bidirectional recurrent neural network with attention and on a convolutional neural network.
To improve the quality of ranking, we propose using additional information from version control system annotations.
arXiv Detail & Related papers (2022-01-14T00:16:57Z) - DeepDiagnosis: Automatically Diagnosing Faults and Recommending
Actionable Fixes in Deep Learning Programs [12.917211542949786]
We propose DeepDiagnosis, a novel approach that localizes the faults, reports error symptoms and suggests fixes for DNN programs.
DeepDiagnosis manifests the best capabilities of fault detection, bug localization, and symptoms identification when compared to other approaches.
arXiv Detail & Related papers (2021-12-07T23:15:23Z) - What Stops Learning-based 3D Registration from Working in the Real
World? [53.68326201131434]
This work identifies the sources of 3D point cloud registration failures, analyze the reasons behind them, and propose solutions.
Ultimately, this translates to a best-practice 3D registration network (BPNet), constituting the first learning-based method able to handle previously-unseen objects in real-world data.
Our model generalizes to real data without any fine-tuning, reaching an accuracy of up to 67% on point clouds of unseen objects obtained with a commercial sensor.
arXiv Detail & Related papers (2021-11-19T19:24:27Z) - Distributionally Robust Semi-Supervised Learning Over Graphs [68.29280230284712]
Semi-supervised learning (SSL) over graph-structured data emerges in many network science applications.
To efficiently manage learning over graphs, variants of graph neural networks (GNNs) have been developed recently.
Despite their success in practice, most of existing methods are unable to handle graphs with uncertain nodal attributes.
Challenges also arise due to distributional uncertainties associated with data acquired by noisy measurements.
A distributionally robust learning framework is developed, where the objective is to train models that exhibit quantifiable robustness against perturbations.
arXiv Detail & Related papers (2021-10-20T14:23:54Z) - ALT-MAS: A Data-Efficient Framework for Active Testing of Machine
Learning Algorithms [58.684954492439424]
We propose a novel framework to efficiently test a machine learning model using only a small amount of labeled test data.
The idea is to estimate the metrics of interest for a model-under-test using Bayesian neural network (BNN)
arXiv Detail & Related papers (2021-04-11T12:14:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.