Towards Understanding the Challenges of Bug Localization in Deep
Learning Systems
- URL: http://arxiv.org/abs/2402.01021v1
- Date: Thu, 1 Feb 2024 21:17:42 GMT
- Title: Towards Understanding the Challenges of Bug Localization in Deep
Learning Systems
- Authors: Sigma Jahan, Mehil B. Shah, Mohammad Masudur Rahman
- Abstract summary: We conduct a large-scale empirical study to better understand the challenges of localizing bugs in deep-learning systems.
First, we determine the bug localization performance of four existing techniques using 2,365 bugs from deep-learning systems and 2,913 from traditional software.
Second, we evaluate how different bug types in deep learning systems impact bug localization.
- Score: 2.9312156642007294
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Software bugs cost the global economy billions of dollars annually and claim
~50\% of the programming time from software developers. Locating these bugs is
crucial for their resolution but challenging. It is even more challenging in
deep-learning systems due to their black-box nature. Bugs in these systems are
also hidden not only in the code but also in the models and training data,
which might make traditional debugging methods less effective. In this article,
we conduct a large-scale empirical study to better understand the challenges of
localizing bugs in deep-learning systems. First, we determine the bug
localization performance of four existing techniques using 2,365 bugs from
deep-learning systems and 2,913 from traditional software. We found these
techniques significantly underperform in localizing deep-learning system bugs.
Second, we evaluate how different bug types in deep learning systems impact bug
localization. We found that the effectiveness of localization techniques varies
with bug type due to their unique challenges. For example, tensor bugs were
more accessible to locate due to their structural nature, while all techniques
struggled with GPU bugs due to their external dependencies. Third, we
investigate the impact of bugs' extrinsic nature on localization in
deep-learning systems. We found that deep learning bugs are often extrinsic and
thus connected to artifacts other than source code (e.g., GPU, training data),
contributing to the poor performance of existing localization methods.
Related papers
- RESTOR: Knowledge Recovery through Machine Unlearning [71.75834077528305]
Large language models trained on web-scale corpora can memorize undesirable datapoints.
Many machine unlearning methods have been proposed that aim to 'erase' these datapoints from trained models.
We propose the RESTOR framework for machine unlearning based on the following dimensions.
arXiv Detail & Related papers (2024-10-31T20:54:35Z) - DebugBench: Evaluating Debugging Capability of Large Language Models [80.73121177868357]
DebugBench is a benchmark for Large Language Models (LLMs)
It covers four major bug categories and 18 minor types in C++, Java, and Python.
We evaluate two commercial and four open-source models in a zero-shot scenario.
arXiv Detail & Related papers (2024-01-09T15:46:38Z) - Towards Enhancing the Reproducibility of Deep Learning Bugs: An Empirical Study [13.17302533571231]
This paper examines the critical issue of reproducing deep learning bugs.
We identify edit actions and useful information that could improve the critical issue.
We successfully reproduced 148 out of 165 bugs attempted.
arXiv Detail & Related papers (2024-01-05T21:30:13Z) - Automated Bug Generation in the era of Large Language Models [6.0770779409377775]
BugFarm transforms arbitrary code into multiple complex bugs.
A comprehensive evaluation of 435k+ bugs from over 1.9M mutants generated by BUGFARM.
arXiv Detail & Related papers (2023-10-03T20:01:51Z) - An Empirical Study on Bugs Inside PyTorch: A Replication Study [10.848682558737494]
We characterize bugs in the PyTorch library, a very popular deep learning framework.
Our results highlight that PyTorch bugs are more like traditional software projects bugs, than related to deep learning characteristics.
arXiv Detail & Related papers (2023-07-25T19:23:55Z) - WELL: Applying Bug Detectors to Bug Localization via Weakly Supervised
Learning [37.09621161662761]
This paper proposes a WEakly supervised bug LocaLization (WELL) method to train a bug localization model.
With CodeBERT finetuned on the buggy-or-not binary labeled data, WELL can address bug localization in a weakly supervised manner.
arXiv Detail & Related papers (2023-05-27T06:34:26Z) - BigIssue: A Realistic Bug Localization Benchmark [89.8240118116093]
BigIssue is a benchmark for realistic bug localization.
We provide a general benchmark with a diversity of real and synthetic Java bugs.
We hope to advance the state of the art in bug localization, in turn improving APR performance and increasing its applicability to the modern development cycle.
arXiv Detail & Related papers (2022-07-21T20:17:53Z) - DapStep: Deep Assignee Prediction for Stack Trace Error rePresentation [61.99379022383108]
We propose new deep learning models to solve the bug triage problem.
The models are based on a bidirectional recurrent neural network with attention and on a convolutional neural network.
To improve the quality of ranking, we propose using additional information from version control system annotations.
arXiv Detail & Related papers (2022-01-14T00:16:57Z) - Knowledge as Invariance -- History and Perspectives of
Knowledge-augmented Machine Learning [69.99522650448213]
Research in machine learning is at a turning point.
Research interests are shifting away from increasing the performance of highly parameterized models to exceedingly specific tasks.
This white paper provides an introduction and discussion of this emerging field in machine learning research.
arXiv Detail & Related papers (2020-12-21T15:07:19Z) - Bayesian active learning for production, a systematic study and a
reusable library [85.32971950095742]
In this paper, we analyse the main drawbacks of current active learning techniques.
We do a systematic study on the effects of the most common issues of real-world datasets on the deep active learning process.
We derive two techniques that can speed up the active learning loop such as partial uncertainty sampling and larger query size.
arXiv Detail & Related papers (2020-06-17T14:51:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.