An Anatomy of 488 Faults from Defects4J Based on the Control- and Data-Flow Graph Representations of Programs
- URL: http://arxiv.org/abs/2502.02299v2
- Date: Mon, 28 Apr 2025 14:13:53 GMT
- Title: An Anatomy of 488 Faults from Defects4J Based on the Control- and Data-Flow Graph Representations of Programs
- Authors: Alexandra van der Spuy, Bernd Fischer,
- Abstract summary: Software fault datasets such as Defects4J provide for each individual fault its location and repair, but do not characterize the faults.<n>We propose a new, direct fault classification scheme based on the control- and data-flow graph representations of programs.
- Score: 49.38684825106323
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Software fault datasets such as Defects4J provide for each individual fault its location and repair, but do not characterize the faults. Current classifications use the repairs as proxies, but these do not capture the intrinsic nature of the fault. In this paper, we propose a new, direct fault classification scheme based on the control- and data-flow graph representations of programs. Our scheme comprises six control-flow and two data-flow fault classes. We manually apply this scheme to 488 faults from seven projects in the Defects4J dataset. We find that the majority of the faults are assigned between one and three classes. We also find that one of the data-flow fault classes (definition fault) is the most common individual class but that the majority of faults are classified with at least one control-flow fault class. Our proposed classification can be applied to other fault datasets and can be used to improve fault localization and automated program repair techniques for specific fault classes.
Related papers
- Where's the Bug? Attention Probing for Scalable Fault Localization [18.699014321422023]
We present Bug Attention Probe (BAP), a method which learns state-of-the-art fault localization without any direct localization labels.
BAP is significantly more efficient than prompting, outperforming large open-weight models at a small fraction of the computational cost.
arXiv Detail & Related papers (2025-02-19T18:59:32Z) - Rethinking Early Stopping: Refine, Then Calibrate [49.966899634962374]
We show that calibration error and refinement error are not minimized simultaneously during training.<n>We introduce a new metric for early stopping and hyper parameter tuning that makes it possible to minimize refinement error during training.<n>Our method integrates seamlessly with any architecture and consistently improves performance across diverse classification tasks.
arXiv Detail & Related papers (2025-01-31T15:03:54Z) - Classification Error Bound for Low Bayes Error Conditions in Machine Learning [50.25063912757367]
We study the relationship between the error mismatch and the Kullback-Leibler divergence in machine learning.<n>Motivated by recent observations of low model-based classification errors in many machine learning tasks, we propose a linear approximation of the classification error bound for low Bayes error conditions.
arXiv Detail & Related papers (2025-01-27T11:57:21Z) - Parameter-tuning-free data entry error unlearning with adaptive
selective synaptic dampening [51.34904967046097]
We introduce an extension to the selective synaptic dampening unlearning method that removes the need for parameter tuning.
We demonstrate the performance of this extension, adaptive selective synaptic dampening (ASSD) on various ResNet18 and Vision Transformer unlearning tasks.
The application of this approach is particularly compelling in industrial settings, such as supply chain management.
arXiv Detail & Related papers (2024-02-06T14:04:31Z) - Making Binary Classification from Multiple Unlabeled Datasets Almost
Free of Supervision [128.6645627461981]
We propose a new problem setting, i.e., binary classification from multiple unlabeled datasets with only one pairwise numerical relationship of class priors.
In MU-OPPO, we do not need the class priors for all unlabeled datasets.
We show that our framework brings smaller estimation errors of class priors and better performance of binary classification.
arXiv Detail & Related papers (2023-06-12T11:33:46Z) - Deep Reinforcement Learning for Online Error Detection in Cyber-Physical
Systems [1.2074552857379273]
This paper proposes a new error detection approach based on Deep Reinforcement Learning (DRL)
The proposed approach can categorize different types of errors from normal data and predict whether the system will fail.
The evaluation results illustrate that the proposed approach has improved more than 2x in terms of accuracy and more than 5x in terms of inference time compared to other approaches.
arXiv Detail & Related papers (2023-02-03T06:28:54Z) - Understanding Factual Errors in Summarization: Errors, Summarizers,
Datasets, Error Detectors [105.12462629663757]
In this work, we aggregate factuality error annotations from nine existing datasets and stratify them according to the underlying summarization model.
We compare performance of state-of-the-art factuality metrics, including recent ChatGPT-based metrics, on this stratified benchmark and show that their performance varies significantly across different types of summarization models.
arXiv Detail & Related papers (2022-05-25T15:26:48Z) - Automatic Classification of Error Types in Solutions to Programming
Assignments at Online Learning Platform [4.028503203417233]
We apply machine learning methods to improve the feedback of automated verification systems for programming assignments.
We detect frequent error types by clustering previously submitted incorrect solutions, label these clusters and use this labeled dataset to identify the type of an error in a new submission.
arXiv Detail & Related papers (2021-07-13T11:59:57Z) - Data-Driven Fault Diagnosis Analysis and Open-Set Classification of
Time-Series Data [1.0152838128195467]
A framework for data-driven analysis and open-set classification is developed for fault diagnosis applications.
A data-driven fault classification algorithm is proposed which can handle imbalanced datasets, class overlapping, and unknown faults.
An algorithm is proposed to estimate the size of the fault when training data contains information from known fault realizations.
arXiv Detail & Related papers (2020-09-10T09:53:13Z) - Self-Learning with Rectification Strategy for Human Parsing [73.06197841003048]
We propose a trainable graph reasoning method to correct two typical errors in the pseudo-labels.
The reconstructed features have a stronger ability to represent the topology structure of the human body.
Our method outperforms other state-of-the-art methods in supervised human parsing tasks.
arXiv Detail & Related papers (2020-04-17T03:51:30Z) - Structured Prediction with Partial Labelling through the Infimum Loss [85.4940853372503]
The goal of weak supervision is to enable models to learn using only forms of labelling which are cheaper to collect.
This is a type of incomplete annotation where, for each datapoint, supervision is cast as a set of labels containing the real one.
This paper provides a unified framework based on structured prediction and on the concept of infimum loss to deal with partial labelling.
arXiv Detail & Related papers (2020-03-02T13:59:41Z) - Implicit supervision for fault detection and segmentation of emerging
fault types with Deep Variational Autoencoders [1.160208922584163]
We propose a variational autoencoder (VAE) with labeled and unlabeled samples while inducing implicit supervision on the latent representation of the healthy conditions.
This creates a compact and informative latent representation that allows good detection and segmentation of unseen fault types.
In an extensive comparison, we demonstrate that the proposed method outperforms other learning strategies.
arXiv Detail & Related papers (2019-12-28T18:40:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.