WDD: Weighted Delta Debugging
- URL: http://arxiv.org/abs/2411.19410v2
- Date: Mon, 16 Dec 2024 23:19:36 GMT
- Title: WDD: Weighted Delta Debugging
- Authors: Xintong Zhou, Zhenyang Xu, Mengxiao Zhang, Yongqiang Tian, Chengnian Sun,
- Abstract summary: Delta Weighted Delta (WDD) is a novel concept to help prior delta debug algorithms overcome limitations.<n>WDD assigns each element in the list a weight according to its size, and distinguish different elements based on their weights during partitioning.<n>We extensively evaluated WDD in two representative applications, HDD and Perses, on 62 benchmarks across two languages.
- Score: 6.393194328016689
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Delta Debugging is a widely used family of algorithms (e.g., ddmin and ProbDD) to automatically minimize bug-triggering test inputs, thus to facilitate debugging. It takes a list of elements with each element representing a fragment of the test input, systematically partitions the list at different granularities, identifies and deletes bug-irrelevant partitions. Prior delta debugging algorithms assume there are no differences among the elements in the list, and thus treat them uniformly during partitioning. However, in practice, this assumption usually does not hold, because the size (referred to as weight) of the fragment represented by each element can vary significantly. For example, a single element representing 50% of the test input is much more likely to be bug-relevant than elements representing only 1%. This assumption inevitably impairs the efficiency or even effectiveness of these delta debugging algorithms. This paper proposes Weighted Delta Debugging (WDD), a novel concept to help prior delta debugging algorithms overcome the limitation mentioned above. The key insight of WDD is to assign each element in the list a weight according to its size, and distinguish different elements based on their weights during partitioning. We designed two new minimization algorithms, Wddmin and WProbDD, by applying WDD to ddmin and ProbDD respectively. We extensively evaluated Wddmin and WProbDD in two representative applications, HDD and Perses, on 62 benchmarks across two languages. The results strongly demonstrate the value of WDD. We firmly believe that WDD opens up a new dimension to improve test input minimization techniques.
Related papers
- FairDD: Fair Dataset Distillation via Synchronized Matching [13.60524473223155]
We propose a novel fair dataset distillation (FDD) framework, namely FairDD, which can be seamlessly applied to diverse matching-based DD approaches.<n>The key innovation of FairDD lies in synchronously matching synthetic datasets to PA-wise groups of original datasets.<n>We show that FairDD significantly improves fairness compared to vanilla DD methods, without sacrificing classification accuracy.
arXiv Detail & Related papers (2024-11-29T11:22:20Z) - Dataset Distillation from First Principles: Integrating Core Information Extraction and Purposeful Learning [10.116674195405126]
We argue that a precise characterization of the underlying optimization problem must specify the inference task associated with the application of interest.
Our formalization reveals novel applications of DD across different modeling environments.
We present numerical results for two case studies important in contemporary settings.
arXiv Detail & Related papers (2024-09-02T18:11:15Z) - Deep Dive into Probabilistic Delta Debugging: Insights and Simplifications [6.393194328016689]
ProbDD, an advanced variant of ddmin, has been proposed and achieved state-of-the-art performance.
We conduct the first in-depth theoretical analysis of ProbDD, clarifying trends in probability and subset size changes.
We propose CDD, a simplified version of ProbDD, reducing complexity in both theory and implementation.
arXiv Detail & Related papers (2024-08-08T19:30:03Z) - ZeroDDI: A Zero-Shot Drug-Drug Interaction Event Prediction Method with Semantic Enhanced Learning and Dual-Modal Uniform Alignment [11.290096354280111]
Previously unobserved/unseen DDIEs have been emerging, posing a new classification task when unseen classes have no labelled instances in the training stage.
Existing computational methods are not directly applicable to ZS-DDIE, which has two primary challenges: obtaining suitable DDIE representations and handling the class imbalance issue.
We propose a novel method named ZeroDDI for the ZS-DDIE task, which emphasizes the key biological semantics and distills discriminative molecular substructure-related semantics for DDIE representation learning.
arXiv Detail & Related papers (2024-07-01T01:28:14Z) - Theoretically Achieving Continuous Representation of Oriented Bounding Boxes [64.15627958879053]
This paper endeavors to completely solve the issue of discontinuity in Oriented Bounding Box representation.
We propose a novel representation method called Continuous OBB (COBB) which can be readily integrated into existing detectors.
For fairness and transparency of experiments, we have developed a modularized benchmark based on the open-source deep learning framework Jittor's detection toolbox JDet for OOD evaluation.
arXiv Detail & Related papers (2024-02-29T09:27:40Z) - AdaNPC: Exploring Non-Parametric Classifier for Test-Time Adaptation [64.9230895853942]
Domain generalization can be arbitrarily hard without exploiting target domain information.
Test-time adaptive (TTA) methods are proposed to address this issue.
In this work, we adopt Non-Parametric to perform the test-time Adaptation (AdaNPC)
arXiv Detail & Related papers (2023-04-25T04:23:13Z) - Absolute Wrong Makes Better: Boosting Weakly Supervised Object Detection
via Negative Deterministic Information [54.35679298764169]
Weakly supervised object detection (WSOD) is a challenging task, in which image-level labels are used to train an object detector.
This paper focuses on identifying and fully exploiting the deterministic information in WSOD.
We propose a negative deterministic information (NDI) based method for improving WSOD, namely NDI-WSOD.
arXiv Detail & Related papers (2022-04-21T12:55:27Z) - Benchmarking Deep Models for Salient Object Detection [67.07247772280212]
We construct a general SALient Object Detection (SALOD) benchmark to conduct a comprehensive comparison among several representative SOD methods.
In the above experiments, we find that existing loss functions usually specialized in some metrics but reported inferior results on the others.
We propose a novel Edge-Aware (EA) loss that promotes deep networks to learn more discriminative features by integrating both pixel- and image-level supervision signals.
arXiv Detail & Related papers (2022-02-07T03:43:16Z) - Contextualized Semantic Distance between Highly Overlapped Texts [85.1541170468617]
Overlapping frequently occurs in paired texts in natural language processing tasks like text editing and semantic similarity evaluation.
This paper aims to address the issue with a mask-and-predict strategy.
We take the words in the longest common sequence as neighboring words and use masked language modeling (MLM) to predict the distributions on their positions.
Experiments on Semantic Textual Similarity show NDD to be more sensitive to various semantic differences, especially on highly overlapped paired texts.
arXiv Detail & Related papers (2021-10-04T03:59:15Z) - ADAPT: Mitigating Idling Errors in Qubits via Adaptive Dynamical
Decoupling [3.2505361717998227]
A qubit is susceptible to idling errors, which occur when the qubit is idle and not actively undergoing any operations.
Existing Dynamical Decoupling protocols have been primarily studied for individual qubits.
We propose Adaptive Dynamical Decoupling (ADAPT), a software framework that estimates the efficacy of DD for each qubit combination.
arXiv Detail & Related papers (2021-09-11T16:15:24Z) - AdamP: Slowing Down the Slowdown for Momentum Optimizers on
Scale-invariant Weights [53.8489656709356]
Normalization techniques are a boon for modern deep learning.
It is often overlooked, however, that the additional introduction of momentum results in a rapid reduction in effective step sizes for scale-invariant weights.
In this paper, we verify that the widely-adopted combination of the two ingredients lead to the premature decay of effective step sizes and sub-optimal model performances.
arXiv Detail & Related papers (2020-06-15T08:35:15Z) - Simple and Effective Prevention of Mode Collapse in Deep One-Class
Classification [93.2334223970488]
We propose two regularizers to prevent hypersphere collapse in deep SVDD.
The first regularizer is based on injecting random noise via the standard cross-entropy loss.
The second regularizer penalizes the minibatch variance when it becomes too small.
arXiv Detail & Related papers (2020-01-24T03:44:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.