Test case prioritization using test case diversification and
fault-proneness estimations
- URL: http://arxiv.org/abs/2106.10524v3
- Date: Fri, 17 Nov 2023 21:27:03 GMT
- Title: Test case prioritization using test case diversification and
fault-proneness estimations
- Authors: Mostafa Mahdieh, Seyed-Hassan Mirian-Hosseinabadi, Mohsen Mahdieh
- Abstract summary: We propose an approach for TCP that takes into account test case coverage data, bug history, and test case diversification.
The diversification of test cases is preserved by incorporating fault-proneness on a clustering-based approach scheme.
The experiments show that the proposed methods are superior to coverage-based TCP methods.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Regression testing activities greatly reduce the risk of faulty software
release. However, the size of the test suites grows throughout the development
process, resulting in time-consuming execution of the test suite and delayed
feedback to the software development team. This has urged the need for
approaches such as test case prioritization (TCP) and test-suite reduction to
reach better results in case of limited resources. In this regard, proposing
approaches that use auxiliary sources of data such as bug history can be
interesting. We aim to propose an approach for TCP that takes into account test
case coverage data, bug history, and test case diversification. To evaluate
this approach we study its performance on real-world open-source projects. The
bug history is used to estimate the fault-proneness of source code areas. The
diversification of test cases is preserved by incorporating fault-proneness on
a clustering-based approach scheme. The proposed methods are evaluated on
datasets collected from the development history of five real-world projects
including 357 versions in total. The experiments show that the proposed methods
are superior to coverage-based TCP methods. The proposed approach shows that
improvement of coverage-based and fault-proneness-based methods is possible by
using a combination of diversification and fault-proneness incorporation.
Related papers
- BoostAdapter: Improving Vision-Language Test-Time Adaptation via Regional Bootstrapping [64.8477128397529]
We propose a training-required and training-free test-time adaptation framework.
We maintain a light-weight key-value memory for feature retrieval from instance-agnostic historical samples and instance-aware boosting samples.
We theoretically justify the rationality behind our method and empirically verify its effectiveness on both the out-of-distribution and the cross-domain datasets.
arXiv Detail & Related papers (2024-10-20T15:58:43Z) - TestBench: Evaluating Class-Level Test Case Generation Capability of Large Language Models [8.22619177301814]
We introduce TestBench, a benchmark for class-level LLM-based test case generation.
We construct a dataset of 108 Java programs from 9 real-world, large-scale projects on GitHub.
We propose a fine-grained evaluation framework that considers five aspects of test cases: syntactic correctness, compilation correctness, test correctness, code coverage rate, and defect detection rate.
arXiv Detail & Related papers (2024-09-26T06:18:06Z) - Source-Free Domain-Invariant Performance Prediction [68.39031800809553]
We propose a source-free approach centred on uncertainty-based estimation, using a generative model for calibration in the absence of source data.
Our experiments on benchmark object recognition datasets reveal that existing source-based methods fall short with limited source sample availability.
Our approach significantly outperforms the current state-of-the-art source-free and source-based methods, affirming its effectiveness in domain-invariant performance estimation.
arXiv Detail & Related papers (2024-08-05T03:18:58Z) - Segment-Based Test Case Prioritization: A Multi-objective Approach [8.972346309150199]
Test case prioritization ( TCP) is a cost-efficient solution to schedule test cases in an execution order that maximizes an objective function.
We introduce a multi-objective optimization approach to prioritize UI test cases using evolutionary search algorithms and four coverage criteria.
Our approach significantly outperforms other methods in terms of Average Percentage of Faults Detected (APFD) and APFD with Cost.
arXiv Detail & Related papers (2024-08-01T16:51:01Z) - Testing for Fault Diversity in Reinforcement Learning [13.133263651395865]
We argue that policy testing should not find as many failures as possible (e.g., inputs that trigger similar car crashes) but rather aim at revealing as informative and diverse faults as possible in the model.
We show that QD optimisation, while being conceptually simple and generally applicable, finds effectively more diverse faults in the decision model.
arXiv Detail & Related papers (2024-03-22T09:46:30Z) - Effective Fault Localization using Probabilistic and Grouping Approach [0.7673339435080445]
The aim of this paper is to use the conception of conditional probability to design an effective fault localization technique.
We present a fault localization technique that derives the association between statement coverage information and test case execution result.
We evaluate the effectiveness of proposed method over eleven open-source data sets.
arXiv Detail & Related papers (2024-03-08T03:55:09Z) - Test Case Recommendations with Distributed Representation of Code
Syntactic Features [2.225268436173329]
We propose an automated approach which exploits both structural and semantic properties of source code methods and test cases.
The proposed approach initially trains a neural network to transform method-level source code, as well as unit tests, into distributed representations.
The model computes cosine similarity between the method's embedding and the previously-embedded training instances.
arXiv Detail & Related papers (2023-10-04T21:42:01Z) - On Pitfalls of Test-Time Adaptation [82.8392232222119]
Test-Time Adaptation (TTA) has emerged as a promising approach for tackling the robustness challenge under distribution shifts.
We present TTAB, a test-time adaptation benchmark that encompasses ten state-of-the-art algorithms, a diverse array of distribution shifts, and two evaluation protocols.
arXiv Detail & Related papers (2023-06-06T09:35:29Z) - Improved Policy Evaluation for Randomized Trials of Algorithmic Resource
Allocation [54.72195809248172]
We present a new estimator leveraging our proposed novel concept, that involves retrospective reshuffling of participants across experimental arms at the end of an RCT.
We prove theoretically that such an estimator is more accurate than common estimators based on sample means.
arXiv Detail & Related papers (2023-02-06T05:17:22Z) - Validation Diagnostics for SBI algorithms based on Normalizing Flows [55.41644538483948]
This work proposes easy to interpret validation diagnostics for multi-dimensional conditional (posterior) density estimators based on NF.
It also offers theoretical guarantees based on results of local consistency.
This work should help the design of better specified models or drive the development of novel SBI-algorithms.
arXiv Detail & Related papers (2022-11-17T15:48:06Z) - A Discriminative Technique for Multiple-Source Adaptation [55.5865665284915]
We present a new discriminative technique for the multiple-source adaptation, MSA, problem.
Our solution only requires conditional probabilities that can easily be accurately estimated from unlabeled data from the source domains.
Our experiments with real-world applications further demonstrate that our new discriminative MSA algorithm outperforms the previous generative solution.
arXiv Detail & Related papers (2020-08-25T14:06:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.