DeepOrder: Deep Learning for Test Case Prioritization in Continuous
Integration Testing
- URL: http://arxiv.org/abs/2110.07443v1
- Date: Thu, 14 Oct 2021 15:10:38 GMT
- Title: DeepOrder: Deep Learning for Test Case Prioritization in Continuous
Integration Testing
- Authors: Aizaz Sharif, Dusica Marijan, Marius Liaaen
- Abstract summary: This work introduces DeepOrder, a deep learning-based model that works on the basis of regression machine learning.
DeepOrder ranks test cases based on the historical record of test executions from any number of previous test cycles.
We experimentally show that deep neural networks, as a simple regression model, can be efficiently used for test case prioritization in continuous integration testing.
- Score: 6.767885381740952
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Continuous integration testing is an important step in the modern software
engineering life cycle. Test prioritization is a method that can improve the
efficiency of continuous integration testing by selecting test cases that can
detect faults in the early stage of each cycle. As continuous integration
testing produces voluminous test execution data, test history is a commonly
used artifact in test prioritization. However, existing test prioritization
techniques for continuous integration either cannot handle large test history
or are optimized for using a limited number of historical test cycles. We show
that such a limitation can decrease fault detection effectiveness of
prioritized test suites.
This work introduces DeepOrder, a deep learning-based model that works on the
basis of regression machine learning. DeepOrder ranks test cases based on the
historical record of test executions from any number of previous test cycles.
DeepOrder learns failed test cases based on multiple factors including the
duration and execution status of test cases. We experimentally show that deep
neural networks, as a simple regression model, can be efficiently used for test
case prioritization in continuous integration testing. DeepOrder is evaluated
with respect to time-effectiveness and fault detection effectiveness in
comparison with an industry practice and the state of the art approaches. The
results show that DeepOrder outperforms the industry practice and
state-of-the-art test prioritization approaches in terms of these two metrics.
Related papers
- Active Test-Time Adaptation: Theoretical Analyses and An Algorithm [51.84691955495693]
Test-time adaptation (TTA) addresses distribution shifts for streaming test data in unsupervised settings.
We propose the novel problem setting of active test-time adaptation (ATTA) that integrates active learning within the fully TTA setting.
arXiv Detail & Related papers (2024-04-07T22:31:34Z) - Precise Error Rates for Computationally Efficient Testing [75.63895690909241]
We revisit the question of simple-versus-simple hypothesis testing with an eye towards computational complexity.
An existing test based on linear spectral statistics achieves the best possible tradeoff curve between type I and type II error rates.
arXiv Detail & Related papers (2023-11-01T04:41:16Z) - Towards Automatic Generation of Amplified Regression Test Oracles [44.45138073080198]
We propose a test oracle derivation approach to amplify regression test oracles.
The approach monitors the object state during test execution and compares it to the previous version to detect any changes in relation to the SUT's intended behaviour.
arXiv Detail & Related papers (2023-07-28T12:38:44Z) - Validation of massively-parallel adaptive testing using dynamic control
matching [0.0]
Modern businesses often run many A/B/n tests at the same time and in parallel, and package many content variations into the same messages.
This paper presents a method for disentangling the causal effects of the various tests under conditions of continuous test adaptation.
arXiv Detail & Related papers (2023-05-02T11:28:12Z) - Sequential Kernelized Independence Testing [101.22966794822084]
We design sequential kernelized independence tests inspired by kernelized dependence measures.
We demonstrate the power of our approaches on both simulated and real data.
arXiv Detail & Related papers (2022-12-14T18:08:42Z) - Hybrid Intelligent Testing in Simulation-Based Verification [0.0]
Several millions of tests may be required to achieve coverage goals.
Coverage-Directed Test Selection learns from coverage feedback to bias testing towards the most effective tests.
Novelty-Driven Verification learns to identify and simulate stimuli that differ from previous stimuli.
arXiv Detail & Related papers (2022-05-19T13:22:08Z) - Comparative Study of Machine Learning Test Case Prioritization for
Continuous Integration Testing [3.8073142980733]
We show that different machine learning models have different performance for different size of test history used for model training and for different time budget available for test case execution.
Our results imply that machine learning approaches for test prioritization in continuous integration testing should be carefully configured to achieve optimal performance.
arXiv Detail & Related papers (2022-04-22T19:20:49Z) - TestRank: Bringing Order into Unlabeled Test Instances for Deep Learning
Tasks [14.547623982073475]
Deep learning systems are notoriously difficult to test and debug.
It is essential to conduct test selection and label only those selected "high quality" bug-revealing test inputs for test cost reduction.
We propose a novel test prioritization technique that brings order into the unlabeled test instances according to their bug-revealing capabilities, namely TestRank.
arXiv Detail & Related papers (2021-05-21T03:41:10Z) - Cross-validation Confidence Intervals for Test Error [83.67415139421448]
This work develops central limit theorems for crossvalidation and consistent estimators of its variance under weak stability conditions on the learning algorithm.
Results are the first of their kind for the popular choice of leave-one-out cross-validation.
arXiv Detail & Related papers (2020-07-24T17:40:06Z) - Noisy Adaptive Group Testing using Bayesian Sequential Experimental
Design [63.48989885374238]
When the infection prevalence of a disease is low, Dorfman showed 80 years ago that testing groups of people can prove more efficient than testing people individually.
Our goal in this paper is to propose new group testing algorithms that can operate in a noisy setting.
arXiv Detail & Related papers (2020-04-26T23:41:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.