Improving the Effectiveness of Traceability Link Recovery using
Hierarchical Bayesian Networks
- URL: http://arxiv.org/abs/2005.09046v2
- Date: Mon, 11 Apr 2022 15:17:25 GMT
- Title: Improving the Effectiveness of Traceability Link Recovery using
Hierarchical Bayesian Networks
- Authors: Kevin Moran, David N. Palacio, Carlos Bernal-C\'ardenas, Daniel
McCrystal, Denys Poshyvanyk, Chris Shenefiel, Jeff Johnson
- Abstract summary: We implement a HierarchiCal PrObabilistic Model for SoftwarE Traceability (Comet)
Comet is capable of modeling relationships between artifacts by combining the complementary observational prowess of multiple measures of textual similarity.
We conduct a comprehensive empirical evaluation of Comet that illustrates an improvement over a set of optimally configured baselines.
- Score: 21.15456830607455
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Traceability is a fundamental component of the modern software development
process that helps to ensure properly functioning, secure programs. Due to the
high cost of manually establishing trace links, researchers have developed
automated approaches that draw relationships between pairs of textual software
artifacts using similarity measures. However, the effectiveness of such
techniques are often limited as they only utilize a single measure of artifact
similarity and cannot simultaneously model (implicit and explicit)
relationships across groups of diverse development artifacts.
In this paper, we illustrate how these limitations can be overcome through
the use of a tailored probabilistic model. To this end, we design and implement
a HierarchiCal PrObabilistic Model for SoftwarE Traceability (Comet) that is
able to infer candidate trace links. Comet is capable of modeling relationships
between artifacts by combining the complementary observational prowess of
multiple measures of textual similarity. Additionally, our model can
holistically incorporate information from a diverse set of sources, including
developer feedback and transitive (often implicit) relationships among groups
of software artifacts, to improve inference accuracy. We conduct a
comprehensive empirical evaluation of Comet that illustrates an improvement
over a set of optimally configured baselines of $\approx$14% in the best case
and $\approx$5% across all subjects in terms of average precision. The
comparative effectiveness of Comet in practice, where optimal configuration is
typically not possible, is likely to be higher. Finally, we illustrate Comets
potential for practical applicability in a survey with developers from Cisco
Systems who used a prototype Comet Jenkins plugin.
Related papers
- Implicitly Guided Design with PropEn: Match your Data to Follow the Gradient [52.2669490431145]
PropEn is inspired by'matching', which enables implicit guidance without training a discriminator.
We show that training with a matched dataset approximates the gradient of the property of interest while remaining within the data distribution.
arXiv Detail & Related papers (2024-05-28T11:30:19Z) - Balancing Similarity and Complementarity for Federated Learning [91.65503655796603]
Federated Learning (FL) is increasingly important in mobile and IoT systems.
One key challenge in FL is managing statistical heterogeneity, such as non-i.i.d. data.
We introduce a novel framework, textttFedSaC, which balances similarity and complementarity in FL cooperation.
arXiv Detail & Related papers (2024-05-16T08:16:19Z) - Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching [53.05954114863596]
We propose a brand-new Deep Boosting Learning (DBL) algorithm for image-text matching.
An anchor branch is first trained to provide insights into the data properties.
A target branch is concurrently tasked with more adaptive margin constraints to further enlarge the relative distance between matched and unmatched samples.
arXiv Detail & Related papers (2024-04-28T08:44:28Z) - TRIAD: Automated Traceability Recovery based on Biterm-enhanced
Deduction of Transitive Links among Artifacts [53.92293118080274]
Traceability allows stakeholders to extract and comprehend the trace links among software artifacts introduced across the software life cycle.
Most rely on textual similarities among software artifacts, such as those based on Information Retrieval (IR)
arXiv Detail & Related papers (2023-12-28T06:44:24Z) - An Empirical Study on Data Leakage and Generalizability of Link
Prediction Models for Issues and Commits [7.061740334417124]
LinkFormer preserves and improves the accuracy of existing predictions.
Our findings support that to simulate real-world scenarios effectively, researchers must maintain the temporal flow of data.
arXiv Detail & Related papers (2022-11-01T10:54:26Z) - Decorrelative Network Architecture for Robust Electrocardiogram
Classification [4.808817930937323]
It is not possible to train networks that are accurate in all scenarios.
Deep learning methods sample the model parameter space to estimate uncertainty.
These parameters are often subject to the same vulnerabilities, which can be exploited by adversarial attacks.
We propose a novel ensemble approach based on feature decorrelation and Fourier partitioning for teaching networks diverse complementary features.
arXiv Detail & Related papers (2022-07-19T02:36:36Z) - Data Fusion with Latent Map Gaussian Processes [0.0]
Multi-fidelity modeling and calibration are data fusion tasks that ubiquitously arise in engineering design.
We introduce a novel approach based on latent-map Gaussian processes (LMGPs) that enables efficient and accurate data fusion.
arXiv Detail & Related papers (2021-12-04T00:54:19Z) - PDC-Net+: Enhanced Probabilistic Dense Correspondence Network [161.76275845530964]
Enhanced Probabilistic Dense Correspondence Network, PDC-Net+, capable of estimating accurate dense correspondences.
We develop an architecture and an enhanced training strategy tailored for robust and generalizable uncertainty prediction.
Our approach obtains state-of-the-art results on multiple challenging geometric matching and optical flow datasets.
arXiv Detail & Related papers (2021-09-28T17:56:41Z) - Comparative Code Structure Analysis using Deep Learning for Performance
Prediction [18.226950022938954]
This paper aims to assess the feasibility of using purely static information (e.g., abstract syntax tree or AST) of applications to predict performance change based on the change in code structure.
Our evaluations of several deep embedding learning methods demonstrate that tree-based Long Short-Term Memory (LSTM) models can leverage the hierarchical structure of source-code to discover latent representations and achieve up to 84% (individual problem) and 73% (combined dataset with multiple of problems) accuracy in predicting the change in performance.
arXiv Detail & Related papers (2021-02-12T16:59:12Z) - Centralized Information Interaction for Salient Object Detection [68.8587064889475]
The U-shape structure has shown its advantage in salient object detection for efficiently combining multi-scale features.
This paper shows that by centralizing these connections, we can achieve the cross-scale information interaction among them.
Our approach can cooperate with various existing U-shape-based salient object detection methods by substituting the connections between the bottom-up and top-down pathways.
arXiv Detail & Related papers (2020-12-21T12:42:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.