Can Transformer Models Effectively Detect Software Aspects in
StackOverflow Discussion?
- URL: http://arxiv.org/abs/2209.12065v1
- Date: Sat, 24 Sep 2022 18:28:14 GMT
- Title: Can Transformer Models Effectively Detect Software Aspects in
StackOverflow Discussion?
- Authors: Nibir Chandra Mandal, Tashreef Muhammad and G. M. Shahariar
- Abstract summary: Developers are constantly searching for all of the benefits and drawbacks of each API, framework, tool, and so on.
One of the typical approaches is to examine all of the features through official documentation and discussion.
In this paper, we have used a benchmark API aspects dataset (Opiner) collected from StackOverflow posts.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Dozens of new tools and technologies are being incorporated to help
developers, which is becoming a source of consternation as they struggle to
choose one over the others. For example, there are at least ten frameworks
available to developers for developing web applications, posing a conundrum in
selecting the best one that meets their needs. As a result, developers are
continuously searching for all of the benefits and drawbacks of each API,
framework, tool, and so on. One of the typical approaches is to examine all of
the features through official documentation and discussion. This approach is
time-consuming, often makes it difficult to determine which aspects are the
most important to a particular developer and whether a particular aspect is
important to the community at large. In this paper, we have used a benchmark
API aspects dataset (Opiner) collected from StackOverflow posts and observed
how Transformer models (BERT, RoBERTa, DistilBERT, and XLNet) perform in
detecting software aspects in textual developer discussion with respect to the
baseline Support Vector Machine (SVM) model. Through extensive experimentation,
we have found that transformer models improve the performance of baseline SVM
for most of the aspects, i.e., `Performance', `Security', `Usability',
`Documentation', `Bug', `Legal', `OnlySentiment', and `Others'. However, the
models fail to apprehend some of the aspects (e.g., `Community' and
`Potability') and their performance varies depending on the aspects. Also,
larger architectures like XLNet are ineffective in interpreting software
aspects compared to smaller architectures like DistilBERT.
Related papers
- BabelBench: An Omni Benchmark for Code-Driven Analysis of Multimodal and Multistructured Data [61.936320820180875]
Large language models (LLMs) have become increasingly pivotal across various domains.
BabelBench is an innovative benchmark framework that evaluates the proficiency of LLMs in managing multimodal multistructured data with code execution.
Our experimental findings on BabelBench indicate that even cutting-edge models like ChatGPT 4 exhibit substantial room for improvement.
arXiv Detail & Related papers (2024-10-01T15:11:24Z) - CodeRAG-Bench: Can Retrieval Augment Code Generation? [78.37076502395699]
We conduct a systematic, large-scale analysis of code generation using retrieval-augmented generation.
We first curate a comprehensive evaluation benchmark, CodeRAG-Bench, encompassing three categories of code generation tasks.
We examine top-performing models on CodeRAG-Bench by providing contexts retrieved from one or multiple sources.
arXiv Detail & Related papers (2024-06-20T16:59:52Z) - Transformer Architecture for NetsDB [0.0]
We create an end-to-end implementation of a transformer for deep learning model serving in NetsDB.
We load out weights from our model for distributed processing, deployment, and efficient inferencing.
arXiv Detail & Related papers (2024-05-08T04:38:36Z) - State Space Model for New-Generation Network Alternative to Transformers: A Survey [52.812260379420394]
In the post-deep learning era, the Transformer architecture has demonstrated its powerful performance across pre-trained big models and various downstream tasks.
To further reduce the complexity of attention models, numerous efforts have been made to design more efficient methods.
Among them, the State Space Model (SSM), as a possible replacement for the self-attention based Transformer model, has drawn more and more attention in recent years.
arXiv Detail & Related papers (2024-04-15T07:24:45Z) - Dissecting Multimodality in VideoQA Transformer Models by Impairing Modality Fusion [54.33764537135906]
VideoQA Transformer models demonstrate competitive performance on standard benchmarks.
Do these models capture the rich multimodal structures and dynamics from video and text jointly?
Are they achieving high scores by exploiting biases and spurious features?
arXiv Detail & Related papers (2023-06-15T06:45:46Z) - Representation Learning for Stack Overflow Posts: How Far are We? [14.520780251680586]
State-of-the-art Stack Overflow post representation models are Post2Vec and BERTOverflow.
Despite their promising results, these representation methods have not been evaluated in the same experimental setting.
We propose SOBERT, which employs a simple-yet-effective strategy to improve the best-performing model.
arXiv Detail & Related papers (2023-03-13T04:49:06Z) - A Comparison of Decision Forest Inference Platforms from A Database
Perspective [4.873098180823506]
Decision forest is one of the most popular machine learning techniques used in many industrial scenarios, such as credit card fraud detection, ranking, and business intelligence.
A number of frameworks were developed and dedicated for decision forest inference, such as ONNX, TreeLite from Amazon, Decision Forest from Google, HummingBird from Microsoft, Nvidia FIL, and lleaves.
arXiv Detail & Related papers (2023-02-09T04:07:50Z) - Simple and Efficient Architectures for Semantic Segmentation [50.1563637917129]
We show that a simple encoder-decoder architecture with a ResNet-like backbone and a small multi-scale head, performs on-par or better than complex semantic segmentation architectures such as HRNet, FANet and DDRNet.
We present a family of such simple architectures for desktop as well as mobile targets, which match or exceed the performance of complex models on the Cityscapes dataset.
arXiv Detail & Related papers (2022-06-16T15:08:34Z) - Fashionformer: A simple, Effective and Unified Baseline for Human
Fashion Segmentation and Recognition [80.74495836502919]
In this work, we focus on joint human fashion segmentation and attribute recognition.
We introduce the object query for segmentation and the attribute query for attribute prediction.
For attribute stream, we design a novel Multi-Layer Rendering module to explore more fine-grained features.
arXiv Detail & Related papers (2022-04-10T11:11:10Z) - Mapping the Internet: Modelling Entity Interactions in Complex
Heterogeneous Networks [0.0]
We propose a versatile, unified framework called HMill' for sample representation, model definition and training.
We show an extension of the universal approximation theorem to the set of all functions realized by models implemented in the framework.
We solve three different problems from the cybersecurity domain using the framework.
arXiv Detail & Related papers (2021-04-19T21:32:44Z) - Patching as Translation: the Data and the Metaphor [18.22949296398319]
We show that "software patching is like language translation"
We show how a more principled approach to model design, based on our empirical findings and general knowledge of software development, can lead to better solutions.
We implement such models ourselves as "proof-of-concept" tools and empirically confirm that they behave in a fundamentally different, more effective way than the studied translation-based architectures.
arXiv Detail & Related papers (2020-08-24T21:05:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.