Improving Quality of a Post's Set of Answers in Stack Overflow
- URL: http://arxiv.org/abs/2006.00341v1
- Date: Sat, 30 May 2020 19:40:19 GMT
- Title: Improving Quality of a Post's Set of Answers in Stack Overflow
- Authors: Mohammadrezar Tavakoli, Maliheh Izadi, Abbas Heydarnoori
- Abstract summary: A large number of low-quality posts on Stack Overflow require improvement.
We propose an approach to automate the identification process of such posts and boost their set of answers.
- Score: 2.0625936401496237
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Community Question Answering platforms such as Stack Overflow help a wide
range of users solve their challenges online. As the popularity of these
communities has grown over the years, both the number of members and posts have
escalated. Also, due to the diverse backgrounds, skills, expertise, and
viewpoints of users, each question may obtain more than one answers. Therefore,
the focus has changed toward producing posts that have a set of answers more
valuable for the community as a whole, not just one accepted-answer aimed at
satisfying only the question-asker. Same as every universal community, a large
number of low-quality posts on Stack Overflow require improvement. We call
these posts deficient and define them as posts with questions that either have
no answer yet or can be improved by other ones. In this paper, we propose an
approach to automate the identification process of such posts and boost their
set of answers, utilizing the help of related experts. With the help of 60
participants, we trained a classification model to identify deficient posts by
investigating the relationship between characteristics of 3075 questions posted
on Stack Overflow and their need for better answers set. Then, we developed an
Eclipse plugin named SOPI and integrated the prediction model in the plugin to
link these deficient posts to related developers and help them improve the
answer set. We evaluated both the functionality of our plugin and the impact of
answers submitted to Stack Overflow with the help of 10 and 15 expert
industrial developers, respectively. Our results indicate that decision trees,
specifically J48, predicts a deficient question better than the other methods
with 0.945 precision and 0.903 recall. We conclude that not only our plugin
helps programmers contribute more easily to Stack Overflow, but also it
improves the quality of answers.
Related papers
- An exploratory analysis of Community-based Question-Answering Platforms and GPT-3-driven Generative AI: Is it the end of online community-based learning? [0.6749750044497732]
ChatGPT offers software engineers an interactive alternative to community question-answering platforms like Stack Overflow.
We analyze 2564 Python and JavaScript questions from StackOverflow that were asked between January 2022 and December 2022.
Our analysis indicates that ChatGPT's responses are 66% shorter and share 35% more words with the questions, showing a 25% increase in positive sentiment compared to human responses.
arXiv Detail & Related papers (2024-09-26T02:17:30Z) - Towards Better Answers: Automated Stack Overflow Post Updating [11.85319691188159]
We introduce a novel framework, named Soup (Stack Overflow Updator for Post) for this task.
Soup addresses two key tasks: Valid Comment-Edit Prediction (VCP) and Automatic Post Updating (APU)
arXiv Detail & Related papers (2024-08-17T04:48:53Z) - Answering Ambiguous Questions with a Database of Questions, Answers, and
Revisions [95.92276099234344]
We present a new state-of-the-art for answering ambiguous questions that exploits a database of unambiguous questions generated from Wikipedia.
Our method improves performance by 15% on recall measures and 10% on measures which evaluate disambiguating questions from predicted outputs.
arXiv Detail & Related papers (2023-08-16T20:23:16Z) - Answer ranking in Community Question Answering: a deep learning approach [0.0]
This work tries to advance the state of the art on answer ranking for community Question Answering by proceeding with a deep learning approach.
We created a large data set of questions and answers posted to the Stack Overflow website.
We leveraged the natural language processing capabilities of dense embeddings and LSTM networks to produce a prediction for the accepted answer attribute.
arXiv Detail & Related papers (2022-10-16T18:47:41Z) - Mining Duplicate Questions of Stack Overflow [5.924018537171331]
We propose two neural network based architectures for duplicate question detection on Stack Overflow.
We also propose explicitly modeling the code present in questions to achieve results that surpass the state of the art.
arXiv Detail & Related papers (2022-10-04T14:34:59Z) - Improving the Question Answering Quality using Answer Candidate
Filtering based on Natural-Language Features [117.44028458220427]
We address the problem of how the Question Answering (QA) quality of a given system can be improved.
Our main contribution is an approach capable of identifying wrong answers provided by a QA system.
In particular, our approach has shown its potential while removing in many cases the majority of incorrect answers.
arXiv Detail & Related papers (2021-12-10T11:09:44Z) - AnswerSumm: A Manually-Curated Dataset and Pipeline for Answer
Summarization [73.91543616777064]
Community Question Answering (CQA) fora such as Stack Overflow and Yahoo! Answers contain a rich resource of answers to a wide range of community-based questions.
One goal of answer summarization is to produce a summary that reflects the range of answer perspectives.
This work introduces a novel dataset of 4,631 CQA threads for answer summarization, curated by professional linguists.
arXiv Detail & Related papers (2021-11-11T21:48:02Z) - Features that Predict the Acceptability of Java and JavaScript Answers
on Stack Overflow [5.332217496693262]
We studied the Stack Overflow dataset by analyzing questions and answers for the two most popular tags (Java and JavaScript)
Our findings reveal that the length of code in answers, reputation of users, similarity of the text between questions and answers, and the time lag between questions and answers have the highest predictive power for differentiating accepted and unaccepted answers.
arXiv Detail & Related papers (2021-01-08T03:09:38Z) - Few-Shot Complex Knowledge Base Question Answering via Meta
Reinforcement Learning [55.08037694027792]
Complex question-answering (CQA) involves answering complex natural-language questions on a knowledge base (KB)
The conventional neural program induction (NPI) approach exhibits uneven performance when the questions have different types.
This paper proposes a meta-reinforcement learning approach to program induction in CQA to tackle the potential distributional bias in questions.
arXiv Detail & Related papers (2020-10-29T18:34:55Z) - Retrieve, Program, Repeat: Complex Knowledge Base Question Answering via
Alternate Meta-learning [56.771557756836906]
We present a novel method that automatically learns a retrieval model alternately with the programmer from weak supervision.
Our system leads to state-of-the-art performance on a large-scale task for complex question answering over knowledge bases.
arXiv Detail & Related papers (2020-10-29T18:28:16Z) - Unsupervised Question Decomposition for Question Answering [102.56966847404287]
We propose an algorithm for One-to-N Unsupervised Sequence Sequence (ONUS) that learns to map one hard, multi-hop question to many simpler, single-hop sub-questions.
We show large QA improvements on HotpotQA over a strong baseline on the original, out-of-domain, and multi-hop dev sets.
arXiv Detail & Related papers (2020-02-22T19:40:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.