Demystifying Feature Requests: Leveraging LLMs to Refine Feature Requests in Open-Source Software
- URL: http://arxiv.org/abs/2507.13555v1
- Date: Thu, 17 Jul 2025 22:16:13 GMT
- Title: Demystifying Feature Requests: Leveraging LLMs to Refine Feature Requests in Open-Source Software
- Authors: Pragyan K C, Rambod Ghandiparsi, Thomas Herron, John Heaps, Mitra Bokaei Hosseini,
- Abstract summary: This paper proposes a novel approach leveraging Large Language Models (LLMs) to detect and refine NL defects in feature requests.<n>Our approach automates the identification of ambiguous and incomplete requests and generates clarification questions (CQs) to enhance their usefulness for developers.<n>In addition, we conduct interviews with GitHub developers to gain deeper insights into their perceptions of NL defects.
- Score: 0.2748831616311481
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The growing popularity and widespread use of software applications (apps) across various domains have driven rapid industry growth. Along with this growth, fast-paced market changes have led to constantly evolving software requirements. Such requirements are often grounded in feature requests and enhancement suggestions, typically provided by users in natural language (NL). However, these requests often suffer from defects such as ambiguity and incompleteness, making them challenging to interpret. Traditional validation methods (e.g., interviews and workshops) help clarify such defects but are impractical in decentralized environments like open-source software (OSS), where change requests originate from diverse users on platforms like GitHub. This paper proposes a novel approach leveraging Large Language Models (LLMs) to detect and refine NL defects in feature requests. Our approach automates the identification of ambiguous and incomplete requests and generates clarification questions (CQs) to enhance their usefulness for developers. To evaluate its effectiveness, we apply our method to real-world OSS feature requests and compare its performance against human annotations. In addition, we conduct interviews with GitHub developers to gain deeper insights into their perceptions of NL defects, the strategies they use to address these defects, and the impact of defects on downstream software engineering (SE) tasks.
Related papers
- Towards Better Requirements from the Crowd: Developer Engagement with Feature Requests in Open Source Software [0.2748831616311481]
This study investigates how feature requests are prone to NL defects (i.e. ambiguous or incomplete) and the conversational dynamics of clarification in open-source software development.<n>Our findings suggest that feature requests published on the OSS platforms do possess ambiguity and incompleteness, and in some cases, both.<n>When clarification occurs, it emphasizes understanding user intent/goal and feasibility, rather than technical details.
arXiv Detail & Related papers (2025-07-17T22:04:29Z) - ONLY: One-Layer Intervention Sufficiently Mitigates Hallucinations in Large Vision-Language Models [67.75439511654078]
Large Vision-Language Models (LVLMs) have introduced a new paradigm for understanding and reasoning about image input through textual responses.<n>They face the persistent challenge of hallucination, which introduces practical weaknesses and raises concerns about their reliable deployment in real-world applications.<n>We propose ONLY, a training-free decoding approach that requires only a single query and a one-layer intervention during decoding, enabling efficient real-time deployment.
arXiv Detail & Related papers (2025-07-01T16:01:08Z) - Automated Repair of Ambiguous Natural Language Requirements [9.379494157034083]
Large language models (LLMs) in software engineering have amplified the role of natural language (NL)<n>We introduce automated repair of ambiguous NL requirements, which we approach by reducing code generation uncertainty.<n>Our results show that SpecFix modifies 23.93% of the requirements, leading to a 33.66% improvement in model Pass@1 on the modified requirements.
arXiv Detail & Related papers (2025-05-12T06:47:53Z) - SweRank: Software Issue Localization with Code Ranking [109.3289316191729]
SweRank is an efficient retrieve-and-rerank framework for software issue localization.<n>We construct SweLoc, a large-scale dataset curated from public GitHub repositories.<n>We show that SweRank achieves state-of-the-art performance, outperforming both prior ranking models and costly agent-based systems.
arXiv Detail & Related papers (2025-05-07T19:44:09Z) - Thinking Longer, Not Larger: Enhancing Software Engineering Agents via Scaling Test-Time Compute [61.00662702026523]
We propose a unified Test-Time Compute scaling framework that leverages increased inference-time instead of larger models.<n>Our framework incorporates two complementary strategies: internal TTC and external TTC.<n>We demonstrate our textbf32B model achieves a 46% issue resolution rate, surpassing significantly larger models such as DeepSeek R1 671B and OpenAI o1.
arXiv Detail & Related papers (2025-03-31T07:31:32Z) - From Bugs to Benefits: Improving User Stories by Leveraging Crowd Knowledge with CrUISE-AC [0.0]
We present CrUISE-AC as a fully automated method that investigates issues and generates non-trivial additional acceptance criteria for a given user story.<n>Our evaluation shows that 80-82% of the generated acceptance criteria add relevant requirements to the user stories.
arXiv Detail & Related papers (2025-01-25T11:44:24Z) - Seeker: Towards Exception Safety Code Generation with Intermediate Language Agents Framework [58.36391985790157]
In real world software development, improper or missing exception handling can severely impact the robustness and reliability of code.<n>We explore the use of large language models (LLMs) to improve exception handling in code.<n>We propose Seeker, a multi-agent framework inspired by expert developer strategies for exception handling.
arXiv Detail & Related papers (2024-12-16T12:35:29Z) - LLMs as Continuous Learners: Improving the Reproduction of Defective Code in Software Issues [62.12404317786005]
EvoCoder is a continuous learning framework for issue code reproduction.
Our results show a 20% improvement in issue reproduction rates over existing SOTA methods.
arXiv Detail & Related papers (2024-11-21T08:49:23Z) - The Transformative Influence of LLMs on Software Development & Developer Productivity [0.0]
Generalized Large Language Models (LLMs) have found their way into diverse domains.<n>With LLMs increasingly serving as AI Pair Programming Assistants, it also presents critical challenges and open problems.<n>Preliminary findings underscore pressing concerns about data privacy, bias, and misinformation.<n>We identify 12 open problems that we have identified through our survey, covering these various domains.
arXiv Detail & Related papers (2023-11-28T02:18:54Z) - Automatically Correcting Large Language Models: Surveying the landscape
of diverse self-correction strategies [104.32199881187607]
Large language models (LLMs) have demonstrated remarkable performance across a wide array of NLP tasks.
A promising approach to rectify these flaws is self-correction, where the LLM itself is prompted or guided to fix problems in its own output.
This paper presents a comprehensive review of this emerging class of techniques.
arXiv Detail & Related papers (2023-08-06T18:38:52Z) - Mining Reddit Data to Elicit Students' Requirements During COVID-19
Pandemic [2.5475486924467075]
We propose a shift in requirements elicitation, focusing on gathering feedback related to the problem itself.
We conducted a case study on student requirements during the COVID-19 pandemic in a higher education institution.
We employed multiple machine-learning and natural language processing techniques to identify requirement sentences.
arXiv Detail & Related papers (2023-07-26T14:26:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.