A Grounded Theory of Debugging in Professional Software Engineering Practice
- URL: http://arxiv.org/abs/2602.11435v1
- Date: Wed, 11 Feb 2026 23:19:45 GMT
- Title: A Grounded Theory of Debugging in Professional Software Engineering Practice
- Authors: Haolin Li, Michael Coblenz,
- Abstract summary: We conducted a study using a grounded theory approach.<n>We observed seven professional developers and five professional live-coding streamers working on 17 debug tasks.
- Score: 5.88596290266904
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Debugging is a central yet complex activity in software engineering. Prior studies have documented debugging strategies and tool usage, but little theory explains how experienced developers reason about bugs in large, real-world codebases. We conducted a qualitative study using a grounded theory approach. We observed seven professional developers and five professional live-coding streamers working on 17 debugging tasks in their own codebases, capturing diverse contexts of debugging. We theorize debugging as a structured, iterative diagnostic process in which programmers update a mental model of the system to guide information gathering. Developers gather information by alternating between navigation and execution strategies, employing forward and backward tracing modes of reasoning and adapting these approaches according to codebase context, complexity, and familiarity. Developers also gather external resources to complement code-based evidence, with their experience enabling them to systematically construct a mental model. We contribute a grounded theory of professional debugging that surfaces the human-centered dimensions of the practice, with implications for tool design and software engineering education.
Related papers
- Debug2Fix: Supercharging Coding Agents with Interactive Debugging Capabilities [3.506382476101256]
We introduce Debug2Fix, a novel framework that incorporates interactive Debug2Fix as a core component of a software engineering agent via a subagent architecture.<n>We evaluate against GitBug-Java and SWE-Bench-Live and achieve >20% improvement in performance compared to the baseline for certain models.<n>Using our framework, we're able to make weaker models like GPT-5 and Claude Haiku 4.5 match or exceed the performances of stronger models like Claude Sonnet 4.5.
arXiv Detail & Related papers (2026-02-20T19:24:16Z) - Coding With AI: From a Reflection on Industrial Practices to Future Computer Science and Software Engineering Education [1.5516092077598485]
Recent advances in large language models (LLMs) have introduced new paradigms in software development.<n>This paper investigates how LLM coding tools are used in professional practice, associated concerns and risks, and the resulting transformations in development.<n>Building on these insights, we argue for curricular shifts toward problem-solving, architectural thinking, code review, and early project-based learning.
arXiv Detail & Related papers (2025-12-30T04:39:59Z) - Building Software by Rolling the Dice: A Qualitative Study of Vibe Coding [15.145249560710377]
"vibe coders" build software primarily through prompts rather than writing code.<n>We conducted a theory study of 20 vibe-coding videos, including 7 live-streamed coding sessions and 13 opinion videos.<n>Our findings reveal a spectrum of behaviors: some vibe coders rely almost entirely on AI, while others examine and adapt generated outputs.
arXiv Detail & Related papers (2025-12-27T00:38:37Z) - Simple Agents Outperform Experts in Biomedical Imaging Workflow Optimization [69.36509281190662]
Adapting production-level computer vision tools to bespoke scientific datasets is a critical "last mile" bottleneck.<n>We consider using AI agents to automate this manual coding, and focus on the open question of optimal agent design.<n>We demonstrate that a simple agent framework consistently generates adaptation code that outperforms human-expert solutions.
arXiv Detail & Related papers (2025-12-02T18:42:26Z) - From Code Foundation Models to Agents and Applications: A Practical Guide to Code Intelligence [150.3696990310269]
Large language models (LLMs) have transformed automated software development by enabling direct translation of natural language descriptions into functional code.<n>We provide a comprehensive synthesis and practical guide (a series of analytic and probing experiments) about code LLMs.<n>We analyze the code capability of the general LLMs (GPT-4, Claude, LLaMA) and code-specialized LLMs (StarCoder, Code LLaMA, DeepSeek-Coder, and QwenCoder)
arXiv Detail & Related papers (2025-11-23T17:09:34Z) - A Survey of Vibe Coding with Large Language Models [93.88284590533242]
"Vibe Coding" is a development methodology where developers validate AI-generated implementations through outcome observation.<n>Despite its transformative potential, the effectiveness of this emergent paradigm remains under-explored.<n>This survey provides the first comprehensive and systematic review of Vibe Coding with large language models.
arXiv Detail & Related papers (2025-10-14T11:26:56Z) - Prompting in Practice: Investigating Software Developers' Use of Generative AI Tools [17.926187565860232]
The integration of generative artificial intelligence (GenAI) tools has fundamentally transformed software development.<n>This study presents a systematic investigation of how software engineers integrate GenAI tools into their professional practice.<n>We surveyed 91 software engineers, including 72 active GenAI users, to understand AI usage patterns throughout the development process.
arXiv Detail & Related papers (2025-10-07T15:02:22Z) - Towards Effective Code-Integrated Reasoning [89.47213509714578]
We investigate code-integrated reasoning, where models generate code when necessary and integrate feedback by executing it through a code interpreter.<n>Tool-augmented reinforcement learning can still suffer from potential instability in the learning dynamics.<n>We develop enhanced training strategies that balance exploration and stability, progressively building tool-use capabilities while improving reasoning performance.
arXiv Detail & Related papers (2025-05-30T11:30:18Z) - ToolCoder: A Systematic Code-Empowered Tool Learning Framework for Large Language Models [81.12673534903979]
Tool learning has emerged as a crucial capability for large language models (LLMs) to solve complex real-world tasks through interaction with external tools.<n>We propose ToolCoder, a novel framework that reformulates tool learning as a code generation task.
arXiv Detail & Related papers (2025-02-17T03:42:28Z) - Code Compass: A Study on the Challenges of Navigating Unfamiliar Codebases [2.808331566391181]
We propose a novel tool, Code, to address these issues.
Our study highlights a significant gap in current tools and methodologies.
Our formative study demonstrates how effectively the tool reduces the time developers spend navigating documentation.
arXiv Detail & Related papers (2024-05-10T06:58:31Z) - Collaborative, Code-Proximal Dynamic Software Visualization within Code
Editors [55.57032418885258]
This paper introduces the design and proof-of-concept implementation for a software visualization approach that can be embedded into code editors.
Our contribution differs from related work in that we use dynamic analysis of a software system's runtime behavior.
Our visualization approach enhances common remote pair programming tools and is collaboratively usable by employing shared code cities.
arXiv Detail & Related papers (2023-08-30T06:35:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.