Why Do We Code? A Theory on Motivations and Challenges in Software Engineering from Education to Practice
- URL: http://arxiv.org/abs/2511.14711v1
- Date: Tue, 18 Nov 2025 17:54:36 GMT
- Title: Why Do We Code? A Theory on Motivations and Challenges in Software Engineering from Education to Practice
- Authors: Aaliyah Chang, Mariam Guizani, Brittany Johnson,
- Abstract summary: Motivations and challenges jointly shape how individuals enter, persist, and evolve within software engineering (SE)<n>Our findings show how unmet motivations and recurring challenges influence persistence, career shifts, or departure from the field.
- Score: 6.2843244159918505
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Motivations and challenges jointly shape how individuals enter, persist, and evolve within software engineering (SE), yet their interplay remains underexplored across the transition from education to professional practice. We conducted 15 semi-structured interviews and employed the Gioia Methodology, an adapted grounded theory methodology from organizational behavior, to inductively derive taxonomies of motivations and challenges, and build the Exposure-Pursuit-Evaluation (EPE) Process Model. Our findings reveal that impactful early exposure triggers intrinsic motivations, while non-impactful exposure requires an extrinsic push (e.g., career/ personal goals, external validation). We identify curiosity and avoiding alternatives as a distinct educational drivers, and barriers to belonging as the only challenge persisting across education and career. Our findings show that career progression challenges (e.g., navigating the corporate world) constrain extrinsic fulfillment while technical training challenges, barriers to belonging and threats to motivation constrain intrinsic fulfillment. The theory shows how unmet motivations and recurring challenges influence persistence, career shifts, or departure from the field. Our results provide a grounded model for designing interventions that strengthen intrinsic fulfillment and reduce systemic barriers in SE education and practice.
Related papers
- Capability-Oriented Training Induced Alignment Risk [101.37328448441208]
We investigate whether language models, when trained with reinforcement learning, will spontaneously learn to exploit flaws to maximize their reward.<n>Our experiments show that models consistently learn to exploit these vulnerabilities, discovering opportunistic strategies that significantly increase their reward at the expense of task correctness or safety.<n>Our findings suggest that future AI safety work must extend beyond content moderation to rigorously auditing and securing the training environments and reward mechanisms themselves.
arXiv Detail & Related papers (2026-02-12T16:13:14Z) - Exit the Code: A Model for Understanding Career Abandonment Intention Among Software Developers [0.6117371161379209]
Career abandonment involves frustration with the lost investment and emotional and financial costs.<n>Previous studies have identified work-related motivators for career abandonment, such as the threat of obsolescence.<n>This article investigates the relationship between these motivators and the intention to abandon among currently active developers.
arXiv Detail & Related papers (2025-03-06T14:13:10Z) - Teleology-Driven Affective Computing: A Causal Framework for Sustained Well-Being [0.1636303041090359]
We propose a teleology-driven affective computing framework that unifies major emotion theories.<n>We advocate for creating a "dataverse" of personal affective events.<n>We introduce a meta-reinforcement learning paradigm to train agents in simulated environments.
arXiv Detail & Related papers (2025-02-24T14:07:53Z) - ReVISE: Learning to Refine at Test-Time via Intrinsic Self-Verification [53.80183105328448]
Refine via Intrinsic Self-Verification (ReVISE) is an efficient framework that enables LLMs to self-correct their outputs through self-verification.<n>Our experiments on various reasoning tasks demonstrate that ReVISE achieves efficient self-correction and significantly improves reasoning performance.
arXiv Detail & Related papers (2025-02-20T13:50:02Z) - Intrinsic Motivation in Model-based Reinforcement Learning: A Brief
Review [77.34726150561087]
This review considers the existing methods for determining intrinsic motivation based on the world model obtained by the agent.
The proposed unified framework describes the architecture of agents using a world model and intrinsic motivation to improve learning.
arXiv Detail & Related papers (2023-01-24T15:13:02Z) - CIM: Constrained Intrinsic Motivation for Sparse-Reward Continuous
Control [25.786085434943338]
Intrinsic motivation is a promising technique for solving reinforcement learning tasks with sparse or absent extrinsic rewards.
There exist two technical challenges in implementing intrinsic motivation.
We propose Constrained Intrinsic Motivation (CIM) to leverage readily attainable task priors to construct a constrained intrinsic objective.
We empirically show, that our CIM approach achieves greatly improved performance and sample efficiency over state-of-the-art methods.
arXiv Detail & Related papers (2022-11-28T10:23:56Z) - Remote Teaching and Learning in Applied Engineering: A Post-Pandemic
Perspective [0.548253258922555]
We provide a rationale for a variety of course delivery models at different stages of the pandemic.
We discuss how we ensured that hands-on learning remains an integral part of engineering curricula.
arXiv Detail & Related papers (2021-08-05T16:28:05Z) - Variational Dynamic for Self-Supervised Exploration in Deep Reinforcement Learning [12.76337275628074]
In this work, we propose a variational dynamic model based on the conditional variational inference to model the multimodality andgenerativeity.
We derive an upper bound of the negative log-likelihood of the environmental transition and use such an upper bound as the intrinsic reward for exploration.
Our method outperforms several state-of-the-art environment model-based exploration approaches.
arXiv Detail & Related papers (2020-10-17T09:54:51Z) - Tracking Emotions: Intrinsic Motivation Grounded on Multi-Level
Prediction Error Dynamics [68.8204255655161]
We discuss how emotions arise when differences between expected and actual rates of progress towards a goal are experienced.
We present an intrinsic motivation architecture that generates behaviors towards self-generated and dynamic goals.
arXiv Detail & Related papers (2020-07-29T06:53:13Z) - Learning with AMIGo: Adversarially Motivated Intrinsic Goals [63.680207855344875]
AMIGo is a goal-generating teacher that proposes Adversarially Motivated Intrinsic Goals.
We show that our method generates a natural curriculum of self-proposed goals which ultimately allows the agent to solve challenging procedurally-generated tasks.
arXiv Detail & Related papers (2020-06-22T10:22:08Z) - Intrinsic Motivation for Encouraging Synergistic Behavior [55.10275467562764]
We study the role of intrinsic motivation as an exploration bias for reinforcement learning in sparse-reward synergistic tasks.
Our key idea is that a good guiding principle for intrinsic motivation in synergistic tasks is to take actions which affect the world in ways that would not be achieved if the agents were acting on their own.
arXiv Detail & Related papers (2020-02-12T19:34:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.