Reinforcement Learning-based Task Offloading in the Internet of Wearable Things
- URL: http://arxiv.org/abs/2510.07487v1
- Date: Wed, 08 Oct 2025 19:36:35 GMT
- Title: Reinforcement Learning-based Task Offloading in the Internet of Wearable Things
- Authors: Waleed Bin Qaim, Aleksandr Ometov, Claudia Campolo, Antonella Molinaro, Elena Simona Lohan, Jari Nurmi,
- Abstract summary: This paper proposes a framework for Reinforcement Learning (RL)-based task offloading in the Internet of Wearable Things (IoWT)<n>We formulate the task offloading process considering the tradeoff between energy consumption and task accomplishment time.<n>We utilize the Q-learning technique to enable the wearable device to make optimal task offloading decisions without prior knowledge.
- Score: 35.87301608329458
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Over the years, significant contributions have been made by the research and industrial sectors to improve wearable devices towards the Internet of Wearable Things (IoWT) paradigm. However, wearables are still facing several challenges. Many stem from the limited battery power and insufficient computation resources available on wearable devices. On the other hand, with the popularity of smart wearables, there is a consistent increase in the development of new computationally intensive and latency-critical applications. In such a context, task offloading allows wearables to leverage the resources available on nearby edge devices to enhance the overall user experience. This paper proposes a framework for Reinforcement Learning (RL)-based task offloading in the IoWT. We formulate the task offloading process considering the tradeoff between energy consumption and task accomplishment time. Moreover, we model the task offloading problem as a Markov Decision Process (MDP) and utilize the Q-learning technique to enable the wearable device to make optimal task offloading decisions without prior knowledge. We evaluate the performance of the proposed framework through extensive simulations for various applications and system configurations conducted in the ns-3 network simulator. We also show how varying the main system parameters of the Q-learning algorithm affects the overall performance in terms of average task accomplishment time, average energy consumption, and percentage of tasks offloaded.
Related papers
- Joint Resource Management for Energy-efficient UAV-assisted SWIPT-MEC: A Deep Reinforcement Learning Approach [50.52139512096988]
6G Internet of Things (IoT) networks face challenges in remote areas and disaster scenarios where ground infrastructure is unavailable.<n>This paper proposes a novel aerial unmanned vehicle (UAV)-assisted computing (MEC) system enhanced by directional antennas to provide both computational and energy support for ground edge terminals.
arXiv Detail & Related papers (2025-05-06T06:46:19Z) - Intelligent Task Offloading in VANETs: A Hybrid AI-Driven Approach for Low-Latency and Energy Efficiency [2.1877558143992184]
Vehicular Ad-hoc Networks (VANETs) are integral to intelligent transportation systems.<n>VANETs enable vehicles to offload computational tasks to nearby roadside units (RSUs) and mobile edge computing (MEC) servers for real-time processing.<n>This research proposes a hybrid AI framework that integrates supervised learning, reinforcement learning, and Particle Swarm Optimization (PSO) for intelligent task offloading and resource allocation.
arXiv Detail & Related papers (2025-04-29T13:20:02Z) - Secure Resource Allocation via Constrained Deep Reinforcement Learning [49.15061461220109]
We present SARMTO, a framework that balances resource allocation, task offloading, security, and performance.<n>SARMTO consistently outperforms five baseline approaches, achieving up to a 40% reduction in system costs.<n>These enhancements highlight SARMTO's potential to revolutionize resource management in intricate distributed computing environments.
arXiv Detail & Related papers (2025-01-20T15:52:43Z) - Task Graph offloading via Deep Reinforcement Learning in Mobile Edge Computing [6.872434270841794]
This paper investigates the task graph offloading in MEC, considering the time-varying capabilities of edge computing devices.
To adapt to environmental changes, we model the task graph scheduling for computation offloading as a Markov Decision Process.
Then, we design a deep reinforcement learning algorithm (SATA-DRL) to learn the task scheduling strategy from the interaction with the environment.
arXiv Detail & Related papers (2023-09-19T12:26:56Z) - FELARE: Fair Scheduling of Machine Learning Applications on
Heterogeneous Edge Systems [5.165692107696155]
Edge computing enables smart IoT-based systems via concurrent and continuous execution of latency-sensitive machine learning (ML) applications.
We study and analyze resource allocation solutions that can increase the on-time task completion rate while considering the energy constraint.
We observed 8.9% improvement in on-time task completion rate and 12.6% in energy-saving without imposing any significant overhead on the edge system.
arXiv Detail & Related papers (2022-05-31T19:19:40Z) - Energy-Efficient Multi-Orchestrator Mobile Edge Learning [54.28419430315478]
Mobile Edge Learning (MEL) is a collaborative learning paradigm that features distributed training of Machine Learning (ML) models over edge devices.
In MEL, possible coexistence of multiple learning tasks with different datasets may arise.
We propose lightweight algorithms that can achieve near-optimal performance and facilitate the trade-offs between energy consumption, accuracy, and solution complexity.
arXiv Detail & Related papers (2021-09-02T07:37:10Z) - Deep Reinforcement Learning for Delay-Oriented IoT Task Scheduling in
Space-Air-Ground Integrated Network [24.022108191145527]
We investigate a computing task scheduling problem in space-air-ground integrated network (SAGIN) for delay-oriented Internet of Things (IoT) services.
In the considered scenario, an unmanned aerial vehicle (UAV) collects computing tasks from IoT devices and then makes online offloading decisions.
Our objective is to design a task scheduling policy that minimizes offloading and computing delay of all tasks given the UAV energy capacity constraint.
arXiv Detail & Related papers (2020-10-04T02:58:03Z) - Green Offloading in Fog-Assisted IoT Systems: An Online Perspective
Integrating Learning and Control [20.68436820937947]
In fog-assisted IoT systems, it is a common practice to offload tasks from IoT devices to their nearby fog nodes to reduce task processing latencies and energy consumptions.
In this paper, we formulate such a task offloading problem with unknown system dynamics as a multi-armed bandit (CMAB) problem with long-term constraints on time-averaged energy consumptions.
Through an effective integration of online learning and online control, we propose a textitLearning-Aided Green Offloading (LAGO) scheme.
arXiv Detail & Related papers (2020-08-01T07:27:24Z) - Federated Learning for Task and Resource Allocation in Wireless High
Altitude Balloon Networks [160.96150373385768]
The problem of minimizing energy and time consumption for task computation and transmission is studied in a mobile edge computing (MEC)-enabled balloon network.
A support vector machine (SVM)-based federated learning (FL) algorithm is proposed to determine the user association proactively.
The proposed SVM-based FL method enables each HAB to cooperatively build an SVM model that can determine all user associations.
arXiv Detail & Related papers (2020-03-19T14:18:25Z) - Risk-Aware Energy Scheduling for Edge Computing with Microgrid: A
Multi-Agent Deep Reinforcement Learning Approach [82.6692222294594]
We study a risk-aware energy scheduling problem for a microgrid-powered MEC network.
We derive the solution by applying a multi-agent deep reinforcement learning (MADRL)-based advantage actor-critic (A3C) algorithm with shared neural networks.
arXiv Detail & Related papers (2020-02-21T02:14:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.