Design and Comparison of Reward Functions in Reinforcement Learning for
Energy Management of Sensor Nodes
- URL: http://arxiv.org/abs/2106.01114v1
- Date: Wed, 2 Jun 2021 12:23:47 GMT
- Title: Design and Comparison of Reward Functions in Reinforcement Learning for
Energy Management of Sensor Nodes
- Authors: Yohann Rioual (1), Yannick Le Moullec (2), Johann Laurent (1), Muhidul
Islam Khan (2) and Jean-Philippe Diguet (3) ((1) Lab-STICC, University
Bretagne Sud, (2) Thomas Johann Seebeck Department of Electronics, Tallinn
University of Technology, (3) IRL CNRS CROSSING)
- Abstract summary: Interest in remote monitoring has grown thanks to recent advancements in Internet-of-Things (IoT) paradigms.
New applications have emerged, using small devices called sensor nodes capable of collecting data from the environment and processing it.
Battery technologies have not improved fast enough to cope with these increasing needs.
Miniature energy harvesting devices have emerged to complement traditional energy sources.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Interest in remote monitoring has grown thanks to recent advancements in
Internet-of-Things (IoT) paradigms. New applications have emerged, using small
devices called sensor nodes capable of collecting data from the environment and
processing it. However, more and more data are processed and transmitted with
longer operational periods. At the same, the battery technologies have not
improved fast enough to cope with these increasing needs. This makes the energy
consumption issue increasingly challenging and thus, miniaturized energy
harvesting devices have emerged to complement traditional energy sources.
Nevertheless, the harvested energy fluctuates significantly during the node
operation, increasing uncertainty in actually available energy resources.
Recently, approaches in energy management have been developed, in particular
using reinforcement learning approaches. However, in reinforcement learning,
the algorithm's performance relies greatly on the reward function. In this
paper, we present two contributions. First, we explore five different reward
functions to identify the most suitable variables to use in such functions to
obtain the desired behaviour. Experiments were conducted using the Q-learning
algorithm to adjust the energy consumption depending on the energy harvested.
Results with the five reward functions illustrate how the choice thereof
impacts the energy consumption of the node. Secondly, we propose two additional
reward functions able to find the compromise between energy consumption and a
node performance using a non-fixed balancing parameter. Our simulation results
show that the proposed reward functions adjust the node's performance depending
on the battery level and reduce the learning time.
Related papers
- Energy-Aware Dynamic Neural Inference [39.04688735618206]
We introduce an on-device adaptive inference system equipped with an energy-harvester and finite-capacity energy storage.
We show that, as the rate of the ambient energy increases, energy- and confidence-aware control schemes show approximately 5% improvement in accuracy.
We derive a principled policy with theoretical guarantees for confidence-aware and -agnostic controllers.
arXiv Detail & Related papers (2024-11-04T16:51:22Z) - Learning Iterative Reasoning through Energy Diffusion [90.24765095498392]
We introduce iterative reasoning through energy diffusion (IRED), a novel framework for learning to reason for a variety of tasks.
IRED learns energy functions to represent the constraints between input conditions and desired outputs.
We show IRED outperforms existing methods in continuous-space reasoning, discrete-space reasoning, and planning tasks.
arXiv Detail & Related papers (2024-06-17T03:36:47Z) - PolyThrottle: Energy-efficient Neural Network Inference on Edge Devices [10.01838504586422]
The continuous operation of ML-powered systems leads to significant energy use during inference.
This paper investigates how the configuration of on-device hardware-elements such as GPU, memory, and CPU frequency, affects energy consumption for NN inference with regular fine-tuning.
We propose PolyThrottle, a solution that optimize configurations across individual hardware components using Constrained Bayesian Optimization in an energy-conserving manner.
arXiv Detail & Related papers (2023-10-30T20:19:41Z) - Unveiling Energy Efficiency in Deep Learning: Measurement, Prediction, and Scoring across Edge Devices [8.140572894424208]
We conduct a threefold study, including energy measurement, prediction, and efficiency scoring.
Firstly, we present a detailed, first-of-its-kind measurement study that uncovers the energy consumption characteristics of on-device deep learning.
Secondly, we design and implement the first kernel-level energy predictors for edge devices based on our kernel-level energy dataset.
arXiv Detail & Related papers (2023-10-19T23:55:00Z) - Enhancing Energy-Awareness in Deep Learning through Fine-Grained Energy
Measurement [11.37120215795946]
This paper introduces FECoM (Fine-grained Energy Consumption Meter), a framework for fine-grained Deep Learning energy consumption measurement.
FECoM addresses the challenges of measuring energy consumption at fine-grained level by using static instrumentation and considering various factors, including computational load stability and temperature.
arXiv Detail & Related papers (2023-08-23T17:32:06Z) - Precise Energy Consumption Measurements of Heterogeneous Artificial
Intelligence Workloads [0.534434568021034]
We present measurements of the energy consumption of two typical applications of deep learning models on different types of compute nodes.
One advantage of our approach is that the information on energy consumption is available to all users of the supercomputer.
arXiv Detail & Related papers (2022-12-03T21:40:55Z) - Energy-based Latent Aligner for Incremental Learning [83.0135278697976]
Deep learning models tend to forget their earlier knowledge while incrementally learning new tasks.
This behavior emerges because the parameter updates optimized for the new tasks may not align well with the updates suitable for older tasks.
We propose ELI: Energy-based Latent Aligner for Incremental Learning.
arXiv Detail & Related papers (2022-03-28T17:57:25Z) - Deep Reinforcement Learning Based Multidimensional Resource Management
for Energy Harvesting Cognitive NOMA Communications [64.1076645382049]
Combination of energy harvesting (EH), cognitive radio (CR), and non-orthogonal multiple access (NOMA) is a promising solution to improve energy efficiency.
In this paper, we study the spectrum, energy, and time resource management for deterministic-CR-NOMA IoT systems.
arXiv Detail & Related papers (2021-09-17T08:55:48Z) - Compute and Energy Consumption Trends in Deep Learning Inference [67.32875669386488]
We study relevant models in the areas of computer vision and natural language processing.
For a sustained increase in performance we see a much softer growth in energy consumption than previously anticipated.
arXiv Detail & Related papers (2021-09-12T09:40:18Z) - To Talk or to Work: Flexible Communication Compression for Energy
Efficient Federated Learning over Heterogeneous Mobile Edge Devices [78.38046945665538]
federated learning (FL) over massive mobile edge devices opens new horizons for numerous intelligent mobile applications.
FL imposes huge communication and computation burdens on participating devices due to periodical global synchronization and continuous local training.
We develop a convergence-guaranteed FL algorithm enabling flexible communication compression.
arXiv Detail & Related papers (2020-12-22T02:54:18Z) - Energy Aware Deep Reinforcement Learning Scheduling for Sensors
Correlated in Time and Space [62.39318039798564]
We propose a scheduling mechanism capable of taking advantage of correlated information.
The proposed mechanism is capable of determining the frequency with which sensors should transmit their updates.
We show that our solution can significantly extend the sensors' lifetime.
arXiv Detail & Related papers (2020-11-19T09:53:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.