An Energy-Aware Online Learning Framework for Resource Management in
Heterogeneous Platforms
- URL: http://arxiv.org/abs/2003.09526v1
- Date: Fri, 20 Mar 2020 22:59:35 GMT
- Title: An Energy-Aware Online Learning Framework for Resource Management in
Heterogeneous Platforms
- Authors: Sumit K. Mandal, Ganapati Bhat, Janardhan Rao Doppa, Partha Pratim
Pande, Umit Y. Ogras
- Abstract summary: Mobile platforms must satisfy the contradictory requirements of fast response time and minimum energy consumption.
To address this need, system-on-chips provide a variety of control knobs, such as the number of active cores and their voltage/frequency levels.
Control policies designed offline are at best sub-optimal since many potential new applications are unknown at design-time.
- Score: 16.94738988958929
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Mobile platforms must satisfy the contradictory requirements of fast response
time and minimum energy consumption as a function of dynamically changing
applications. To address this need, system-on-chips (SoC) that are at the heart
of these devices provide a variety of control knobs, such as the number of
active cores and their voltage/frequency levels. Controlling these knobs
optimally at runtime is challenging for two reasons. First, the large
configuration space prohibits exhaustive solutions. Second, control policies
designed offline are at best sub-optimal since many potential new applications
are unknown at design-time. We address these challenges by proposing an online
imitation learning approach. Our key idea is to construct an offline policy and
adapt it online to new applications to optimize a given metric (e.g., energy).
The proposed methodology leverages the supervision enabled by power-performance
models learned at runtime. We demonstrate its effectiveness on a commercial
mobile platform with 16 diverse benchmarks. Our approach successfully adapts
the control policy to an unknown application after executing less than 25% of
its instructions.
Related papers
- Optimizing Load Scheduling in Power Grids Using Reinforcement Learning and Markov Decision Processes [0.0]
This paper proposes a reinforcement learning (RL) approach to address the challenges of dynamic load scheduling.
Our results show that the RL-based method provides a robust and scalable solution for real-time load scheduling.
arXiv Detail & Related papers (2024-10-23T09:16:22Z) - Action-Quantized Offline Reinforcement Learning for Robotic Skill
Learning [68.16998247593209]
offline reinforcement learning (RL) paradigm provides recipe to convert static behavior datasets into policies that can perform better than the policy that collected the data.
In this paper, we propose an adaptive scheme for action quantization.
We show that several state-of-the-art offline RL methods such as IQL, CQL, and BRAC improve in performance on benchmarks when combined with our proposed discretization scheme.
arXiv Detail & Related papers (2023-10-18T06:07:10Z) - Policy Search for Model Predictive Control with Application to Agile
Drone Flight [56.24908013905407]
We propose a policy-search-for-model-predictive-control framework for MPC.
Specifically, we formulate the MPC as a parameterized controller, where the hard-to-optimize decision variables are represented as high-level policies.
Experiments show that our controller achieves robust and real-time control performance in both simulation and the real world.
arXiv Detail & Related papers (2021-12-07T17:39:24Z) - Towards Automatic Actor-Critic Solutions to Continuous Control [7.312692481631664]
This paper creates an evolutionary approach that tunes actor-critic algorithms to new domains.
Our design is sample efficient and provides practical advantages over baseline approaches.
We then apply it to new control tasks to find high-performance solutions with minimal compute and research effort.
arXiv Detail & Related papers (2021-06-16T16:18:20Z) - Enforcing Policy Feasibility Constraints through Differentiable
Projection for Energy Optimization [57.88118988775461]
We propose PROjected Feasibility (PROF) to enforce convex operational constraints within neural policies.
We demonstrate PROF on two applications: energy-efficient building operation and inverter control.
arXiv Detail & Related papers (2021-05-19T01:58:10Z) - Non-stationary Online Learning with Memory and Non-stochastic Control [71.14503310914799]
We study the problem of Online Convex Optimization (OCO) with memory, which allows loss functions to depend on past decisions.
In this paper, we introduce dynamic policy regret as the performance measure to design algorithms robust to non-stationary environments.
We propose a novel algorithm for OCO with memory that provably enjoys an optimal dynamic policy regret in terms of time horizon, non-stationarity measure, and memory length.
arXiv Detail & Related papers (2021-02-07T09:45:15Z) - MLComp: A Methodology for Machine Learning-based Performance Estimation
and Adaptive Selection of Pareto-Optimal Compiler Optimization Sequences [10.200899224740871]
We propose a novel Reinforcement Learning-based policy methodology for embedded software optimization.
We show that different Machine Learning models are automatically tested to choose the best-fitting one.
We also show that our framework can be trained efficiently for any target platform and application domain.
arXiv Detail & Related papers (2020-12-09T19:13:39Z) - Neural Dynamic Policies for End-to-End Sensorimotor Learning [51.24542903398335]
The current dominant paradigm in sensorimotor control, whether imitation or reinforcement learning, is to train policies directly in raw action spaces.
We propose Neural Dynamic Policies (NDPs) that make predictions in trajectory distribution space.
NDPs outperform the prior state-of-the-art in terms of either efficiency or performance across several robotic control tasks.
arXiv Detail & Related papers (2020-12-04T18:59:32Z) - Scheduling and Power Control for Wireless Multicast Systems via Deep
Reinforcement Learning [33.737301955006345]
Multicasting in wireless systems is a way to exploit the redundancy in user requests in a Content Centric Network.
Power control and optimal scheduling can significantly improve the wireless multicast network's performance under fading.
We show that power control policy can be learnt for reasonably large systems via this approach.
arXiv Detail & Related papers (2020-09-27T15:59:44Z) - Online Adaptive Learning for Runtime Resource Management of
Heterogeneous SoCs [15.523246628432654]
This paper describes the use of predictive models for online control using two modern approaches: imitation learning (IL) and an explicit nonlinear model predictive control (NMPC)
Evaluations on a commercial mobile platform with 16 benchmarks show that the IL approach successfully adapts the control policy to unknown applications.
The explicit NMPC provides 25% energy savings compared to a state-of-the-art algorithm for multi-variable power management of modern GPU sub-systems.
arXiv Detail & Related papers (2020-08-22T01:39:32Z) - Learning High-Level Policies for Model Predictive Control [54.00297896763184]
Model Predictive Control (MPC) provides robust solutions to robot control tasks.
We propose a self-supervised learning algorithm for learning a neural network high-level policy.
We show that our approach can handle situations that are difficult for standard MPC.
arXiv Detail & Related papers (2020-07-20T17:12:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.