Source Code Classification for Energy Efficiency in Parallel Ultra
Low-Power Microcontrollers
- URL: http://arxiv.org/abs/2012.06836v1
- Date: Sat, 12 Dec 2020 15:12:03 GMT
- Title: Source Code Classification for Energy Efficiency in Parallel Ultra
Low-Power Microcontrollers
- Authors: Emanuele Parisi, Francesco Barchi, Andrea Bartolini, Giuseppe
Tagliavini, Andrea Acquaviva
- Abstract summary: This paper aims at increasing smartness in the software toolchain to exploit modern architectures in the best way.
In the case of low-power, parallel embedded architectures, this means finding the configuration, for instance in terms of the number of cores, leading to minimum energy consumption.
Experiments show that using machine learning models on the source code to select the best energy scaling configuration automatically is viable and has the potential to be used in the context of automatic system configuration for energy minimisation.
- Score: 5.4352987210173955
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The analysis of source code through machine learning techniques is an
increasingly explored research topic aiming at increasing smartness in the
software toolchain to exploit modern architectures in the best possible way. In
the case of low-power, parallel embedded architectures, this means finding the
configuration, for instance in terms of the number of cores, leading to minimum
energy consumption. Depending on the kernel to be executed, the energy optimal
scaling configuration is not trivial. While recent work has focused on
general-purpose systems to learn and predict the best execution target in terms
of the execution time of a snippet of code or kernel (e.g. offload OpenCL
kernel on multicore CPU or GPU), in this work we focus on static compile-time
features to assess if they can be successfully used to predict the minimum
energy configuration on PULP, an ultra-low-power architecture featuring an
on-chip cluster of RISC-V processors. Experiments show that using machine
learning models on the source code to select the best energy scaling
configuration automatically is viable and has the potential to be used in the
context of automatic system configuration for energy minimisation.
Related papers
- Kernel Approximation using Analog In-Memory Computing [3.5231018007564203]
Kernel functions are vital ingredients of several machine learning algorithms, but often incur significant memory and computational costs.
We introduce an approach to kernel approximation in machine learning algorithms suitable for mixed-signal Analog In-Memory Computing (AIMC) architectures.
arXiv Detail & Related papers (2024-11-05T16:18:47Z) - Harnessing Deep Learning and HPC Kernels via High-Level Loop and Tensor Abstractions on CPU Architectures [67.47328776279204]
This work introduces a framework to develop efficient, portable Deep Learning and High Performance Computing kernels.
We decompose the kernel development in two steps: 1) Expressing the computational core using Processing Primitives (TPPs) and 2) Expressing the logical loops around TPPs in a high-level, declarative fashion.
We demonstrate the efficacy of our approach using standalone kernels and end-to-end workloads that outperform state-of-the-art implementations on diverse CPU platforms.
arXiv Detail & Related papers (2023-04-25T05:04:44Z) - Power Constrained Autotuning using Graph Neural Networks [1.7188280334580197]
We propose a novel Graph Neural Network based auto-tuning approach to improve the performance, power, and energy efficiency of scientific applications on modern processors.
Our approach identifies OpenMP configurations at different power constraints that yield a mean geometric performance improvement of more than $25%$ and $13%$ over the default OpenMP configuration.
arXiv Detail & Related papers (2023-02-22T16:06:00Z) - Towards making the most of NLP-based device mapping optimization for
OpenCL kernels [5.6596607119831575]
We extend the work of Cummins et al., namely Deeptune, that tackles the problem of optimal device selection ( CPU or GPU) for accelerated OpenCL kernels.
We propose four different models that provide enhanced contextual information of source codes.
Experimental results show that our proposed methodology surpasses that of Cummins et al. work, providing up to 4% improvement in prediction accuracy.
arXiv Detail & Related papers (2022-08-30T10:20:55Z) - Towards Optimal VPU Compiler Cost Modeling by using Neural Networks to
Infer Hardware Performances [58.720142291102135]
'VPUNN' is a neural network-based cost model trained on low-level task profiling.
It consistently outperforms the state-of-the-art cost modeling in Intel's line of VPU processors.
arXiv Detail & Related papers (2022-05-09T22:48:39Z) - An Adaptive Device-Edge Co-Inference Framework Based on Soft
Actor-Critic [72.35307086274912]
High-dimension parameter model and large-scale mathematical calculation restrict execution efficiency, especially for Internet of Things (IoT) devices.
We propose a new Deep Reinforcement Learning (DRL)-Soft Actor Critic for discrete (SAC-d), which generates the emphexit point, emphexit point, and emphcompressing bits by soft policy iterations.
Based on the latency and accuracy aware reward design, such an computation can well adapt to the complex environment like dynamic wireless channel and arbitrary processing, and is capable of supporting the 5G URL
arXiv Detail & Related papers (2022-01-09T09:31:50Z) - Towards High Performance Relativistic Electronic Structure Modelling:
The EXP-T Program Package [68.8204255655161]
We present a new implementation of the FS-RCC method designed for modern parallel computers.
The performance and scaling features of the implementation are analyzed.
The software developed allows to achieve a completely new level of accuracy for prediction of properties of atoms and molecules containing heavy and superheavy nuclei.
arXiv Detail & Related papers (2020-04-07T20:08:30Z) - Towards automated kernel selection in machine learning systems: A SYCL
case study [0.0]
We present initial results using machine learning to select kernels in a case study deploying high performance SYCL kernels in libraries.
By combining auto-tuning and machine learning these kernel selection processes can be deployed with little developer effort to achieve high performance on new hardware.
arXiv Detail & Related papers (2020-03-15T11:23:36Z) - On Coresets for Support Vector Machines [61.928187390362176]
A coreset is a small, representative subset of the original data points.
We show that our algorithm can be used to extend the applicability of any off-the-shelf SVM solver to streaming, distributed, and dynamic data settings.
arXiv Detail & Related papers (2020-02-15T23:25:12Z) - PolyScientist: Automatic Loop Transformations Combined with Microkernels
for Optimization of Deep Learning Primitives [55.79741270235602]
We develop a hybrid solution to the development of deep learning kernels.
We use the advanced polyhedral technology to automatically tune the outer loops for performance.
arXiv Detail & Related papers (2020-02-06T08:02:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.