MAPLE-Edge: A Runtime Latency Predictor for Edge Devices
- URL: http://arxiv.org/abs/2204.12950v1
- Date: Wed, 27 Apr 2022 14:00:48 GMT
- Title: MAPLE-Edge: A Runtime Latency Predictor for Edge Devices
- Authors: Saeejith Nair, Saad Abbasi, Alexander Wong, Mohammad Javad Shafiee
- Abstract summary: We propose MAPLE-Edge, an edge device-oriented extension of MAPLE, the state-of-the-art latency predictor for general purpose hardware.
Compared to MAPLE, MAPLE-Edge can describe the runtime and target device platform using a much smaller set of CPU performance counters.
We also demonstrate that unlike MAPLE which performs best when trained on a pool of devices sharing a common runtime, MAPLE-Edge can effectively generalize across runtimes.
- Score: 80.01591186546793
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural Architecture Search (NAS) has enabled automatic discovery of more
efficient neural network architectures, especially for mobile and embedded
vision applications. Although recent research has proposed ways of quickly
estimating latency on unseen hardware devices with just a few samples, little
focus has been given to the challenges of estimating latency on runtimes using
optimized graphs, such as TensorRT and specifically for edge devices. In this
work, we propose MAPLE-Edge, an edge device-oriented extension of MAPLE, the
state-of-the-art latency predictor for general purpose hardware, where we train
a regression network on architecture-latency pairs in conjunction with a
hardware-runtime descriptor to effectively estimate latency on a diverse pool
of edge devices. Compared to MAPLE, MAPLE-Edge can describe the runtime and
target device platform using a much smaller set of CPU performance counters
that are widely available on all Linux kernels, while still achieving up to
+49.6% accuracy gains against previous state-of-the-art baseline methods on
optimized edge device runtimes, using just 10 measurements from an unseen
target device. We also demonstrate that unlike MAPLE which performs best when
trained on a pool of devices sharing a common runtime, MAPLE-Edge can
effectively generalize across runtimes by applying a trick of normalizing
performance counters by the operator latency, in the measured hardware-runtime
descriptor. Lastly, we show that for runtimes exhibiting lower than desired
accuracy, performance can be boosted by collecting additional samples from the
target device, with an extra 90 samples translating to gains of nearly +40%.
Related papers
- AccEPT: An Acceleration Scheme for Speeding Up Edge Pipeline-parallel
Training [22.107070114339038]
We propose AccEPT, an acceleration scheme for accelerating the edge collaborative pipeline-parallel training.
In particular, we propose a light-weight adaptive latency predictor to accurately estimate the latency of each layer at different devices.
Our numerical results demonstrate that our proposed acceleration approach is able to significantly speed up edge pipeline parallel training up to 3 times faster.
arXiv Detail & Related papers (2023-11-10T02:18:33Z) - FLEdge: Benchmarking Federated Machine Learning Applications in Edge Computing Systems [61.335229621081346]
Federated Learning (FL) has become a viable technique for realizing privacy-enhancing distributed deep learning on the network edge.
In this paper, we propose FLEdge, which complements existing FL benchmarks by enabling a systematic evaluation of client capabilities.
arXiv Detail & Related papers (2023-06-08T13:11:20Z) - MAPLE-X: Latency Prediction with Explicit Microprocessor Prior Knowledge [87.41163540910854]
Deep neural network (DNN) latency characterization is a time-consuming process.
We propose MAPLE-X which extends MAPLE by incorporating explicit prior knowledge of hardware devices and DNN architecture latency.
arXiv Detail & Related papers (2022-05-25T11:08:20Z) - Lightweight Jet Reconstruction and Identification as an Object Detection
Task [5.071565475111431]
We apply convolutional techniques to end-to-end jet identification and reconstruction tasks encountered at the CERN Large Hadron Collider.
PFJet-SSD performs simultaneous localization, classification and regression tasks to cluster jets and reconstruct their features.
We show that the ternary network closely matches the performance of its full-precision equivalent and outperforms the state-of-the-art rule-based algorithm.
arXiv Detail & Related papers (2022-02-09T15:01:53Z) - MAPLE: Microprocessor A Priori for Latency Estimation [81.91509153539566]
Modern deep neural networks must demonstrate state-of-the-art accuracy while exhibiting low latency and energy consumption.
Measuring the latency of every evaluated architecture adds a significant amount of time to the NAS process.
We propose Microprocessor A Priori for Estimation Estimation MAPLE that does not rely on transfer learning or domain adaptation.
arXiv Detail & Related papers (2021-11-30T03:52:15Z) - HELP: Hardware-Adaptive Efficient Latency Predictor for NAS via
Meta-Learning [43.751220068642624]
Hardware-adaptive Predictor (HELP) is a device-specific latency estimation problem as a meta-learning problem.
We introduce novel hardware embeddings to embed any devices considering them as black-box functions that output latencies, and meta-learn the hardware-adaptive latency predictor in a device-dependent manner.
We validate the proposed HELP for its latency estimation performance on unseen platforms, on which it achieves high estimation performance with as few as 10 measurement samples, outperforming all relevant baselines.
arXiv Detail & Related papers (2021-06-16T08:36:21Z) - Multi-Exit Semantic Segmentation Networks [78.44441236864057]
We propose a framework for converting state-of-the-art segmentation models to MESS networks.
specially trained CNNs that employ parametrised early exits along their depth to save during inference on easier samples.
We co-optimise the number, placement and architecture of the attached segmentation heads, along with the exit policy, to adapt to the device capabilities and application-specific requirements.
arXiv Detail & Related papers (2021-06-07T11:37:03Z) - PatDNN: Achieving Real-Time DNN Execution on Mobile Devices with
Pattern-based Weight Pruning [57.20262984116752]
We introduce a new dimension, fine-grained pruning patterns inside the coarse-grained structures, revealing a previously unknown point in design space.
With the higher accuracy enabled by fine-grained pruning patterns, the unique insight is to use the compiler to re-gain and guarantee high hardware efficiency.
arXiv Detail & Related papers (2020-01-01T04:52:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.