Selective Task offloading for Maximum Inference Accuracy and Energy
efficient Real-Time IoT Sensing Systems
- URL: http://arxiv.org/abs/2402.16904v1
- Date: Sat, 24 Feb 2024 18:46:06 GMT
- Title: Selective Task offloading for Maximum Inference Accuracy and Energy
efficient Real-Time IoT Sensing Systems
- Authors: Abdelkarim Ben Sada, Amar Khelloufi, Abdenacer Naouri, Huansheng Ning
and Sahraoui Dhelim
- Abstract summary: We propose a lightweight hybrid genetic algorithm (LGSTO) to solve the multidimensional knapsack problem.
Experiment results show that LGSTO performed 3 times faster than the fastest comparable schemes.
- Score: 3.0748861313823
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recent advancements in small-size inference models facilitated AI
deployment on the edge. However, the limited resource nature of edge devices
poses new challenges especially for real-time applications. Deploying multiple
inference models (or a single tunable model) varying in size and therefore
accuracy and power consumption, in addition to an edge server inference model,
can offer a dynamic system in which the allocation of inference models to
inference jobs is performed according to the current resource conditions.
Therefore, in this work, we tackle the problem of selectively allocating
inference models to jobs or offloading them to the edge server to maximize
inference accuracy under time and energy constraints. This problem is shown to
be an instance of the unbounded multidimensional knapsack problem which is
considered a strongly NP-hard problem. We propose a lightweight hybrid genetic
algorithm (LGSTO) to solve this problem. We introduce a termination condition
and neighborhood exploration techniques for faster evolution of populations. We
compare LGSTO with the Naive and Dynamic programming solutions. In addition to
classic genetic algorithms using different reproduction methods including
NSGA-II, and finally we compare to other evolutionary methods such as Particle
swarm optimization (PSO) and Ant colony optimization (ACO). Experiment results
show that LGSTO performed 3 times faster than the fastest comparable schemes
while producing schedules with higher average accuracy.
Related papers
- MARS: Unleashing the Power of Variance Reduction for Training Large Models [56.47014540413659]
Large gradient algorithms like Adam, Adam, and their variants have been central to the development of this type of training.
We propose a framework that reconciles preconditioned gradient optimization methods with variance reduction via a scaled momentum technique.
arXiv Detail & Related papers (2024-11-15T18:57:39Z) - Computation-Aware Gaussian Processes: Model Selection And Linear-Time Inference [55.150117654242706]
We show that model selection for computation-aware GPs trained on 1.8 million data points can be done within a few hours on a single GPU.
As a result of this work, Gaussian processes can be trained on large-scale datasets without significantly compromising their ability to quantify uncertainty.
arXiv Detail & Related papers (2024-11-01T21:11:48Z) - Switchable Decision: Dynamic Neural Generation Networks [98.61113699324429]
We propose a switchable decision to accelerate inference by dynamically assigning resources for each data instance.
Our method benefits from less cost during inference while keeping the same accuracy.
arXiv Detail & Related papers (2024-05-07T17:44:54Z) - Multi-Objective Optimization for Sparse Deep Multi-Task Learning [0.0]
We present a Multi-Objective Optimization algorithm using a modified Weighted Chebyshev scalarization for training Deep Neural Networks (DNNs)
Our work aims to address the (economical and also ecological) sustainability issue of DNN models, with particular focus on Deep Multi-Task models.
arXiv Detail & Related papers (2023-08-23T16:42:27Z) - Scaling Structured Inference with Randomization [64.18063627155128]
We propose a family of dynamic programming (RDP) randomized for scaling structured models to tens of thousands of latent states.
Our method is widely applicable to classical DP-based inference.
It is also compatible with automatic differentiation so can be integrated with neural networks seamlessly.
arXiv Detail & Related papers (2021-12-07T11:26:41Z) - Learning to Fit Morphable Models [12.469605679847085]
We build upon recent advances in learned optimization and propose an update rule inspired by the classic Levenberg-Marquardt algorithm.
We show the effectiveness of the proposed neural on the problems of 3D body surface estimation from a head-mounted device and face fitting from 2D landmarks.
arXiv Detail & Related papers (2021-11-29T18:59:53Z) - Surrogate-Assisted Genetic Algorithm for Wrapper Feature Selection [4.89253144446913]
We propose a novel multi-stage feature selection framework utilizing multiple levels of approximations, or surrogates.
Our experiments show that SAGA can arrive at near-optimal solutions three times faster than a wrapper GA, on average.
arXiv Detail & Related papers (2021-11-17T12:33:18Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z) - Fast and stable MAP-Elites in noisy domains using deep grids [1.827510863075184]
Deep-Grid MAP-Elites is a variant of the MAP-Elites algorithm that uses an archive of similar previously encountered solutions to approximate the performance of a solution.
We show that this simple approach is significantly more resilient to noise on the behavioural descriptors, while achieving competitive performances in terms of fitness optimisation.
arXiv Detail & Related papers (2020-06-25T08:47:23Z) - Communication-Efficient Distributed Stochastic AUC Maximization with
Deep Neural Networks [50.42141893913188]
We study a distributed variable for large-scale AUC for a neural network as with a deep neural network.
Our model requires a much less number of communication rounds and still a number of communication rounds in theory.
Our experiments on several datasets show the effectiveness of our theory and also confirm our theory.
arXiv Detail & Related papers (2020-05-05T18:08:23Z) - GeneCAI: Genetic Evolution for Acquiring Compact AI [36.04715576228068]
Deep Neural Networks (DNNs) are evolving towards more complex architectures to achieve higher inference accuracy.
Model compression techniques can be leveraged to efficiently deploy such compute-intensive architectures on resource-limited mobile devices.
This paper introduces GeneCAI, a novel optimization method that automatically learns how to tune per-layer compression hyper- parameters.
arXiv Detail & Related papers (2020-04-08T20:56:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.