Enabling Resource-efficient AIoT System with Cross-level Optimization: A
survey
- URL: http://arxiv.org/abs/2309.15467v1
- Date: Wed, 27 Sep 2023 08:04:24 GMT
- Title: Enabling Resource-efficient AIoT System with Cross-level Optimization: A
survey
- Authors: Sicong Liu, Bin Guo, Cheng Fang, Ziqi Wang, Shiyan Luo, Zimu Zhou,
Zhiwen Yu
- Abstract summary: This survey aims to provide a broader optimization space for more free resource-performance tradeoffs.
By consolidating problems and techniques scattered over diverse levels, we aim to help readers understand their connections and stimulate further discussions.
- Score: 20.360136850102833
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The emerging field of artificial intelligence of things (AIoT, AI+IoT) is
driven by the widespread use of intelligent infrastructures and the impressive
success of deep learning (DL). With the deployment of DL on various intelligent
infrastructures featuring rich sensors and weak DL computing capabilities, a
diverse range of AIoT applications has become possible. However, DL models are
notoriously resource-intensive. Existing research strives to realize
near-/realtime inference of AIoT live data and low-cost training using AIoT
datasets on resource-scare infrastructures. Accordingly, the accuracy and
responsiveness of DL models are bounded by resource availability. To this end,
the algorithm-system co-design that jointly optimizes the resource-friendly DL
models and model-adaptive system scheduling improves the runtime resource
availability and thus pushes the performance boundary set by the standalone
level. Unlike previous surveys on resource-friendly DL models or hand-crafted
DL compilers/frameworks with partially fine-tuned components, this survey aims
to provide a broader optimization space for more free resource-performance
tradeoffs. The cross-level optimization landscape involves various granularity,
including the DL model, computation graph, operator, memory schedule, and
hardware instructor in both on-device and distributed paradigms. Furthermore,
due to the dynamic nature of AIoT context, which includes heterogeneous
hardware, agnostic sensing data, varying user-specified performance demands,
and resource constraints, this survey explores the context-aware
inter-/intra-device controllers for automatic cross-level adaptation.
Additionally, we identify some potential directions for resource-efficient AIoT
systems. By consolidating problems and techniques scattered over diverse
levels, we aim to help readers understand their connections and stimulate
further discussions.
Related papers
- DLBacktrace: A Model Agnostic Explainability for any Deep Learning Models [1.747623282473278]
Deep learning models operate as opaque 'black boxes' with limited transparency in their decision-making processes.
This study addresses the pressing need for interpretability in AI systems, emphasizing its role in fostering trust, ensuring accountability, and promoting responsible deployment in mission-critical fields.
We introduce DLBacktrace, an innovative technique developed by the AryaXAI team to illuminate model decisions across a wide array of domains.
arXiv Detail & Related papers (2024-11-19T16:54:30Z) - Automatic Speech Recognition using Advanced Deep Learning Approaches: A survey [2.716339075963185]
Recent advancements in deep learning (DL) have posed a significant challenge for automatic speech recognition (ASR)
ASR relies on extensive training datasets, including confidential ones, and demands substantial computational and storage resources.
Advanced DL techniques like deep transfer learning (DTL), federated learning (FL), and reinforcement learning (RL) address these issues.
arXiv Detail & Related papers (2024-03-02T16:25:42Z) - GISTEmbed: Guided In-sample Selection of Training Negatives for Text
Embedding Fine-tuning [0.0]
GISTEmbed is a novel strategy that enhances in-batch negative selection during contrastive training through a guide model.
Benchmarked against the Massive Text Embedding Benchmark (MTEB), GISTEmbed showcases consistent performance improvements across various model sizes.
arXiv Detail & Related papers (2024-02-26T18:55:15Z) - When Parameter-efficient Tuning Meets General-purpose Vision-language
Models [65.19127815275307]
PETAL revolutionizes the training process by requiring only 0.5% of the total parameters, achieved through a unique mode approximation technique.
Our experiments reveal that PETAL not only outperforms current state-of-the-art methods in most scenarios but also surpasses full fine-tuning models in effectiveness.
arXiv Detail & Related papers (2023-12-16T17:13:08Z) - Filling the Missing: Exploring Generative AI for Enhanced Federated
Learning over Heterogeneous Mobile Edge Devices [72.61177465035031]
We propose a generative AI-empowered federated learning to address these challenges by leveraging the idea of FIlling the MIssing (FIMI) portion of local data.
Experiment results demonstrate that FIMI can save up to 50% of the device-side energy to achieve the target global test accuracy.
arXiv Detail & Related papers (2023-10-21T12:07:04Z) - Serving Deep Learning Model in Relational Databases [70.53282490832189]
Serving deep learning (DL) models on relational data has become a critical requirement across diverse commercial and scientific domains.
We highlight three pivotal paradigms: The state-of-the-art DL-centric architecture offloads DL computations to dedicated DL frameworks.
The potential UDF-centric architecture encapsulates one or more tensor computations into User Defined Functions (UDFs) within the relational database management system (RDBMS)
arXiv Detail & Related papers (2023-10-07T06:01:35Z) - Deep Transfer Learning for Automatic Speech Recognition: Towards Better
Generalization [3.6393183544320236]
Speech recognition has become an important challenge when using deep learning (DL)
It requires large-scale training datasets and high computational and storage resources.
Deep transfer learning (DTL) has been introduced to overcome these issues.
arXiv Detail & Related papers (2023-04-27T21:08:05Z) - DISTREAL: Distributed Resource-Aware Learning in Heterogeneous Systems [2.1506382989223782]
We study the problem of distributed training of neural networks (NNs) on devices with heterogeneous, limited, and time-varying availability of computational resources.
We present an adaptive, resource-aware, on-device learning mechanism, DISTREAL, which is able to fully and efficiently utilize the available resources.
arXiv Detail & Related papers (2021-12-16T10:15:31Z) - Dynamic Network-Assisted D2D-Aided Coded Distributed Learning [59.29409589861241]
We propose a novel device-to-device (D2D)-aided coded federated learning method (D2D-CFL) for load balancing across devices.
We derive an optimal compression rate for achieving minimum processing time and establish its connection with the convergence time.
Our proposed method is beneficial for real-time collaborative applications, where the users continuously generate training data.
arXiv Detail & Related papers (2021-11-26T18:44:59Z) - Reconfigurable Intelligent Surface Assisted Mobile Edge Computing with
Heterogeneous Learning Tasks [53.1636151439562]
Mobile edge computing (MEC) provides a natural platform for AI applications.
We present an infrastructure to perform machine learning tasks at an MEC with the assistance of a reconfigurable intelligent surface (RIS)
Specifically, we minimize the learning error of all participating users by jointly optimizing transmit power of mobile users, beamforming vectors of the base station, and the phase-shift matrix of the RIS.
arXiv Detail & Related papers (2020-12-25T07:08:50Z) - Resource Allocation via Model-Free Deep Learning in Free Space Optical
Communications [119.81868223344173]
The paper investigates the general problem of resource allocation for mitigating channel fading effects in Free Space Optical (FSO) communications.
Under this framework, we propose two algorithms that solve FSO resource allocation problems.
arXiv Detail & Related papers (2020-07-27T17:38:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.