MetaNetwork: A Task-agnostic Network Parameters Generation Framework for
Improving Device Model Generalization
- URL: http://arxiv.org/abs/2209.05227v1
- Date: Mon, 12 Sep 2022 13:26:26 GMT
- Title: MetaNetwork: A Task-agnostic Network Parameters Generation Framework for
Improving Device Model Generalization
- Authors: Zheqi Lv, Feng Wang, Kun Kuang, Yongwei Wang, Zhengyu Chen, Tao Shen,
Hongxia Yang, Fei Wu
- Abstract summary: We propose a novel task-agnostic framework, named MetaNetwork, for generating adaptive device model parameters from cloud without on-device training.
The MetaGenerator is designed to learn a mapping function from samples to model parameters, and it can generate and deliver the adaptive parameters to the device based on samples uploaded from the device to the cloud.
The MetaStabilizer aims to reduce the oscillation of the MetaGenerator, accelerate the convergence and improve the model performance during both training and inference.
- Score: 65.02542875281233
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Deploying machine learning models on mobile devices has gained increasing
attention. To tackle the model generalization problem with the limitations of
hardware resources on the device, the device model needs to be lightweight by
techniques such as model compression from the cloud model. However, the major
obstacle to improve the device model generalization is the distribution shift
between the data of cloud and device models, since the data distribution on
device model often changes over time (e.g., users might have different
preferences in recommendation system). Although real-time fine-tuning and
distillation method take this situation into account, these methods require
on-device training, which are practically infeasible due to the low
computational power and a lack of real-time labeled samples on the device.
In this paper, we propose a novel task-agnostic framework, named MetaNetwork,
for generating adaptive device model parameters from cloud without on-device
training. Specifically, our MetaNetwork is deployed on cloud and consists of
MetaGenerator and MetaStabilizer modules. The MetaGenerator is designed to
learn a mapping function from samples to model parameters, and it can generate
and deliver the adaptive parameters to the device based on samples uploaded
from the device to the cloud. The MetaStabilizer aims to reduce the oscillation
of the MetaGenerator, accelerate the convergence and improve the model
performance during both training and inference. We evaluate our method on two
tasks with three datasets. Extensive experiments show that MetaNetwork can
achieve competitive performances in different modalities.
Related papers
- Task-Oriented Real-time Visual Inference for IoVT Systems: A Co-design Framework of Neural Networks and Edge Deployment [61.20689382879937]
Task-oriented edge computing addresses this by shifting data analysis to the edge.
Existing methods struggle to balance high model performance with low resource consumption.
We propose a novel co-design framework to optimize neural network architecture.
arXiv Detail & Related papers (2024-10-29T19:02:54Z) - Dual-Model Distillation for Efficient Action Classification with Hybrid Edge-Cloud Solution [1.8029479474051309]
We design a hybrid edge-cloud solution that leverages the efficiency of smaller models for local processing while deferring to larger, more accurate cloud-based models when necessary.
Specifically, we propose a novel unsupervised data generation method, Dual-Model Distillation (DMD), to train a lightweight switcher model that can predict when the edge model's output is uncertain.
Experimental results on the action classification task show that our framework not only requires less computational overhead, but also improves accuracy compared to using a large model alone.
arXiv Detail & Related papers (2024-10-16T02:06:27Z) - Towards Robust and Efficient Cloud-Edge Elastic Model Adaptation via Selective Entropy Distillation [56.79064699832383]
We establish a Cloud-Edge Elastic Model Adaptation (CEMA) paradigm in which the edge models only need to perform forward propagation.
In our CEMA, to reduce the communication burden, we devise two criteria to exclude unnecessary samples from uploading to the cloud.
arXiv Detail & Related papers (2024-02-27T08:47:19Z) - ECLM: Efficient Edge-Cloud Collaborative Learning with Continuous
Environment Adaptation [47.35179593006409]
We propose ECLM, an edge-cloud collaborative learning framework for rapid model adaptation for dynamic edge environments.
We show that ECLM significantly improves model performance (e.g., 18.89% accuracy increase) and resource efficiency (e.g. 7.12x communication cost reduction) in adapting models to dynamic edge environments.
arXiv Detail & Related papers (2023-11-18T14:10:09Z) - FLuID: Mitigating Stragglers in Federated Learning using Invariant
Dropout [1.8262547855491458]
Federated Learning allows machine learning models to train locally on individual mobile devices, synchronizing model updates via a shared server.
As a result, straggler devices with lower performance often dictate the overall training time in FL.
We introduce Invariant Dropout, a method that extracts a sub-model based on the weight update threshold.
We develop an adaptive training framework, Federated Learning using Invariant Dropout.
arXiv Detail & Related papers (2023-07-05T19:53:38Z) - Cloud-Device Collaborative Adaptation to Continual Changing Environments
in the Real-world [20.547119604004774]
We propose a new learning paradigm of Cloud-Device Collaborative Continual Adaptation, which encourages collaboration between cloud and device.
We also propose an Uncertainty-based Visual Prompt Adapted (U-VPA) teacher-student model to transfer the generalization capability of the large model on the cloud to the device model.
Our proposed U-VPA teacher-student framework outperforms previous state-of-the-art test time adaptation and device-cloud collaboration methods.
arXiv Detail & Related papers (2022-12-02T05:02:36Z) - p-Meta: Towards On-device Deep Model Adaptation [30.27192953408665]
p-Meta is a new meta learning method that enforces structure-wise partial parameter updates while ensuring fast generalization to unseen tasks.
We show that p-Meta substantially reduces the peak dynamic memory by a factor of 2.5 on average compared to state-of-the-art few-shot adaptation methods.
arXiv Detail & Related papers (2022-06-25T18:36:59Z) - SensiX++: Bringing MLOPs and Multi-tenant Model Serving to Sensory Edge
Devices [69.1412199244903]
We present a multi-tenant runtime for adaptive model execution with integrated MLOps on edge devices, e.g., a camera, a microphone, or IoT sensors.
S SensiX++ operates on two fundamental principles - highly modular componentisation to externalise data operations with clear abstractions and document-centric manifestation for system-wide orchestration.
We report on the overall throughput and quantified benefits of various automation components of SensiX++ and demonstrate its efficacy to significantly reduce operational complexity and lower the effort to deploy, upgrade, reconfigure and serve embedded models on edge devices.
arXiv Detail & Related papers (2021-09-08T22:06:16Z) - Device-Cloud Collaborative Learning for Recommendation [50.01289274123047]
We propose a novel MetaPatch learning approach on the device side to efficiently achieve "thousands of people with thousands of models" given a centralized cloud model.
With billions of updated personalized device models, we propose a "model-over-models" distillation algorithm, namely MoMoDistill, to update the centralized cloud model.
arXiv Detail & Related papers (2021-04-14T05:06:59Z) - Fast-Convergent Federated Learning [82.32029953209542]
Federated learning is a promising solution for distributing machine learning tasks through modern networks of mobile devices.
We propose a fast-convergent federated learning algorithm, called FOLB, which performs intelligent sampling of devices in each round of model training.
arXiv Detail & Related papers (2020-07-26T14:37:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.