MDLdroid: a ChainSGD-reduce Approach to Mobile Deep Learning for
Personal Mobile Sensing
- URL: http://arxiv.org/abs/2002.02897v2
- Date: Sat, 15 Feb 2020 14:34:48 GMT
- Title: MDLdroid: a ChainSGD-reduce Approach to Mobile Deep Learning for
Personal Mobile Sensing
- Authors: Yu Zhang, Tao Gu, Xi Zhang
- Abstract summary: Running deep learning on devices offers several advantages including data privacy preservation and low-latency response for both model robustness and update.
Personal mobile sensing applications are mostly user-specific and highly affected by environment.
We present MDLdroid, a novel decentralized mobile deep learning framework to enable resource-aware on-device collaborative learning.
- Score: 14.574274428615666
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Personal mobile sensing is fast permeating our daily lives to enable activity
monitoring, healthcare and rehabilitation. Combined with deep learning, these
applications have achieved significant success in recent years. Different from
conventional cloud-based paradigms, running deep learning on devices offers
several advantages including data privacy preservation and low-latency response
for both model inference and update. Since data collection is costly in
reality, Google's Federated Learning offers not only complete data privacy but
also better model robustness based on multiple user data. However, personal
mobile sensing applications are mostly user-specific and highly affected by
environment. As a result, continuous local changes may seriously affect the
performance of a global model generated by Federated Learning. In addition,
deploying Federated Learning on a local server, e.g., edge server, may quickly
reach the bottleneck due to resource constraint and serious failure by attacks.
Towards pushing deep learning on devices, we present MDLdroid, a novel
decentralized mobile deep learning framework to enable resource-aware on-device
collaborative learning for personal mobile sensing applications. To address
resource limitation, we propose a ChainSGD-reduce approach which includes a
novel chain-directed Synchronous Stochastic Gradient Descent algorithm to
effectively reduce overhead among multiple devices. We also design an
agent-based multi-goal reinforcement learning mechanism to balance resources in
a fair and efficient manner. Our evaluations show that our model training on
off-the-shelf mobile devices achieves 2x to 3.5x faster than single-device
training, and 1.5x faster than the master-slave approach.
Related papers
- Edge Unlearning is Not "on Edge"! An Adaptive Exact Unlearning System on Resource-Constrained Devices [26.939025828011196]
The right to be forgotten mandates that machine learning models enable the erasure of a data owner's data and information from a trained model.
We propose a Constraint-aware Adaptive Exact Unlearning System at the network Edge (CAUSE) to enable exact unlearning on resource-constrained devices.
arXiv Detail & Related papers (2024-10-14T03:28:09Z) - CRSFL: Cluster-based Resource-aware Split Federated Learning for Continuous Authentication [5.636155173401658]
Split Learning (SL) and Federated Learning (FL) have emerged as promising technologies for training a decentralized Machine Learning (ML) model.
We propose combining these technologies to address the continuous authentication challenge while protecting user privacy.
arXiv Detail & Related papers (2024-05-12T06:08:21Z) - Robot Fine-Tuning Made Easy: Pre-Training Rewards and Policies for
Autonomous Real-World Reinforcement Learning [58.3994826169858]
We introduce RoboFuME, a reset-free fine-tuning system for robotic reinforcement learning.
Our insights are to utilize offline reinforcement learning techniques to ensure efficient online fine-tuning of a pre-trained policy.
Our method can incorporate data from an existing robot dataset and improve on a target task within as little as 3 hours of autonomous real-world experience.
arXiv Detail & Related papers (2023-10-23T17:50:08Z) - Federated Learning and Meta Learning: Approaches, Applications, and
Directions [94.68423258028285]
In this tutorial, we present a comprehensive review of FL, meta learning, and federated meta learning (FedMeta)
Unlike other tutorial papers, our objective is to explore how FL, meta learning, and FedMeta methodologies can be designed, optimized, and evolved, and their applications over wireless networks.
arXiv Detail & Related papers (2022-10-24T10:59:29Z) - Online Data Selection for Federated Learning with Limited Storage [53.46789303416799]
Federated Learning (FL) has been proposed to achieve distributed machine learning among networked devices.
The impact of on-device storage on the performance of FL is still not explored.
In this work, we take the first step to consider the online data selection for FL with limited on-device storage.
arXiv Detail & Related papers (2022-09-01T03:27:33Z) - Reducing Impacts of System Heterogeneity in Federated Learning using
Weight Update Magnitudes [0.0]
Federated learning enables machine learning models to train locally on each handheld device while only synchronizing their neuron updates with a server.
This results in the training time of federated learning tasks being dictated by a few low-performance straggler devices.
In this work, we aim to mitigate the performance bottleneck of federated learning by dynamically forming sub-models for stragglers based on their performance and accuracy feedback.
arXiv Detail & Related papers (2022-08-30T00:39:06Z) - Federated Split GANs [12.007429155505767]
We propose an alternative approach to train ML models in user's devices themselves.
We focus on GANs (generative adversarial networks) and leverage their inherent privacy-preserving attribute.
Our system preserves data privacy, keeps a short training time, and yields same accuracy of model training in unconstrained devices.
arXiv Detail & Related papers (2022-07-04T23:53:47Z) - Applied Federated Learning: Architectural Design for Robust and
Efficient Learning in Privacy Aware Settings [0.8454446648908585]
The classical machine learning paradigm requires the aggregation of user data in a central location.
Centralization of data poses risks, including a heightened risk of internal and external security incidents.
Federated learning with differential privacy is designed to avoid the server-side centralization pitfall.
arXiv Detail & Related papers (2022-06-02T00:30:04Z) - Dynamic Network-Assisted D2D-Aided Coded Distributed Learning [59.29409589861241]
We propose a novel device-to-device (D2D)-aided coded federated learning method (D2D-CFL) for load balancing across devices.
We derive an optimal compression rate for achieving minimum processing time and establish its connection with the convergence time.
Our proposed method is beneficial for real-time collaborative applications, where the users continuously generate training data.
arXiv Detail & Related papers (2021-11-26T18:44:59Z) - Federated Learning-based Active Authentication on Mobile Devices [98.23904302910022]
User active authentication on mobile devices aims to learn a model that can correctly recognize the enrolled user based on device sensor information.
We propose a novel user active authentication training, termed as Federated Active Authentication (FAA)
We show that existing FL/SL methods are suboptimal for FAA as they rely on the data to be distributed homogeneously.
arXiv Detail & Related papers (2021-04-14T22:59:08Z) - Toward Multiple Federated Learning Services Resource Sharing in Mobile
Edge Networks [88.15736037284408]
We study a new model of multiple federated learning services at the multi-access edge computing server.
We propose a joint resource optimization and hyper-learning rate control problem, namely MS-FEDL.
Our simulation results demonstrate the convergence performance of our proposed algorithms.
arXiv Detail & Related papers (2020-11-25T01:29:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.