On the Energy and Communication Efficiency Tradeoffs in Federated and
Multi-Task Learning
- URL: http://arxiv.org/abs/2212.01049v1
- Date: Fri, 2 Dec 2022 09:40:17 GMT
- Title: On the Energy and Communication Efficiency Tradeoffs in Federated and
Multi-Task Learning
- Authors: Stefano Savazzi, Vittorio Rampa, Sanaz Kianoush and Mehdi Bennis
- Abstract summary: Multi-Task Learning (MTL) exploits relevant commonalities across tasks to improve efficiency compared with traditional transfer learning approaches.
This article provides a first look into the energy costs of MTL processes driven by the Model-Agnostic Meta-Learning (MAML) paradigm and implemented in distributed wireless networks.
- Score: 42.37180749113699
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recent advances in Federated Learning (FL) have paved the way towards the
design of novel strategies for solving multiple learning tasks simultaneously,
by leveraging cooperation among networked devices. Multi-Task Learning (MTL)
exploits relevant commonalities across tasks to improve efficiency compared
with traditional transfer learning approaches. By learning multiple tasks
jointly, significant reduction in terms of energy footprints can be obtained.
This article provides a first look into the energy costs of MTL processes
driven by the Model-Agnostic Meta-Learning (MAML) paradigm and implemented in
distributed wireless networks. The paper targets a clustered multi-task network
setup where autonomous agents learn different but related tasks. The MTL
process is carried out in two stages: the optimization of a meta-model that can
be quickly adapted to learn new tasks, and a task-specific model adaptation
stage where the learned meta-model is transferred to agents and tailored for a
specific task. This work analyzes the main factors that influence the MTL
energy balance by considering a multi-task Reinforcement Learning (RL) setup in
a robotized environment. Results show that the MAML method can reduce the
energy bill by at least 2 times compared with traditional approaches without
inductive transfer. Moreover, it is shown that the optimal energy balance in
wireless networks depends on uplink/downlink and sidelink communication
efficiencies.
Related papers
- Meta-Sparsity: Learning Optimal Sparse Structures in Multi-task Networks through Meta-learning [4.462334751640166]
meta-sparsity is a framework for learning model sparsity that allows deep neural networks (DNNs) to generate optimal sparse shared structures in multi-task learning setting.
Inspired by Model Agnostic Meta-Learning (MAML), the emphasis is on learning shared and optimally sparse parameters in multi-task scenarios.
The effectiveness of meta-sparsity is rigorously evaluated by extensive experiments on two datasets.
arXiv Detail & Related papers (2025-01-21T13:25:32Z) - Transforming Vision Transformer: Towards Efficient Multi-Task Asynchronous Learning [59.001091197106085]
Multi-Task Learning (MTL) for Vision Transformer aims at enhancing the model capability by tackling multiple tasks simultaneously.
Most recent works have predominantly focused on designing Mixture-of-Experts (MoE) structures and in tegrating Low-Rank Adaptation (LoRA) to efficiently perform multi-task learning.
We propose a novel approach dubbed Efficient Multi-Task Learning (EMTAL) by transforming a pre-trained Vision Transformer into an efficient multi-task learner.
arXiv Detail & Related papers (2025-01-12T17:41:23Z) - R-MTLLMF: Resilient Multi-Task Large Language Model Fusion at the Wireless Edge [78.26352952957909]
Multi-task large language models (MTLLMs) are important for many applications at the wireless edge, where users demand specialized models to handle multiple tasks efficiently.
The concept of model fusion via task vectors has emerged as an efficient approach for combining fine-tuning parameters to produce an MTLLM.
In this paper, the problem of enabling edge users to collaboratively craft such MTLMs via tasks vectors is studied, under the assumption of worst-case adversarial attacks.
arXiv Detail & Related papers (2024-11-27T10:57:06Z) - Dynamic Transformer Architecture for Continual Learning of Multimodal
Tasks [27.59758964060561]
Transformer neural networks are increasingly replacing prior architectures in a wide range of applications in different data modalities.
Continual learning (CL) emerges as a solution by facilitating the transfer of knowledge across tasks that arrive sequentially for an autonomously learning agent.
We propose a transformer-based CL framework focusing on learning tasks that involve both vision and language.
arXiv Detail & Related papers (2024-01-27T03:03:30Z) - M$^3$ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task
Learning with Model-Accelerator Co-design [95.41238363769892]
Multi-task learning (MTL) encapsulates multiple learned tasks in a single model and often lets those tasks learn better jointly.
Current MTL regimes have to activate nearly the entire model even to just execute a single task.
We present a model-accelerator co-design framework to enable efficient on-device MTL.
arXiv Detail & Related papers (2022-10-26T15:40:24Z) - Attentive Task Interaction Network for Multi-Task Learning [4.1372815372396525]
ATI-Net employs knowledge distillation of the latent features for each task, then combines the feature maps to provide improved contextualized information to the decoder.
This novel approach to introducing knowledge distillation into an attention based multitask network outperforms state of the art MTL baselines.
arXiv Detail & Related papers (2022-01-25T22:03:20Z) - Energy-Efficient Multi-Orchestrator Mobile Edge Learning [54.28419430315478]
Mobile Edge Learning (MEL) is a collaborative learning paradigm that features distributed training of Machine Learning (ML) models over edge devices.
In MEL, possible coexistence of multiple learning tasks with different datasets may arise.
We propose lightweight algorithms that can achieve near-optimal performance and facilitate the trade-offs between energy consumption, accuracy, and solution complexity.
arXiv Detail & Related papers (2021-09-02T07:37:10Z) - Reparameterizing Convolutions for Incremental Multi-Task Learning
without Task Interference [75.95287293847697]
Two common challenges in developing multi-task models are often overlooked in literature.
First, enabling the model to be inherently incremental, continuously incorporating information from new tasks without forgetting the previously learned ones (incremental learning)
Second, eliminating adverse interactions amongst tasks, which has been shown to significantly degrade the single-task performance in a multi-task setup (task interference)
arXiv Detail & Related papers (2020-07-24T14:44:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.