PaRT: Parallel Learning Towards Robust and Transparent AI
- URL: http://arxiv.org/abs/2201.09534v1
- Date: Mon, 24 Jan 2022 09:03:28 GMT
- Title: PaRT: Parallel Learning Towards Robust and Transparent AI
- Authors: Mahsa Paknezhad, Hamsawardhini Rengarajan, Chenghao Yuan, Sujanya
Suresh, Manas Gupta, Savitha Ramasamy, Lee Hwee Kuan
- Abstract summary: This paper takes a parallel learning approach for robust and transparent AI.
A deep neural network is trained in parallel on multiple tasks, where each task is trained only on a subset of the network resources.
We show that the network does indeed use learned knowledge from some tasks in other tasks, through shared representations.
- Score: 4.160969852186451
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper takes a parallel learning approach for robust and transparent AI.
A deep neural network is trained in parallel on multiple tasks, where each task
is trained only on a subset of the network resources. Each subset consists of
network segments, that can be combined and shared across specific tasks. Tasks
can share resources with other tasks, while having independent task-related
network resources. Therefore, the trained network can share similar
representations across various tasks, while also enabling independent
task-related representations. The above allows for some crucial outcomes. (1)
The parallel nature of our approach negates the issue of catastrophic
forgetting. (2) The sharing of segments uses network resources more
efficiently. (3) We show that the network does indeed use learned knowledge
from some tasks in other tasks, through shared representations. (4) Through
examination of individual task-related and shared representations, the model
offers transparency in the network and in the relationships across tasks in a
multi-task setting. Evaluation of the proposed approach against complex
competing approaches such as Continual Learning, Neural Architecture Search,
and Multi-task learning shows that it is capable of learning robust
representations. This is the first effort to train a DL model on multiple tasks
in parallel. Our code is available at https://github.com/MahsaPaknezhad/PaRT
Related papers
- OmniVec: Learning robust representations with cross modal sharing [28.023214572340336]
We present an approach to learn multiple tasks, in multiple modalities, with a unified architecture.
The proposed network is composed of task specific encoders, a common trunk in the middle, followed by task specific prediction heads.
We train the network on all major modalities, e.g. visual, audio, text and 3D, and report results on $22$ diverse and challenging public benchmarks.
arXiv Detail & Related papers (2023-11-07T14:00:09Z) - Provable Multi-Task Representation Learning by Two-Layer ReLU Neural Networks [69.38572074372392]
We present the first results proving that feature learning occurs during training with a nonlinear model on multiple tasks.
Our key insight is that multi-task pretraining induces a pseudo-contrastive loss that favors representations that align points that typically have the same label across tasks.
arXiv Detail & Related papers (2023-07-13T16:39:08Z) - Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners [67.5865966762559]
We study whether sparsely activated Mixture-of-Experts (MoE) improve multi-task learning.
We devise task-aware gating functions to route examples from different tasks to specialized experts.
This results in a sparsely activated multi-task model with a large number of parameters, but with the same computational cost as that of a dense model.
arXiv Detail & Related papers (2022-04-16T00:56:12Z) - Active Multi-Task Representation Learning [50.13453053304159]
We give the first formal study on resource task sampling by leveraging the techniques from active learning.
We propose an algorithm that iteratively estimates the relevance of each source task to the target task and samples from each source task based on the estimated relevance.
arXiv Detail & Related papers (2022-02-02T08:23:24Z) - MultiTask-CenterNet (MCN): Efficient and Diverse Multitask Learning
using an Anchor Free Approach [0.13764085113103217]
Multitask learning is a common approach in machine learning.
In this paper we augment the CenterNet anchor-free approach for training multiple perception related tasks together.
arXiv Detail & Related papers (2021-08-11T06:57:04Z) - Exploring Relational Context for Multi-Task Dense Prediction [76.86090370115]
We consider a multi-task environment for dense prediction tasks, represented by a common backbone and independent task-specific heads.
We explore various attention-based contexts, such as global and local, in the multi-task setting.
We propose an Adaptive Task-Relational Context module, which samples the pool of all available contexts for each task pair.
arXiv Detail & Related papers (2021-04-28T16:45:56Z) - One Network Fits All? Modular versus Monolithic Task Formulations in
Neural Networks [36.07011014271394]
We show that a single neural network is capable of simultaneously learning multiple tasks from a combined data set.
We study how the complexity of learning such combined tasks grows with the complexity of the task codes.
arXiv Detail & Related papers (2021-03-29T01:16:42Z) - Navigating the Trade-Off between Multi-Task Learning and Learning to
Multitask in Deep Neural Networks [9.278739724750343]
Multi-task learning refers to a paradigm in machine learning in which a network is trained on various related tasks to facilitate the acquisition of tasks.
multitasking is used to indicate, especially in the cognitive science literature, the ability to execute multiple tasks simultaneously.
We show that the same tension arises in deep networks and discuss a meta-learning algorithm for an agent to manage this trade-off in an unfamiliar environment.
arXiv Detail & Related papers (2020-07-20T23:26:16Z) - Auxiliary Learning by Implicit Differentiation [54.92146615836611]
Training neural networks with auxiliary tasks is a common practice for improving the performance on a main task of interest.
Here, we propose a novel framework, AuxiLearn, that targets both challenges based on implicit differentiation.
First, when useful auxiliaries are known, we propose learning a network that combines all losses into a single coherent objective function.
Second, when no useful auxiliary task is known, we describe how to learn a network that generates a meaningful, novel auxiliary task.
arXiv Detail & Related papers (2020-06-22T19:35:07Z) - MTI-Net: Multi-Scale Task Interaction Networks for Multi-Task Learning [82.62433731378455]
We show that tasks with high affinity at a certain scale are not guaranteed to retain this behaviour at other scales.
We propose a novel architecture, namely MTI-Net, that builds upon this finding.
arXiv Detail & Related papers (2020-01-19T21:02:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.