Provable Pathways: Learning Multiple Tasks over Multiple Paths
- URL: http://arxiv.org/abs/2303.04338v1
- Date: Wed, 8 Mar 2023 02:25:28 GMT
- Title: Provable Pathways: Learning Multiple Tasks over Multiple Paths
- Authors: Yingcong Li, Samet Oymak
- Abstract summary: We develop novel generalization bounds for empirical risk minimization problems learning multiple tasks over multiple paths.
In conjunction, we formalize the benefits of resulting multipath representation when adapting to new downstream tasks.
- Score: 31.43753806123382
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Constructing useful representations across a large number of tasks is a key
requirement for sample-efficient intelligent systems. A traditional idea in
multitask learning (MTL) is building a shared representation across tasks which
can then be adapted to new tasks by tuning last layers. A desirable refinement
of using a shared one-fits-all representation is to construct task-specific
representations. To this end, recent PathNet/muNet architectures represent
individual tasks as pathways within a larger supernet. The subnetworks induced
by pathways can be viewed as task-specific representations that are composition
of modules within supernet's computation graph. This work explores the pathways
proposal from the lens of statistical learning: We first develop novel
generalization bounds for empirical risk minimization problems learning
multiple tasks over multiple paths (Multipath MTL). In conjunction, we
formalize the benefits of resulting multipath representation when adapting to
new downstream tasks. Our bounds are expressed in terms of Gaussian complexity,
lead to tangible guarantees for the class of linear representations, and
provide novel insights into the quality and benefits of a multipath
representation. When computation graph is a tree, Multipath MTL hierarchically
clusters the tasks and builds cluster-specific representations. We provide
further discussion and experiments for hierarchical MTL and rigorously identify
the conditions under which Multipath MTL is provably superior to traditional
MTL approaches with shallow supernets.
Related papers
- Distribution Matching for Multi-Task Learning of Classification Tasks: a
Large-Scale Study on Faces & Beyond [62.406687088097605]
Multi-Task Learning (MTL) is a framework, where multiple related tasks are learned jointly and benefit from a shared representation space.
We show that MTL can be successful with classification tasks with little, or non-overlapping annotations.
We propose a novel approach, where knowledge exchange is enabled between the tasks via distribution matching.
arXiv Detail & Related papers (2024-01-02T14:18:11Z) - Provable Benefits of Multi-task RL under Non-Markovian Decision Making
Processes [56.714690083118406]
In multi-task reinforcement learning (RL) under Markov decision processes (MDPs), the presence of shared latent structures has been shown to yield significant benefits to the sample efficiency compared to single-task RL.
We investigate whether such a benefit can extend to more general sequential decision making problems, such as partially observable MDPs (POMDPs) and more general predictive state representations (PSRs)
We propose a provably efficient algorithm UMT-PSR for finding near-optimal policies for all PSRs, and demonstrate that the advantage of multi-task learning manifests if the joint model class of PSR
arXiv Detail & Related papers (2023-10-20T14:50:28Z) - Low-Rank Multitask Learning based on Tensorized SVMs and LSSVMs [65.42104819071444]
Multitask learning (MTL) leverages task-relatedness to enhance performance.
We employ high-order tensors, with each mode corresponding to a task index, to naturally represent tasks referenced by multiple indices.
We propose a general framework of low-rank MTL methods with tensorized support vector machines (SVMs) and least square support vector machines (LSSVMs)
arXiv Detail & Related papers (2023-08-30T14:28:26Z) - Deep Task-specific Bottom Representation Network for Multi-Task
Recommendation [36.128708266100645]
We propose the Deep Task-specific Bottom Representation Network (DTRN) to alleviate the negative transfer problem.
The two proposed modules can achieve the purpose of getting task-specific bottom representation to relieve tasks' mutual interference.
arXiv Detail & Related papers (2023-08-11T08:04:43Z) - Task Aware Feature Extraction Framework for Sequential Dependence
Multi-Task Learning [1.0765359420035392]
We analyze sequential dependence MTL from rigorous mathematical perspective.
We propose a Task Aware Feature Extraction (TAFE) framework for sequential dependence MTL.
arXiv Detail & Related papers (2023-01-06T13:12:59Z) - Task Adaptive Parameter Sharing for Multi-Task Learning [114.80350786535952]
Adaptive Task Adapting Sharing (TAPS) is a method for tuning a base model to a new task by adaptively modifying a small, task-specific subset of layers.
Compared to other methods, TAPS retains high accuracy on downstream tasks while introducing few task-specific parameters.
We evaluate our method on a suite of fine-tuning tasks and architectures (ResNet, DenseNet, ViT) and show that it achieves state-of-the-art performance while being simple to implement.
arXiv Detail & Related papers (2022-03-30T23:16:07Z) - A Tree-Structured Multi-Task Model Recommender [25.445073413243925]
Tree-structured multi-task architectures have been employed to tackle multiple vision tasks in the context of multi-task learning (MTL)
This paper proposes a recommender that automatically suggests tree-structured multi-task architectures that could achieve a high task performance while meeting a user-specified computation budget without performing model training.
Extensive evaluations on popular MTL benchmarks show that the recommended architectures could achieve competitive task accuracy and computation efficiency compared with state-of-the-art MTL methods.
arXiv Detail & Related papers (2022-03-10T00:09:43Z) - Multi-Task Learning with Deep Neural Networks: A Survey [0.0]
Multi-task learning (MTL) is a subfield of machine learning in which multiple tasks are simultaneously learned by a shared model.
We give an overview of multi-task learning methods for deep neural networks, with the aim of summarizing both the well-established and most recent directions within the field.
arXiv Detail & Related papers (2020-09-10T19:31:04Z) - Boosting Share Routing for Multi-task Learning [0.12891210250935145]
Multi-task learning (MTL) aims to make full use of the knowledge contained in multi-task supervision signals to improve the overall performance.
How to make the knowledge of multiple tasks shared appropriately is an open problem for MTL.
We propose a general framework called Multi-Task Neural Architecture Search (MTNAS) to efficiently find a suitable sharing route for a given MTL problem.
arXiv Detail & Related papers (2020-09-01T12:37:19Z) - Multi-Task Learning for Dense Prediction Tasks: A Survey [87.66280582034838]
Multi-task learning (MTL) techniques have shown promising results w.r.t. performance, computations and/or memory footprint.
We provide a well-rounded view on state-of-the-art deep learning approaches for MTL in computer vision.
arXiv Detail & Related papers (2020-04-28T09:15:50Z) - MTL-NAS: Task-Agnostic Neural Architecture Search towards
General-Purpose Multi-Task Learning [71.90902837008278]
We propose to incorporate neural architecture search (NAS) into general-purpose multi-task learning (GP-MTL)
In order to adapt to different task combinations, we disentangle the GP-MTL networks into single-task backbones.
We also propose a novel single-shot gradient-based search algorithm that closes the performance gap between the searched architectures.
arXiv Detail & Related papers (2020-03-31T09:49:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.