Equitable Multi-task Learning
- URL: http://arxiv.org/abs/2306.09373v2
- Date: Mon, 19 Jun 2023 01:49:46 GMT
- Title: Equitable Multi-task Learning
- Authors: Jun Yuan and Rui Zhang
- Abstract summary: Multi-task learning (MTL) has achieved great success in various research domains, such as CV, NLP and IR.
We propose a novel multi-task optimization method, named EMTL, to achieve equitable MTL.
Our method stably outperforms state-of-the-art methods on the public benchmark datasets of two different research domains.
- Score: 18.65048321820911
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Multi-task learning (MTL) has achieved great success in various research
domains, such as CV, NLP and IR etc. Due to the complex and competing task
correlation, naive training all tasks may lead to inequitable learning, i.e.
some tasks are learned well while others are overlooked. Multi-task
optimization (MTO) aims to improve all tasks at same time, but conventional
methods often perform poor when tasks with large loss scale or gradient norm
magnitude difference. To solve the issue, we in-depth investigate the equity
problem for MTL and find that regularizing relative contribution of different
tasks (i.e. value of task-specific loss divides its raw gradient norm) in
updating shared parameter can improve generalization performance of MTL. Based
on our theoretical analysis, we propose a novel multi-task optimization method,
named EMTL, to achieve equitable MTL. Specifically, we efficiently add variance
regularization to make different tasks' relative contribution closer. Extensive
experiments have been conduct to evaluate EMTL, our method stably outperforms
state-of-the-art methods on the public benchmark datasets of two different
research domains. Furthermore, offline and online A/B test on multi-task
recommendation are conducted too. EMTL improves multi-task recommendation
significantly, demonstrating the superiority and practicability of our method
in industrial landscape.
Related papers
- Can Optimization Trajectories Explain Multi-Task Transfer? [19.797036312370185]
We study how multi-task learning affects generalization in deep learning.
We find that MTL results in a generalization gap-a gap in generalization at comparable training loss-between single-task and multi-task trajectories.
Our work sheds light on the underlying causes for failures in MTL and, importantly, raises questions about the role of general purpose multi-task optimization algorithms.
arXiv Detail & Related papers (2024-08-26T22:57:01Z) - Sample Efficient Myopic Exploration Through Multitask Reinforcement
Learning with Diverse Tasks [53.44714413181162]
This paper shows that when an agent is trained on a sufficiently diverse set of tasks, a generic policy-sharing algorithm with myopic exploration design can be sample-efficient.
To the best of our knowledge, this is the first theoretical demonstration of the "exploration benefits" of MTRL.
arXiv Detail & Related papers (2024-03-03T22:57:44Z) - Fair Resource Allocation in Multi-Task Learning [12.776767874217663]
Multi-task learning (MTL) can leverage the shared knowledge across tasks, resulting in improved data efficiency and generalization performance.
A major challenge in MTL lies in the presence of conflicting gradients, which can hinder the fair optimization of some tasks.
Inspired by fair resource allocation in communication networks, we propose FairGrad, a novel MTL optimization method.
arXiv Detail & Related papers (2024-02-23T22:46:14Z) - Multi-Task Cooperative Learning via Searching for Flat Minima [8.835287696319641]
We propose to formulate MTL as a multi/bi-level optimization problem, and therefore force features to learn from each task in a cooperative approach.
Specifically, we update the sub-model for each task alternatively taking advantage of the learned sub-models of the other tasks.
To alleviate the negative transfer problem during the optimization, we search for flat minima for the current objective function.
arXiv Detail & Related papers (2023-09-21T14:00:11Z) - When Multi-Task Learning Meets Partial Supervision: A Computer Vision Review [7.776434991976473]
Multi-Task Learning (MTL) aims to learn multiple tasks simultaneously while exploiting their mutual relationships.
This review focuses on how MTL could be utilised under different partial supervision settings to address these challenges.
arXiv Detail & Related papers (2023-07-25T20:08:41Z) - Multi-Task Learning as a Bargaining Game [63.49888996291245]
In Multi-task learning (MTL), a joint model is trained to simultaneously make predictions for several tasks.
Since the gradients of these different tasks may conflict, training a joint model for MTL often yields lower performance than its corresponding single-task counterparts.
We propose viewing the gradients combination step as a bargaining game, where tasks negotiate to reach an agreement on a joint direction of parameter update.
arXiv Detail & Related papers (2022-02-02T13:21:53Z) - In Defense of the Unitary Scalarization for Deep Multi-Task Learning [121.76421174107463]
We present a theoretical analysis suggesting that many specialized multi-tasks can be interpreted as forms of regularization.
We show that, when coupled with standard regularization and stabilization techniques, unitary scalarization matches or improves upon the performance of complex multitasks.
arXiv Detail & Related papers (2022-01-11T18:44:17Z) - Variational Multi-Task Learning with Gumbel-Softmax Priors [105.22406384964144]
Multi-task learning aims to explore task relatedness to improve individual tasks.
We propose variational multi-task learning (VMTL), a general probabilistic inference framework for learning multiple related tasks.
arXiv Detail & Related papers (2021-11-09T18:49:45Z) - Semi-supervised Multi-task Learning for Semantics and Depth [88.77716991603252]
Multi-Task Learning (MTL) aims to enhance the model generalization by sharing representations between related tasks for better performance.
We propose the Semi-supervised Multi-Task Learning (MTL) method to leverage the available supervisory signals from different datasets.
We present a domain-aware discriminator structure with various alignment formulations to mitigate the domain discrepancy issue among datasets.
arXiv Detail & Related papers (2021-10-14T07:43:39Z) - Small Towers Make Big Differences [59.243296878666285]
Multi-task learning aims at solving multiple machine learning tasks at the same time.
A good solution to a multi-task learning problem should be generalizable in addition to being Pareto optimal.
We propose a method of under- parameterized self-auxiliaries for multi-task models to achieve the best of both worlds.
arXiv Detail & Related papers (2020-08-13T10:45:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.