MetaBalance: Improving Multi-Task Recommendations via Adapting Gradient
Magnitudes of Auxiliary Tasks
- URL: http://arxiv.org/abs/2203.06801v1
- Date: Mon, 14 Mar 2022 01:08:31 GMT
- Title: MetaBalance: Improving Multi-Task Recommendations via Adapting Gradient
Magnitudes of Auxiliary Tasks
- Authors: Yun He, Xue Feng, Cheng Cheng, Geng Ji, Yunsong Guo, James Caverlee
- Abstract summary: We propose MetaBalance to balance auxiliary losses via manipulating their gradients in the multi-task network.
Our proposed method achieves a significant improvement of 8.34% in terms of NDCG@10 upon the strongest baseline on two real-world datasets.
- Score: 19.606256087873028
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In many personalized recommendation scenarios, the generalization ability of
a target task can be improved via learning with additional auxiliary tasks
alongside this target task on a multi-task network. However, this method often
suffers from a serious optimization imbalance problem. On the one hand, one or
more auxiliary tasks might have a larger influence than the target task and
even dominate the network weights, resulting in worse recommendation accuracy
for the target task. On the other hand, the influence of one or more auxiliary
tasks might be too weak to assist the target task. More challenging is that
this imbalance dynamically changes throughout the training process and varies
across the parts of the same network. We propose a new method: MetaBalance to
balance auxiliary losses via directly manipulating their gradients w.r.t the
shared parameters in the multi-task network. Specifically, in each training
iteration and adaptively for each part of the network, the gradient of an
auxiliary loss is carefully reduced or enlarged to have a closer magnitude to
the gradient of the target loss, preventing auxiliary tasks from being so
strong that dominate the target task or too weak to help the target task.
Moreover, the proximity between the gradient magnitudes can be flexibly
adjusted to adapt MetaBalance to different scenarios. The experiments show that
our proposed method achieves a significant improvement of 8.34% in terms of
NDCG@10 upon the strongest baseline on two real-world datasets. The code of our
approach can be found at here: https://github.com/facebookresearch/MetaBalance
Related papers
- Perturbing the Gradient for Alleviating Meta Overfitting [6.6157730528755065]
This paper proposes a number of solutions to tackle meta-overfitting on few-shot learning settings.
Our proposed approaches demonstrate improved generalization performance compared to state-of-the-art baselines for learning in a non-mutually exclusive task setting.
arXiv Detail & Related papers (2024-05-20T18:04:59Z) - ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning [59.08197876733052]
Auxiliary-Task Learning (ATL) aims to improve the performance of the target task by leveraging the knowledge obtained from related tasks.
Sometimes, learning multiple tasks simultaneously results in lower accuracy than learning only the target task, known as negative transfer.
ForkMerge is a novel approach that periodically forks the model into multiple branches, automatically searches the varying task weights.
arXiv Detail & Related papers (2023-01-30T02:27:02Z) - Learning Action Translator for Meta Reinforcement Learning on
Sparse-Reward Tasks [56.63855534940827]
This work introduces a novel objective function to learn an action translator among training tasks.
We theoretically verify that the value of the transferred policy with the action translator can be close to the value of the source policy.
We propose to combine the action translator with context-based meta-RL algorithms for better data collection and more efficient exploration during meta-training.
arXiv Detail & Related papers (2022-07-19T04:58:06Z) - Meta-learning with an Adaptive Task Scheduler [93.63502984214918]
Existing meta-learning algorithms randomly sample meta-training tasks with a uniform probability.
It is likely that tasks are detrimental with noise or imbalanced given a limited number of meta-training tasks.
We propose an adaptive task scheduler (ATS) for the meta-training process.
arXiv Detail & Related papers (2021-10-26T22:16:35Z) - Conflict-Averse Gradient Descent for Multi-task Learning [56.379937772617]
A major challenge in optimizing a multi-task model is the conflicting gradients.
We introduce Conflict-Averse Gradient descent (CAGrad) which minimizes the average loss function.
CAGrad balances the objectives automatically and still provably converges to a minimum over the average loss.
arXiv Detail & Related papers (2021-10-26T22:03:51Z) - Adaptive Transfer Learning on Graph Neural Networks [4.233435459239147]
Graph neural networks (GNNs) are widely used to learn a powerful representation of graph-structured data.
Recent work demonstrates that transferring knowledge from self-supervised tasks to downstream tasks could further improve graph representation.
We propose a new transfer learning paradigm on GNNs which could effectively leverage self-supervised tasks as auxiliary tasks to help the target task.
arXiv Detail & Related papers (2021-07-19T11:46:28Z) - Automatic Curriculum Learning through Value Disagreement [95.19299356298876]
Continually solving new, unsolved tasks is the key to learning diverse behaviors.
In the multi-task domain, where an agent needs to reach multiple goals, the choice of training goals can largely affect sample efficiency.
We propose setting up an automatic curriculum for goals that the agent needs to solve.
We evaluate our method across 13 multi-goal robotic tasks and 5 navigation tasks, and demonstrate performance gains over current state-of-the-art methods.
arXiv Detail & Related papers (2020-06-17T03:58:25Z) - Gradient Surgery for Multi-Task Learning [119.675492088251]
Multi-task learning has emerged as a promising approach for sharing structure across multiple tasks.
The reasons why multi-task learning is so challenging compared to single-task learning are not fully understood.
We propose a form of gradient surgery that projects a task's gradient onto the normal plane of the gradient of any other task that has a conflicting gradient.
arXiv Detail & Related papers (2020-01-19T06:33:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.