Unprejudiced Training Auxiliary Tasks Makes Primary Better: A Multi-Task Learning Perspective
- URL: http://arxiv.org/abs/2412.19547v1
- Date: Fri, 27 Dec 2024 09:27:18 GMT
- Title: Unprejudiced Training Auxiliary Tasks Makes Primary Better: A Multi-Task Learning Perspective
- Authors: Yuanze Li, Chun-Mei Feng, Qilong Wang, Guanglei Yang, Wangmeng Zuo,
- Abstract summary: Multi-task learning methods suggest using auxiliary tasks to enhance a neural network's performance on a specific primary task.
Previous methods often select auxiliary tasks carefully but treat them as secondary during training.
We propose an uncertainty-based impartial learning method that ensures balanced training across all tasks.
- Score: 55.531894882776726
- License:
- Abstract: Human beings can leverage knowledge from relative tasks to improve learning on a primary task. Similarly, multi-task learning methods suggest using auxiliary tasks to enhance a neural network's performance on a specific primary task. However, previous methods often select auxiliary tasks carefully but treat them as secondary during training. The weights assigned to auxiliary losses are typically smaller than the primary loss weight, leading to insufficient training on auxiliary tasks and ultimately failing to support the main task effectively. To address this issue, we propose an uncertainty-based impartial learning method that ensures balanced training across all tasks. Additionally, we consider both gradients and uncertainty information during backpropagation to further improve performance on the primary task. Extensive experiments show that our method achieves performance comparable to or better than state-of-the-art approaches. Moreover, our weighting strategy is effective and robust in enhancing the performance of the primary task regardless the noise auxiliary tasks' pseudo labels.
Related papers
- Reinforcement Learning via Auxiliary Task Distillation [24.87090247662755]
We present Reinforcement Learning via Auxiliary Task Distillation (AuxDistill)
AuxDistill enables reinforcement learning to perform long-horizon robot control problems by distilling behaviors from auxiliary tasks.
We demonstrate that AuxDistill can learn a pixels-to-actions policy for a challenging multi-stage embodied object rearrangement task without demonstrations, a learning curriculum, or pre-trained skills.
arXiv Detail & Related papers (2024-06-24T23:02:18Z) - Instruction Matters: A Simple yet Effective Task Selection for Optimized Instruction Tuning of Specific Tasks [51.15473776489712]
We introduce a simple yet effective task selection method that leverages instruction information alone to identify relevant tasks.
Our method is significantly more efficient than traditional approaches, which require complex measurements of pairwise transferability between tasks or the creation of data samples for the target task.
Experimental results demonstrate that training on a small set of tasks, chosen solely on the instructions, results in substantial improvements in performance on benchmarks such as P3, Big-Bench, NIV2, and Big-Bench Hard.
arXiv Detail & Related papers (2024-04-25T08:49:47Z) - Auxiliary Learning as an Asymmetric Bargaining Game [50.826710465264505]
We propose a novel approach, named AuxiNash, for balancing tasks in auxiliary learning.
We describe an efficient procedure for learning the bargaining power of tasks based on their contribution to the performance of the main task.
We evaluate AuxiNash on multiple multi-task benchmarks and find that it consistently outperforms competing methods.
arXiv Detail & Related papers (2023-01-31T09:41:39Z) - Auxiliary task discovery through generate-and-test [7.800263769988046]
Auxiliary tasks improve data efficiency by forcing the agent to learn auxiliary prediction and control objectives.
In this paper, we explore an approach to auxiliary task discovery in reinforcement learning based on ideas from representation learning.
We introduce a new measure of auxiliary tasks' usefulness based on how useful the features induced by them are for the main task.
arXiv Detail & Related papers (2022-10-25T22:04:37Z) - Transfer Learning in Conversational Analysis through Reusing
Preprocessing Data as Supervisors [52.37504333689262]
Using noisy labels in single-task learning increases the risk of over-fitting.
Auxiliary tasks could improve the performance of the primary task learning during the same training.
arXiv Detail & Related papers (2021-12-02T08:40:42Z) - Should We Be Pre-training? An Argument for End-task Aware Training as an
Alternative [88.11465517304515]
In general, the pre-training step relies on little to no direct knowledge of the task on which the model will be fine-tuned.
We show that multi-tasking the end-task and auxiliary objectives results in significantly better downstream task performance.
arXiv Detail & Related papers (2021-09-15T17:13:18Z) - Auxiliary Task Reweighting for Minimum-data Learning [118.69683270159108]
Supervised learning requires a large amount of training data, limiting its application where labeled data is scarce.
To compensate for data scarcity, one possible method is to utilize auxiliary tasks to provide additional supervision for the main task.
We propose a method to automatically reweight auxiliary tasks in order to reduce the data requirement on the main task.
arXiv Detail & Related papers (2020-10-16T08:45:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.