Distill Knowledge in Multi-task Reinforcement Learning with
  Optimal-Transport Regularization
        - URL: http://arxiv.org/abs/2309.15603v1
 - Date: Wed, 27 Sep 2023 12:06:34 GMT
 - Title: Distill Knowledge in Multi-task Reinforcement Learning with
  Optimal-Transport Regularization
 - Authors: Bang Giang Le, Viet Cuong Ta
 - Abstract summary: In multi-task reinforcement learning, it is possible to improve the data efficiency of training agents by transferring knowledge from other different but related tasks.
Traditional methods rely on Kullback-Leibler regularization to stabilize the transfer of knowledge from one task to the others.
In this work, we explore the direction of replacing the Kullback-Leibler divergence with a novel Optimal transport-based regularization.
 - Score: 0.24475591916185496
 - License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
 - Abstract:   In multi-task reinforcement learning, it is possible to improve the data
efficiency of training agents by transferring knowledge from other different
but related tasks. Because the experiences from different tasks are usually
biased toward the specific task goals. Traditional methods rely on
Kullback-Leibler regularization to stabilize the transfer of knowledge from one
task to the others. In this work, we explore the direction of replacing the
Kullback-Leibler divergence with a novel Optimal transport-based
regularization. By using the Sinkhorn mapping, we can approximate the Optimal
transport distance between the state distribution of tasks. The distance is
then used as an amortized reward to regularize the amount of sharing
information. We experiment our frameworks on several grid-based navigation
multi-goal to validate the effectiveness of the approach. The results show that
our added Optimal transport-based rewards are able to speed up the learning
process of agents and outperforms several baselines on multi-task learning.
 
       
      
        Related papers
        - PEMT: Multi-Task Correlation Guided Mixture-of-Experts Enables   Parameter-Efficient Transfer Learning [28.353530290015794]
We propose PEMT, a novel parameter-efficient fine-tuning framework based on multi-task transfer learning.
We conduct experiments on a broad range of tasks over 17 datasets.
arXiv  Detail & Related papers  (2024-02-23T03:59:18Z) - Sharing Knowledge in Multi-Task Deep Reinforcement Learning [57.38874587065694]
We study the benefit of sharing representations among tasks to enable the effective use of deep neural networks in Multi-Task Reinforcement Learning.
We prove this by providing theoretical guarantees that highlight the conditions for which is convenient to share representations among tasks.
arXiv  Detail & Related papers  (2024-01-17T19:31:21Z) - Similarity-based Knowledge Transfer for Cross-Domain Reinforcement
  Learning [3.3148826359547523]
We develop a semi-supervised alignment loss to match different spaces with a set of encoder-decoders.
In comparison to prior works, our method does not require data to be aligned, paired or collected by expert policies.
arXiv  Detail & Related papers  (2023-12-05T19:26:01Z) - Visual Exemplar Driven Task-Prompting for Unified Perception in
  Autonomous Driving [100.3848723827869]
We present an effective multi-task framework, VE-Prompt, which introduces visual exemplars via task-specific prompting.
 Specifically, we generate visual exemplars based on bounding boxes and color-based markers, which provide accurate visual appearances of target categories.
We bridge transformer-based encoders and convolutional layers for efficient and accurate unified perception in autonomous driving.
arXiv  Detail & Related papers  (2023-03-03T08:54:06Z) - ForkMerge: Mitigating Negative Transfer in Auxiliary-Task Learning [59.08197876733052]
Auxiliary-Task Learning (ATL) aims to improve the performance of the target task by leveraging the knowledge obtained from related tasks.
Sometimes, learning multiple tasks simultaneously results in lower accuracy than learning only the target task, known as negative transfer.
ForkMerge is a novel approach that periodically forks the model into multiple branches, automatically searches the varying task weights.
arXiv  Detail & Related papers  (2023-01-30T02:27:02Z) - Learning Multi-Task Transferable Rewards via Variational Inverse
  Reinforcement Learning [10.782043595405831]
We extend an empowerment-based regularization technique to situations with multiple tasks based on the framework of a generative adversarial network.
Under the multitask environments with unknown dynamics, we focus on learning a reward and policy from unlabeled expert examples.
Our proposed method derives the variational lower bound of the situational mutual information to optimize it.
arXiv  Detail & Related papers  (2022-06-19T22:32:41Z) - Measuring and Harnessing Transference in Multi-Task Learning [58.48659733262734]
Multi-task learning can leverage information learned by one task to benefit the training of other tasks.
We analyze the dynamics of information transfer, or transference, across tasks throughout training.
arXiv  Detail & Related papers  (2020-10-29T08:25:43Z) - Unsupervised Transfer Learning for Spatiotemporal Predictive Networks [90.67309545798224]
We study how to transfer knowledge from a zoo of unsupervisedly learned models towards another network.
Our motivation is that models are expected to understand complex dynamics from different sources.
Our approach yields significant improvements on three benchmarks fortemporal prediction, and benefits the target even from less relevant ones.
arXiv  Detail & Related papers  (2020-09-24T15:40:55Z) - Uniform Priors for Data-Efficient Transfer [65.086680950871]
We show that features that are most transferable have high uniformity in the embedding space.
We evaluate the regularization on its ability to facilitate adaptation to unseen tasks and data.
arXiv  Detail & Related papers  (2020-06-30T04:39:36Z) - Inter- and Intra-domain Knowledge Transfer for Related Tasks in Deep
  Character Recognition [2.320417845168326]
Pre-training a deep neural network on the ImageNet dataset is a common practice for training deep learning models.
The technique of pre-training on one task and then retraining on a new one is called transfer learning.
In this paper we analyse the effectiveness of using deep transfer learning for character recognition tasks.
arXiv  Detail & Related papers  (2020-01-02T14:18:25Z) 
        This list is automatically generated from the titles and abstracts of the papers in this site.
       
     
           This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.