Meta-Reinforcement Learning via Exploratory Task Clustering
- URL: http://arxiv.org/abs/2302.07958v1
- Date: Wed, 15 Feb 2023 21:42:38 GMT
- Title: Meta-Reinforcement Learning via Exploratory Task Clustering
- Authors: Zhendong Chu, Hongning Wang
- Abstract summary: We develop a dedicated exploratory policy to discover task structures via divide-and-conquer.
The knowledge of the identified clusters helps to narrow the search space of task-specific information.
Experiments on various MuJoCo tasks showed the proposed method can unravel cluster structures effectively in both rewards and state dynamics.
- Score: 43.936406999765886
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Meta-reinforcement learning (meta-RL) aims to quickly solve new tasks by
leveraging knowledge from prior tasks. However, previous studies often assume a
single mode homogeneous task distribution, ignoring possible structured
heterogeneity among tasks. Leveraging such structures can better facilitate
knowledge sharing among related tasks and thus improve sample efficiency. In
this paper, we explore the structured heterogeneity among tasks via clustering
to improve meta-RL. We develop a dedicated exploratory policy to discover task
structures via divide-and-conquer. The knowledge of the identified clusters
helps to narrow the search space of task-specific information, leading to more
sample efficient policy adaptation. Experiments on various MuJoCo tasks showed
the proposed method can unravel cluster structures effectively in both rewards
and state dynamics, proving strong advantages against a set of state-of-the-art
baselines.
Related papers
- Learning Action Translator for Meta Reinforcement Learning on
Sparse-Reward Tasks [56.63855534940827]
This work introduces a novel objective function to learn an action translator among training tasks.
We theoretically verify that the value of the transferred policy with the action translator can be close to the value of the source policy.
We propose to combine the action translator with context-based meta-RL algorithms for better data collection and more efficient exploration during meta-training.
arXiv Detail & Related papers (2022-07-19T04:58:06Z) - EEML: Ensemble Embedded Meta-learning [5.9514420658483935]
We propose an ensemble embedded meta-learning algorithm (EEML) that explicitly utilizes multi-model-ensemble to organize prior knowledge into diverse specific experts.
We rely on a task embedding cluster mechanism to deliver diverse tasks to matching experts in training process and instruct how experts collaborate in test phase.
The experimental results show that the proposed method outperforms recent state-of-the-arts easily in few-shot learning problem.
arXiv Detail & Related papers (2022-06-18T12:37:17Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Meta-Learning with Fewer Tasks through Task Interpolation [67.03769747726666]
Current meta-learning algorithms require a large number of meta-training tasks, which may not be accessible in real-world scenarios.
By meta-learning with task gradient (MLTI), our approach effectively generates additional tasks by randomly sampling a pair of tasks and interpolating the corresponding features and labels.
Empirically, in our experiments on eight datasets from diverse domains, we find that the proposed general MLTI framework is compatible with representative meta-learning algorithms and consistently outperforms other state-of-the-art strategies.
arXiv Detail & Related papers (2021-06-04T20:15:34Z) - MetaCURE: Meta Reinforcement Learning with Empowerment-Driven
Exploration [52.48362697163477]
Experimental evaluation shows that our meta-RL method significantly outperforms state-of-the-art baselines on sparse-reward tasks.
We model an exploration policy learning problem for meta-RL, which is separated from exploitation policy learning.
We develop a new off-policy meta-RL framework, which efficiently learns separate context-aware exploration and exploitation policies.
arXiv Detail & Related papers (2020-06-15T06:56:18Z) - Automated Relational Meta-learning [95.02216511235191]
We propose an automated relational meta-learning framework that automatically extracts the cross-task relations and constructs the meta-knowledge graph.
We conduct extensive experiments on 2D toy regression and few-shot image classification and the results demonstrate the superiority of ARML over state-of-the-art baselines.
arXiv Detail & Related papers (2020-01-03T07:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.