CompoSuite: A Compositional Reinforcement Learning Benchmark
- URL: http://arxiv.org/abs/2207.04136v1
- Date: Fri, 8 Jul 2022 22:01:52 GMT
- Title: CompoSuite: A Compositional Reinforcement Learning Benchmark
- Authors: Jorge A. Mendez, Marcel Hussing, Meghna Gummadi, Eric Eaton
- Abstract summary: We present CompoSuite, an open-source benchmark for compositional multi-task reinforcement learning (RL)
Each CompoSuite task requires a particular robot arm to manipulate one individual object to achieve a task objective while avoiding an obstacle.
We benchmark existing single-task, multi-task, and compositional learning algorithms on various training settings, and assess their capability to compositionally generalize to unseen tasks.
- Score: 20.89464587308586
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We present CompoSuite, an open-source simulated robotic manipulation
benchmark for compositional multi-task reinforcement learning (RL). Each
CompoSuite task requires a particular robot arm to manipulate one individual
object to achieve a task objective while avoiding an obstacle. This
compositional definition of the tasks endows CompoSuite with two remarkable
properties. First, varying the robot/object/objective/obstacle elements leads
to hundreds of RL tasks, each of which requires a meaningfully different
behavior. Second, RL approaches can be evaluated specifically for their ability
to learn the compositional structure of the tasks. This latter capability to
functionally decompose problems would enable intelligent agents to identify and
exploit commonalities between learning tasks to handle large varieties of
highly diverse problems. We benchmark existing single-task, multi-task, and
compositional learning algorithms on various training settings, and assess
their capability to compositionally generalize to unseen tasks. Our evaluation
exposes the shortcomings of existing RL approaches with respect to
compositionality and opens new avenues for investigation.
Related papers
- Guiding Multi-agent Multi-task Reinforcement Learning by a Hierarchical Framework with Logical Reward Shaping [16.5526277899717]
This study aims to design a multi-agent cooperative algorithm with logic reward shaping.
Experiments have been conducted on various types of tasks in the Minecraft-like environment.
arXiv Detail & Related papers (2024-11-02T09:03:23Z) - MmAP : Multi-modal Alignment Prompt for Cross-domain Multi-task Learning [29.88567810099265]
Multi-task learning is designed to train multiple correlated tasks simultaneously.
To tackle this challenge, we integrate the decoder-free vision-language model CLIP.
We propose Multi-modal Alignment Prompt (MmAP) for CLIP, which aligns text and visual modalities during fine-tuning process.
arXiv Detail & Related papers (2023-12-14T03:33:02Z) - Zero-Shot Policy Transfer with Disentangled Task Representation of
Meta-Reinforcement Learning [30.633075584454275]
In this work, we aim to achieve zero-shot policy generalization of Reinforcement Learning (RL) agents by leveraging the task compositionality.
Our proposed method is a meta- RL algorithm with disentangled task representation, explicitly encoding different aspects of the tasks.
Policy generalization is then performed by inferring unseen compositional task representations via the obtained disentanglement.
arXiv Detail & Related papers (2022-10-01T19:31:46Z) - Task-Agnostic Continual Reinforcement Learning: Gaining Insights and
Overcoming Challenges [27.474011433615317]
Continual learning (CL) enables the development of models and agents that learn from a sequence of tasks.
We investigate the factors that contribute to the performance differences between task-agnostic CL and multi-task (MTL) agents.
arXiv Detail & Related papers (2022-05-28T17:59:00Z) - Fast Inference and Transfer of Compositional Task Structures for
Few-shot Task Generalization [101.72755769194677]
We formulate it as a few-shot reinforcement learning problem where a task is characterized by a subtask graph.
Our multi-task subtask graph inferencer (MTSGI) first infers the common high-level task structure in terms of the subtask graph from the training tasks.
Our experiment results on 2D grid-world and complex web navigation domains show that the proposed method can learn and leverage the common underlying structure of the tasks for faster adaptation to the unseen tasks.
arXiv Detail & Related papers (2022-05-25T10:44:25Z) - Combining Modular Skills in Multitask Learning [149.8001096811708]
A modular design encourages neural models to disentangle and recombine different facets of knowledge to generalise more systematically to new tasks.
In this work, we assume each task is associated with a subset of latent discrete skills from a (potentially small) inventory.
We find that the modular design of a network significantly increases sample efficiency in reinforcement learning and few-shot generalisation in supervised learning.
arXiv Detail & Related papers (2022-02-28T16:07:19Z) - Efficiently Identifying Task Groupings for Multi-Task Learning [55.80489920205404]
Multi-task learning can leverage information learned by one task to benefit the training of other tasks.
We suggest an approach to select which tasks should train together in multi-task learning models.
Our method determines task groupings in a single training run by co-training all tasks together and quantifying the effect to which one task's gradient would affect another task's loss.
arXiv Detail & Related papers (2021-09-10T02:01:43Z) - Exploring Relational Context for Multi-Task Dense Prediction [76.86090370115]
We consider a multi-task environment for dense prediction tasks, represented by a common backbone and independent task-specific heads.
We explore various attention-based contexts, such as global and local, in the multi-task setting.
We propose an Adaptive Task-Relational Context module, which samples the pool of all available contexts for each task pair.
arXiv Detail & Related papers (2021-04-28T16:45:56Z) - CausalWorld: A Robotic Manipulation Benchmark for Causal Structure and
Transfer Learning [138.40338621974954]
CausalWorld is a benchmark for causal structure and transfer learning in a robotic manipulation environment.
Tasks consist of constructing 3D shapes from a given set of blocks - inspired by how children learn to build complex structures.
arXiv Detail & Related papers (2020-10-08T23:01:13Z) - LEMMA: A Multi-view Dataset for Learning Multi-agent Multi-task
Activities [119.88381048477854]
We introduce the LEMMA dataset to provide a single home to address missing dimensions with meticulously designed settings.
We densely annotate the atomic-actions with human-object interactions to provide ground-truths of the compositionality, scheduling, and assignment of daily activities.
We hope this effort would drive the machine vision community to examine goal-directed human activities and further study the task scheduling and assignment in the real world.
arXiv Detail & Related papers (2020-07-31T00:13:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.