Abstract-to-Executable Trajectory Translation for One-Shot Task
Generalization
- URL: http://arxiv.org/abs/2210.07658v2
- Date: Tue, 30 May 2023 23:44:17 GMT
- Title: Abstract-to-Executable Trajectory Translation for One-Shot Task
Generalization
- Authors: Stone Tao, Xiaochen Li, Tongzhou Mu, Zhiao Huang, Yuzhe Qin and Hao Su
- Abstract summary: We propose to achieve one-shot task generalization by decoupling plan generation and plan execution.
Our method solves complex long-horizon tasks in three steps: build a paired abstract environment, generate abstract trajectories, and solve the original task by an abstract-to-executable trajectory translator.
- Score: 21.709054087028946
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Training long-horizon robotic policies in complex physical environments is
essential for many applications, such as robotic manipulation. However,
learning a policy that can generalize to unseen tasks is challenging. In this
work, we propose to achieve one-shot task generalization by decoupling plan
generation and plan execution. Specifically, our method solves complex
long-horizon tasks in three steps: build a paired abstract environment by
simplifying geometry and physics, generate abstract trajectories, and solve the
original task by an abstract-to-executable trajectory translator. In the
abstract environment, complex dynamics such as physical manipulation are
removed, making abstract trajectories easier to generate. However, this
introduces a large domain gap between abstract trajectories and the actual
executed trajectories as abstract trajectories lack low-level details and are
not aligned frame-to-frame with the executed trajectory. In a manner
reminiscent of language translation, our approach leverages a seq-to-seq model
to overcome the large domain gap between the abstract and executable
trajectories, enabling the low-level policy to follow the abstract trajectory.
Experimental results on various unseen long-horizon tasks with different robot
embodiments demonstrate the practicability of our methods to achieve one-shot
task generalization.
Related papers
- RT-Affordance: Affordances are Versatile Intermediate Representations for Robot Manipulation [52.14638923430338]
We propose conditioning policies on affordances, which capture the pose of the robot at key stages of the task.
Our method, RT-Affordance, is a hierarchical model that first proposes an affordance plan given the task language.
We show on a diverse set of novel tasks how RT-Affordance exceeds the performance of existing methods by over 50%.
arXiv Detail & Related papers (2024-11-05T01:02:51Z) - HACMan++: Spatially-Grounded Motion Primitives for Manipulation [28.411361363637006]
We introduce spatially-grounded parameterized motion primitives in our method HACMan++.
By grounding the primitives on a spatial location in the environment, our method is able to effectively generalize across object shape and pose variations.
Our approach significantly outperforms existing methods, particularly in complex scenarios demanding both high-level sequential reasoning and object generalization.
arXiv Detail & Related papers (2024-07-11T15:10:14Z) - Unified Task and Motion Planning using Object-centric Abstractions of
Motion Constraints [56.283944756315066]
We propose an alternative TAMP approach that unifies task and motion planning into a single search.
Our approach is based on an object-centric abstraction of motion constraints that permits leveraging the computational efficiency of off-the-shelf AI search to yield physically feasible plans.
arXiv Detail & Related papers (2023-12-29T14:00:20Z) - RT-Trajectory: Robotic Task Generalization via Hindsight Trajectory
Sketches [74.300116260004]
Generalization remains one of the most important desiderata for robust robot learning systems.
We propose a policy conditioning method using rough trajectory sketches.
We show that RT-Trajectory is able to perform a wider range of tasks compared to language-conditioned and goal-conditioned policies.
arXiv Detail & Related papers (2023-11-03T15:31:51Z) - Planning Immediate Landmarks of Targets for Model-Free Skill Transfer
across Agents [34.56191646231944]
We propose PILoT, i.e., Planning Immediate Landmarks of Targets.
PILoT learns a goal-conditioned state planner and distills a goal-planner to plan immediate landmarks in a model-free style.
We show the power of PILoT on various transferring challenges, including few-shot transferring across action spaces and dynamics.
arXiv Detail & Related papers (2022-12-18T08:03:21Z) - Inventing Relational State and Action Abstractions for Effective and
Efficient Bilevel Planning [26.715198108255162]
We develop a novel framework for learning state and action abstractions.
We learn relational, neuro-symbolic abstractions that generalize over object identities and numbers.
We show that our learned abstractions are able to quickly solve held-out tasks of longer horizons.
arXiv Detail & Related papers (2022-03-17T22:13:09Z) - Learning to Shift Attention for Motion Generation [55.61994201686024]
One challenge of motion generation using robot learning from demonstration techniques is that human demonstrations follow a distribution with multiple modes for one task query.
Previous approaches fail to capture all modes or tend to average modes of the demonstrations and thus generate invalid trajectories.
We propose a motion generation model with extrapolation ability to overcome this problem.
arXiv Detail & Related papers (2021-02-24T09:07:52Z) - ReLMoGen: Leveraging Motion Generation in Reinforcement Learning for
Mobile Manipulation [99.2543521972137]
ReLMoGen is a framework that combines a learned policy to predict subgoals and a motion generator to plan and execute the motion needed to reach these subgoals.
Our method is benchmarked on a diverse set of seven robotics tasks in photo-realistic simulation environments.
ReLMoGen shows outstanding transferability between different motion generators at test time, indicating a great potential to transfer to real robots.
arXiv Detail & Related papers (2020-08-18T08:05:15Z) - Learning Abstract Models for Strategic Exploration and Fast Reward
Transfer [85.19766065886422]
We learn an accurate Markov Decision Process (MDP) over abstract states to avoid compounding errors.
Our approach achieves strong results on three of the hardest Arcade Learning Environment games.
We can reuse the learned abstract MDP for new reward functions, achieving higher reward in 1000x fewer samples than model-free methods trained from scratch.
arXiv Detail & Related papers (2020-07-12T03:33:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.