Generalizing Object-Centric Task-Axes Controllers using Keypoints
- URL: http://arxiv.org/abs/2103.10524v1
- Date: Thu, 18 Mar 2021 21:08:00 GMT
- Title: Generalizing Object-Centric Task-Axes Controllers using Keypoints
- Authors: Mohit Sharma, Oliver Kroemer
- Abstract summary: We learn modular task policies which compose object-centric task-axes controllers.
These task-axes controllers are parameterized by properties associated with underlying objects in the scene.
Our overall approach provides a simple, modular and yet powerful framework for learning manipulation tasks.
- Score: 15.427056235112152
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To perform manipulation tasks in the real world, robots need to operate on
objects with various shapes, sizes and without access to geometric models. It
is often unfeasible to train monolithic neural network policies across such
large variance in object properties. Towards this generalization challenge, we
propose to learn modular task policies which compose object-centric task-axes
controllers. These task-axes controllers are parameterized by properties
associated with underlying objects in the scene. We infer these controller
parameters directly from visual input using multi-view dense correspondence
learning. Our overall approach provides a simple, modular and yet powerful
framework for learning manipulation tasks. We empirically evaluate our approach
on multiple different manipulation tasks and show its ability to generalize to
large variance in object size, shape and geometry.
Related papers
- Keypoint Abstraction using Large Models for Object-Relative Imitation Learning [78.92043196054071]
Generalization to novel object configurations and instances across diverse tasks and environments is a critical challenge in robotics.
Keypoint-based representations have been proven effective as a succinct representation for essential object capturing features.
We propose KALM, a framework that leverages large pre-trained vision-language models to automatically generate task-relevant and cross-instance consistent keypoints.
arXiv Detail & Related papers (2024-10-30T17:37:31Z) - Entity-Centric Reinforcement Learning for Object Manipulation from Pixels [22.104757862869526]
Reinforcement Learning (RL) offers a general approach to learn object manipulation.
In practice, domains with more than a few objects are difficult for RL agents due to the curse of dimensionality.
We propose a structured approach for visual RL that is suitable for representing multiple objects and their interaction.
arXiv Detail & Related papers (2024-04-01T16:25:08Z) - Learning Reusable Manipulation Strategies [86.07442931141634]
Humans demonstrate an impressive ability to acquire and generalize manipulation "tricks"
We present a framework that enables machines to acquire such manipulation skills through a single demonstration and self-play.
These learned mechanisms and samplers can be seamlessly integrated into standard task and motion planners.
arXiv Detail & Related papers (2023-11-06T17:35:42Z) - Learning Extrinsic Dexterity with Parameterized Manipulation Primitives [8.7221770019454]
We learn a sequence of actions that utilize the environment to change the object's pose.
Our approach can control the object's state through exploiting interactions between the object, the gripper, and the environment.
We evaluate our approach on picking box-shaped objects of various weight, shape, and friction properties from a constrained table-top workspace.
arXiv Detail & Related papers (2023-10-26T21:28:23Z) - GAMMA: Generalizable Articulation Modeling and Manipulation for
Articulated Objects [53.965581080954905]
We propose a novel framework of Generalizable Articulation Modeling and Manipulating for Articulated Objects (GAMMA)
GAMMA learns both articulation modeling and grasp pose affordance from diverse articulated objects with different categories.
Results show that GAMMA significantly outperforms SOTA articulation modeling and manipulation algorithms in unseen and cross-category articulated objects.
arXiv Detail & Related papers (2023-09-28T08:57:14Z) - Universal Instance Perception as Object Discovery and Retrieval [90.96031157557806]
UNI reformulates diverse instance perception tasks into a unified object discovery and retrieval paradigm.
It can flexibly perceive different types of objects by simply changing the input prompts.
UNI shows superior performance on 20 challenging benchmarks from 10 instance-level tasks.
arXiv Detail & Related papers (2023-03-12T14:28:24Z) - Decoupling Skill Learning from Robotic Control for Generalizable Object
Manipulation [35.34044822433743]
Recent works in robotic manipulation have shown potential for tackling a range of tasks.
We conjecture that this is due to the high-dimensional action space for joint control.
In this paper, we take an alternative approach and separate the task of learning 'what to do' from 'how to do it'
The whole-body robotic kinematic control is optimized to execute the high-dimensional joint motion to reach the goals in the workspace.
arXiv Detail & Related papers (2023-03-07T16:31:13Z) - Generalization in Dexterous Manipulation via Geometry-Aware Multi-Task
Learning [108.08083976908195]
We show that policies learned by existing reinforcement learning algorithms can in fact be generalist.
We show that a single generalist policy can perform in-hand manipulation of over 100 geometrically-diverse real-world objects.
Interestingly, we find that multi-task learning with object point cloud representations not only generalizes better but even outperforms single-object specialist policies.
arXiv Detail & Related papers (2021-11-04T17:59:56Z) - Learning to Compose Hierarchical Object-Centric Controllers for Robotic
Manipulation [26.24940293693809]
We propose using reinforcement learning to compose hierarchical object-centric controllers for manipulation tasks.
Experiments in both simulation and real world show how the proposed approach leads to improved sample efficiency, zero-shot generalization, and simulation-to-reality transfer without fine-tuning.
arXiv Detail & Related papers (2020-11-09T18:38:29Z) - Goal-Conditioned End-to-End Visuomotor Control for Versatile Skill
Primitives [89.34229413345541]
We propose a conditioning scheme which avoids pitfalls by learning the controller and its conditioning in an end-to-end manner.
Our model predicts complex action sequences based directly on a dynamic image representation of the robot motion.
We report significant improvements in task success over representative MPC and IL baselines.
arXiv Detail & Related papers (2020-03-19T15:04:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.