Visuomotor Mechanical Search: Learning to Retrieve Target Objects in
Clutter
- URL: http://arxiv.org/abs/2008.06073v1
- Date: Thu, 13 Aug 2020 18:23:00 GMT
- Title: Visuomotor Mechanical Search: Learning to Retrieve Target Objects in
Clutter
- Authors: Andrey Kurenkov, Joseph Taglic, Rohun Kulkarni, Marcus
Dominguez-Kuhne, Animesh Garg, Roberto Mart\'in-Mart\'in, Silvio Savarese
- Abstract summary: We present a novel Deep RL procedure that combines teacher-aided exploration, ii) a critic with privileged information, andiii) mid-level representations.
Our approach trains faster and converges to more efficient uncovering solutions than baselines and ablations, and that our uncovering policies lead to an average improvement in the graspability of the target object.
- Score: 43.668395529368354
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When searching for objects in cluttered environments, it is often necessary
to perform complex interactions in order to move occluding objects out of the
way and fully reveal the object of interest and make it graspable. Due to the
complexity of the physics involved and the lack of accurate models of the
clutter, planning and controlling precise predefined interactions with accurate
outcome is extremely hard, when not impossible. In problems where accurate
(forward) models are lacking, Deep Reinforcement Learning (RL) has shown to be
a viable solution to map observations (e.g. images) to good interactions in the
form of close-loop visuomotor policies. However, Deep RL is sample inefficient
and fails when applied directly to the problem of unoccluding objects based on
images. In this work we present a novel Deep RL procedure that combines i)
teacher-aided exploration, ii) a critic with privileged information, and iii)
mid-level representations, resulting in sample efficient and effective learning
for the problem of uncovering a target object occluded by a heap of unknown
objects. Our experiments show that our approach trains faster and converges to
more efficient uncovering solutions than baselines and ablations, and that our
uncovering policies lead to an average improvement in the graspability of the
target object, facilitating downstream retrieval applications.
Related papers
- Learning Extrinsic Dexterity with Parameterized Manipulation Primitives [8.7221770019454]
We learn a sequence of actions that utilize the environment to change the object's pose.
Our approach can control the object's state through exploiting interactions between the object, the gripper, and the environment.
We evaluate our approach on picking box-shaped objects of various weight, shape, and friction properties from a constrained table-top workspace.
arXiv Detail & Related papers (2023-10-26T21:28:23Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - Cycle Consistency Driven Object Discovery [75.60399804639403]
We introduce a method that explicitly optimize the constraint that each object in a scene should be associated with a distinct slot.
By integrating these consistency objectives into various existing slot-based object-centric methods, we showcase substantial improvements in object-discovery performance.
Our results suggest that the proposed approach not only improves object discovery, but also provides richer features for downstream tasks.
arXiv Detail & Related papers (2023-06-03T21:49:06Z) - Bridging the Gap to Real-World Object-Centric Learning [66.55867830853803]
We show that reconstructing features from models trained in a self-supervised manner is a sufficient training signal for object-centric representations to arise in a fully unsupervised way.
Our approach, DINOSAUR, significantly out-performs existing object-centric learning models on simulated data.
arXiv Detail & Related papers (2022-09-29T15:24:47Z) - Curious Exploration via Structured World Models Yields Zero-Shot Object
Manipulation [19.840186443344]
We propose to use structured world models to incorporate inductive biases in the control loop to achieve sample-efficient exploration.
Our method generates free-play behavior that starts to interact with objects early on and develops more complex behavior over time.
arXiv Detail & Related papers (2022-06-22T22:08:50Z) - Towards Optimal Correlational Object Search [25.355936023640506]
Correlational Object Search POMDP can be solved to produce search strategies that use correlational information.
We conduct experiments using AI2-THOR, a realistic simulator of household environments, as well as YOLOv5, a widely-used object detector.
arXiv Detail & Related papers (2021-10-19T14:03:43Z) - Progressive Self-Guided Loss for Salient Object Detection [102.35488902433896]
We present a progressive self-guided loss function to facilitate deep learning-based salient object detection in images.
Our framework takes advantage of adaptively aggregated multi-scale features to locate and detect salient objects effectively.
arXiv Detail & Related papers (2021-01-07T07:33:38Z) - Occlusion-Aware Search for Object Retrieval in Clutter [4.693170687870612]
We address the manipulation task of retrieving a target object from a cluttered shelf.
When the target object is hidden, the robot must search through the clutter for retrieving it.
We present a data-driven hybrid planner for generating occlusion-aware actions in closed-loop.
arXiv Detail & Related papers (2020-11-06T13:15:27Z) - Reinforcement Learning for Sparse-Reward Object-Interaction Tasks in a
First-person Simulated 3D Environment [73.9469267445146]
First-person object-interaction tasks in high-fidelity, 3D, simulated environments such as the AI2Thor pose significant sample-efficiency challenges for reinforcement learning agents.
We show that one can learn object-interaction tasks from scratch without supervision by learning an attentive object-model as an auxiliary task.
arXiv Detail & Related papers (2020-10-28T19:27:26Z) - Relevance-Guided Modeling of Object Dynamics for Reinforcement Learning [0.0951828574518325]
Current deep reinforcement learning (RL) approaches incorporate minimal prior knowledge about the environment.
We propose a framework for reasoning about object dynamics and behavior to rapidly determine minimal and task-specific object representations.
We also highlight the potential of this framework on several Atari games, using our object representation and standard RL and planning algorithms to learn dramatically faster than existing deep RL algorithms.
arXiv Detail & Related papers (2020-03-03T08:18:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.