Emergence of Different Modes of Tool Use in a Reaching and Dragging Task
- URL: http://arxiv.org/abs/2012.04700v1
- Date: Tue, 8 Dec 2020 19:37:58 GMT
- Title: Emergence of Different Modes of Tool Use in a Reaching and Dragging Task
- Authors: Khuong Nguyen and Yoonsuck Choe
- Abstract summary: In this paper, we investigate different modes of tool use that emerge in a reaching and dragging task.
We trained a deep-reinforcement learning based controller with minimal reward shaping information to tackle this task.
We observed the emergence of a wide range of unexpected behaviors, not directly encoded in the motor primitives or reward functions.
- Score: 7.40839907166763
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Tool use is an important milestone in the evolution of intelligence. In this
paper, we investigate different modes of tool use that emerge in a reaching and
dragging task. In this task, a jointed arm with a gripper must grab a tool (T,
I, or L-shaped) and drag an object down to the target location (the bottom of
the arena). The simulated environment had real physics such as gravity and
friction. We trained a deep-reinforcement learning based controller (with raw
visual and proprioceptive input) with minimal reward shaping information to
tackle this task. We observed the emergence of a wide range of unexpected
behaviors, not directly encoded in the motor primitives or reward functions.
Examples include hitting the object to the target location, correcting error of
initial contact, throwing the tool toward the object, as well as normal
expected behavior such as wide sweep. Also, we further analyzed these behaviors
based on the type of tool and the initial position of the target object. Our
results show a rich repertoire of behaviors, beyond the basic built-in
mechanisms of the deep reinforcement learning method we used.
Related papers
- Dynamic object goal pushing with mobile manipulators through model-free constrained reinforcement learning [9.305146484955296]
We develop a learning-based controller for a mobile manipulator to move an unknown object to a desired position and yaw orientation through a sequence of pushing actions.
The proposed controller for the robotic arm and the mobile base motion is trained using a constrained Reinforcement Learning (RL) formulation.
The learned policy achieves a success rate of 91.35% in simulation and at least 80% on hardware in challenging scenarios.
arXiv Detail & Related papers (2025-02-03T17:28:35Z) - Articulated Object Manipulation using Online Axis Estimation with SAM2-Based Tracking [59.87033229815062]
Articulated object manipulation requires precise object interaction, where the object's axis must be carefully considered.
Previous research employed interactive perception for manipulating articulated objects, but typically, open-loop approaches often suffer from overlooking the interaction dynamics.
We present a closed-loop pipeline integrating interactive perception with online axis estimation from segmented 3D point clouds.
arXiv Detail & Related papers (2024-09-24T17:59:56Z) - Learning Extrinsic Dexterity with Parameterized Manipulation Primitives [8.7221770019454]
We learn a sequence of actions that utilize the environment to change the object's pose.
Our approach can control the object's state through exploiting interactions between the object, the gripper, and the environment.
We evaluate our approach on picking box-shaped objects of various weight, shape, and friction properties from a constrained table-top workspace.
arXiv Detail & Related papers (2023-10-26T21:28:23Z) - Localizing Active Objects from Egocentric Vision with Symbolic World
Knowledge [62.981429762309226]
The ability to actively ground task instructions from an egocentric view is crucial for AI agents to accomplish tasks or assist humans virtually.
We propose to improve phrase grounding models' ability on localizing the active objects by: learning the role of objects undergoing change and extracting them accurately from the instructions.
We evaluate our framework on Ego4D and Epic-Kitchens datasets.
arXiv Detail & Related papers (2023-10-23T16:14:05Z) - Is an Object-Centric Video Representation Beneficial for Transfer? [86.40870804449737]
We introduce a new object-centric video recognition model on a transformer architecture.
We show that the object-centric model outperforms prior video representations.
arXiv Detail & Related papers (2022-07-20T17:59:44Z) - DiffSkill: Skill Abstraction from Differentiable Physics for Deformable
Object Manipulations with Tools [96.38972082580294]
DiffSkill is a novel framework that uses a differentiable physics simulator for skill abstraction to solve deformable object manipulation tasks.
In particular, we first obtain short-horizon skills using individual tools from a gradient-based simulator.
We then learn a neural skill abstractor from the demonstration trajectories which takes RGBD images as input.
arXiv Detail & Related papers (2022-03-31T17:59:38Z) - Object Manipulation via Visual Target Localization [64.05939029132394]
Training agents to manipulate objects, poses many challenges.
We propose an approach that explores the environment in search for target objects, computes their 3D coordinates once they are located, and then continues to estimate their 3D locations even when the objects are not visible.
Our evaluations show a massive 3x improvement in success rate over a model that has access to the same sensory suite.
arXiv Detail & Related papers (2022-03-15T17:59:01Z) - Task-Focused Few-Shot Object Detection for Robot Manipulation [1.8275108630751844]
We develop a manipulation method based solely on detection then introduce task-focused few-shot object detection to learn new objects and settings.
In experiments for our interactive approach to few-shot learning, we train a robot to manipulate objects directly from detection (ClickBot)
arXiv Detail & Related papers (2022-01-28T21:52:05Z) - How to select and use tools? : Active Perception of Target Objects Using
Multimodal Deep Learning [9.677391628613025]
We focus on active perception using multimodal sensorimotor data while a robot interacts with objects.
We construct a deep neural networks (DNN) model that learns to recognize object characteristics.
We also examine the contributions of images, force, and tactile data and show that learning a variety of multimodal information results in rich perception for tool use.
arXiv Detail & Related papers (2021-06-04T12:49:30Z) - TANGO: Commonsense Generalization in Predicting Tool Interactions for
Mobile Manipulators [15.61285199988595]
We introduce TANGO, a novel neural model for predicting task-specific tool interactions.
TANGO encodes the world state comprising of objects and symbolic relationships between them using a graph neural network.
We show that by augmenting the representation of the environment with pre-trained embeddings derived from a knowledge-base, the model can generalize effectively to novel environments.
arXiv Detail & Related papers (2021-05-05T18:11:57Z) - Learning Dexterous Grasping with Object-Centric Visual Affordances [86.49357517864937]
Dexterous robotic hands are appealing for their agility and human-like morphology.
We introduce an approach for learning dexterous grasping.
Our key idea is to embed an object-centric visual affordance model within a deep reinforcement learning loop.
arXiv Detail & Related papers (2020-09-03T04:00:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.