DexVIP: Learning Dexterous Grasping with Human Hand Pose Priors from
Video
- URL: http://arxiv.org/abs/2202.00164v1
- Date: Tue, 1 Feb 2022 00:45:57 GMT
- Title: DexVIP: Learning Dexterous Grasping with Human Hand Pose Priors from
Video
- Authors: Priyanka Mandikal and Kristen Grauman
- Abstract summary: We propose DexVIP, an approach to learn dexterous robotic grasping from human-object interaction videos.
We do this by curating grasp images from human-object interaction videos and imposing a prior over the agent's hand pose.
We demonstrate that DexVIP compares favorably to existing approaches that lack a hand pose prior or rely on specialized tele-operation equipment.
- Score: 86.49357517864937
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Dexterous multi-fingered robotic hands have a formidable action space, yet
their morphological similarity to the human hand holds immense potential to
accelerate robot learning. We propose DexVIP, an approach to learn dexterous
robotic grasping from human-object interactions present in in-the-wild YouTube
videos. We do this by curating grasp images from human-object interaction
videos and imposing a prior over the agent's hand pose when learning to grasp
with deep reinforcement learning. A key advantage of our method is that the
learned policy is able to leverage free-form in-the-wild visual data. As a
result, it can easily scale to new objects, and it sidesteps the standard
practice of collecting human demonstrations in a lab -- a much more expensive
and indirect way to capture human expertise. Through experiments on 27 objects
with a 30-DoF simulated robot hand, we demonstrate that DexVIP compares
favorably to existing approaches that lack a hand pose prior or rely on
specialized tele-operation equipment to obtain human demonstrations, while also
being faster to train. Project page:
https://vision.cs.utexas.edu/projects/dexvip-dexterous-grasp-pose-prior
Related papers
- Bridging the Human to Robot Dexterity Gap through Object-Oriented Rewards [15.605887551756934]
Training robots directly from human videos is an emerging area in robotics and computer vision.
A key reason for this difficulty is that a policy trained on human hands may not directly transfer to a robot hand due to morphology differences.
We present HuDOR, a technique that enables online fine-tuning of policies by directly computing rewards from human videos.
arXiv Detail & Related papers (2024-10-30T17:59:41Z) - RealDex: Towards Human-like Grasping for Robotic Dexterous Hand [64.47045863999061]
We introduce RealDex, a pioneering dataset capturing authentic dexterous hand grasping motions infused with human behavioral patterns.
RealDex holds immense promise in advancing humanoid robot for automated perception, cognition, and manipulation in real-world scenarios.
arXiv Detail & Related papers (2024-02-21T14:59:46Z) - Giving Robots a Hand: Learning Generalizable Manipulation with
Eye-in-Hand Human Video Demonstrations [66.47064743686953]
Eye-in-hand cameras have shown promise in enabling greater sample efficiency and generalization in vision-based robotic manipulation.
Videos of humans performing tasks, on the other hand, are much cheaper to collect since they eliminate the need for expertise in robotic teleoperation.
In this work, we augment narrow robotic imitation datasets with broad unlabeled human video demonstrations to greatly enhance the generalization of eye-in-hand visuomotor policies.
arXiv Detail & Related papers (2023-07-12T07:04:53Z) - VideoDex: Learning Dexterity from Internet Videos [27.49510986378025]
We propose leveraging the next best thing as real-world experience: internet videos of humans using their hands.
Visual priors, such as visual features, are often learned from videos, but more information from videos can be utilized as a stronger prior.
We build a learning algorithm, VideoDex, that leverages visual, action, and physical priors from human video datasets to guide robot behavior.
arXiv Detail & Related papers (2022-12-08T18:59:59Z) - HERD: Continuous Human-to-Robot Evolution for Learning from Human
Demonstration [57.045140028275036]
We show that manipulation skills can be transferred from a human to a robot through the use of micro-evolutionary reinforcement learning.
We propose an algorithm for multi-dimensional evolution path searching that allows joint optimization of both the robot evolution path and the policy.
arXiv Detail & Related papers (2022-12-08T15:56:13Z) - From One Hand to Multiple Hands: Imitation Learning for Dexterous
Manipulation from Single-Camera Teleoperation [26.738893736520364]
We introduce a novel single-camera teleoperation system to collect the 3D demonstrations efficiently with only an iPad and a computer.
We construct a customized robot hand for each user in the physical simulator, which is a manipulator resembling the same kinematics structure and shape of the operator's hand.
With imitation learning using our data, we show large improvement over baselines with multiple complex manipulation tasks.
arXiv Detail & Related papers (2022-04-26T17:59:51Z) - DexMV: Imitation Learning for Dexterous Manipulation from Human Videos [11.470141313103465]
We propose a new platform and pipeline, DexMV, for imitation learning to bridge the gap between computer vision and robot learning.
We design a platform with: (i) a simulation system for complex dexterous manipulation tasks with a multi-finger robot hand and (ii) a computer vision system to record large-scale demonstrations of a human hand conducting the same tasks.
We show that the demonstrations can indeed improve robot learning by a large margin and solve the complex tasks which reinforcement learning alone cannot solve.
arXiv Detail & Related papers (2021-08-12T17:51:18Z) - Learning Dexterous Grasping with Object-Centric Visual Affordances [86.49357517864937]
Dexterous robotic hands are appealing for their agility and human-like morphology.
We introduce an approach for learning dexterous grasping.
Our key idea is to embed an object-centric visual affordance model within a deep reinforcement learning loop.
arXiv Detail & Related papers (2020-09-03T04:00:40Z) - Human Grasp Classification for Reactive Human-to-Robot Handovers [50.91803283297065]
We propose an approach for human-to-robot handovers in which the robot meets the human halfway.
We collect a human grasp dataset which covers typical ways of holding objects with various hand shapes and poses.
We present a planning and execution approach that takes the object from the human hand according to the detected grasp and hand position.
arXiv Detail & Related papers (2020-03-12T19:58:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.