A System for Traded Control Teleoperation of Manipulation Tasks using
Intent Prediction from Hand Gestures
- URL: http://arxiv.org/abs/2107.01829v1
- Date: Mon, 5 Jul 2021 07:37:17 GMT
- Title: A System for Traded Control Teleoperation of Manipulation Tasks using
Intent Prediction from Hand Gestures
- Authors: Yoojin Oh, Marc Toussaint, Jim Mainprice
- Abstract summary: This paper presents a teleoperation system that includes robot perception and intent prediction from hand gestures.
The perception module identifies the objects present in the robot workspace and the intent prediction module which object the user likely wants to grasp.
- Score: 20.120263332724438
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper presents a teleoperation system that includes robot perception and
intent prediction from hand gestures. The perception module identifies the
objects present in the robot workspace and the intent prediction module which
object the user likely wants to grasp. This architecture allows the approach to
rely on traded control instead of direct control: we use hand gestures to
specify the goal objects for a sequential manipulation task, the robot then
autonomously generates a grasping or a retrieving motion using trajectory
optimization. The perception module relies on the model-based tracker to
precisely track the 6D pose of the objects and makes use of a state of the art
learning-based object detection and segmentation method, to initialize the
tracker by automatically detecting objects in the scene. Goal objects are
identified from user hand gestures using a trained a multi-layer perceptron
classifier. After presenting all the components of the system and their
empirical evaluation, we present experimental results comparing our pipeline to
a direct traded control approach (i.e., one that does not use prediction) which
shows that using intent prediction allows to bring down the overall task
execution time.
Related papers
- SeMoLi: What Moves Together Belongs Together [51.72754014130369]
We tackle semi-supervised object detection based on motion cues.
Recent results suggest that motion-based clustering methods can be used to pseudo-label instances of moving objects.
We re-think this approach and suggest that both, object detection, as well as motion-inspired pseudo-labeling, can be tackled in a data-driven manner.
arXiv Detail & Related papers (2024-02-29T18:54:53Z) - Modular Neural Network Policies for Learning In-Flight Object Catching
with a Robot Hand-Arm System [55.94648383147838]
We present a modular framework designed to enable a robot hand-arm system to learn how to catch flying objects.
Our framework consists of five core modules: (i) an object state estimator that learns object trajectory prediction, (ii) a catching pose quality network that learns to score and rank object poses for catching, (iii) a reaching control policy trained to move the robot hand to pre-catch poses, and (iv) a grasping control policy trained to perform soft catching motions.
We conduct extensive evaluations of our framework in simulation for each module and the integrated system, to demonstrate high success rates of in-flight
arXiv Detail & Related papers (2023-12-21T16:20:12Z) - H-SAUR: Hypothesize, Simulate, Act, Update, and Repeat for Understanding
Object Articulations from Interactions [62.510951695174604]
"Hypothesize, Simulate, Act, Update, and Repeat" (H-SAUR) is a probabilistic generative framework that generates hypotheses about how objects articulate given input observations.
We show that the proposed model significantly outperforms the current state-of-the-art articulated object manipulation framework.
We further improve the test-time efficiency of H-SAUR by integrating a learned prior from learning-based vision models.
arXiv Detail & Related papers (2022-10-22T18:39:33Z) - Intention estimation from gaze and motion features for human-robot
shared-control object manipulation [1.128708201885454]
Shared control can help in teleoperated object manipulation by assisting with the execution of the user's intention.
An intention estimation framework is presented, which uses natural gaze and motion features to predict the current action and the target object.
arXiv Detail & Related papers (2022-08-18T07:53:19Z) - Object Manipulation via Visual Target Localization [64.05939029132394]
Training agents to manipulate objects, poses many challenges.
We propose an approach that explores the environment in search for target objects, computes their 3D coordinates once they are located, and then continues to estimate their 3D locations even when the objects are not visible.
Our evaluations show a massive 3x improvement in success rate over a model that has access to the same sensory suite.
arXiv Detail & Related papers (2022-03-15T17:59:01Z) - Task-Focused Few-Shot Object Detection for Robot Manipulation [1.8275108630751844]
We develop a manipulation method based solely on detection then introduce task-focused few-shot object detection to learn new objects and settings.
In experiments for our interactive approach to few-shot learning, we train a robot to manipulate objects directly from detection (ClickBot)
arXiv Detail & Related papers (2022-01-28T21:52:05Z) - Trajectory-based Reinforcement Learning of Non-prehensile Manipulation
Skills for Semi-Autonomous Teleoperation [18.782289957834475]
We present a semi-autonomous teleoperation framework for a pick-and-place task using an RGB-D sensor.
A trajectory-based reinforcement learning is utilized for learning the non-prehensile manipulation to rearrange the objects.
We show that the proposed method outperforms manual keyboard control in terms of the time duration for the grasping.
arXiv Detail & Related papers (2021-09-27T14:27:28Z) - Neural Network Based Lidar Gesture Recognition for Realtime Robot
Teleoperation [0.0]
We propose a novel low-complexity lidar gesture recognition system for mobile robot control.
The system is lightweight and suitable for mobile robot control with limited computing power.
The use of lidar contributes to the robustness of the system, allowing it to operate in most outdoor conditions.
arXiv Detail & Related papers (2021-09-17T00:49:31Z) - Dynamic Modeling of Hand-Object Interactions via Tactile Sensing [133.52375730875696]
In this work, we employ a high-resolution tactile glove to perform four different interactive activities on a diversified set of objects.
We build our model on a cross-modal learning framework and generate the labels using a visual processing pipeline to supervise the tactile model.
This work takes a step on dynamics modeling in hand-object interactions from dense tactile sensing.
arXiv Detail & Related papers (2021-09-09T16:04:14Z) - "What's This?" -- Learning to Segment Unknown Objects from Manipulation
Sequences [27.915309216800125]
We present a novel framework for self-supervised grasped object segmentation with a robotic manipulator.
We propose a single, end-to-end trainable architecture which jointly incorporates motion cues and semantic knowledge.
Our method neither depends on any visual registration of a kinematic robot or 3D object models, nor on precise hand-eye calibration or any additional sensor data.
arXiv Detail & Related papers (2020-11-06T10:55:28Z) - Goal-Conditioned End-to-End Visuomotor Control for Versatile Skill
Primitives [89.34229413345541]
We propose a conditioning scheme which avoids pitfalls by learning the controller and its conditioning in an end-to-end manner.
Our model predicts complex action sequences based directly on a dynamic image representation of the robot motion.
We report significant improvements in task success over representative MPC and IL baselines.
arXiv Detail & Related papers (2020-03-19T15:04:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.