TeleOpBench: A Simulator-Centric Benchmark for Dual-Arm Dexterous Teleoperation
- URL: http://arxiv.org/abs/2505.12748v2
- Date: Mon, 15 Sep 2025 08:42:29 GMT
- Title: TeleOpBench: A Simulator-Centric Benchmark for Dual-Arm Dexterous Teleoperation
- Authors: Hangyu Li, Qin Zhao, Haoran Xu, Xinyu Jiang, Qingwei Ben, Feiyu Jia, Haoyu Zhao, Liang Xu, Jia Zeng, Hanqing Wang, Bo Dai, Junting Dong, Jiangmiao Pang,
- Abstract summary: We introduce TeleOpBench, a simulator-centric benchmark tailored to bimanual dexterous teleoperation.<n>Within this benchmark we implement four representative teleoperation modalities-(i) MoCap, (ii) VR device, (iii) arm-hand exoskeletons, and (iv) monocular vision tracking.
- Score: 50.261933845325636
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Teleoperation is a cornerstone of embodied-robot learning, and bimanual dexterous teleoperation in particular provides rich demonstrations that are difficult to obtain with fully autonomous systems. While recent studies have proposed diverse hardware pipelines-ranging from inertial motion-capture gloves to exoskeletons and vision-based interfaces-there is still no unified benchmark that enables fair, reproducible comparison of these systems. In this paper, we introduce TeleOpBench, a simulator-centric benchmark tailored to bimanual dexterous teleoperation. TeleOpBench contains 30 high-fidelity task environments that span pick-and-place, tool use, and collaborative manipulation, covering a broad spectrum of kinematic and force-interaction difficulty. Within this benchmark we implement four representative teleoperation modalities-(i) MoCap, (ii) VR device, (iii) arm-hand exoskeletons, and (iv) monocular vision tracking-and evaluate them with a common protocol and metric suite. To validate that performance in simulation is predictive of real-world behavior, we conduct mirrored experiments on a physical dual-arm platform equipped with two 6-DoF dexterous hands. Across 10 held-out tasks we observe a strong correlation between simulator and hardware performance, confirming the external validity of TeleOpBench. TeleOpBench establishes a common yardstick for teleoperation research and provides an extensible platform for future algorithmic and hardware innovation. Codes is now available at https://github.com/cyjdlhy/TeleOpBench .
Related papers
- Isaac Lab: A GPU-Accelerated Simulation Framework for Multi-Modal Robot Learning [72.43357471969564]
Isaac Lab combines high-fidelity GPU parallel physics, rendering, and a modular, composable architecture for designing environments and training robot policies.<n>We highlight its application to a diverse set of challenges, including whole-body control, cross-embodiment mobility, contact-rich and dexterous manipulation, and the integration of human demonstrations for skill acquisition.<n>We believe Isaac Lab's combination of advanced simulation capabilities, rich sensing, and data-center scale execution will help unlock the next generation of breakthroughs in robotics research.
arXiv Detail & Related papers (2025-11-06T21:43:02Z) - VT-Refine: Learning Bimanual Assembly with Visuo-Tactile Feedback via Simulation Fine-Tuning [39.49846628626501]
Humans excel at bimanual assembly tasks by adapting to rich tactile feedback.<n>We present VT-Refine, a visuo-tactile policy learning framework that combines real-world demonstrations, high-fidelity tactile simulation, and reinforcement learning.
arXiv Detail & Related papers (2025-10-16T17:41:36Z) - The Role of Embodiment in Intuitive Whole-Body Teleoperation for Mobile Manipulation [20.65893345441958]
A strong sense of embodiment combined with minimal physical and cognitive demands helps maintain data quality over extended periods.<n>We evaluate two visual feedback mechanisms: immersive virtual reality and conventional screen-based visualization of the robot's field of view.<n>Our results show that the use of VR as a feedback modality increases task completion time, cognitive workload, and perceived effort of the teleoperator.
arXiv Detail & Related papers (2025-09-03T11:25:36Z) - XRoboToolkit: A Cross-Platform Framework for Robot Teleoperation [1.0522824606408765]
XRoboToolkit is a cross-platform framework for extended reality based robot teleoperation built on the OpenXR standard.<n>System features low-latency stereoscopic visual feedback, optimization-based inverse kinematics, and support for diverse tracking modalities.<n>We demonstrate the framework's effectiveness through precision manipulation tasks and validate data quality by training VLA models that exhibit robust autonomous performance.
arXiv Detail & Related papers (2025-07-31T18:45:13Z) - Casper: Inferring Diverse Intents for Assistive Teleoperation with Vision Language Models [50.19518681574399]
A central challenge in real-world assistive teleoperation is for the robot to infer a wide range of human intentions from user control inputs.<n>We introduce Casper, an assistive teleoperation system that leverages commonsense knowledge embedded in pre-trained visual language models.<n>We show that Casper improves task performance, reduces human cognitive load, and achieves higher user satisfaction than direct teleoperation and assistive teleoperation baselines.
arXiv Detail & Related papers (2025-06-17T17:06:43Z) - Open-TeleVision: Teleoperation with Immersive Active Visual Feedback [17.505318269362512]
Open-TeleVision allows operators to actively perceive the robot's surroundings in a stereoscopic manner.
The system mirrors the operator's arm and hand movements on the robot, creating an immersive experience.
We validate the effectiveness of our system by collecting data and training imitation learning policies on four long-horizon, precise tasks.
arXiv Detail & Related papers (2024-07-01T17:55:35Z) - Learning Visuotactile Skills with Two Multifingered Hands [80.99370364907278]
We explore learning from human demonstrations using a bimanual system with multifingered hands and visuotactile data.
Our results mark a promising step forward in bimanual multifingered manipulation from visuotactile data.
arXiv Detail & Related papers (2024-04-25T17:59:41Z) - AnyTeleop: A General Vision-Based Dexterous Robot Arm-Hand Teleoperation System [51.48191418148764]
Vision-based teleoperation can endow robots with human-level intelligence to interact with the environment.
Current vision-based teleoperation systems are designed and engineered towards a particular robot model and deploy environment.
We propose AnyTeleop, a unified and general teleoperation system to support multiple different arms, hands, realities, and camera configurations within a single system.
arXiv Detail & Related papers (2023-07-10T14:11:07Z) - Orbit: A Unified Simulation Framework for Interactive Robot Learning
Environments [38.23943905182543]
We present Orbit, a unified and modular framework for robot learning powered by NVIDIA Isaac Sim.
It offers a modular design to create robotic environments with photo-realistic scenes and high-fidelity rigid and deformable body simulation.
We aim to support various research areas, including representation learning, reinforcement learning, imitation learning, and task and motion planning.
arXiv Detail & Related papers (2023-01-10T20:19:17Z) - DeXtreme: Transfer of Agile In-hand Manipulation from Simulation to
Reality [64.51295032956118]
We train a policy that can perform robust dexterous manipulation on an anthropomorphic robot hand.
Our work reaffirms the possibilities of sim-to-real transfer for dexterous manipulation in diverse kinds of hardware and simulator setups.
arXiv Detail & Related papers (2022-10-25T01:51:36Z) - Visual Imitation Made Easy [102.36509665008732]
We present an alternate interface for imitation that simplifies the data collection process while allowing for easy transfer to robots.
We use commercially available reacher-grabber assistive tools both as a data collection device and as the robot's end-effector.
We experimentally evaluate on two challenging tasks: non-prehensile pushing and prehensile stacking, with 1000 diverse demonstrations for each task.
arXiv Detail & Related papers (2020-08-11T17:58:50Z) - A Mobile Robot Hand-Arm Teleoperation System by Vision and IMU [25.451864296962288]
We present a novel vision-based hand pose regression network (Transteleop) and an IMU-based arm tracking method.
Transteleop observes the human hand through a low-cost depth camera and generates depth images of paired robot hand poses.
A wearable camera holder enables simultaneous hand-arm control and facilitates the mobility of the whole teleoperation system.
arXiv Detail & Related papers (2020-03-11T10:57:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.