Enhancing Generalizable 6D Pose Tracking of an In-Hand Object with
Tactile Sensing
- URL: http://arxiv.org/abs/2210.04026v2
- Date: Sat, 23 Dec 2023 14:54:29 GMT
- Title: Enhancing Generalizable 6D Pose Tracking of an In-Hand Object with
Tactile Sensing
- Authors: Yun Liu, Xiaomeng Xu, Weihang Chen, Haocheng Yuan, He Wang, Jing Xu,
Rui Chen, Li Yi
- Abstract summary: TEG-Track is a tactile-enhanced 6D pose tracking system.
It can track previously unseen objects held in hand.
Results demonstrate that TEG-Track consistently enhances state-of-the-art generalizable 6D pose trackers.
- Score: 31.49529551069215
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: When manipulating an object to accomplish complex tasks, humans rely on both
vision and touch to keep track of the object's 6D pose. However, most existing
object pose tracking systems in robotics rely exclusively on visual signals,
which hinder a robot's ability to manipulate objects effectively. To address
this limitation, we introduce TEG-Track, a tactile-enhanced 6D pose tracking
system that can track previously unseen objects held in hand. From consecutive
tactile signals, TEG-Track optimizes object velocities from marker flows when
slippage does not occur, or regresses velocities using a slippage estimation
network when slippage is detected. The estimated object velocities are
integrated into a geometric-kinematic optimization scheme to enhance existing
visual pose trackers. To evaluate our method and to facilitate future research,
we construct a real-world dataset for visual-tactile in-hand object pose
tracking. Experimental results demonstrate that TEG-Track consistently enhances
state-of-the-art generalizable 6D pose trackers in synthetic and real-world
scenarios. Our code and dataset are available at
https://github.com/leolyliu/TEG-Track.
Related papers
- TrackAgent: 6D Object Tracking via Reinforcement Learning [24.621588217873395]
We propose to simplify object tracking to a reinforced point cloud (depth only) alignment task.
This allows us to train a streamlined approach from scratch with limited amounts of sparse 3D point clouds.
We also show that the RL agent's uncertainty and a rendering-based mask propagation are effective reinitialization triggers.
arXiv Detail & Related papers (2023-07-28T17:03:00Z) - MotionTrack: Learning Motion Predictor for Multiple Object Tracking [68.68339102749358]
We introduce a novel motion-based tracker, MotionTrack, centered around a learnable motion predictor.
Our experimental results demonstrate that MotionTrack yields state-of-the-art performance on datasets such as Dancetrack and SportsMOT.
arXiv Detail & Related papers (2023-06-05T04:24:11Z) - Once Detected, Never Lost: Surpassing Human Performance in Offline LiDAR
based 3D Object Detection [50.959453059206446]
This paper aims for high-performance offline LiDAR-based 3D object detection.
We first observe that experienced human annotators annotate objects from a track-centric perspective.
We propose a high-performance offline detector in a track-centric perspective instead of the conventional object-centric perspective.
arXiv Detail & Related papers (2023-04-24T17:59:05Z) - ROFT: Real-Time Optical Flow-Aided 6D Object Pose and Velocity Tracking [7.617467911329272]
We introduce ROFT, a Kalman filtering approach for 6D object pose and velocity tracking from a stream of RGB-D images.
By leveraging real-time optical flow, ROFT synchronizes delayed outputs of low frame rate Convolutional Neural Networks for instance segmentation and 6D object pose estimation.
Results demonstrate that our approach outperforms state-of-the-art methods for 6D object pose tracking, while also providing 6D object velocity tracking.
arXiv Detail & Related papers (2021-11-06T07:30:00Z) - EagerMOT: 3D Multi-Object Tracking via Sensor Fusion [68.8204255655161]
Multi-object tracking (MOT) enables mobile robots to perform well-informed motion planning and navigation by localizing surrounding objects in 3D space and time.
Existing methods rely on depth sensors (e.g., LiDAR) to detect and track targets in 3D space, but only up to a limited sensing range due to the sparsity of the signal.
We propose EagerMOT, a simple tracking formulation that integrates all available object observations from both sensor modalities to obtain a well-informed interpretation of the scene dynamics.
arXiv Detail & Related papers (2021-04-29T22:30:29Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - Learning to Track with Object Permanence [61.36492084090744]
We introduce an end-to-end trainable approach for joint object detection and tracking.
Our model, trained jointly on synthetic and real data, outperforms the state of the art on KITTI, and MOT17 datasets.
arXiv Detail & Related papers (2021-03-26T04:43:04Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - Seeing Behind Objects for 3D Multi-Object Tracking in RGB-D Sequences [46.65702220573459]
We infer the complete geometry of objects as well as track them, for rigidly moving objects over time.
From a sequence of RGB-D frames, we detect objects in each frame and learn to predict their complete object geometry.
Experiments on both synthetic and real-world RGB-D data demonstrate that we achieve state-of-the-art performance on dynamic object tracking.
arXiv Detail & Related papers (2020-12-15T10:33:21Z) - se(3)-TrackNet: Data-driven 6D Pose Tracking by Calibrating Image
Residuals in Synthetic Domains [12.71983073907091]
This work proposes a data-driven optimization approach for long-term, 6D pose tracking.
It aims to identify the optimal relative pose given the current RGB-D observation and a synthetic image conditioned on the previous best estimate and the object's model.
The proposed approach achieves consistently robust estimates and outperforms alternatives, even though they have been trained with real images.
arXiv Detail & Related papers (2020-07-27T21:09:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.