Seeing Behind Objects for 3D Multi-Object Tracking in RGB-D Sequences
- URL: http://arxiv.org/abs/2012.08197v2
- Date: Wed, 16 Dec 2020 14:39:51 GMT
- Title: Seeing Behind Objects for 3D Multi-Object Tracking in RGB-D Sequences
- Authors: Norman M\"uller, Yu-Shiang Wong, Niloy J. Mitra, Angela Dai and
Matthias Nie{\ss}ner
- Abstract summary: We infer the complete geometry of objects as well as track them, for rigidly moving objects over time.
From a sequence of RGB-D frames, we detect objects in each frame and learn to predict their complete object geometry.
Experiments on both synthetic and real-world RGB-D data demonstrate that we achieve state-of-the-art performance on dynamic object tracking.
- Score: 46.65702220573459
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-object tracking from RGB-D video sequences is a challenging problem due
to the combination of changing viewpoints, motion, and occlusions over time. We
observe that having the complete geometry of objects aids in their tracking,
and thus propose to jointly infer the complete geometry of objects as well as
track them, for rigidly moving objects over time. Our key insight is that
inferring the complete geometry of the objects significantly helps in tracking.
By hallucinating unseen regions of objects, we can obtain additional
correspondences between the same instance, thus providing robust tracking even
under strong change of appearance. From a sequence of RGB-D frames, we detect
objects in each frame and learn to predict their complete object geometry as
well as a dense correspondence mapping into a canonical space. This allows us
to derive 6DoF poses for the objects in each frame, along with their
correspondence between frames, providing robust object tracking across the
RGB-D sequence. Experiments on both synthetic and real-world RGB-D data
demonstrate that we achieve state-of-the-art performance on dynamic object
tracking. Furthermore, we show that our object completion significantly helps
tracking, providing an improvement of $6.5\%$ in mean MOTA.
Related papers
- Tracking by 3D Model Estimation of Unknown Objects in Videos [122.56499878291916]
We argue that this representation is limited and instead propose to guide and improve 2D tracking with an explicit object representation.
Our representation tackles a complex long-term dense correspondence problem between all 3D points on the object for all video frames.
The proposed optimization minimizes a novel loss function to estimate the best 3D shape, texture, and 6DoF pose.
arXiv Detail & Related papers (2023-04-13T11:32:36Z) - BundleSDF: Neural 6-DoF Tracking and 3D Reconstruction of Unknown
Objects [89.2314092102403]
We present a near real-time method for 6-DoF tracking of an unknown object from a monocular RGBD video sequence.
Our method works for arbitrary rigid objects, even when visual texture is largely absent.
arXiv Detail & Related papers (2023-03-24T17:13:49Z) - SpOT: Spatiotemporal Modeling for 3D Object Tracking [68.12017780034044]
3D multi-object tracking aims to consistently identify all mobile time.
Current 3D tracking methods rely on abstracted information and limited history.
We develop a holistic representation of scenes that leverage both spatial and temporal information.
arXiv Detail & Related papers (2022-07-12T21:45:49Z) - 3D Multi-Object Tracking with Differentiable Pose Estimation [0.0]
We propose a novel approach for joint 3D multi-object tracking and reconstruction from RGB-D sequences in indoor environments.
We leverage those correspondences to inform a graph neural network to solve for the optimal, temporally-consistent 7-DoF pose trajectories of all objects.
Our method improves the accumulated MOTA score for all test sequences by 24.8% over existing state-of-the-art methods.
arXiv Detail & Related papers (2022-06-28T06:46:32Z) - Learning to Track with Object Permanence [61.36492084090744]
We introduce an end-to-end trainable approach for joint object detection and tracking.
Our model, trained jointly on synthetic and real data, outperforms the state of the art on KITTI, and MOT17 datasets.
arXiv Detail & Related papers (2021-03-26T04:43:04Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking
from View Aggregation [8.854112907350624]
3D multi-object tracking plays a vital role in autonomous navigation.
Many approaches detect objects in 2D RGB sequences for tracking, which is lack of reliability when localizing objects in 3D space.
We propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames.
arXiv Detail & Related papers (2020-11-25T16:14:40Z) - e-TLD: Event-based Framework for Dynamic Object Tracking [23.026432675020683]
This paper presents a long-term object tracking framework with a moving event camera under general tracking conditions.
The framework uses a discriminative representation for the object with online learning, and detects and re-tracks the object when it comes back into the field-of-view.
arXiv Detail & Related papers (2020-09-02T07:08:56Z) - How to track your dragon: A Multi-Attentional Framework for real-time
RGB-D 6-DOF Object Pose Tracking [35.21561169636035]
We present a novel multi-attentional convolutional architecture to tackle the problem of real-time RGB-D 6D object pose tracking.
We consider the special geometrical properties of both the object's 3D model and the pose space, and we use a more sophisticated approach for data augmentation during training.
arXiv Detail & Related papers (2020-04-21T23:00:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.