JRMOT: A Real-Time 3D Multi-Object Tracker and a New Large-Scale Dataset
- URL: http://arxiv.org/abs/2002.08397v4
- Date: Wed, 22 Jul 2020 07:29:39 GMT
- Title: JRMOT: A Real-Time 3D Multi-Object Tracker and a New Large-Scale Dataset
- Authors: Abhijeet Shenoi, Mihir Patel, JunYoung Gwak, Patrick Goebel, Amir
Sadeghian, Hamid Rezatofighi, Roberto Mart\'in-Mart\'in, Silvio Savarese
- Abstract summary: We present JRMOT, a novel 3D MOT system that integrates information from RGB images and 3D point clouds to achieve real-time tracking performance.
As part of our work, we release the JRDB dataset, a novel large scale 2D+3D dataset and benchmark.
The presented 3D MOT system demonstrates state-of-the-art performance against competing methods on the popular 2D tracking KITTI benchmark.
- Score: 34.609125601292
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Robots navigating autonomously need to perceive and track the motion of
objects and other agents in its surroundings. This information enables planning
and executing robust and safe trajectories. To facilitate these processes, the
motion should be perceived in 3D Cartesian space. However, most recent
multi-object tracking (MOT) research has focused on tracking people and moving
objects in 2D RGB video sequences. In this work we present JRMOT, a novel 3D
MOT system that integrates information from RGB images and 3D point clouds to
achieve real-time, state-of-the-art tracking performance. Our system is built
with recent neural networks for re-identification, 2D and 3D detection and
track description, combined into a joint probabilistic data-association
framework within a multi-modal recursive Kalman architecture. As part of our
work, we release the JRDB dataset, a novel large scale 2D+3D dataset and
benchmark, annotated with over 2 million boxes and 3500 time consistent 2D+3D
trajectories across 54 indoor and outdoor scenes. JRDB contains over 60 minutes
of data including 360 degree cylindrical RGB video and 3D pointclouds in social
settings that we use to develop, train and evaluate JRMOT. The presented 3D MOT
system demonstrates state-of-the-art performance against competing methods on
the popular 2D tracking KITTI benchmark and serves as first 3D tracking
solution for our benchmark. Real-robot tests on our social robot JackRabbot
indicate that the system is capable of tracking multiple pedestrians fast and
reliably. We provide the ROS code of our tracker at
https://sites.google.com/view/jrmot.
Related papers
- TAPVid-3D: A Benchmark for Tracking Any Point in 3D [63.060421798990845]
We introduce a new benchmark, TAPVid-3D, for evaluating the task of Tracking Any Point in 3D.
This benchmark will serve as a guidepost to improve our ability to understand precise 3D motion and surface deformation from monocular video.
arXiv Detail & Related papers (2024-07-08T13:28:47Z) - ByteTrackV2: 2D and 3D Multi-Object Tracking by Associating Every
Detection Box [81.45219802386444]
Multi-object tracking (MOT) aims at estimating bounding boxes and identities of objects across video frames.
We propose a hierarchical data association strategy to mine the true objects in low-score detection boxes.
In 3D scenarios, it is much easier for the tracker to predict object velocities in the world coordinate.
arXiv Detail & Related papers (2023-03-27T15:35:21Z) - A Lightweight and Detector-free 3D Single Object Tracker on Point Clouds [50.54083964183614]
It is non-trivial to perform accurate target-specific detection since the point cloud of objects in raw LiDAR scans is usually sparse and incomplete.
We propose DMT, a Detector-free Motion prediction based 3D Tracking network that totally removes the usage of complicated 3D detectors.
arXiv Detail & Related papers (2022-03-08T17:49:07Z) - 3D Visual Tracking Framework with Deep Learning for Asteroid Exploration [22.808962211830675]
In this paper, we focus on the studied accurate and real-time method for 3D tracking.
A new large-scale 3D asteroid tracking dataset is presented, including binocular video sequences, depth maps, and point clouds of diverse asteroids.
We propose a deep-learning based 3D tracking framework, named as Track3D, which involves 2D monocular tracker and a novel light-weight amodal axis-aligned bounding-box network, A3BoxNet.
arXiv Detail & Related papers (2021-11-21T04:14:45Z) - PC-DAN: Point Cloud based Deep Affinity Network for 3D Multi-Object
Tracking (Accepted as an extended abstract in JRDB-ACT Workshop at CVPR21) [68.12101204123422]
A point cloud is a dense compilation of spatial data in 3D coordinates.
We propose a PointNet-based approach for 3D Multi-Object Tracking (MOT)
arXiv Detail & Related papers (2021-06-03T05:36:39Z) - Learnable Online Graph Representations for 3D Multi-Object Tracking [156.58876381318402]
We propose a unified and learning based approach to the 3D MOT problem.
We employ a Neural Message Passing network for data association that is fully trainable.
We show the merit of the proposed approach on the publicly available nuScenes dataset by achieving state-of-the-art performance of 65.6% AMOTA and 58% fewer ID-switches.
arXiv Detail & Related papers (2021-04-23T17:59:28Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - Relation3DMOT: Exploiting Deep Affinity for 3D Multi-Object Tracking
from View Aggregation [8.854112907350624]
3D multi-object tracking plays a vital role in autonomous navigation.
Many approaches detect objects in 2D RGB sequences for tracking, which is lack of reliability when localizing objects in 3D space.
We propose a novel convolutional operation, named RelationConv, to better exploit the correlation between each pair of objects in the adjacent frames.
arXiv Detail & Related papers (2020-11-25T16:14:40Z) - F-Siamese Tracker: A Frustum-based Double Siamese Network for 3D Single
Object Tracking [12.644452175343059]
A main challenge in 3D single object tracking is how to reduce search space for generating appropriate 3D candidates.
Instead of relying on 3D proposals, we produce 2D region proposals which are then extruded into 3D viewing frustums.
We perform an online accuracy validation on the 3D frustum to generate refined point cloud searching space.
arXiv Detail & Related papers (2020-10-22T08:01:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.