MMPTRACK: Large-scale Densely Annotated Multi-camera Multiple People
Tracking Benchmark
- URL: http://arxiv.org/abs/2111.15157v1
- Date: Tue, 30 Nov 2021 06:29:14 GMT
- Title: MMPTRACK: Large-scale Densely Annotated Multi-camera Multiple People
Tracking Benchmark
- Authors: Xiaotian Han, Quanzeng You, Chunyu Wang, Zhizheng Zhang, Peng Chu,
Houdong Hu, Jiang Wang, Zicheng Liu
- Abstract summary: We provide a large-scale densely-labeled multi-camera tracking dataset in five different environments with the help of an auto-annotation system.
The 3D tracking results are projected to each RGB camera view using camera parameters to create 2D tracking results.
This dataset provides a more reliable benchmark of multi-camera, multi-object tracking systems in cluttered and crowded environments.
- Score: 40.363608495563305
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Multi-camera tracking systems are gaining popularity in applications that
demand high-quality tracking results, such as frictionless checkout because
monocular multi-object tracking (MOT) systems often fail in cluttered and
crowded environments due to occlusion. Multiple highly overlapped cameras can
significantly alleviate the problem by recovering partial 3D information.
However, the cost of creating a high-quality multi-camera tracking dataset with
diverse camera settings and backgrounds has limited the dataset scale in this
domain. In this paper, we provide a large-scale densely-labeled multi-camera
tracking dataset in five different environments with the help of an
auto-annotation system. The system uses overlapped and calibrated depth and RGB
cameras to build a high-performance 3D tracker that automatically generates the
3D tracking results. The 3D tracking results are projected to each RGB camera
view using camera parameters to create 2D tracking results. Then, we manually
check and correct the 3D tracking results to ensure the label quality, which is
much cheaper than fully manual annotation. We have conducted extensive
experiments using two real-time multi-camera trackers and a person
re-identification (ReID) model with different settings. This dataset provides a
more reliable benchmark of multi-camera, multi-object tracking systems in
cluttered and crowded environments. Also, our results demonstrate that adapting
the trackers and ReID models on this dataset significantly improves their
performance. Our dataset will be publicly released upon the acceptance of this
work.
Related papers
- RockTrack: A 3D Robust Multi-Camera-Ken Multi-Object Tracking Framework [28.359633046753228]
We propose RockTrack, a 3D MOT method for multi-camera detectors.
RockTrack incorporates a confidence-guided preprocessing module to extract reliable motion and image observations.
RockTrack achieves state-of-the-art performance on the nuScenes vision-only tracking leaderboard with 59.1% AMOTA.
arXiv Detail & Related papers (2024-09-18T07:08:08Z) - TAPVid-3D: A Benchmark for Tracking Any Point in 3D [63.060421798990845]
We introduce a new benchmark, TAPVid-3D, for evaluating the task of Tracking Any Point in 3D.
This benchmark will serve as a guidepost to improve our ability to understand precise 3D motion and surface deformation from monocular video.
arXiv Detail & Related papers (2024-07-08T13:28:47Z) - MTMMC: A Large-Scale Real-World Multi-Modal Camera Tracking Benchmark [63.878793340338035]
Multi-target multi-camera tracking is a crucial task that involves identifying and tracking individuals over time using video streams from multiple cameras.
Existing datasets for this task are either synthetically generated or artificially constructed within a controlled camera network setting.
We present MTMMC, a real-world, large-scale dataset that includes long video sequences captured by 16 multi-modal cameras in two different environments.
arXiv Detail & Related papers (2024-03-29T15:08:37Z) - A Unified Multi-view Multi-person Tracking Framework [5.143965432709092]
This study presents a Unified Multi-view Multi-person Tracking framework to bridge the gap between footprint tracking and pose tracking.
Without additional modifications, the framework can adopt monocular 2D bounding boxes and 2D poses as the input to produce robust 3D trajectories for multiple persons.
arXiv Detail & Related papers (2023-02-08T01:08:02Z) - MUTR3D: A Multi-camera Tracking Framework via 3D-to-2D Queries [18.70932813595532]
3D tracking from multiple cameras is a key component in a vision-based autonomous driving system.
We propose an end-to-end textbfMUlti-camera textbfTRacking framework called MUTR3D.
MUTR3D does not explicitly rely on the spatial and appearance similarity of objects.
It outperforms state-of-the-art methods by 5.3 AMOTA on the nuScenes dataset.
arXiv Detail & Related papers (2022-05-02T01:45:41Z) - Scalable and Real-time Multi-Camera Vehicle Detection,
Re-Identification, and Tracking [58.95210121654722]
We propose a real-time city-scale multi-camera vehicle tracking system that handles real-world, low-resolution CCTV instead of idealized and curated video streams.
Our method is ranked among the top five performers on the public leaderboard.
arXiv Detail & Related papers (2022-04-15T12:47:01Z) - LMGP: Lifted Multicut Meets Geometry Projections for Multi-Camera
Multi-Object Tracking [42.87953709286856]
Multi-Camera Multi-Object Tracking is currently drawing attention in the computer vision field due to its superior performance in real-world applications.
We propose a mathematically elegant multi-camera multiple object tracking approach based on a spatial-temporal lifted multicut formulation.
arXiv Detail & Related papers (2021-11-23T14:09:47Z) - MetaPose: Fast 3D Pose from Multiple Views without 3D Supervision [72.5863451123577]
We show how to train a neural model that can perform accurate 3D pose and camera estimation.
Our method outperforms both classical bundle adjustment and weakly-supervised monocular 3D baselines.
arXiv Detail & Related papers (2021-08-10T18:39:56Z) - Monocular Quasi-Dense 3D Object Tracking [99.51683944057191]
A reliable and accurate 3D tracking framework is essential for predicting future locations of surrounding objects and planning the observer's actions in numerous applications such as autonomous driving.
We propose a framework that can effectively associate moving objects over time and estimate their full 3D bounding box information from a sequence of 2D images captured on a moving platform.
arXiv Detail & Related papers (2021-03-12T15:30:02Z) - Real-time 3D Deep Multi-Camera Tracking [13.494550690138775]
We propose a novel end-to-end tracking pipeline, Deep Multi-Camera Tracking (DMCT), which achieves reliable real-time multi-camera people tracking.
Our system achieves the state-of-the-art tracking results while maintaining real-time performance.
arXiv Detail & Related papers (2020-03-26T06:08:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.