Semi-automatic Data Annotation System for Multi-Target Multi-Camera
Vehicle Tracking
- URL: http://arxiv.org/abs/2209.09606v1
- Date: Tue, 20 Sep 2022 10:37:38 GMT
- Title: Semi-automatic Data Annotation System for Multi-Target Multi-Camera
Vehicle Tracking
- Authors: Haohong Liao, Silin Zheng, Xuelin Shen, Mark Junjie Li and Xu Wang
- Abstract summary: Multi-target multi-camera tracking (MTMCT) plays an important role in intelligent video analysis, surveillance video retrieval, and other application scenarios.
Deep-learning-based MTMCT has been the mainstream and has achieved fascinating improvements regarding tracking accuracy and efficiency.
This paper presents a semi-automatic data annotation system to facilitate the real-world MTMCT dataset establishment.
- Score: 5.479834571773961
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-target multi-camera tracking (MTMCT) plays an important role in
intelligent video analysis, surveillance video retrieval, and other application
scenarios. Nowadays, the deep-learning-based MTMCT has been the mainstream and
has achieved fascinating improvements regarding tracking accuracy and
efficiency. However, according to our investigation, the lacking of datasets
focusing on real-world application scenarios limits the further improvements
for current learning-based MTMCT models. Specifically, the learning-based MTMCT
models training by common datasets usually cannot achieve satisfactory results
in real-world application scenarios. Motivated by this, this paper presents a
semi-automatic data annotation system to facilitate the real-world MTMCT
dataset establishment. The proposed system first employs a deep-learning-based
single-camera trajectory generation method to automatically extract
trajectories from surveillance videos. Subsequently, the system provides a
recommendation list in the following manual cross-camera trajectory matching
process. The recommendation list is generated based on side information,
including camera location, timestamp relation, and background scene. In the
experimental stage, extensive results further demonstrate the efficiency of the
proposed system.
Related papers
- MCTrack: A Unified 3D Multi-Object Tracking Framework for Autonomous Driving [10.399817864597347]
This paper introduces MCTrack, a new 3D multi-object tracking method that achieves state-of-the-art (SOTA) performance across KITTI, nuScenes, and datasets.
arXiv Detail & Related papers (2024-09-23T11:26:01Z) - UdeerLID+: Integrating LiDAR, Image, and Relative Depth with Semi-Supervised [12.440461420762265]
Road segmentation is a critical task for autonomous driving systems.
Our work introduces an innovative approach that integrates LiDAR point cloud data, visual image, and relative depth maps.
One of the primary challenges is the scarcity of large-scale, accurately labeled datasets.
arXiv Detail & Related papers (2024-09-10T03:57:30Z) - MTMMC: A Large-Scale Real-World Multi-Modal Camera Tracking Benchmark [63.878793340338035]
Multi-target multi-camera tracking is a crucial task that involves identifying and tracking individuals over time using video streams from multiple cameras.
Existing datasets for this task are either synthetically generated or artificially constructed within a controlled camera network setting.
We present MTMMC, a real-world, large-scale dataset that includes long video sequences captured by 16 multi-modal cameras in two different environments.
arXiv Detail & Related papers (2024-03-29T15:08:37Z) - Self-Supervised Representation Learning from Temporal Ordering of
Automated Driving Sequences [49.91741677556553]
We propose TempO, a temporal ordering pretext task for pre-training region-level feature representations for perception tasks.
We embed each frame by an unordered set of proposal feature vectors, a representation that is natural for object detection or tracking systems.
Extensive evaluations on the BDD100K, nuImages, and MOT17 datasets show that our TempO pre-training approach outperforms single-frame self-supervised learning methods.
arXiv Detail & Related papers (2023-02-17T18:18:27Z) - Unifying Tracking and Image-Video Object Detection [54.91658924277527]
TrIVD (Tracking and Image-Video Detection) is the first framework that unifies image OD, video OD, and MOT within one end-to-end model.
To handle the discrepancies and semantic overlaps of category labels, TrIVD formulates detection/tracking as grounding and reasons about object categories.
arXiv Detail & Related papers (2022-11-20T20:30:28Z) - End-to-end Tracking with a Multi-query Transformer [96.13468602635082]
Multiple-object tracking (MOT) is a challenging task that requires simultaneous reasoning about location, appearance, and identity of the objects in the scene over time.
Our aim in this paper is to move beyond tracking-by-detection approaches, to class-agnostic tracking that performs well also for unknown object classes.
arXiv Detail & Related papers (2022-10-26T10:19:37Z) - Know Your Surroundings: Panoramic Multi-Object Tracking by Multimodality
Collaboration [56.01625477187448]
We propose a MultiModality PAnoramic multi-object Tracking framework (MMPAT)
It takes both 2D panorama images and 3D point clouds as input and then infers target trajectories using the multimodality data.
We evaluate the proposed method on the JRDB dataset, where the MMPAT achieves the top performance in both the detection and tracking tasks.
arXiv Detail & Related papers (2021-05-31T03:16:38Z) - Multi-Target Multi-Camera Tracking of Vehicles using Metadata-Aided
Re-ID and Trajectory-Based Camera Link Model [32.01329933787149]
We propose a novel framework for multi-target multi-camera tracking of vehicles based on metadata-aided re-identification (MA-ReID) and the trajectory-based camera link model (TCLM)
The proposed method is evaluated on the CityFlow dataset, achieving IDF1 76.77%, which outperforms the state-of-the-art MTMCT methods.
arXiv Detail & Related papers (2021-05-03T23:20:37Z) - Traffic-Aware Multi-Camera Tracking of Vehicles Based on ReID and Camera
Link Model [43.850588717944916]
Multi-target multi-camera tracking (MTMCT) is a crucial technique for smart city applications.
We propose an effective and reliable MTMCT framework for vehicles.
Our proposed MTMCT is evaluated on the CityFlow dataset and achieves a new state-of-the-art performance with IDF1 of 74.93%.
arXiv Detail & Related papers (2020-08-22T08:54:47Z) - Auto-Rectify Network for Unsupervised Indoor Depth Estimation [119.82412041164372]
We establish that the complex ego-motions exhibited in handheld settings are a critical obstacle for learning depth.
We propose a data pre-processing method that rectifies training images by removing their relative rotations for effective learning.
Our results outperform the previous unsupervised SOTA method by a large margin on the challenging NYUv2 dataset.
arXiv Detail & Related papers (2020-06-04T08:59:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.