Tracking Pedestrian Heads in Dense Crowd
- URL: http://arxiv.org/abs/2103.13516v1
- Date: Wed, 24 Mar 2021 22:51:17 GMT
- Title: Tracking Pedestrian Heads in Dense Crowd
- Authors: Ramana Sundararaman, Cedric De Almeida Braga, Eric Marchand, Julien
Pettre
- Abstract summary: We propose to revitalize head tracking with Crowd of Heads dataset (CroHD)
CroHD consists of 9 sequences of 11,463 frames with over 2,276,838 heads and 5,230 tracks annotated in diverse scenes.
We also propose a new head detector, HeadHunter, which is designed for small head detection in crowded scenes.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Tracking humans in crowded video sequences is an important constituent of
visual scene understanding. Increasing crowd density challenges visibility of
humans, limiting the scalability of existing pedestrian trackers to higher
crowd densities. For that reason, we propose to revitalize head tracking with
Crowd of Heads Dataset (CroHD), consisting of 9 sequences of 11,463 frames with
over 2,276,838 heads and 5,230 tracks annotated in diverse scenes. For
evaluation, we proposed a new metric, IDEucl, to measure an algorithm's
efficacy in preserving a unique identity for the longest stretch in image
coordinate space, thus building a correspondence between pedestrian crowd
motion and the performance of a tracking algorithm. Moreover, we also propose a
new head detector, HeadHunter, which is designed for small head detection in
crowded scenes. We extend HeadHunter with a Particle Filter and a color
histogram based re-identification module for head tracking. To establish this
as a strong baseline, we compare our tracker with existing state-of-the-art
pedestrian trackers on CroHD and demonstrate superiority, especially in
identity preserving tracking metrics. With a light-weight head detector and a
tracker which is efficient at identity preservation, we believe our
contributions will serve useful in advancement of pedestrian tracking in dense
crowds.
Related papers
- DynamicTrack: Advancing Gigapixel Tracking in Crowded Scenes [29.98165509387273]
We introduce DynamicTrack, a dynamic tracking framework designed to address gigapixel tracking challenges in crowded scenes.
In particular, we propose a dynamic detector that utilizes contrastive learning to jointly detect the head and body of pedestrians.
arXiv Detail & Related papers (2024-07-26T10:08:01Z) - DenseTrack: Drone-based Crowd Tracking via Density-aware Motion-appearance Synergy [33.57923199717605]
Drone-based crowd tracking faces difficulties in accurately identifying and monitoring objects from an aerial perspective.
To address these challenges, we present the Density-aware Tracking (DenseTrack) framework.
DenseTrack capitalizes on crowd counting to precisely determine object locations, blending visual and motion cues to improve the tracking of small-scale objects.
arXiv Detail & Related papers (2024-07-24T13:39:07Z) - OneTracker: Unifying Visual Object Tracking with Foundation Models and Efficient Tuning [33.521077115333696]
We present a general framework to unify various tracking tasks, termed as OneTracker.
OneTracker first performs a large-scale pre-training on a RGB tracker called Foundation Tracker.
Then we regard other modality information as prompt and build Prompt Tracker upon Foundation Tracker.
arXiv Detail & Related papers (2024-03-14T17:59:13Z) - CoTracker: It is Better to Track Together [70.63040730154984]
CoTracker is a transformer-based model that tracks a large number of 2D points in long video sequences.
We show that joint tracking significantly improves tracking accuracy and robustness, and allows CoTracker to track occluded points and points outside of the camera view.
arXiv Detail & Related papers (2023-07-14T21:13:04Z) - Handling Heavy Occlusion in Dense Crowd Tracking by Focusing on the
Heads [29.80438304958294]
In this work, we have designed a joint head and body detector in an anchor-free style to boost the detection recall and precision performance of pedestrians.
Our model does not require information on the statistical head-body ratio for common pedestrians detection for training.
We evaluate the model with extensive experiments on different datasets, including MOT20, Crowdhuman, and HT21 datasets.
arXiv Detail & Related papers (2023-04-16T06:00:35Z) - Unsupervised Learning of Accurate Siamese Tracking [68.58171095173056]
We present a novel unsupervised tracking framework, in which we can learn temporal correspondence both on the classification branch and regression branch.
Our tracker outperforms preceding unsupervised methods by a substantial margin, performing on par with supervised methods on large-scale datasets such as TrackingNet and LaSOT.
arXiv Detail & Related papers (2022-04-04T13:39:43Z) - STCrowd: A Multimodal Dataset for Pedestrian Perception in Crowded
Scenes [78.95447086305381]
Accurately detecting and tracking pedestrians in 3D space is challenging due to large variations in rotations, poses and scales.
Existing benchmarks either only provide 2D annotations, or have limited 3D annotations with low-density pedestrian distribution.
We introduce a large-scale multimodal dataset, STCrowd, to better evaluate pedestrian perception algorithms in crowded scenarios.
arXiv Detail & Related papers (2022-04-03T08:26:07Z) - Tracking-by-Counting: Using Network Flows on Crowd Density Maps for
Tracking Multiple Targets [96.98888948518815]
State-of-the-art multi-object tracking(MOT) methods follow the tracking-by-detection paradigm.
We propose a new MOT paradigm, tracking-by-counting, tailored for crowded scenes.
arXiv Detail & Related papers (2020-07-18T19:51:53Z) - TAO: A Large-Scale Benchmark for Tracking Any Object [95.87310116010185]
Tracking Any Object dataset consists of 2,907 high resolution videos, captured in diverse environments, which are half a minute long on average.
We ask annotators to label objects that move at any point in the video, and give names to them post factum.
Our vocabulary is both significantly larger and qualitatively different from existing tracking datasets.
arXiv Detail & Related papers (2020-05-20T21:07:28Z) - MOT20: A benchmark for multi object tracking in crowded scenes [73.92443841487503]
We present our MOT20benchmark, consisting of 8 new sequences depicting very crowded challenging scenes.
The benchmark was presented first at the 4thBMTT MOT Challenge Workshop at the Computer Vision and Pattern Recognition Conference (CVPR)
arXiv Detail & Related papers (2020-03-19T20:08:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.