Kalman Filter Based Multiple Person Head Tracking
- URL: http://arxiv.org/abs/2006.06134v1
- Date: Thu, 11 Jun 2020 00:54:45 GMT
- Title: Kalman Filter Based Multiple Person Head Tracking
- Authors: Mohib Ullah, Maqsood Mahmud, Habib Ullah, Kashif Ahmad, Ali Shariq
Imran, Faouzi Alaya Cheikh
- Abstract summary: State-of-the-art approaches rely on the deep learning-based visual representation.
We come up with a simple yet effective target representation for human tracking.
- Score: 10.235946073664818
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: For multi-target tracking, target representation plays a crucial rule in
performance. State-of-the-art approaches rely on the deep learning-based visual
representation that gives an optimal performance at the cost of high
computational complexity. In this paper, we come up with a simple yet effective
target representation for human tracking. Our inspiration comes from the fact
that the human body goes through severe deformation and inter/intra occlusion
over the passage of time. So, instead of tracking the whole body part, a
relative rigid organ tracking is selected for tracking the human over an
extended period of time. Hence, we followed the tracking-by-detection paradigm
and generated the target hypothesis of only the spatial locations of heads in
every frame. After the localization of head location, a Kalman filter with a
constant velocity motion model is instantiated for each target that follows the
temporal evolution of the targets in the scene. For associating the targets in
the consecutive frames, combinatorial optimization is used that associates the
corresponding targets in a greedy fashion. Qualitative results are evaluated on
four challenging video surveillance dataset and promising results has been
achieved.
Related papers
- AlphaPose: Whole-Body Regional Multi-Person Pose Estimation and Tracking
in Real-Time [47.19339667836196]
We present AlphaPose, a system that can perform accurate whole-body pose estimation and tracking jointly while running in realtime.
We show a significant improvement over current state-of-the-art methods in both speed and accuracy on COCO-wholebody, COCO, PoseTrack, and our proposed Halpe-FullBody pose estimation dataset.
arXiv Detail & Related papers (2022-11-07T09:15:38Z) - Tracking by Joint Local and Global Search: A Target-aware Attention
based Approach [63.50045332644818]
We propose a novel target-aware attention mechanism (termed TANet) to conduct joint local and global search for robust tracking.
Specifically, we extract the features of target object patch and continuous video frames, then we track and feed them into a decoder network to generate target-aware global attention maps.
In the tracking procedure, we integrate the target-aware attention with multiple trackers by exploring candidate search regions for robust tracking.
arXiv Detail & Related papers (2021-06-09T06:54:15Z) - Learning Dynamics via Graph Neural Networks for Human Pose Estimation
and Tracking [98.91894395941766]
We propose a novel online approach to learning the pose dynamics, which are independent of pose detections in current fame.
Specifically, we derive this prediction of dynamics through a graph neural network(GNN) that explicitly accounts for both spatial-temporal and visual information.
Experiments on PoseTrack 2017 and PoseTrack 2018 datasets demonstrate that the proposed method achieves results superior to the state of the art on both human pose estimation and tracking tasks.
arXiv Detail & Related papers (2021-06-07T16:36:50Z) - TRiPOD: Human Trajectory and Pose Dynamics Forecasting in the Wild [77.59069361196404]
TRiPOD is a novel method for predicting body dynamics based on graph attentional networks.
To incorporate a real-world challenge, we learn an indicator representing whether an estimated body joint is visible/invisible at each frame.
Our evaluation shows that TRiPOD outperforms all prior work and state-of-the-art specifically designed for each of the trajectory and pose forecasting tasks.
arXiv Detail & Related papers (2021-04-08T20:01:00Z) - Learning Target Candidate Association to Keep Track of What Not to Track [100.80610986625693]
We propose to keep track of distractor objects in order to continue tracking the target.
To tackle the problem of lacking ground-truth correspondences between distractor objects in visual tracking, we propose a training strategy that combines partial annotations with self-supervision.
Our tracker sets a new state-of-the-art on six benchmarks, achieving an AUC score of 67.2% on LaSOT and a +6.1% absolute gain on the OxUvA long-term dataset.
arXiv Detail & Related papers (2021-03-30T17:58:02Z) - Dynamic Attention guided Multi-Trajectory Analysis for Single Object
Tracking [62.13213518417047]
We propose to introduce more dynamics by devising a dynamic attention-guided multi-trajectory tracking strategy.
In particular, we construct dynamic appearance model that contains multiple target templates, each of which provides its own attention for locating the target in the new frame.
After spanning the whole sequence, we introduce a multi-trajectory selection network to find the best trajectory that delivers improved tracking performance.
arXiv Detail & Related papers (2021-03-30T05:36:31Z) - Learning to Track with Object Permanence [61.36492084090744]
We introduce an end-to-end trainable approach for joint object detection and tracking.
Our model, trained jointly on synthetic and real data, outperforms the state of the art on KITTI, and MOT17 datasets.
arXiv Detail & Related papers (2021-03-26T04:43:04Z) - Blending of Learning-based Tracking and Object Detection for Monocular
Camera-based Target Following [2.578242050187029]
We present a real-time approach which fuses a generic target tracker and object detection module with a target re-identification module.
Our work focuses on improving the performance of Convolutional Recurrent Neural Network-based object trackers in cases where the object of interest belongs to the category of emphfamiliar objects.
arXiv Detail & Related papers (2020-08-21T18:44:35Z) - Applying r-spatiogram in object tracking for occlusion handling [16.36552899280708]
The aim of video tracking is to accurately locate a moving target in a video sequence and discriminate target from non-targets in the feature space of the sequence.
In this paper, we use the basic idea of many trackers which consists of three main components of the reference model, i.e. object modeling, object detection and localization, and model updating.
arXiv Detail & Related papers (2020-03-18T02:42:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.