Tackling Occlusion in Siamese Tracking with Structured Dropouts
- URL: http://arxiv.org/abs/2006.16571v1
- Date: Tue, 30 Jun 2020 07:09:33 GMT
- Title: Tackling Occlusion in Siamese Tracking with Structured Dropouts
- Authors: Deepak K. Gupta, Efstratios Gavves and Arnold W. M. Smeulders
- Abstract summary: Occlusion is one of the most difficult challenges in object tracking to model.
We present structured dropout to mimick the change in latent codes under occlusion.
Experiments on several tracking benchmarks show the benefits of structured dropouts.
- Score: 42.303946665229965
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Occlusion is one of the most difficult challenges in object tracking to
model. This is because unlike other challenges, where data augmentation can be
of help, occlusion is hard to simulate as the occluding object can be anything
in any shape. In this paper, we propose a simple solution to simulate the
effects of occlusion in the latent space. Specifically, we present structured
dropout to mimick the change in latent codes under occlusion. We present three
forms of dropout (channel dropout, segment dropout and slice dropout) with the
various forms of occlusion in mind. To demonstrate its effectiveness, the
dropouts are incorporated into two modern Siamese trackers (SiamFC and
SiamRPN++). The outputs from multiple dropouts are combined using an encoder
network to obtain the final prediction. Experiments on several tracking
benchmarks show the benefits of structured dropouts, while due to their
simplicity requiring only small changes to the existing tracker models.
Related papers
- Occlusion Resilient 3D Human Pose Estimation [52.49366182230432]
Occlusions remain one of the key challenges in 3D body pose estimation from single-camera video sequences.
We demonstrate the effectiveness of this approach compared to state-of-the-art techniques that infer poses from single-camera sequences.
arXiv Detail & Related papers (2024-02-16T19:29:43Z) - CharacterGAN: Few-Shot Keypoint Character Animation and Reposing [64.19520387536741]
We introduce CharacterGAN, a generative model that can be trained on only a few samples of a given character.
Our model generates novel poses based on keypoint locations, which can be modified in real time while providing interactive feedback.
We show that our approach outperforms recent baselines and creates realistic animations for diverse characters.
arXiv Detail & Related papers (2021-02-05T12:38:15Z) - Recurrent Multi-view Alignment Network for Unsupervised Surface
Registration [79.72086524370819]
Learning non-rigid registration in an end-to-end manner is challenging due to the inherent high degrees of freedom and the lack of labeled training data.
We propose to represent the non-rigid transformation with a point-wise combination of several rigid transformations.
We also introduce a differentiable loss function that measures the 3D shape similarity on the projected multi-view 2D depth images.
arXiv Detail & Related papers (2020-11-24T14:22:42Z) - Exploring Simple Siamese Representation Learning [68.37628268182185]
We show that simple Siamese networks can learn meaningful representations even using none of the following: (i) negative sample pairs, (ii) large batches, (iii) momentum encoders.
Our experiments show that collapsing solutions do exist for the loss and structure, but a stop-gradient operation plays an essential role in preventing collapsing.
arXiv Detail & Related papers (2020-11-20T18:59:33Z) - Cascaded Regression Tracking: Towards Online Hard Distractor
Discrimination [202.2562153608092]
We propose a cascaded regression tracker with two sequential stages.
In the first stage, we filter out abundant easily-identified negative candidates.
In the second stage, a discrete sampling based ridge regression is designed to double-check the remaining ambiguous hard samples.
arXiv Detail & Related papers (2020-06-18T07:48:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.