Adaptive Siamese Tracking with a Compact Latent Network
- URL: http://arxiv.org/abs/2302.00930v2
- Date: Wed, 14 Jun 2023 09:59:34 GMT
- Title: Adaptive Siamese Tracking with a Compact Latent Network
- Authors: Xingping Dong, Jianbing Shen, Fatih Porikli, Jiebo Luo, and Ling Shao
- Abstract summary: We present an intuitive viewing to simplify the Siamese-based trackers by converting the tracking task to a classification.
Under this viewing, we perform an in-depth analysis for them through visual simulations and real tracking examples.
We apply it to adjust three classical Siamese-based trackers, namely SiamRPN++, SiamFC, and SiamBAN.
- Score: 219.38172719948048
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In this paper, we provide an intuitive viewing to simplify the Siamese-based
trackers by converting the tracking task to a classification. Under this
viewing, we perform an in-depth analysis for them through visual simulations
and real tracking examples, and find that the failure cases in some challenging
situations can be regarded as the issue of missing decisive samples in offline
training. Since the samples in the initial (first) frame contain rich
sequence-specific information, we can regard them as the decisive samples to
represent the whole sequence. To quickly adapt the base model to new scenes, a
compact latent network is presented via fully using these decisive samples.
Specifically, we present a statistics-based compact latent feature for fast
adjustment by efficiently extracting the sequence-specific information.
Furthermore, a new diverse sample mining strategy is designed for training to
further improve the discrimination ability of the proposed compact latent
network. Finally, a conditional updating strategy is proposed to efficiently
update the basic models to handle scene variation during the tracking phase. To
evaluate the generalization ability and effectiveness and of our method, we
apply it to adjust three classical Siamese-based trackers, namely SiamRPN++,
SiamFC, and SiamBAN. Extensive experimental results on six recent datasets
demonstrate that all three adjusted trackers obtain the superior performance in
terms of the accuracy, while having high running speed.
Related papers
- Improving Network Interpretability via Explanation Consistency Evaluation [56.14036428778861]
We propose a framework that acquires more explainable activation heatmaps and simultaneously increase the model performance.
Specifically, our framework introduces a new metric, i.e., explanation consistency, to reweight the training samples adaptively in model learning.
Our framework then promotes the model learning by paying closer attention to those training samples with a high difference in explanations.
arXiv Detail & Related papers (2024-08-08T17:20:08Z) - Boosting Low-Data Instance Segmentation by Unsupervised Pre-training
with Saliency Prompt [103.58323875748427]
This work offers a novel unsupervised pre-training solution for low-data regimes.
Inspired by the recent success of the Prompting technique, we introduce a new pre-training method that boosts QEIS models.
Experimental results show that our method significantly boosts several QEIS models on three datasets.
arXiv Detail & Related papers (2023-02-02T15:49:03Z) - Dynamic Template Selection Through Change Detection for Adaptive Siamese
Tracking [7.662745552551165]
Single object tracking (SOT) remains a challenging task in real-world application due to changes and deformations in a target object's appearance.
We propose a new method for dynamic sample selection and memory replay, preventing template corruption.
Our proposed method can be integrated into any object tracking algorithm that leverages online learning for model adaptation.
arXiv Detail & Related papers (2022-03-07T07:27:02Z) - An Informative Tracking Benchmark [133.0931262969931]
We develop a small and informative tracking benchmark (ITB) with 7% out of 1.2 M frames of existing and newly collected datasets.
We select the most informative sequences from existing benchmarks taking into account 1) challenging level, 2) discriminative strength, 3) and density of appearance variations.
By analyzing the results of 15 state-of-the-art trackers re-trained on the same data, we determine the effective methods for robust tracking under each scenario.
arXiv Detail & Related papers (2021-12-13T07:56:16Z) - Video Annotation for Visual Tracking via Selection and Refinement [74.08109740917122]
We present a new framework to facilitate bounding box annotations for video sequences.
A temporal assessment network is proposed which is able to capture the temporal coherence of target locations.
A visual-geometry refinement network is also designed to further enhance the selected tracking results.
arXiv Detail & Related papers (2021-08-09T05:56:47Z) - Simpler Certified Radius Maximization by Propagating Covariances [39.851641822878996]
We show an algorithm for maximizing the certified radius on datasets including Cifar-10, ImageNet, and Places365.
We show how satisfying these criteria yields an algorithm for maximizing the certified radius on datasets with moderate depth, with a small compromise in overall accuracy.
arXiv Detail & Related papers (2021-04-13T01:38:36Z) - Dynamic Scale Training for Object Detection [111.33112051962514]
We propose a Dynamic Scale Training paradigm (abbreviated as DST) to mitigate scale variation challenge in object detection.
Experimental results demonstrate the efficacy of our proposed DST towards scale variation handling.
It does not introduce inference overhead and could serve as a free lunch for general detection configurations.
arXiv Detail & Related papers (2020-04-26T16:48:17Z) - Efficient Deep Representation Learning by Adaptive Latent Space Sampling [16.320898678521843]
Supervised deep learning requires a large amount of training samples with annotations, which are expensive and time-consuming to obtain.
We propose a novel training framework which adaptively selects informative samples that are fed to the training process.
arXiv Detail & Related papers (2020-03-19T22:17:02Z) - Object-Adaptive LSTM Network for Real-time Visual Tracking with
Adversarial Data Augmentation [31.842910084312265]
We propose a novel real-time visual tracking method, which adopts an object-adaptive LSTM network to effectively capture the video sequential dependencies and adaptively learn the object appearance variations.
Experiments on four visual tracking benchmarks demonstrate the state-of-the-art performance of our method in terms of both tracking accuracy and speed.
arXiv Detail & Related papers (2020-02-07T03:06:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.