RPT++: Customized Feature Representation for Siamese Visual Tracking
- URL: http://arxiv.org/abs/2110.12194v1
- Date: Sat, 23 Oct 2021 10:58:57 GMT
- Title: RPT++: Customized Feature Representation for Siamese Visual Tracking
- Authors: Ziang Ma, Haitao Zhang, Linyuan Wang and Jun Yin
- Abstract summary: We argue that the performance gain of visual tracking is limited since features extracted from the salient area provide more recognizable visual patterns for classification.
We propose two customized feature extractors, named polar pooling and extreme pooling to capture task-specific visual patterns.
We demonstrate the effectiveness of the task-specific feature representation by integrating it into the recent and advanced tracker RPT.
- Score: 16.305972000224358
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: While recent years have witnessed remarkable progress in the feature
representation of visual tracking, the problem of feature misalignment between
the classification and regression tasks is largely overlooked. The approaches
of feature extraction make no difference for these two tasks in most of
advanced trackers. We argue that the performance gain of visual tracking is
limited since features extracted from the salient area provide more
recognizable visual patterns for classification, while these around the
boundaries contribute to accurately estimating the target state.
We address this problem by proposing two customized feature extractors, named
polar pooling and extreme pooling to capture task-specific visual patterns.
Polar pooling plays the role of enriching information collected from the
semantic keypoints for stronger classification, while extreme pooling
facilitates explicit visual patterns of the object boundary for accurate target
state estimation. We demonstrate the effectiveness of the task-specific feature
representation by integrating it into the recent and advanced tracker RPT.
Extensive experiments on several benchmarks show that our Customized Features
based RPT (RPT++) achieves new state-of-the-art performances on OTB-100,
VOT2018, VOT2019, GOT-10k, TrackingNet and LaSOT.
Related papers
- LEAP-VO: Long-term Effective Any Point Tracking for Visual Odometry [52.131996528655094]
We present the Long-term Effective Any Point Tracking (LEAP) module.
LEAP innovatively combines visual, inter-track, and temporal cues with mindfully selected anchors for dynamic track estimation.
Based on these traits, we develop LEAP-VO, a robust visual odometry system adept at handling occlusions and dynamic scenes.
arXiv Detail & Related papers (2024-01-03T18:57:27Z) - AVisT: A Benchmark for Visual Object Tracking in Adverse Visibility [125.77396380698639]
AVisT is a benchmark for visual tracking in diverse scenarios with adverse visibility.
AVisT comprises 120 challenging sequences with 80k annotated frames, spanning 18 diverse scenarios.
We benchmark 17 popular and recent trackers on AVisT with detailed analysis of their tracking performance across attributes.
arXiv Detail & Related papers (2022-08-14T17:49:37Z) - Visible-Thermal UAV Tracking: A Large-Scale Benchmark and New Baseline [80.13652104204691]
In this paper, we construct a large-scale benchmark with high diversity for visible-thermal UAV tracking (VTUAV)
We provide a coarse-to-fine attribute annotation, where frame-level attributes are provided to exploit the potential of challenge-specific trackers.
In addition, we design a new RGB-T baseline, named Hierarchical Multi-modal Fusion Tracker (HMFT), which fuses RGB-T data in various levels.
arXiv Detail & Related papers (2022-04-08T15:22:33Z) - Correlation-Aware Deep Tracking [83.51092789908677]
We propose a novel target-dependent feature network inspired by the self-/cross-attention scheme.
Our network deeply embeds cross-image feature correlation in multiple layers of the feature network.
Our model can be flexibly pre-trained on abundant unpaired images, leading to notably faster convergence than the existing methods.
arXiv Detail & Related papers (2022-03-03T11:53:54Z) - Learning Dynamic Compact Memory Embedding for Deformable Visual Object
Tracking [82.34356879078955]
We propose a compact memory embedding to enhance the discrimination of the segmentation-based deformable visual tracking method.
Our method outperforms the excellent segmentation-based trackers, i.e., D3S and SiamMask on DAVIS 2017 benchmark.
arXiv Detail & Related papers (2021-11-23T03:07:12Z) - Coarse-to-Fine Object Tracking Using Deep Features and Correlation
Filters [2.3526458707956643]
This paper presents a novel deep learning tracking algorithm.
We exploit the generalization ability of deep features to coarsely estimate target translation.
Then, we capitalize on the discriminative power of correlation filters to precisely localize the tracked object.
arXiv Detail & Related papers (2020-12-23T16:43:21Z) - RPT: Learning Point Set Representation for Siamese Visual Tracking [15.04182251944942]
We propose an effcient visual tracking framework to accurately estimate the target state with a finer representation as a set of representative points.
Our method achieves new state-of-the-art performance while running at over 20 FPS.
arXiv Detail & Related papers (2020-08-08T07:42:58Z) - Robust Visual Object Tracking with Two-Stream Residual Convolutional
Networks [62.836429958476735]
We propose a Two-Stream Residual Convolutional Network (TS-RCN) for visual tracking.
Our TS-RCN can be integrated with existing deep learning based visual trackers.
To further improve the tracking performance, we adopt a "wider" residual network ResNeXt as its feature extraction backbone.
arXiv Detail & Related papers (2020-05-13T19:05:42Z) - Rethinking Convolutional Features in Correlation Filter Based Tracking [0.0]
We revisit a hierarchical deep feature-based visual tracker and find that both the performance and efficiency of the deep tracker are limited by the poor feature quality.
After removing redundant features, our proposed tracker achieves significant improvements in both performance and efficiency.
arXiv Detail & Related papers (2019-12-30T04:39:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.