FEAR: Fast, Efficient, Accurate and Robust Visual Tracker
- URL: http://arxiv.org/abs/2112.07957v1
- Date: Wed, 15 Dec 2021 08:28:55 GMT
- Title: FEAR: Fast, Efficient, Accurate and Robust Visual Tracker
- Authors: Vasyl Borsuk, Roman Vei, Orest Kupyn, Tetiana Martyniuk, Igor
Krashenyi, Ji\v{r}i Matas
- Abstract summary: We present FEAR, a novel, fast, efficient, accurate, and robust Siamese visual tracker.
FEAR-XS tracker is 2.4x smaller and 4.3x faster than LightTrack [62] with superior accuracy.
- Score: 2.544539499281093
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present FEAR, a novel, fast, efficient, accurate, and robust Siamese
visual tracker. We introduce an architecture block for object model adaption,
called dual-template representation, and a pixel-wise fusion block to achieve
extra flexibility and efficiency of the model. The dual-template module
incorporates temporal information with only a single learnable parameter, while
the pixel-wise fusion block encodes more discriminative features with fewer
parameters compared to standard correlation modules. By plugging-in
sophisticated backbones with the novel modules, FEAR-M and FEAR-L trackers
surpass most Siamesetrackers on several academic benchmarks in both accuracy
and efficiencies. Employed with the lightweight backbone, the optimized version
FEAR-XS offers more than 10 times faster tracking than current Siamese trackers
while maintaining near state-of-the-art results. FEAR-XS tracker is 2.4x
smaller and 4.3x faster than LightTrack [62] with superior accuracy. In
addition, we expand the definition of the model efficiency by introducing a
benchmark on energy consumption and execution speed. Source code, pre-trained
models, and evaluation protocol will be made available upon request
Related papers
- LITE: A Paradigm Shift in Multi-Object Tracking with Efficient ReID Feature Integration [0.3277163122167433]
Lightweight Integrated Tracking-Feature Extraction paradigm is introduced as a novel multi-object tracking (MOT) approach.
It enhances ReID-based trackers by eliminating inference, pre-processing, post-processing, and ReID model training costs.
arXiv Detail & Related papers (2024-09-06T11:05:12Z) - Temporal Correlation Meets Embedding: Towards a 2nd Generation of JDE-based Real-Time Multi-Object Tracking [52.04679257903805]
Joint Detection and Embedding (JDE) trackers have demonstrated excellent performance in Multi-Object Tracking (MOT) tasks.
Our tracker, named TCBTrack, achieves state-of-the-art performance on multiple public benchmarks.
arXiv Detail & Related papers (2024-07-19T07:48:45Z) - Exploring Dynamic Transformer for Efficient Object Tracking [58.120191254379854]
We propose DyTrack, a dynamic transformer framework for efficient tracking.
DyTrack automatically learns to configure proper reasoning routes for various inputs, gaining better utilization of the available computational budget.
Experiments on multiple benchmarks demonstrate that DyTrack achieves promising speed-precision trade-offs with only a single model.
arXiv Detail & Related papers (2024-03-26T12:31:58Z) - Tracking Meets LoRA: Faster Training, Larger Model, Stronger Performance [87.19164603145056]
We propose LoRAT, a method that unveils the power of large ViT model for tracking within laboratory-level resources.
The essence of our work lies in adapting LoRA, a technique that fine-tunes a small subset of model parameters without adding inference latency.
We design an anchor-free head solely based on to adapt PETR, enabling better performance with less computational overhead.
arXiv Detail & Related papers (2024-03-08T11:41:48Z) - Lightweight Full-Convolutional Siamese Tracker [4.903759699116597]
This paper proposes a lightweight full-convolutional Siamese tracker called LightFC.
LightFC employs a novel efficient cross-correlation module and a novel efficient rep-center head.
Experiments show that LightFC achieves the optimal balance between performance, parameters, Flops and FPS.
arXiv Detail & Related papers (2023-10-09T04:07:35Z) - Exploring Lightweight Hierarchical Vision Transformers for Efficient
Visual Tracking [69.89887818921825]
HiT is a new family of efficient tracking models that can run at high speed on different devices.
HiT achieves 64.6% AUC on the LaSOT benchmark, surpassing all previous efficient trackers.
arXiv Detail & Related papers (2023-08-14T02:51:34Z) - Joint Feature Learning and Relation Modeling for Tracking: A One-Stream
Framework [76.70603443624012]
We propose a novel one-stream tracking (OSTrack) framework that unifies feature learning and relation modeling.
In this way, discriminative target-oriented features can be dynamically extracted by mutual guidance.
OSTrack achieves state-of-the-art performance on multiple benchmarks, in particular, it shows impressive results on the one-shot tracking benchmark GOT-10k.
arXiv Detail & Related papers (2022-03-22T18:37:11Z) - Distractor-Aware Fast Tracking via Dynamic Convolutions and MOT
Philosophy [63.91005999481061]
A practical long-term tracker typically contains three key properties, i.e. an efficient model design, an effective global re-detection strategy and a robust distractor awareness mechanism.
We propose a two-task tracking frame work (named DMTrack) to achieve distractor-aware fast tracking via Dynamic convolutions (d-convs) and Multiple object tracking (MOT) philosophy.
Our tracker achieves state-of-the-art performance on the LaSOT, OxUvA, TLP, VOT2018LT and VOT 2019LT benchmarks and runs in real-time (3x faster
arXiv Detail & Related papers (2021-04-25T00:59:53Z) - STMTrack: Template-free Visual Tracking with Space-time Memory Networks [42.06375415765325]
Existing trackers with template updating mechanisms rely on time-consuming numerical optimization and complex hand-designed strategies to achieve competitive performance.
We propose a novel tracking framework built on top of a space-time memory network that is competent to make full use of historical information related to the target.
Specifically, a novel memory mechanism is introduced, which stores the historical information of the target to guide the tracker to focus on the most informative regions in the current frame.
arXiv Detail & Related papers (2021-04-01T08:10:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.