A Survey for Deep RGBT Tracking
- URL: http://arxiv.org/abs/2201.09296v1
- Date: Sun, 23 Jan 2022 15:52:26 GMT
- Title: A Survey for Deep RGBT Tracking
- Authors: Zhangyong Tang (1), Tianyang Xu (1) and Xiao-Jun Wu (1) ((1) Jiangnan
University, China)
- Abstract summary: Visual object tracking with the visible (RGB) and thermal infrared (TIR) electromagnetic waves, shorted in RGBT tracking, recently draws increasing attention in the tracking community.
Considering the rapid development of deep learning, a survey for the recent deep neural network based RGBT trackers is presented.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Visual object tracking with the visible (RGB) and thermal infrared (TIR)
electromagnetic waves, shorted in RGBT tracking, recently draws increasing
attention in the tracking community. Considering the rapid development of deep
learning, a survey for the recent deep neural network based RGBT trackers is
presented in this paper. Firstly, we give brief introduction for the RGBT
trackers concluded into this category. Then, a comparison among the existing
RGBT trackers on several challenging benchmarks is given statistically.
Specifically, MDNet and Siamese architectures are the two mainstream frameworks
in the RGBT community, especially the former. Trackers based on MDNet achieve
higher performance while Siamese-based trackers satisfy the real-time
requirement. In summary, since the large-scale dataset LasHeR is published, the
integration of end-to-end framework, e.g., Siamese and Transformer, should be
further considered to fulfil the real-time as well as more robust performance.
Furthermore, the mathematical meaning should be more considered during
designing the network. This survey can be treated as a look-up-table for
researchers who are concerned about RGBT tracking.
Related papers
- Long-term Frame-Event Visual Tracking: Benchmark Dataset and Baseline [37.06330707742272]
We first propose a new long-term and large-scale frame-event single object tracking dataset, termed FELT.
It contains 742 videos and 1,594,474 RGB frames and event stream pairs and has become the largest frame-event tracking dataset to date.
We propose a novel associative memory Transformer network as a unified backbone by introducing modern Hopfield layers into multi-head self-attention blocks to fuse both RGB and event data.
arXiv Detail & Related papers (2024-03-09T08:49:50Z) - EANet: Enhanced Attribute-based RGBT Tracker Network [0.0]
We propose a deep learning-based image tracking approach that fuses RGB and thermal images (RGBT)
The proposed model consists of two main components: a feature extractor and a tracker.
The proposed methods are evaluated on the RGBT234 citeLiCLiang 2018 and LasHeR citeLiLasher 2021 datasets.
arXiv Detail & Related papers (2023-07-04T19:34:53Z) - Visible-Thermal UAV Tracking: A Large-Scale Benchmark and New Baseline [80.13652104204691]
In this paper, we construct a large-scale benchmark with high diversity for visible-thermal UAV tracking (VTUAV)
We provide a coarse-to-fine attribute annotation, where frame-level attributes are provided to exploit the potential of challenge-specific trackers.
In addition, we design a new RGB-T baseline, named Hierarchical Multi-modal Fusion Tracker (HMFT), which fuses RGB-T data in various levels.
arXiv Detail & Related papers (2022-04-08T15:22:33Z) - RGBD Object Tracking: An In-depth Review [89.96221353160831]
We firstly review RGBD object trackers from different perspectives, including RGBD fusion, depth usage, and tracking framework.
We benchmark a representative set of RGBD trackers, and give detailed analyses based on their performances.
arXiv Detail & Related papers (2022-03-26T18:53:51Z) - Temporal Aggregation for Adaptive RGBT Tracking [14.00078027541162]
We propose an RGBT tracker which takes clues into account for robust appearance model learning.
Unlike most existing RGBT trackers that implement object tracking tasks with only spatial information included, temporal information is further considered in this method.
arXiv Detail & Related papers (2022-01-22T02:31:56Z) - LasHeR: A Large-scale High-diversity Benchmark for RGBT Tracking [27.00930976353204]
LasHeR consists of 1224 visible and thermal infrared video pairs with more than 730K frame pairs in total.
LasHeR is highly diverse capturing from a broad range of object categories, camera viewpoints, scene complexities and environmental factors.
We conduct a comprehensive performance evaluation of 12 RGBT tracking algorithms on the LasHeR dataset.
arXiv Detail & Related papers (2021-04-27T14:04:23Z) - Multi-modal Visual Tracking: Review and Experimental Comparison [85.20414397784937]
We summarize the multi-modal tracking algorithms, especially visible-depth (RGB-D) tracking and visible-thermal (RGB-T) tracking.
We conduct experiments to analyze the effectiveness of trackers on five datasets.
arXiv Detail & Related papers (2020-12-08T02:39:38Z) - LSOTB-TIR:A Large-Scale High-Diversity Thermal Infrared Object Tracking
Benchmark [51.1506855334948]
This paper presents a Large-Scale and high-diversity general Thermal InfraRed (TIR) Object Tracking Benchmark, called LSOTBTIR.
We annotate the bounding box of objects in every frame of all sequences and generate over 730K bounding boxes in total.
We evaluate and analyze more than 30 trackers on LSOTB-TIR to provide a series of baselines, and the results show that deep trackers achieve promising performance.
arXiv Detail & Related papers (2020-08-03T12:36:06Z) - Jointly Modeling Motion and Appearance Cues for Robust RGB-T Tracking [85.333260415532]
We develop a novel late fusion method to infer the fusion weight maps of both RGB and thermal (T) modalities.
When the appearance cue is unreliable, we take motion cues into account to make the tracker robust.
Numerous results on three recent RGB-T tracking datasets show that the proposed tracker performs significantly better than other state-of-the-art algorithms.
arXiv Detail & Related papers (2020-07-04T08:11:33Z) - Object Tracking through Residual and Dense LSTMs [67.98948222599849]
Deep learning-based trackers based on LSTMs (Long Short-Term Memory) recurrent neural networks have emerged as a powerful alternative.
DenseLSTMs outperform Residual and regular LSTM, and offer a higher resilience to nuisances.
Our case study supports the adoption of residual-based RNNs for enhancing the robustness of other trackers.
arXiv Detail & Related papers (2020-06-22T08:20:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.