EANet: Enhanced Attribute-based RGBT Tracker Network
- URL: http://arxiv.org/abs/2307.01893v1
- Date: Tue, 4 Jul 2023 19:34:53 GMT
- Title: EANet: Enhanced Attribute-based RGBT Tracker Network
- Authors: Abbas T\"urko\u{g}lu, Erdem Akag\"und\"uz
- Abstract summary: We propose a deep learning-based image tracking approach that fuses RGB and thermal images (RGBT)
The proposed model consists of two main components: a feature extractor and a tracker.
The proposed methods are evaluated on the RGBT234 citeLiCLiang 2018 and LasHeR citeLiLasher 2021 datasets.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Tracking objects can be a difficult task in computer vision, especially when
faced with challenges such as occlusion, changes in lighting, and motion blur.
Recent advances in deep learning have shown promise in challenging these
conditions. However, most deep learning-based object trackers only use visible
band (RGB) images. Thermal infrared electromagnetic waves (TIR) can provide
additional information about an object, including its temperature, when faced
with challenging conditions. We propose a deep learning-based image tracking
approach that fuses RGB and thermal images (RGBT). The proposed model consists
of two main components: a feature extractor and a tracker. The feature
extractor encodes deep features from both the RGB and the TIR images. The
tracker then uses these features to track the object using an enhanced
attribute-based architecture. We propose a fusion of attribute-specific feature
selection with an aggregation module. The proposed methods are evaluated on the
RGBT234 \cite{LiCLiang2018} and LasHeR \cite{LiLasher2021} datasets, which are
the most widely used RGBT object-tracking datasets in the literature. The
results show that the proposed system outperforms state-of-the-art RGBT object
trackers on these datasets, with a relatively smaller number of parameters.
Related papers
- BihoT: A Large-Scale Dataset and Benchmark for Hyperspectral Camouflaged Object Tracking [22.533682363532403]
We provide a new task called hyperspectral camouflaged object tracking (HCOT)
We meticulously construct a large-scale HCOT dataset, termed BihoT, which consists of 41,912 hyperspectral images covering 49 video sequences.
A simple but effective baseline model, named spectral prompt-based distractor-aware network (SPDAN), is proposed.
arXiv Detail & Related papers (2024-08-22T09:07:51Z) - Spherical Space Feature Decomposition for Guided Depth Map
Super-Resolution [123.04455334124188]
Guided depth map super-resolution (GDSR) aims to upsample low-resolution (LR) depth maps with additional information involved in high-resolution (HR) RGB images from the same scene.
In this paper, we propose the Spherical Space feature Decomposition Network (SSDNet) to solve the above issues.
Our method can achieve state-of-the-art results on four test datasets, as well as successfully generalize to real-world scenes.
arXiv Detail & Related papers (2023-03-15T21:22:21Z) - Learning Dual-Fused Modality-Aware Representations for RGBD Tracking [67.14537242378988]
Compared with the traditional RGB object tracking, the addition of the depth modality can effectively solve the target and background interference.
Some existing RGBD trackers use the two modalities separately and thus some particularly useful shared information between them is ignored.
We propose a novel Dual-fused Modality-aware Tracker (termed DMTracker) which aims to learn informative and discriminative representations of the target objects for robust RGBD tracking.
arXiv Detail & Related papers (2022-11-06T07:59:07Z) - Mirror Complementary Transformer Network for RGB-thermal Salient Object
Detection [16.64781797503128]
RGB-thermal object detection (RGB-T SOD) aims to locate the common prominent objects of an aligned visible and thermal infrared image pair.
In this paper, we propose a novel mirror complementary Transformer network (MCNet) for RGB-T SOD.
Experiments on benchmark and VT723 datasets show that the proposed method outperforms state-of-the-art approaches.
arXiv Detail & Related papers (2022-07-07T20:26:09Z) - Visible-Thermal UAV Tracking: A Large-Scale Benchmark and New Baseline [80.13652104204691]
In this paper, we construct a large-scale benchmark with high diversity for visible-thermal UAV tracking (VTUAV)
We provide a coarse-to-fine attribute annotation, where frame-level attributes are provided to exploit the potential of challenge-specific trackers.
In addition, we design a new RGB-T baseline, named Hierarchical Multi-modal Fusion Tracker (HMFT), which fuses RGB-T data in various levels.
arXiv Detail & Related papers (2022-04-08T15:22:33Z) - Temporal Aggregation for Adaptive RGBT Tracking [14.00078027541162]
We propose an RGBT tracker which takes clues into account for robust appearance model learning.
Unlike most existing RGBT trackers that implement object tracking tasks with only spatial information included, temporal information is further considered in this method.
arXiv Detail & Related papers (2022-01-22T02:31:56Z) - MFGNet: Dynamic Modality-Aware Filter Generation for RGB-T Tracking [72.65494220685525]
We propose a new dynamic modality-aware filter generation module (named MFGNet) to boost the message communication between visible and thermal data.
We generate dynamic modality-aware filters with two independent networks. The visible and thermal filters will be used to conduct a dynamic convolutional operation on their corresponding input feature maps respectively.
To address issues caused by heavy occlusion, fast motion, and out-of-view, we propose to conduct a joint local and global search by exploiting a new direction-aware target-driven attention mechanism.
arXiv Detail & Related papers (2021-07-22T03:10:51Z) - MobileSal: Extremely Efficient RGB-D Salient Object Detection [62.04876251927581]
This paper introduces a novel network, methodname, which focuses on efficient RGB-D salient object detection (SOD)
We propose an implicit depth restoration (IDR) technique to strengthen the feature representation capability of mobile networks for RGB-D SOD.
With IDR and CPR incorporated, methodnameperforms favorably against sArt methods on seven challenging RGB-D SOD datasets.
arXiv Detail & Related papers (2020-12-24T04:36:42Z) - RGBT Salient Object Detection: A Large-scale Dataset and Benchmark [12.14043884641457]
Taking advantage of RGB and thermal infrared images becomes a new research direction for detecting salient object in complex scenes.
This work contributes such a RGBT image dataset named VT5000, including 5000 spatially aligned RGBT image pairs with ground truth annotations.
We propose a powerful baseline approach, which extracts multi-level features within each modality and aggregates these features of all modalities with the attention mechanism.
arXiv Detail & Related papers (2020-07-07T07:58:14Z) - Is Depth Really Necessary for Salient Object Detection? [50.10888549190576]
We make the first attempt in realizing an unified depth-aware framework with only RGB information as input for inference.
Not only surpasses the state-of-the-art performances on five public RGB SOD benchmarks, but also surpasses the RGBD-based methods on five benchmarks by a large margin.
arXiv Detail & Related papers (2020-05-30T13:40:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.