Beyond Background-Aware Correlation Filters: Adaptive Context Modeling
by Hand-Crafted and Deep RGB Features for Visual Tracking
- URL: http://arxiv.org/abs/2004.02932v2
- Date: Wed, 29 Sep 2021 07:53:59 GMT
- Title: Beyond Background-Aware Correlation Filters: Adaptive Context Modeling
by Hand-Crafted and Deep RGB Features for Visual Tracking
- Authors: Seyed Mojtaba Marvasti-Zadeh, Hossein Ghanei-Yakhdan, Shohreh Kasaei
- Abstract summary: An adaptive background-aware correlation filter-based tracker is proposed in this paper.
It effectively models the target appearance by using either the histogram of oriented gradients (HOG) or convolutional neural network (CNN) feature maps.
The proposed method exploits the fast 2D non-maximum suppression (NMS) algorithm and the semantic information comparison to detect challenging situations.
- Score: 16.439797365064003
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, the background-aware correlation filters have achie-ved a
lot of research interest in the visual target tracking. However, these methods
cannot suitably model the target appearance due to the exploitation of
hand-crafted features. On the other hand, the recent deep learning-based visual
tracking methods have provided a competitive performance along with extensive
computations. In this paper, an adaptive background-aware correlation
filter-based tracker is proposed that effectively models the target appearance
by using either the histogram of oriented gradients (HOG) or convolutional
neural network (CNN) feature maps. The proposed method exploits the fast 2D
non-maximum suppression (NMS) algorithm and the semantic information comparison
to detect challenging situations. When the HOG-based response map is not
reliable, or the context region has a low semantic similarity with prior
regions, the proposed method constructs the CNN context model to improve the
target region estimation. Furthermore, the rejection option allows the proposed
method to update the CNN context model only on valid regions. Comprehensive
experimental results demonstrate that the proposed adaptive method clearly
outperforms the accuracy and robustness of visual target tracking compared to
the state-of-the-art methods on the OTB-50, OTB-100, TC-128, UAV-123, and
VOT-2015 datasets.
Related papers
- SatDM: Synthesizing Realistic Satellite Image with Semantic Layout
Conditioning using Diffusion Models [0.0]
Denoising Diffusion Probabilistic Models (DDPMs) have demonstrated significant promise in synthesizing realistic images from semantic layouts.
In this paper, a conditional DDPM model capable of taking a semantic map and generating high-quality, diverse, and correspondingly accurate satellite images is implemented.
The effectiveness of our proposed model is validated using a meticulously labeled dataset introduced within the context of this study.
arXiv Detail & Related papers (2023-09-28T19:39:13Z) - Reconstructing Spatiotemporal Data with C-VAEs [49.1574468325115]
Conditional continuous representation of moving regions is commonly used.
In this work, we explore the capabilities of Conditional Varitemporal Autoencoder (C-VAE) models to generate realistic representations of regions' evolution.
arXiv Detail & Related papers (2023-07-12T15:34:10Z) - TSG: Target-Selective Gradient Backprop for Probing CNN Visual Saliency [72.9106103283475]
We study the visual saliency, a.k.a. visual explanation, to interpret convolutional neural networks.
Inspired by those observations, we propose a novel visual saliency framework, termed Target-Selective Gradient (TSG) backprop.
The proposed TSG consists of two components, namely, TSG-Conv and TSG-FC, which rectify the gradients for convolutional layers and fully-connected layers, respectively.
arXiv Detail & Related papers (2021-10-11T12:00:20Z) - MFGNet: Dynamic Modality-Aware Filter Generation for RGB-T Tracking [72.65494220685525]
We propose a new dynamic modality-aware filter generation module (named MFGNet) to boost the message communication between visible and thermal data.
We generate dynamic modality-aware filters with two independent networks. The visible and thermal filters will be used to conduct a dynamic convolutional operation on their corresponding input feature maps respectively.
To address issues caused by heavy occlusion, fast motion, and out-of-view, we propose to conduct a joint local and global search by exploiting a new direction-aware target-driven attention mechanism.
arXiv Detail & Related papers (2021-07-22T03:10:51Z) - An Adaptive Framework for Learning Unsupervised Depth Completion [59.17364202590475]
We present a method to infer a dense depth map from a color image and associated sparse depth measurements.
We show that regularization and co-visibility are related via the fitness of the model to data and can be unified into a single framework.
arXiv Detail & Related papers (2021-06-06T02:27:55Z) - Coarse-to-Fine Object Tracking Using Deep Features and Correlation
Filters [2.3526458707956643]
This paper presents a novel deep learning tracking algorithm.
We exploit the generalization ability of deep features to coarsely estimate target translation.
Then, we capitalize on the discriminative power of correlation filters to precisely localize the tracked object.
arXiv Detail & Related papers (2020-12-23T16:43:21Z) - Adaptive Exploitation of Pre-trained Deep Convolutional Neural Networks
for Robust Visual Tracking [14.627458410954628]
This paper provides a comprehensive analysis of four commonly used CNN models to determine the best feature maps of each model.
With the aid of analysis results as attribute dictionaries, adaptive exploitation of deep features is proposed to improve the accuracy and robustness of visual trackers.
arXiv Detail & Related papers (2020-08-29T17:09:43Z) - Bayesian Graph Neural Networks with Adaptive Connection Sampling [62.51689735630133]
We propose a unified framework for adaptive connection sampling in graph neural networks (GNNs)
The proposed framework not only alleviates over-smoothing and over-fitting tendencies of deep GNNs, but also enables learning with uncertainty in graph analytic tasks with GNNs.
arXiv Detail & Related papers (2020-06-07T07:06:35Z) - Efficient Scale Estimation Methods using Lightweight Deep Convolutional
Neural Networks for Visual Tracking [16.439797365064003]
This paper exploits pre-trained lightweight CNNs models to propose two efficient scale estimation methods.
The proposed methods are formulated based on either holistic or region representation of convolutional feature maps.
They exploit proposed one-pass feature extraction processes that significantly improve the computational efficiency.
arXiv Detail & Related papers (2020-04-06T18:49:37Z) - Effective Fusion of Deep Multitasking Representations for Robust Visual
Tracking [34.09763324745818]
This paper aims to evaluate the performance of twelve state-of-the-art ResNet-based FENs in a DCF-based framework.
It ranks their best feature maps and explores the generalized adoption of the best ResNet-based FEN into another DCF-based method.
The proposed method extracts deep semantic information from a fully convolutional FEN and fuses it with the best ResNet-based feature maps to strengthen the target representation.
arXiv Detail & Related papers (2020-04-03T05:33:59Z) - Deep Semantic Matching with Foreground Detection and Cycle-Consistency [103.22976097225457]
We address weakly supervised semantic matching based on a deep network.
We explicitly estimate the foreground regions to suppress the effect of background clutter.
We develop cycle-consistent losses to enforce the predicted transformations across multiple images to be geometrically plausible and consistent.
arXiv Detail & Related papers (2020-03-31T22:38:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.