RSINet: Rotation-Scale Invariant Network for Online Visual Tracking
- URL: http://arxiv.org/abs/2011.09153v1
- Date: Wed, 18 Nov 2020 08:19:14 GMT
- Title: RSINet: Rotation-Scale Invariant Network for Online Visual Tracking
- Authors: Yang Fang, Geun-Sik Jo and Chang-Hee Lee
- Abstract summary: Most network-based trackers perform the tracking process without model update, and cannot learn targetspecific variation adaptively.
In this paper, we propose a novel Rotation-Scale Invariant Network (RSINet) to address the above problem.
Our RSINet tracker consists of a target-distractor discrimination branch and a rotation-scale estimation branch, the rotation and scale knowledge can be explicitly learned by a multi-task learning method in an end-to-end manner.
In addtion, the tracking model is adaptively optimized and updated undertemporal energy control, which ensures model stability and reliability, as well as high tracking
- Score: 7.186849714896344
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most Siamese network-based trackers perform the tracking process without
model update, and cannot learn targetspecific variation adaptively. Moreover,
Siamese-based trackers infer the new state of tracked objects by generating
axis-aligned bounding boxes, which contain extra background noise, and are
unable to accurately estimate the rotation and scale transformation of moving
objects, thus potentially reducing tracking performance. In this paper, we
propose a novel Rotation-Scale Invariant Network (RSINet) to address the above
problem. Our RSINet tracker consists of a target-distractor discrimination
branch and a rotation-scale estimation branch, the rotation and scale knowledge
can be explicitly learned by a multi-task learning method in an end-to-end
manner. In addtion, the tracking model is adaptively optimized and updated
under spatio-temporal energy control, which ensures model stability and
reliability, as well as high tracking efficiency. Comprehensive experiments on
OTB-100, VOT2018, and LaSOT benchmarks demonstrate that our proposed RSINet
tracker yields new state-of-the-art performance compared with recent trackers,
while running at real-time speed about 45 FPS.
Related papers
- Temporal Correlation Meets Embedding: Towards a 2nd Generation of JDE-based Real-Time Multi-Object Tracking [52.04679257903805]
Joint Detection and Embedding (JDE) trackers have demonstrated excellent performance in Multi-Object Tracking (MOT) tasks.
Our tracker, named TCBTrack, achieves state-of-the-art performance on multiple public benchmarks.
arXiv Detail & Related papers (2024-07-19T07:48:45Z) - Robust Visual Tracking via Iterative Gradient Descent and Threshold Selection [4.978166837959101]
We introduce a novel robust linear regression estimator, which achieves favorable performance when the error vector follows i.i.d Gaussian-Laplacian distribution.
In addition, we expend IGDTS to a generative tracker, and apply IGDTS-distance to measure the deviation between the sample and the model.
Experimental results on several challenging image sequences show that the proposed tracker outperformance existing trackers.
arXiv Detail & Related papers (2024-06-02T01:51:09Z) - Exploring Dynamic Transformer for Efficient Object Tracking [58.120191254379854]
We propose DyTrack, a dynamic transformer framework for efficient tracking.
DyTrack automatically learns to configure proper reasoning routes for various inputs, gaining better utilization of the available computational budget.
Experiments on multiple benchmarks demonstrate that DyTrack achieves promising speed-precision trade-offs with only a single model.
arXiv Detail & Related papers (2024-03-26T12:31:58Z) - Autoregressive Queries for Adaptive Tracking with Spatio-TemporalTransformers [55.46413719810273]
rich-temporal information is crucial to the complicated target appearance in visual tracking.
Our method improves the tracker's performance on six popular tracking benchmarks.
arXiv Detail & Related papers (2024-03-15T02:39:26Z) - ACTrack: Adding Spatio-Temporal Condition for Visual Object Tracking [0.5371337604556311]
Efficiently modeling-temporal relations of objects is a key challenge in visual object tracking (VOT)
Existing methods track by appearance-based similarity or long-term relation modeling, resulting in rich temporal contexts between consecutive frames being easily overlooked.
In this paper we present ACTrack, a new framework with additive pre-temporal tracking framework with large memory conditions. It preserves the quality and capabilities of the pre-trained backbone by freezing its parameters, and makes a trainable lightweight additive net to model temporal relations in tracking.
We design an additive siamese convolutional network to ensure the integrity of spatial features and temporal sequence
arXiv Detail & Related papers (2024-02-27T07:34:08Z) - A Bayesian Detect to Track System for Robust Visual Object Tracking and
Semi-Supervised Model Learning [1.7268829007643391]
We ad-dress problems in a Bayesian tracking and detection framework parameterized by neural network outputs.
We propose a particle filter-based approximate sampling algorithm for tracking object state estimation.
Based on our particle filter inference algorithm, a semi-supervised learn-ing algorithm is utilized for learning tracking network on intermittent labeled frames.
arXiv Detail & Related papers (2022-05-05T00:18:57Z) - Rotation Equivariant Siamese Networks for Tracking [26.8787636300794]
We present rotation-equivariant Siamese networks (RE-SiamNets) for object tracking.
SiamNets allow estimating the change in orientation of the object in an unsupervised manner.
We show that RE-SiamNets handle the problem of rotation very well and out-perform their regular counterparts.
arXiv Detail & Related papers (2020-12-24T03:06:47Z) - Object Tracking through Residual and Dense LSTMs [67.98948222599849]
Deep learning-based trackers based on LSTMs (Long Short-Term Memory) recurrent neural networks have emerged as a powerful alternative.
DenseLSTMs outperform Residual and regular LSTM, and offer a higher resilience to nuisances.
Our case study supports the adoption of residual-based RNNs for enhancing the robustness of other trackers.
arXiv Detail & Related papers (2020-06-22T08:20:17Z) - Cascaded Regression Tracking: Towards Online Hard Distractor
Discrimination [202.2562153608092]
We propose a cascaded regression tracker with two sequential stages.
In the first stage, we filter out abundant easily-identified negative candidates.
In the second stage, a discrete sampling based ridge regression is designed to double-check the remaining ambiguous hard samples.
arXiv Detail & Related papers (2020-06-18T07:48:01Z) - Robust Visual Object Tracking with Two-Stream Residual Convolutional
Networks [62.836429958476735]
We propose a Two-Stream Residual Convolutional Network (TS-RCN) for visual tracking.
Our TS-RCN can be integrated with existing deep learning based visual trackers.
To further improve the tracking performance, we adopt a "wider" residual network ResNeXt as its feature extraction backbone.
arXiv Detail & Related papers (2020-05-13T19:05:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.