Siamese Box Adaptive Network for Visual Tracking
- URL: http://arxiv.org/abs/2003.06761v2
- Date: Wed, 22 Apr 2020 09:59:30 GMT
- Title: Siamese Box Adaptive Network for Visual Tracking
- Authors: Zedu Chen, Bineng Zhong, Guorong Li, Shengping Zhang, Rongrong Ji
- Abstract summary: We propose a simple yet effective visual tracking framework (named Siamese Box Adaptive Network, SiamBAN)
SiamBAN directly classifies objects and regresses their bounding boxes in a unified convolutional network (FCN)
SiamBAN achieves state-of-the-art performance and runs at 40 FPS, confirming its effectiveness and efficiency.
- Score: 100.46025199664642
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Most of the existing trackers usually rely on either a multi-scale searching
scheme or pre-defined anchor boxes to accurately estimate the scale and aspect
ratio of a target. Unfortunately, they typically call for tedious and heuristic
configurations. To address this issue, we propose a simple yet effective visual
tracking framework (named Siamese Box Adaptive Network, SiamBAN) by exploiting
the expressive power of the fully convolutional network (FCN). SiamBAN views
the visual tracking problem as a parallel classification and regression
problem, and thus directly classifies objects and regresses their bounding
boxes in a unified FCN. The no-prior box design avoids hyper-parameters
associated with the candidate boxes, making SiamBAN more flexible and general.
Extensive experiments on visual tracking benchmarks including VOT2018, VOT2019,
OTB100, NFS, UAV123, and LaSOT demonstrate that SiamBAN achieves
state-of-the-art performance and runs at 40 FPS, confirming its effectiveness
and efficiency. The code will be available at https://github.com/hqucv/siamban.
Related papers
- Two-stream Beats One-stream: Asymmetric Siamese Network for Efficient Visual Tracking [54.124445709376154]
We propose a novel asymmetric Siamese tracker named textbfAsymTrack for efficient tracking.
Building on this architecture, we devise an efficient template modulation mechanism to inject crucial cues into the search features.
Experiments demonstrate that AsymTrack offers superior speed-precision trade-offs across different platforms.
arXiv Detail & Related papers (2025-03-01T14:44:54Z) - Multi-attention Associate Prediction Network for Visual Tracking [3.9628431811908533]
classification-regression prediction networks have realized impressive success in several modern deep trackers.
There is an inherent difference between classification and regression tasks, so they have diverse even opposite demands for feature matching.
We propose a multi-attention associate prediction network (MAPNet) to tackle the above problems.
arXiv Detail & Related papers (2024-03-25T03:18:58Z) - BEVTrack: A Simple and Strong Baseline for 3D Single Object Tracking in Bird's-Eye View [56.77287041917277]
3D Single Object Tracking (SOT) is a fundamental task of computer vision, proving essential for applications like autonomous driving.
In this paper, we propose BEVTrack, a simple yet effective baseline method.
By estimating the target motion in Bird's-Eye View (BEV) to perform tracking, BEVTrack demonstrates surprising simplicity from various aspects, i.e., network designs, training objectives, and tracking pipeline, while achieving superior performance.
arXiv Detail & Related papers (2023-09-05T12:42:26Z) - Adaptive Siamese Tracking with a Compact Latent Network [219.38172719948048]
We present an intuitive viewing to simplify the Siamese-based trackers by converting the tracking task to a classification.
Under this viewing, we perform an in-depth analysis for them through visual simulations and real tracking examples.
We apply it to adjust three classical Siamese-based trackers, namely SiamRPN++, SiamFC, and SiamBAN.
arXiv Detail & Related papers (2023-02-02T08:06:02Z) - Unsupervised Learning of Accurate Siamese Tracking [68.58171095173056]
We present a novel unsupervised tracking framework, in which we can learn temporal correspondence both on the classification branch and regression branch.
Our tracker outperforms preceding unsupervised methods by a substantial margin, performing on par with supervised methods on large-scale datasets such as TrackingNet and LaSOT.
arXiv Detail & Related papers (2022-04-04T13:39:43Z) - SiamCorners: Siamese Corner Networks for Visual Tracking [39.43480791427431]
We propose a simple yet effective anchor-free tracker (named Siamese corner networks, SiamCorners)
By tracking a target as a pair of corners, we avoid the need to design the anchor boxes.
SiamCorners achieves a 53.7% AUC on NFS30 and a 61.4% AUC on UAV123, while still running at 42 frames per second (FPS)
arXiv Detail & Related papers (2021-04-15T08:23:30Z) - Multiple Convolutional Features in Siamese Networks for Object Tracking [13.850110645060116]
Multiple Features-Siamese Tracker (MFST) is a novel tracking algorithm exploiting several hierarchical feature maps for robust tracking.
MFST achieves high tracking accuracy, while outperforming the standard siamese tracker on object tracking benchmarks.
arXiv Detail & Related papers (2021-03-01T08:02:27Z) - CRACT: Cascaded Regression-Align-Classification for Robust Visual
Tracking [97.84109669027225]
We introduce an improved proposal refinement module, Cascaded Regression-Align- Classification (CRAC)
CRAC yields new state-of-the-art performances on many benchmarks.
In experiments on seven benchmarks including OTB-2015, UAV123, NfS, VOT-2018, TrackingNet, GOT-10k and LaSOT, our CRACT exhibits very promising results in comparison with state-of-the-art competitors.
arXiv Detail & Related papers (2020-11-25T02:18:33Z) - FCOS: A simple and strong anchor-free object detector [111.87691210818194]
We propose a fully convolutional one-stage object detector (FCOS) to solve object detection in a per-pixel prediction fashion.
Almost all state-of-the-art object detectors such as RetinaNet, SSD, YOLOv3, and Faster R-CNN rely on pre-defined anchor boxes.
In contrast, our proposed detector FCOS is anchor box free, as well as proposal free.
arXiv Detail & Related papers (2020-06-14T01:03:39Z) - Accurate Anchor Free Tracking [9.784386353369483]
This paper develops the first Anchor Free Siamese Network (AFSN)
A target object is defined by a bounding box center, tracking offset, and object size.
We compare AFSN to the best anchor-based trackers with source codes available for each benchmark.
arXiv Detail & Related papers (2020-06-13T04:42:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.