Predictive Visual Tracking: A New Benchmark and Baseline Approach
- URL: http://arxiv.org/abs/2103.04508v1
- Date: Mon, 8 Mar 2021 01:50:05 GMT
- Title: Predictive Visual Tracking: A New Benchmark and Baseline Approach
- Authors: Bowen Li, Yiming Li, Junjie Ye, Changhong Fu, and Hang Zhao
- Abstract summary: In the real-world scenarios, the onboard processing time of the image streams inevitably leads to a discrepancy between the tracking results and the real-world states.
Existing visual tracking benchmarks commonly run the trackers offline and ignore such latency in the evaluation.
In this work, we aim to deal with a more realistic problem of latency-aware tracking.
- Score: 27.87099869398515
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: As a crucial robotic perception capability, visual tracking has been
intensively studied recently. In the real-world scenarios, the onboard
processing time of the image streams inevitably leads to a discrepancy between
the tracking results and the real-world states. However, existing visual
tracking benchmarks commonly run the trackers offline and ignore such latency
in the evaluation. In this work, we aim to deal with a more realistic problem
of latency-aware tracking. The state-of-the-art trackers are evaluated in the
aerial scenarios with new metrics jointly assessing the tracking accuracy and
efficiency. Moreover, a new predictive visual tracking baseline is developed to
compensate for the latency stemming from the onboard computation. Our
latency-aware benchmark can provide a more realistic evaluation of the trackers
for the robotic applications. Besides, exhaustive experiments have proven the
effectiveness of the proposed predictive visual tracking baseline approach.
Related papers
- RTracker: Recoverable Tracking via PN Tree Structured Memory [71.05904715104411]
We propose a recoverable tracking framework, RTracker, that uses a tree-structured memory to dynamically associate a tracker and a detector to enable self-recovery.
Specifically, we propose a Positive-Negative Tree-structured memory to chronologically store and maintain positive and negative target samples.
Our core idea is to use the support samples of positive and negative target categories to establish a relative distance-based criterion for a reliable assessment of target loss.
arXiv Detail & Related papers (2024-03-28T08:54:40Z) - LEAP-VO: Long-term Effective Any Point Tracking for Visual Odometry [52.131996528655094]
We present the Long-term Effective Any Point Tracking (LEAP) module.
LEAP innovatively combines visual, inter-track, and temporal cues with mindfully selected anchors for dynamic track estimation.
Based on these traits, we develop LEAP-VO, a robust visual odometry system adept at handling occlusions and dynamic scenes.
arXiv Detail & Related papers (2024-01-03T18:57:27Z) - Towards Real-World Visual Tracking with Temporal Contexts [64.7981374129495]
We propose a two-level framework (TCTrack) that can exploit temporal contexts efficiently.
Based on it, we propose a stronger version for real-world visual tracking, i.e., TCTrack++.
For feature extraction, we propose an attention-based temporally adaptive convolution to enhance the spatial features.
For similarity map refinement, we introduce an adaptive temporal transformer to encode the temporal knowledge efficiently.
arXiv Detail & Related papers (2023-08-20T17:59:40Z) - PVT++: A Simple End-to-End Latency-Aware Visual Tracking Framework [33.7932898514321]
We present a framework for end-to-end latency-aware tracking, i.e., end-to-end predictive visual tracking (PVT++)
Unlike existing solutions that naively append Kalman Filters after trackers, PVT++ can be jointly optimized.
PVT++ can achieve significant performance gain on various trackers and exhibit higher accuracy than prior solutions.
arXiv Detail & Related papers (2022-11-21T16:43:33Z) - A Bayesian Detect to Track System for Robust Visual Object Tracking and
Semi-Supervised Model Learning [1.7268829007643391]
We ad-dress problems in a Bayesian tracking and detection framework parameterized by neural network outputs.
We propose a particle filter-based approximate sampling algorithm for tracking object state estimation.
Based on our particle filter inference algorithm, a semi-supervised learn-ing algorithm is utilized for learning tracking network on intermittent labeled frames.
arXiv Detail & Related papers (2022-05-05T00:18:57Z) - Benchmarking high-fidelity pedestrian tracking systems for research,
real-time monitoring and crowd control [55.41644538483948]
High-fidelity pedestrian tracking in real-life conditions has been an important tool in fundamental crowd dynamics research.
As this technology advances, it is becoming increasingly useful also in society.
To successfully employ pedestrian tracking techniques in research and technology, it is crucial to validate and benchmark them for accuracy.
We present and discuss a benchmark suite, towards an open standard in the community, for privacy-respectful pedestrian tracking techniques.
arXiv Detail & Related papers (2021-08-26T11:45:26Z) - Continuity-Discrimination Convolutional Neural Network for Visual Object
Tracking [150.51667609413312]
This paper proposes a novel model, named Continuity-Discrimination Convolutional Neural Network (CD-CNN) for visual object tracking.
To address this problem, CD-CNN models temporal appearance continuity based on the idea of temporal slowness.
In order to alleviate inaccurate target localization and drifting, we propose a novel notion, object-centroid.
arXiv Detail & Related papers (2021-04-18T06:35:03Z) - Robust Visual Object Tracking with Two-Stream Residual Convolutional
Networks [62.836429958476735]
We propose a Two-Stream Residual Convolutional Network (TS-RCN) for visual tracking.
Our TS-RCN can be integrated with existing deep learning based visual trackers.
To further improve the tracking performance, we adopt a "wider" residual network ResNeXt as its feature extraction backbone.
arXiv Detail & Related papers (2020-05-13T19:05:42Z) - RetinaTrack: Online Single Stage Joint Detection and Tracking [22.351109024452462]
We focus on the tracking-by-detection paradigm for autonomous driving where both tasks are mission critical.
We propose a conceptually simple and efficient joint model of detection and tracking, called RetinaTrack, which modifies the popular single stage RetinaNet approach.
arXiv Detail & Related papers (2020-03-30T23:46:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.