Tracking-by-Trackers with a Distilled and Reinforced Model
- URL: http://arxiv.org/abs/2007.04108v2
- Date: Wed, 30 Sep 2020 13:42:53 GMT
- Title: Tracking-by-Trackers with a Distilled and Reinforced Model
- Authors: Matteo Dunnhofer, Niki Martinel, Christian Micheloni
- Abstract summary: A compact student model is trained via the marriage of knowledge distillation and reinforcement learning.
The proposed algorithms compete with real-time state-of-the-art trackers.
- Score: 24.210580784051277
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Visual object tracking was generally tackled by reasoning independently on
fast processing algorithms, accurate online adaptation methods, and fusion of
trackers. In this paper, we unify such goals by proposing a novel tracking
methodology that takes advantage of other visual trackers, offline and online.
A compact student model is trained via the marriage of knowledge distillation
and reinforcement learning. The first allows to transfer and compress tracking
knowledge of other trackers. The second enables the learning of evaluation
measures which are then exploited online. After learning, the student can be
ultimately used to build (i) a very fast single-shot tracker, (ii) a tracker
with a simple and effective online adaptation mechanism, (iii) a tracker that
performs fusion of other trackers. Extensive validation shows that the proposed
algorithms compete with real-time state-of-the-art trackers.
Related papers
- CoTracker3: Simpler and Better Point Tracking by Pseudo-Labelling Real Videos [63.90674869153876]
We introduce CoTracker3, comprising a new tracking model and a new semi-supervised training recipe.
This allows real videos without annotations to be used during training by generating pseudo-labels using off-the-shelf teachers.
The model is available in online and offline variants and reliably tracks visible and occluded points.
arXiv Detail & Related papers (2024-10-15T17:56:32Z) - FACT: Feature Adaptive Continual-learning Tracker for Multiple Object Tracking [22.53374351982883]
We propose a new MOT framework called the Feature Adaptive Continual-learning Tracker (FACT)
FACT enables real-time tracking and feature learning for targets by utilizing all past tracking information.
We demonstrate that the framework can be integrated with various state-of-the-art feature-based trackers.
arXiv Detail & Related papers (2024-09-12T10:14:48Z) - Tracking with Human-Intent Reasoning [64.69229729784008]
This work proposes a new tracking task -- Instruction Tracking.
It involves providing implicit tracking instructions that require the trackers to perform tracking automatically in video frames.
TrackGPT is capable of performing complex reasoning-based tracking.
arXiv Detail & Related papers (2023-12-29T03:22:18Z) - TDT: Teaching Detectors to Track without Fully Annotated Videos [2.8292841621378844]
One-stage trackers that predict both detections and appearance embeddings in one forward pass received much attention.
Our proposed one-stage solution matches the two-stage counterpart in quality but is 3 times faster.
arXiv Detail & Related papers (2022-05-11T15:56:17Z) - Recursive Least-Squares Estimator-Aided Online Learning for Visual
Tracking [58.14267480293575]
We propose a simple yet effective online learning approach for few-shot online adaptation without requiring offline training.
It allows an in-built memory retention mechanism for the model to remember the knowledge about the object seen before.
We evaluate our approach based on two networks in the online learning families for tracking, i.e., multi-layer perceptrons in RT-MDNet and convolutional neural networks in DiMP.
arXiv Detail & Related papers (2021-12-28T06:51:18Z) - Learning to Track Objects from Unlabeled Videos [63.149201681380305]
In this paper, we propose to learn an Unsupervised Single Object Tracker (USOT) from scratch.
To narrow the gap between unsupervised trackers and supervised counterparts, we propose an effective unsupervised learning approach composed of three stages.
Experiments show that the proposed USOT learned from unlabeled videos performs well over the state-of-the-art unsupervised trackers by large margins.
arXiv Detail & Related papers (2021-08-28T22:10:06Z) - Unsupervised Deep Representation Learning for Real-Time Tracking [137.69689503237893]
We propose an unsupervised learning method for visual tracking.
The motivation of our unsupervised learning is that a robust tracker should be effective in bidirectional tracking.
We build our framework on a Siamese correlation filter network, and propose a multi-frame validation scheme and a cost-sensitive loss to facilitate unsupervised learning.
arXiv Detail & Related papers (2020-07-22T08:23:12Z) - Tracking by Instance Detection: A Meta-Learning Approach [99.66119903655711]
We propose a principled three-step approach to build a high-performance tracker.
We build two trackers, named Retina-MAML and FCOS-MAML, based on two modern detectors RetinaNet and FCOS.
Both trackers run in real-time at 40 FPS.
arXiv Detail & Related papers (2020-04-02T05:55:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.