DMV: Visual Object Tracking via Part-level Dense Memory and Voting-based
Retrieval
- URL: http://arxiv.org/abs/2003.09171v1
- Date: Fri, 20 Mar 2020 10:05:30 GMT
- Title: DMV: Visual Object Tracking via Part-level Dense Memory and Voting-based
Retrieval
- Authors: Gunhee Nam, Seoung Wug Oh, Joon-Young Lee, Seon Joo Kim
- Abstract summary: We propose a novel memory-based tracker via part-level dense memory and voting-based retrieval, called DMV.
We also propose a novel voting mechanism for the memory reading to filter out unreliable information in the memory.
- Score: 61.366644088881735
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel memory-based tracker via part-level dense memory and
voting-based retrieval, called DMV. Since deep learning techniques have been
introduced to the tracking field, Siamese trackers have attracted many
researchers due to the balance between speed and accuracy. However, most of
them are based on a single template matching, which limits the performance as
it restricts the accessible in-formation to the initial target features. In
this paper, we relieve this limitation by maintaining an external memory that
saves the tracking record. Part-level retrieval from the memory also liberates
the information from the template and allows our tracker to better handle the
challenges such as appearance changes and occlusions. By updating the memory
during tracking, the representative power for the target object can be enhanced
without online learning. We also propose a novel voting mechanism for the
memory reading to filter out unreliable information in the memory. We
comprehensively evaluate our tracker on OTB-100,TrackingNet, GOT-10k, LaSOT,
and UAV123, which show that our method yields comparable results to the
state-of-the-art methods.
Related papers
- A Distractor-Aware Memory for Visual Object Tracking with SAM2 [11.864619292028278]
Memory-based trackers are video object segmentation methods that form the target model by concatenating recently tracked frames into a memory buffer and localize the target by attending the current image to the buffered frames.
SAM2.1++ outperforms SAM2.1 and related SAM memory extensions on seven benchmarks and sets a solid new state-of-the-art on six of them.
arXiv Detail & Related papers (2024-11-26T16:41:09Z) - RTracker: Recoverable Tracking via PN Tree Structured Memory [71.05904715104411]
We propose a recoverable tracking framework, RTracker, that uses a tree-structured memory to dynamically associate a tracker and a detector to enable self-recovery.
Specifically, we propose a Positive-Negative Tree-structured memory to chronologically store and maintain positive and negative target samples.
Our core idea is to use the support samples of positive and negative target categories to establish a relative distance-based criterion for a reliable assessment of target loss.
arXiv Detail & Related papers (2024-03-28T08:54:40Z) - MeMOTR: Long-Term Memory-Augmented Transformer for Multi-Object Tracking [19.173503245000678]
We propose MeMOTR, a long-term memory-augmented Transformer for multi-object tracking.
MeMOTR impressively surpasses the state-of-the-art method by 7.9% and 13.0% on HOTA and AssA metrics.
Our model also outperforms other Transformer-based methods on association performance on MOT17 and generalizes well on BDD100K.
arXiv Detail & Related papers (2023-07-28T17:50:09Z) - MeMOT: Multi-Object Tracking with Memory [97.48960039220823]
Our model, called MeMOT, consists of three main modules that are all Transformer-based.
MeMOT observes very competitive performance on widely adopted MOT datasets.
arXiv Detail & Related papers (2022-03-31T02:33:20Z) - Learning Dynamic Compact Memory Embedding for Deformable Visual Object
Tracking [82.34356879078955]
We propose a compact memory embedding to enhance the discrimination of the segmentation-based deformable visual tracking method.
Our method outperforms the excellent segmentation-based trackers, i.e., D3S and SiamMask on DAVIS 2017 benchmark.
arXiv Detail & Related papers (2021-11-23T03:07:12Z) - STMTrack: Template-free Visual Tracking with Space-time Memory Networks [42.06375415765325]
Existing trackers with template updating mechanisms rely on time-consuming numerical optimization and complex hand-designed strategies to achieve competitive performance.
We propose a novel tracking framework built on top of a space-time memory network that is competent to make full use of historical information related to the target.
Specifically, a novel memory mechanism is introduced, which stores the historical information of the target to guide the tracker to focus on the most informative regions in the current frame.
arXiv Detail & Related papers (2021-04-01T08:10:56Z) - Learning Spatio-Appearance Memory Network for High-Performance Visual
Tracking [79.80401607146987]
Existing object tracking usually learns a bounding-box based template to match visual targets across frames, which cannot accurately learn a pixel-wise representation.
This paper presents a novel segmentation-based tracking architecture, which is equipped with a local-temporal memory network to learn accurate-temporal correspondence.
arXiv Detail & Related papers (2020-09-21T08:12:02Z) - PeTra: A Sparsely Supervised Memory Model for People Tracking [50.98911178059019]
We propose PeTra, a memory-augmented neural network designed to track entities in its memory slots.
We empirically compare key modeling choices, finding that we can simplify several aspects of the design of the memory module while retaining strong performance.
PeTra is highly effective in both evaluations, demonstrating its ability to track people in its memory despite being trained with limited annotation.
arXiv Detail & Related papers (2020-05-06T17:45:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.