1st Place Solution to ECCV-TAO-2020: Detect and Represent Any Object for
Tracking
- URL: http://arxiv.org/abs/2101.08040v2
- Date: Mon, 1 Feb 2021 08:38:31 GMT
- Title: 1st Place Solution to ECCV-TAO-2020: Detect and Represent Any Object for
Tracking
- Authors: Fei Du, Bo Xu, Jiasheng Tang, Yuqi Zhang, Fan Wang, and Hao Li
- Abstract summary: We extend the classical tracking-by-detection paradigm to this tracking-any-object task.
We learn appearance features to represent any object by training feature learning networks.
Simple linking strategies with most similar appearance features and tracklet-level post association module are finally applied to generate final tracking results.
- Score: 19.15537335764895
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We extend the classical tracking-by-detection paradigm to this
tracking-any-object task. Solid detection results are first extracted from TAO
dataset. Some state-of-the-art techniques like \textbf{BA}lanced-\textbf{G}roup
\textbf{S}oftmax (\textbf{BAGS}\cite{li2020overcoming}) and
DetectoRS\cite{qiao2020detectors} are integrated during detection. Then we
learned appearance features to represent any object by training feature
learning networks. We ensemble several models for improving detection and
feature representation. Simple linking strategies with most similar appearance
features and tracklet-level post association module are finally applied to
generate final tracking results. Our method is submitted as \textbf{AOA} on the
challenge website. Code is available at
https://github.com/feiaxyt/Winner_ECCV20_TAO.
Related papers
- UnsMOT: Unified Framework for Unsupervised Multi-Object Tracking with
Geometric Topology Guidance [6.577227592760559]
UnsMOT is a novel framework that combines appearance and motion features of objects with geometric information to provide more accurate tracking.
Experimental results show remarkable performance in terms of HOTA, IDF1, and MOTA metrics in comparison with state-of-the-art methods.
arXiv Detail & Related papers (2023-09-03T04:58:12Z) - Weakly-supervised Contrastive Learning for Unsupervised Object Discovery [52.696041556640516]
Unsupervised object discovery is promising due to its ability to discover objects in a generic manner.
We design a semantic-guided self-supervised learning model to extract high-level semantic features from images.
We introduce Principal Component Analysis (PCA) to localize object regions.
arXiv Detail & Related papers (2023-07-07T04:03:48Z) - SalienDet: A Saliency-based Feature Enhancement Algorithm for Object
Detection for Autonomous Driving [160.57870373052577]
We propose a saliency-based OD algorithm (SalienDet) to detect unknown objects.
Our SalienDet utilizes a saliency-based algorithm to enhance image features for object proposal generation.
We design a dataset relabeling approach to differentiate the unknown objects from all objects in training sample set to achieve Open-World Detection.
arXiv Detail & Related papers (2023-05-11T16:19:44Z) - Unified Visual Relationship Detection with Vision and Language Models [89.77838890788638]
This work focuses on training a single visual relationship detector predicting over the union of label spaces from multiple datasets.
We propose UniVRD, a novel bottom-up method for Unified Visual Relationship Detection by leveraging vision and language models.
Empirical results on both human-object interaction detection and scene-graph generation demonstrate the competitive performance of our model.
arXiv Detail & Related papers (2023-03-16T00:06:28Z) - Learning to Track with Object Permanence [61.36492084090744]
We introduce an end-to-end trainable approach for joint object detection and tracking.
Our model, trained jointly on synthetic and real data, outperforms the state of the art on KITTI, and MOT17 datasets.
arXiv Detail & Related papers (2021-03-26T04:43:04Z) - Global Correlation Network: End-to-End Joint Multi-Object Detection and
Tracking [2.749204052800622]
We present a novel network to realize joint multi-object detection and tracking in an end-to-end way, called Global Correlation Network (GCNet)
GCNet introduces the global correlation layer for regression of absolute size and coordinates of bounding boxes instead of offsets prediction.
The pipeline of detection and tracking by GCNet is conceptually simple, which does not need non-maximum suppression, data association, and other complicated tracking strategies.
arXiv Detail & Related papers (2021-03-23T13:16:42Z) - Structure-Consistent Weakly Supervised Salient Object Detection with
Local Saliency Coherence [14.79639149658596]
We propose a one-round end-to-end training approach for weakly supervised salient object detection via scribble annotations.
Our method achieves a new state-of-the-art performance on six benchmarks.
arXiv Detail & Related papers (2020-12-08T12:49:40Z) - Many-shot from Low-shot: Learning to Annotate using Mixed Supervision
for Object Detection [27.354492072251492]
Online annotation module (OAM) learns to generate a many-shot set of emphreliable annotations from a larger volume of weakly labelled images.
Our OAM can be jointly trained with any fully supervised two-stage object detection method, providing additional training annotations on the fly.
The integration of the OAM with Fast(er) R-CNN improves their performance by $17%$ mAP, $9%$ AP50 on PASCAL VOC 2007 and MS-COCO benchmarks, and significantly outperforms competing methods using mixed supervision.
arXiv Detail & Related papers (2020-08-21T22:06:43Z) - Chained-Tracker: Chaining Paired Attentive Regression Results for
End-to-End Joint Multiple-Object Detection and Tracking [102.31092931373232]
We propose a simple online model named Chained-Tracker (CTracker), which naturally integrates all the three subtasks into an end-to-end solution.
The two major novelties: chained structure and paired attentive regression, make CTracker simple, fast and effective.
arXiv Detail & Related papers (2020-07-29T02:38:49Z) - Attention-based Joint Detection of Object and Semantic Part [4.389917490809522]
Our model is created on top of two Faster-RCNN models that share their features to get enhanced representations of both.
Experiments on the PASCAL-Part 2010 dataset show that joint detection can simultaneously improve both object detection and part detection.
arXiv Detail & Related papers (2020-07-05T18:54:10Z) - One-Shot Object Detection without Fine-Tuning [62.39210447209698]
We introduce a two-stage model consisting of a first stage Matching-FCOS network and a second stage Structure-Aware Relation Module.
We also propose novel training strategies that effectively improve detection performance.
Our method exceeds the state-of-the-art one-shot performance consistently on multiple datasets.
arXiv Detail & Related papers (2020-05-08T01:59:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.