LSOTB-TIR:A Large-Scale High-Diversity Thermal Infrared Object Tracking
Benchmark
- URL: http://arxiv.org/abs/2008.00836v1
- Date: Mon, 3 Aug 2020 12:36:06 GMT
- Title: LSOTB-TIR:A Large-Scale High-Diversity Thermal Infrared Object Tracking
Benchmark
- Authors: Qiao Liu, Xin Li, Zhenyu He, Chenglong Li, Jun Li, Zikun Zhou, Di
Yuan, Jing Li, Kai Yang, Nana Fan, Feng Zheng
- Abstract summary: This paper presents a Large-Scale and high-diversity general Thermal InfraRed (TIR) Object Tracking Benchmark, called LSOTBTIR.
We annotate the bounding box of objects in every frame of all sequences and generate over 730K bounding boxes in total.
We evaluate and analyze more than 30 trackers on LSOTB-TIR to provide a series of baselines, and the results show that deep trackers achieve promising performance.
- Score: 51.1506855334948
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: In this paper, we present a Large-Scale and high-diversity general Thermal
InfraRed (TIR) Object Tracking Benchmark, called LSOTBTIR, which consists of an
evaluation dataset and a training dataset with a total of 1,400 TIR sequences
and more than 600K frames. We annotate the bounding box of objects in every
frame of all sequences and generate over 730K bounding boxes in total. To the
best of our knowledge, LSOTB-TIR is the largest and most diverse TIR object
tracking benchmark to date. To evaluate a tracker on different attributes, we
define 4 scenario attributes and 12 challenge attributes in the evaluation
dataset. By releasing LSOTB-TIR, we encourage the community to develop deep
learning based TIR trackers and evaluate them fairly and comprehensively. We
evaluate and analyze more than 30 trackers on LSOTB-TIR to provide a series of
baselines, and the results show that deep trackers achieve promising
performance. Furthermore, we re-train several representative deep trackers on
LSOTB-TIR, and their results demonstrate that the proposed training dataset
significantly improves the performance of deep TIR trackers. Codes and dataset
are available at https://github.com/QiaoLiuHit/LSOTB-TIR.
Related papers
- Temporal Correlation Meets Embedding: Towards a 2nd Generation of JDE-based Real-Time Multi-Object Tracking [52.04679257903805]
Joint Detection and Embedding (JDE) trackers have demonstrated excellent performance in Multi-Object Tracking (MOT) tasks.
Our tracker, named TCBTrack, achieves state-of-the-art performance on multiple public benchmarks.
arXiv Detail & Related papers (2024-07-19T07:48:45Z) - Cross-domain Learning Framework for Tracking Users in RIS-aided Multi-band ISAC Systems with Sparse Labeled Data [55.70071704247794]
Integrated sensing and communications (ISAC) is pivotal for 6G communications and is boosted by the rapid development of reconfigurable intelligent surfaces (RISs)
This paper proposes the X2Track framework, where we model the tracking function by a hierarchical architecture, jointly utilizing multi-modal CSI indicators across multiple bands, and optimize it in a cross-domain manner.
Under X2Track, we design an efficient deep learning algorithm to minimize tracking errors, based on transformer neural networks and adversarial learning techniques.
arXiv Detail & Related papers (2024-05-10T08:04:27Z) - Long-term Frame-Event Visual Tracking: Benchmark Dataset and Baseline [37.06330707742272]
We first propose a new long-term and large-scale frame-event single object tracking dataset, termed FELT.
It contains 742 videos and 1,594,474 RGB frames and event stream pairs and has become the largest frame-event tracking dataset to date.
We propose a novel associative memory Transformer network as a unified backbone by introducing modern Hopfield layers into multi-head self-attention blocks to fuse both RGB and event data.
arXiv Detail & Related papers (2024-03-09T08:49:50Z) - SOMPT22: A Surveillance Oriented Multi-Pedestrian Tracking Dataset [5.962184741057505]
We introduce SOMPT22 dataset; a new set for multi person tracking with annotated short videos captured from static cameras located on poles with 6-8 meters in height positioned for city surveillance.
We analyze MOT trackers classified as one-shot and two-stage with respect to the way of use of detection and reID networks on this new dataset.
The experimental results of our new dataset indicate that SOTA is still far from high efficiency, and single-shot trackers are good candidates to unify fast execution and accuracy with competitive performance.
arXiv Detail & Related papers (2022-08-04T11:09:19Z) - Visible-Thermal UAV Tracking: A Large-Scale Benchmark and New Baseline [80.13652104204691]
In this paper, we construct a large-scale benchmark with high diversity for visible-thermal UAV tracking (VTUAV)
We provide a coarse-to-fine attribute annotation, where frame-level attributes are provided to exploit the potential of challenge-specific trackers.
In addition, we design a new RGB-T baseline, named Hierarchical Multi-modal Fusion Tracker (HMFT), which fuses RGB-T data in various levels.
arXiv Detail & Related papers (2022-04-08T15:22:33Z) - LasHeR: A Large-scale High-diversity Benchmark for RGBT Tracking [27.00930976353204]
LasHeR consists of 1224 visible and thermal infrared video pairs with more than 730K frame pairs in total.
LasHeR is highly diverse capturing from a broad range of object categories, camera viewpoints, scene complexities and environmental factors.
We conduct a comprehensive performance evaluation of 12 RGBT tracking algorithms on the LasHeR dataset.
arXiv Detail & Related papers (2021-04-27T14:04:23Z) - Benchmarking Deep Trackers on Aerial Videos [5.414308305392762]
In this paper, we compare ten trackers based on deep learning techniques on four aerial datasets.
We choose top performing trackers utilizing different approaches, specifically tracking by detection, discriminative correlation filters, Siamese networks and reinforcement learning.
Our findings indicate that the trackers perform significantly worse in aerial datasets compared to standard ground level videos.
arXiv Detail & Related papers (2021-03-24T01:45:19Z) - Multi-modal Visual Tracking: Review and Experimental Comparison [85.20414397784937]
We summarize the multi-modal tracking algorithms, especially visible-depth (RGB-D) tracking and visible-thermal (RGB-T) tracking.
We conduct experiments to analyze the effectiveness of trackers on five datasets.
arXiv Detail & Related papers (2020-12-08T02:39:38Z) - TAO: A Large-Scale Benchmark for Tracking Any Object [95.87310116010185]
Tracking Any Object dataset consists of 2,907 high resolution videos, captured in diverse environments, which are half a minute long on average.
We ask annotators to label objects that move at any point in the video, and give names to them post factum.
Our vocabulary is both significantly larger and qualitatively different from existing tracking datasets.
arXiv Detail & Related papers (2020-05-20T21:07:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.