RGBD1K: A Large-scale Dataset and Benchmark for RGB-D Object Tracking
- URL: http://arxiv.org/abs/2208.09787v1
- Date: Sun, 21 Aug 2022 03:07:36 GMT
- Title: RGBD1K: A Large-scale Dataset and Benchmark for RGB-D Object Tracking
- Authors: Xue-Feng Zhu, Tianyang Xu, Zhangyong Tang, Zucheng Wu, Haodong Liu,
Xiao Yang, Xiao-Jun Wu, Josef Kittler
- Abstract summary: Given a limited amount of annotated RGB-D tracking data, most state-of-the-art RGB-D trackers are simple extensions of high-performance RGB-only trackers.
To address the dataset deficiency issue, a new RGB-D dataset named RGBD1K is released in this paper.
- Score: 30.448658049744775
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: RGB-D object tracking has attracted considerable attention recently,
achieving promising performance thanks to the symbiosis between visual and
depth channels. However, given a limited amount of annotated RGB-D tracking
data, most state-of-the-art RGB-D trackers are simple extensions of
high-performance RGB-only trackers, without fully exploiting the underlying
potential of the depth channel in the offline training stage. To address the
dataset deficiency issue, a new RGB-D dataset named RGBD1K is released in this
paper. The RGBD1K contains 1,050 sequences with about 2.5M frames in total. To
demonstrate the benefits of training on a larger RGB-D data set in general, and
RGBD1K in particular, we develop a transformer-based RGB-D tracker, named SPT,
as a baseline for future visual object tracking studies using the new dataset.
The results, of extensive experiments using the SPT tracker emonstrate the
potential of the RGBD1K dataset to improve the performance of RGB-D tracking,
inspiring future developments of effective tracker designs. The dataset and
codes will be available on the project homepage:
https://will.be.available.at.this.website.
Related papers
- RGB-Sonar Tracking Benchmark and Spatial Cross-Attention Transformer Tracker [4.235252053339947]
This paper introduces a new challenging RGB-Sonar (RGB-S) tracking task.
It investigates how to achieve efficient tracking of an underwater target through the interaction of RGB and sonar modalities.
arXiv Detail & Related papers (2024-06-11T12:01:11Z) - DFormer: Rethinking RGBD Representation Learning for Semantic
Segmentation [76.81628995237058]
DFormer is a novel framework to learn transferable representations for RGB-D segmentation tasks.
It pretrains the backbone using image-depth pairs from ImageNet-1K.
DFormer achieves new state-of-the-art performance on two popular RGB-D tasks.
arXiv Detail & Related papers (2023-09-18T11:09:11Z) - ARKitTrack: A New Diverse Dataset for Tracking Using Mobile RGB-D Data [75.73063721067608]
We propose a new RGB-D tracking dataset for both static and dynamic scenes captured by consumer-grade LiDAR scanners equipped on Apple's iPhone and iPad.
ARKitTrack contains 300 RGB-D sequences, 455 targets, and 229.7K video frames in total.
In-depth empirical analysis has verified that the ARKitTrack dataset can significantly facilitate RGB-D tracking and that the proposed baseline method compares favorably against the state of the arts.
arXiv Detail & Related papers (2023-03-24T09:51:13Z) - RGBD Object Tracking: An In-depth Review [89.96221353160831]
We firstly review RGBD object trackers from different perspectives, including RGBD fusion, depth usage, and tracking framework.
We benchmark a representative set of RGBD trackers, and give detailed analyses based on their performances.
arXiv Detail & Related papers (2022-03-26T18:53:51Z) - Visual Object Tracking on Multi-modal RGB-D Videos: A Review [16.098468526632473]
The goal of this review is to summarize the relative knowledge of the research filed of RGB-D tracking.
To be specific, we will generalize the related RGB-D tracking benchmarking datasets as well as the corresponding performance measurements.
arXiv Detail & Related papers (2022-01-23T08:02:49Z) - Boosting RGB-D Saliency Detection by Leveraging Unlabeled RGB Images [89.81919625224103]
Training deep models for RGB-D salient object detection (SOD) often requires a large number of labeled RGB-D images.
We present a Dual-Semi RGB-D Salient Object Detection Network (DS-Net) to leverage unlabeled RGB images for boosting RGB-D saliency detection.
arXiv Detail & Related papers (2022-01-01T03:02:27Z) - DepthTrack : Unveiling the Power of RGBD Tracking [29.457114656913944]
This work introduces a new RGBD tracking dataset - Depth-Track.
It has twice as many sequences (200) and scene types (40) than in the largest existing dataset.
The average length of the sequences (1473), the number of deformable objects (16) and the number of tracking attributes (15) have been increased.
arXiv Detail & Related papers (2021-08-31T16:42:38Z) - Synergistic saliency and depth prediction for RGB-D saliency detection [76.27406945671379]
Existing RGB-D saliency datasets are small, which may lead to overfitting and limited generalization for diverse scenarios.
We propose a semi-supervised system for RGB-D saliency detection that can be trained on smaller RGB-D saliency datasets without saliency ground truth.
arXiv Detail & Related papers (2020-07-03T14:24:41Z) - Is Depth Really Necessary for Salient Object Detection? [50.10888549190576]
We make the first attempt in realizing an unified depth-aware framework with only RGB information as input for inference.
Not only surpasses the state-of-the-art performances on five public RGB SOD benchmarks, but also surpasses the RGBD-based methods on five benchmarks by a large margin.
arXiv Detail & Related papers (2020-05-30T13:40:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.