DG-Labeler and DGL-MOTS Dataset: Boost the Autonomous Driving Perception
- URL: http://arxiv.org/abs/2110.07790v1
- Date: Fri, 15 Oct 2021 01:04:31 GMT
- Title: DG-Labeler and DGL-MOTS Dataset: Boost the Autonomous Driving Perception
- Authors: Yiming Cui, Zhiwen Cao, Yixin Xie, Xingyu Jiang, Feng Tao, Yingjie
Chen, Lin Li, Dongfang Liu
- Abstract summary: We introduce the DG-Labeler and DGL-MOTS dataset to facilitate the training data annotation for the MOTS task.
Results on extensive cross-dataset evaluations indicate significant performance improvements for several state-of-the-art methods trained on our dataset.
- Score: 15.988493804970092
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-object tracking and segmentation (MOTS) is a critical task for
autonomous driving applications. The existing MOTS studies face two critical
challenges: 1) the published datasets inadequately capture the real-world
complexity for network training to address various driving settings; 2) the
working pipeline annotation tool is under-studied in the literature to improve
the quality of MOTS learning examples. In this work, we introduce the
DG-Labeler and DGL-MOTS dataset to facilitate the training data annotation for
the MOTS task and accordingly improve network training accuracy and efficiency.
DG-Labeler uses the novel Depth-Granularity Module to depict the instance
spatial relations and produce fine-grained instance masks. Annotated by
DG-Labeler, our DGL-MOTS dataset exceeds the prior effort (i.e., KITTI MOTS and
BDD100K) in data diversity, annotation quality, and temporal representations.
Results on extensive cross-dataset evaluations indicate significant performance
improvements for several state-of-the-art methods trained on our DGL-MOTS
dataset. We believe our DGL-MOTS Dataset and DG-Labeler hold the valuable
potential to boost the visual perception of future transportation.
Related papers
- DA-MoE: Addressing Depth-Sensitivity in Graph-Level Analysis through Mixture of Experts [70.21017141742763]
Graph neural networks (GNNs) are gaining popularity for processing graph-structured data.
Existing methods generally use a fixed number of GNN layers to generate representations for all graphs.
We propose the depth adaptive mixture of expert (DA-MoE) method, which incorporates two main improvements to GNN.
arXiv Detail & Related papers (2024-11-05T11:46:27Z) - Web-Scale Visual Entity Recognition: An LLM-Driven Data Approach [56.55633052479446]
Web-scale visual entity recognition presents significant challenges due to the lack of clean, large-scale training data.
We propose a novel methodology to curate such a dataset, leveraging a multimodal large language model (LLM) for label verification, metadata generation, and rationale explanation.
Experiments demonstrate that models trained on this automatically curated data achieve state-of-the-art performance on web-scale visual entity recognition tasks.
arXiv Detail & Related papers (2024-10-31T06:55:24Z) - EDGE: Enhanced Grounded GUI Understanding with Enriched Multi-Granularity Synthetic Data [15.801018643716437]
This paper aims to enhance the GUI understanding and interacting capabilities of large vision-language models (LVLMs) through a data-driven approach.
We propose EDGE, a general data synthesis framework that automatically generates large-scale, multi-granularity training data from webpages across the Web.
Our approach significantly reduces the dependence on manual annotations, empowering researchers to harness the vast public resources available on the Web to advance their work.
arXiv Detail & Related papers (2024-10-25T10:46:17Z) - UdeerLID+: Integrating LiDAR, Image, and Relative Depth with Semi-Supervised [12.440461420762265]
Road segmentation is a critical task for autonomous driving systems.
Our work introduces an innovative approach that integrates LiDAR point cloud data, visual image, and relative depth maps.
One of the primary challenges is the scarcity of large-scale, accurately labeled datasets.
arXiv Detail & Related papers (2024-09-10T03:57:30Z) - Diffusion Model is an Effective Planner and Data Synthesizer for
Multi-Task Reinforcement Learning [101.66860222415512]
Multi-Task Diffusion Model (textscMTDiff) is a diffusion-based method that incorporates Transformer backbones and prompt learning for generative planning and data synthesis.
For generative planning, we find textscMTDiff outperforms state-of-the-art algorithms across 50 tasks on Meta-World and 8 maps on Maze2D.
arXiv Detail & Related papers (2023-05-29T05:20:38Z) - Self-Supervised Graph Neural Network for Multi-Source Domain Adaptation [51.21190751266442]
Domain adaptation (DA) tries to tackle the scenarios when the test data does not fully follow the same distribution of the training data.
By learning from large-scale unlabeled samples, self-supervised learning has now become a new trend in deep learning.
We propose a novel textbfSelf-textbfSupervised textbfGraph Neural Network (SSG) to enable more effective inter-task information exchange and knowledge sharing.
arXiv Detail & Related papers (2022-04-08T03:37:56Z) - Gram-SLD: Automatic Self-labeling and Detection for Instance Objects [6.512856940779818]
We propose a new framework based on co-training called Gram Self-Labeling and Detection (Gram-SLD)
Gram-SLD can automatically annotate a large amount of data with very limited manually labeled key data and achieve competitive performance.
arXiv Detail & Related papers (2021-12-07T11:34:55Z) - Unsupervised Domain Adaptive Learning via Synthetic Data for Person
Re-identification [101.1886788396803]
Person re-identification (re-ID) has gained more and more attention due to its widespread applications in video surveillance.
Unfortunately, the mainstream deep learning methods still need a large quantity of labeled data to train models.
In this paper, we develop a data collector to automatically generate synthetic re-ID samples in a computer game, and construct a data labeler to simultaneously annotate them.
arXiv Detail & Related papers (2021-09-12T15:51:41Z) - DAGA: Data Augmentation with a Generation Approach for Low-resource
Tagging Tasks [88.62288327934499]
We propose a novel augmentation method with language models trained on the linearized labeled sentences.
Our method is applicable to both supervised and semi-supervised settings.
arXiv Detail & Related papers (2020-11-03T07:49:15Z) - Improving the Performance of Fine-Grain Image Classifiers via Generative
Data Augmentation [0.5161531917413706]
We develop Data Augmentation from Proficient Pre-Training of Robust Generative Adrial Networks (DAPPER GAN)
DAPPER GAN is an ML analytics support tool that automatically generates novel views of training images.
We experimentally evaluate this technique on the Stanford Cars dataset, demonstrating improved vehicle make and model classification accuracy.
arXiv Detail & Related papers (2020-08-12T15:29:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.