Exploring the Impacts from Datasets to Monocular Depth Estimation (MDE)
Models with MineNavi
- URL: http://arxiv.org/abs/2008.08454v2
- Date: Tue, 28 Jun 2022 13:55:03 GMT
- Title: Exploring the Impacts from Datasets to Monocular Depth Estimation (MDE)
Models with MineNavi
- Authors: Xiangtong Wang, Binbin Liang, Menglong Yang and Wei Li
- Abstract summary: Current computer vision tasks based on deep learning require a huge amount of data with annotations for model training or testing.
In practice, manual labeling for dense estimation tasks is very difficult or even impossible, and the scenes of the dataset are often restricted to a small range.
We propose a synthetic dataset generation method to obtain the expandable dataset without burdensome manual workforce.
- Score: 5.689127984415125
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current computer vision tasks based on deep learning require a huge amount of
data with annotations for model training or testing, especially in some dense
estimation tasks, such as optical flow segmentation and depth estimation. In
practice, manual labeling for dense estimation tasks is very difficult or even
impossible, and the scenes of the dataset are often restricted to a small
range, which dramatically limits the development of the community. To overcome
this deficiency, we propose a synthetic dataset generation method to obtain the
expandable dataset without burdensome manual workforce. By this method, we
construct a dataset called MineNavi containing video footages from
first-perspective-view of the aircraft matched with accurate ground truth for
depth estimation in aircraft navigation application. We also provide
quantitative experiments to prove that pre-training via our MineNavi dataset
can improve the performance of depth estimation model and speed up the
convergence of the model on real scene data. Since the synthetic dataset has a
similar effect to the real-world dataset in the training process of deep model,
we also provide additional experiments with monocular depth estimation method
to demonstrate the impact of various factors in our dataset such as lighting
conditions and motion mode.
Related papers
- A New Dataset for Monocular Depth Estimation Under Viewpoint Shifts [6.260553883409459]
This paper introduces a novel dataset and evaluation methodology to quantify the impact of different camera positions and orientations on monocular depth estimation performance.
We propose a ground truth strategy based on homography estimation and object detection, eliminating the need for expensive lidar sensors.
arXiv Detail & Related papers (2024-09-26T13:57:05Z) - TanDepth: Leveraging Global DEMs for Metric Monocular Depth Estimation in UAVs [5.6168844664788855]
This work presents TanDepth, a practical, online scale recovery method for obtaining metric depth results from relative estimations at inference-time.
Tailored for Unmanned Aerial Vehicle (UAV) applications, our method leverages sparse measurements from Global Digital Elevation Models (GDEM) by projecting them to the camera view.
An adaptation to the Cloth Simulation Filter is presented, which allows selecting ground points from the estimated depth map to then correlate with the projected reference points.
arXiv Detail & Related papers (2024-09-08T15:54:43Z) - Domain-Transferred Synthetic Data Generation for Improving Monocular Depth Estimation [9.812476193015488]
We propose a method of data generation in simulation using 3D synthetic environments and CycleGAN domain transfer.
We compare this method of data generation to the popular NYUDepth V2 dataset by training a depth estimation model based on the DenseDepth structure using different training sets of real and simulated data.
We evaluate the performance of the models on newly collected images and LiDAR depth data from a Husky robot to verify the generalizability of the approach and show that GAN-transformed data can serve as an effective alternative to real-world data, particularly in depth estimation.
arXiv Detail & Related papers (2024-05-02T09:21:10Z) - LargeST: A Benchmark Dataset for Large-Scale Traffic Forecasting [65.71129509623587]
Road traffic forecasting plays a critical role in smart city initiatives and has experienced significant advancements thanks to the power of deep learning.
However, the promising results achieved on current public datasets may not be applicable to practical scenarios.
We introduce the LargeST benchmark dataset, which includes a total of 8,600 sensors in California with a 5-year time coverage.
arXiv Detail & Related papers (2023-06-14T05:48:36Z) - Dataset Distillation: A Comprehensive Review [76.26276286545284]
dataset distillation (DD) aims to derive a much smaller dataset containing synthetic samples, based on which the trained models yield performance comparable with those trained on the original dataset.
This paper gives a comprehensive review and summary of recent advances in DD and its application.
arXiv Detail & Related papers (2023-01-17T17:03:28Z) - SC-DepthV3: Robust Self-supervised Monocular Depth Estimation for
Dynamic Scenes [58.89295356901823]
Self-supervised monocular depth estimation has shown impressive results in static scenes.
It relies on the multi-view consistency assumption for training networks, however, that is violated in dynamic object regions.
We introduce an external pretrained monocular depth estimation model for generating single-image depth prior.
Our model can predict sharp and accurate depth maps, even when training from monocular videos of highly-dynamic scenes.
arXiv Detail & Related papers (2022-11-07T16:17:47Z) - Learning to Simulate Realistic LiDARs [66.7519667383175]
We introduce a pipeline for data-driven simulation of a realistic LiDAR sensor.
We show that our model can learn to encode realistic effects such as dropped points on transparent surfaces.
We use our technique to learn models of two distinct LiDAR sensors and use them to improve simulated LiDAR data accordingly.
arXiv Detail & Related papers (2022-09-22T13:12:54Z) - TRoVE: Transforming Road Scene Datasets into Photorealistic Virtual
Environments [84.6017003787244]
This work proposes a synthetic data generation pipeline to address the difficulties and domain-gaps present in simulated datasets.
We show that using annotations and visual cues from existing datasets, we can facilitate automated multi-modal data generation.
arXiv Detail & Related papers (2022-08-16T20:46:08Z) - Joint Prediction of Monocular Depth and Structure using Planar and
Parallax Geometry [4.620624344434533]
Supervised learning depth estimation methods can achieve good performance when trained on high-quality ground-truth, like LiDAR data.
We propose a novel approach combining structure information from a promising Plane and Parallax geometry pipeline with depth information into a U-Net supervised learning network.
Our model has impressive performance on depth prediction of thin objects and edges, and compared to structure prediction baseline, our model performs more robustly.
arXiv Detail & Related papers (2022-07-13T17:04:05Z) - DeepSatData: Building large scale datasets of satellite images for
training machine learning models [77.17638664503215]
This report presents design considerations for automatically generating satellite imagery datasets for training machine learning models.
We discuss issues faced from the point of view of deep neural network training and evaluation.
arXiv Detail & Related papers (2021-04-28T15:13:12Z) - Domain Adaptive Monocular Depth Estimation With Semantic Information [13.387521845596149]
We propose an adversarial training model that leverages semantic information to narrow the domain gap.
The proposed compact model achieves state-of-the-art performance comparable to complex latest models.
arXiv Detail & Related papers (2021-04-12T18:50:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.