Ground material classification and for UAV-based photogrammetric 3D data
A 2D-3D Hybrid Approach
- URL: http://arxiv.org/abs/2109.12221v1
- Date: Fri, 24 Sep 2021 22:29:26 GMT
- Title: Ground material classification and for UAV-based photogrammetric 3D data
A 2D-3D Hybrid Approach
- Authors: Meida Chen, Andrew Feng, Yu Hou, Kyle McCullough, Pratusha Bhuvana
Prasad, Lucio Soibelman
- Abstract summary: In recent years, photogrammetry has been widely used in many areas to create 3D virtual data representing the physical environment.
These cutting-edge technologies have caught the US Army and Navy's attention for the purpose of rapid 3D battlefield reconstruction, virtual training, and simulations.
- Score: 1.3359609092684614
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, photogrammetry has been widely used in many areas to create
photorealistic 3D virtual data representing the physical environment. The
innovation of small unmanned aerial vehicles (sUAVs) has provided additional
high-resolution imaging capabilities with low cost for mapping a relatively
large area of interest. These cutting-edge technologies have caught the US Army
and Navy's attention for the purpose of rapid 3D battlefield reconstruction,
virtual training, and simulations. Our previous works have demonstrated the
importance of information extraction from the derived photogrammetric data to
create semantic-rich virtual environments (Chen et al., 2019). For example, an
increase of simulation realism and fidelity was achieved by segmenting and
replacing photogrammetric trees with game-ready tree models. In this work, we
further investigated the semantic information extraction problem and focused on
the ground material segmentation and object detection tasks. The main
innovation of this work was that we leveraged both the original 2D images and
the derived 3D photogrammetric data to overcome the challenges faced when using
each individual data source. For ground material segmentation, we utilized an
existing convolutional neural network architecture (i.e., 3DMV) which was
originally designed for segmenting RGB-D sensed indoor data. We improved its
performance for outdoor photogrammetric data by introducing a depth pooling
layer in the architecture to take into consideration the distance between the
source images and the reconstructed terrain model. To test the performance of
our improved 3DMV, a ground truth ground material database was created using
data from the One World Terrain (OWT) data repository. Finally, a workflow for
importing the segmented ground materials into a virtual simulation scene was
introduced, and visual results are reported in this paper.
Related papers
- VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection [80.62052650370416]
monocular 3D object detection holds significant importance across various applications, including autonomous driving and robotics.
In this paper, we present VFMM3D, an innovative framework that leverages the capabilities of Vision Foundation Models (VFMs) to accurately transform single-view images into LiDAR point cloud representations.
arXiv Detail & Related papers (2024-04-15T03:12:12Z) - Multi-Modal Dataset Acquisition for Photometrically Challenging Object [56.30027922063559]
This paper addresses the limitations of current datasets for 3D vision tasks in terms of accuracy, size, realism, and suitable imaging modalities for photometrically challenging objects.
We propose a novel annotation and acquisition pipeline that enhances existing 3D perception and 6D object pose datasets.
arXiv Detail & Related papers (2023-08-21T10:38:32Z) - Shape, Pose, and Appearance from a Single Image via Bootstrapped
Radiance Field Inversion [54.151979979158085]
We introduce a principled end-to-end reconstruction framework for natural images, where accurate ground-truth poses are not available.
We leverage an unconditional 3D-aware generator, to which we apply a hybrid inversion scheme where a model produces a first guess of the solution.
Our framework can de-render an image in as few as 10 steps, enabling its use in practical scenarios.
arXiv Detail & Related papers (2022-11-21T17:42:42Z) - Simple and Effective Synthesis of Indoor 3D Scenes [78.95697556834536]
We study the problem of immersive 3D indoor scenes from one or more images.
Our aim is to generate high-resolution images and videos from novel viewpoints.
We propose an image-to-image GAN that maps directly from reprojections of incomplete point clouds to full high-resolution RGB-D images.
arXiv Detail & Related papers (2022-04-06T17:54:46Z) - RandomRooms: Unsupervised Pre-training from Synthetic Shapes and
Randomized Layouts for 3D Object Detection [138.2892824662943]
A promising solution is to make better use of the synthetic dataset, which consists of CAD object models, to boost the learning on real datasets.
Recent work on 3D pre-training exhibits failure when transfer features learned on synthetic objects to other real-world applications.
In this work, we put forward a new method called RandomRooms to accomplish this objective.
arXiv Detail & Related papers (2021-08-17T17:56:12Z) - Unsupervised Learning of 3D Object Categories from Videos in the Wild [75.09720013151247]
We focus on learning a model from multiple views of a large collection of object instances.
We propose a new neural network design, called warp-conditioned ray embedding (WCR), which significantly improves reconstruction.
Our evaluation demonstrates performance improvements over several deep monocular reconstruction baselines on existing benchmarks.
arXiv Detail & Related papers (2021-03-30T17:57:01Z) - Semantic Segmentation and Data Fusion of Microsoft Bing 3D Cities and
Small UAV-based Photogrammetric Data [0.0]
Authors presented a fully automated data segmentation and object information extraction framework for creating simulation terrain using UAV-based photogrammetric data.
Data quality issues in the aircraft-based photogrammetric data are identified.
Authors also proposed a data registration workflow that utilized the traditional iterative closest point (ICP) with the extracted semantic information.
arXiv Detail & Related papers (2020-08-21T18:56:05Z) - Generating synthetic photogrammetric data for training deep learning
based 3D point cloud segmentation models [0.0]
At I/ITSEC 2019, the authors presented a fully-automated workflow to segment 3D photogrammetric point-clouds/meshes and extract object information.
The ultimate goal is to create realistic virtual environments and provide the necessary information for simulation.
arXiv Detail & Related papers (2020-08-21T18:50:42Z) - Fully Automated Photogrammetric Data Segmentation and Object Information
Extraction Approach for Creating Simulation Terrain [0.0]
This research aims to develop a fully automated photogrammetric data segmentation and object information extraction framework.
Considering the use case of the data in creating realistic virtual environments for training and simulations, segmenting the data and extracting object information are essential tasks.
arXiv Detail & Related papers (2020-08-09T09:32:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.