AI-supported Framework of Semi-Automatic Monoplotting for Monocular
Oblique Visual Data Analysis
- URL: http://arxiv.org/abs/2111.14021v1
- Date: Sun, 28 Nov 2021 02:03:43 GMT
- Title: AI-supported Framework of Semi-Automatic Monoplotting for Monocular
Oblique Visual Data Analysis
- Authors: Behzad Golparvar, Ruo-Qian Wang
- Abstract summary: We propose and demonstrate a novel semi-automatic monoplotting framework that provides pixel-level correspondence between photos and Digital Elevation Model (DEM)
A pipeline of analyses was developed including key point detection in images and DEMs, retrieving georeferenced 3D DEMs, regularized pose estimation, gradient-based optimization, and the identification between image pixels and real world coordinates.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In the last decades, the development of smartphones, drones, aerial patrols,
and digital cameras enabled high-quality photographs available to large
populations and, thus, provides an opportunity to collect massive data of the
nature and society with global coverage. However, the data collected with new
photography tools is usually oblique - they are difficult to be georeferenced,
and huge amounts of data is often obsolete. Georeferencing oblique imagery data
may be solved by a technique called monoplotting, which only requires a single
image and Digital Elevation Model (DEM). In traditional monoplotting, a human
user has to manually choose a series of ground control point (GCP) pairs in the
image and DEM and then determine the extrinsic and intrinsic parameters of the
camera to establish a pixel-level correspondence between photos and the DEM to
enable the mapping and georeferencing of objects in photos. This traditional
method is difficult to scale due to several challenges including the
labor-intensive inputs, the need of rich experience to identify well-defined
GCPs, and limitations in camera pose estimation. Therefore, existing
monoplotting methods are rarely used in analyzing large-scale databases or
near-real-time warning systems. In this paper, we propose and demonstrate a
novel semi-automatic monoplotting framework that provides pixel-level
correspondence between photos and DEMs requiring minimal human interventions. A
pipeline of analyses was developed including key point detection in images and
DEM rasters, retrieving georeferenced 3D DEM GCPs, regularized gradient-based
optimization, pose estimation, ray tracing, and the correspondence
identification between image pixels and real world coordinates. Two numerical
experiments show that the framework is superior in georeferencing visual data
in 3-D coordinates, paving a way toward fully automatic monoplotting
methodology.
Related papers
- 3D Human Reconstruction in the Wild with Synthetic Data Using Generative Models [52.96248836582542]
We propose an effective approach based on recent diffusion models, termed HumanWild, which can effortlessly generate human images and corresponding 3D mesh annotations.
By exclusively employing generative models, we generate large-scale in-the-wild human images and high-quality annotations, eliminating the need for real-world data collection.
arXiv Detail & Related papers (2024-03-17T06:31:16Z) - Cross-domain and Cross-dimension Learning for Image-to-Graph
Transformers [50.576354045312115]
Direct image-to-graph transformation is a challenging task that solves object detection and relationship prediction in a single model.
We introduce a set of methods enabling cross-domain and cross-dimension transfer learning for image-to-graph transformers.
We demonstrate our method's utility in cross-domain and cross-dimension experiments, where we pretrain our models on 2D satellite images before applying them to vastly different target domains in 2D and 3D.
arXiv Detail & Related papers (2024-03-11T10:48:56Z) - Towards Hard-pose Virtual Try-on via 3D-aware Global Correspondence
Learning [70.75369367311897]
3D-aware global correspondences are reliable flows that jointly encode global semantic correlations, local deformations, and geometric priors of 3D human bodies.
An adversarial generator takes the garment warped by the 3D-aware flow, and the image of the target person as inputs, to synthesize the photo-realistic try-on result.
arXiv Detail & Related papers (2022-11-25T12:16:21Z) - Towards Model Generalization for Monocular 3D Object Detection [57.25828870799331]
We present an effective unified camera-generalized paradigm (CGP) for Mono3D object detection.
We also propose the 2D-3D geometry-consistent object scaling strategy (GCOS) to bridge the gap via an instance-level augment.
Our method called DGMono3D achieves remarkable performance on all evaluated datasets and surpasses the SoTA unsupervised domain adaptation scheme.
arXiv Detail & Related papers (2022-05-23T23:05:07Z) - Feature matching for multi-epoch historical aerial images [0.0]
We present a fully automatic approach to detecting feature correspondences between historical images taken at different times.
Compared to the state-of-the-art, our method improves the image georeferencing accuracy by a factor of 2.
arXiv Detail & Related papers (2021-12-08T12:28:24Z) - Conditional Generation of Synthetic Geospatial Images from Pixel-level
and Feature-level Inputs [0.0]
We present a conditional generative model, called VAE-Info-cGAN, for synthesizing semantically rich images simultaneously conditioned on a pixel-level condition (PLC) and a feature-level condition (FLC)
The proposed model can accurately generate various forms of macroscopic aggregates across different geographic locations while conditioned only on atemporal representation of the road network.
arXiv Detail & Related papers (2021-09-11T06:58:19Z) - Depth-Aware Multi-Grid Deep Homography Estimation with Contextual
Correlation [38.95610086309832]
Homography estimation is an important task in computer vision, such as image stitching, video stabilization, and camera calibration.
Traditional homography estimation methods depend on the quantity and distribution of feature points, leading to poor robustness in textureless scenes.
We propose a contextual correlation layer, which can capture the long-range correlation on feature maps and flexibly be bridged in a learning framework.
We equip our network with depth perception capability, by introducing a novel depth-aware shape-preserved loss.
arXiv Detail & Related papers (2021-07-06T10:33:12Z) - Self-supervised Human Detection and Segmentation via Multi-view
Consensus [116.92405645348185]
We propose a multi-camera framework in which geometric constraints are embedded in the form of multi-view consistency during training.
We show that our approach outperforms state-of-the-art self-supervised person detection and segmentation techniques on images that visually depart from those of standard benchmarks.
arXiv Detail & Related papers (2020-12-09T15:47:21Z) - VAE-Info-cGAN: Generating Synthetic Images by Combining Pixel-level and
Feature-level Geospatial Conditional Inputs [0.0]
We present a conditional generative model for synthesizing semantically rich images simultaneously conditioned on a pixellevel (PLC) and a featurelevel condition (FLC)
Experiments on a GPS dataset show that the proposed model can accurately generate various forms of macroscopic aggregates across different geographic locations.
arXiv Detail & Related papers (2020-12-08T03:46:19Z) - Single View Metrology in the Wild [94.7005246862618]
We present a novel approach to single view metrology that can recover the absolute scale of a scene represented by 3D heights of objects or camera height above the ground.
Our method relies on data-driven priors learned by a deep network specifically designed to imbibe weakly supervised constraints from the interplay of the unknown camera with 3D entities such as object heights.
We demonstrate state-of-the-art qualitative and quantitative results on several datasets as well as applications including virtual object insertion.
arXiv Detail & Related papers (2020-07-18T22:31:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.