Sim2Air - Synthetic aerial dataset for UAV monitoring
- URL: http://arxiv.org/abs/2110.05145v1
- Date: Mon, 11 Oct 2021 10:36:33 GMT
- Title: Sim2Air - Synthetic aerial dataset for UAV monitoring
- Authors: Antonella Barisic and Frano Petric and Stjepan Bogdan
- Abstract summary: We propose to accentuate shape-based object representation by applying texture randomization.
A diverse dataset with photorealism in all parameters is created in a 3D modelling software Blender.
- Score: 2.1638817206926855
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper we propose a novel approach to generate a synthetic aerial
dataset for application in UAV monitoring. We propose to accentuate shape-based
object representation by applying texture randomization. A diverse dataset with
photorealism in all parameters such as shape, pose, lighting, scale, viewpoint,
etc. except for atypical textures is created in a 3D modelling software
Blender. Our approach specifically targets two conditions in aerial images
where texture of objects is difficult to detect, namely illumination changes
and objects occupying only a small portion of the image. Experimental
evaluation confirmed our approach by increasing the mAP value by 17 and 3.7
points on two test datasets of real images. In analysing domain similarity, we
conclude that the more the generalisation capability is put to the test, the
more obvious are the advantages of the shape-based representation.
Related papers
- PT43D: A Probabilistic Transformer for Generating 3D Shapes from Single Highly-Ambiguous RGB Images [26.900974153235456]
We propose a transformer-based autoregressive model to generate 3D shapes from single RGB images.
To handle realistic scenarios such as field-of-view truncation, we create simulated image-to-shape training pairs.
We then adopt cross-attention to effectively identify the most relevant region of interest from the input image for shape generation.
arXiv Detail & Related papers (2024-05-20T09:49:13Z) - VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection [80.62052650370416]
monocular 3D object detection holds significant importance across various applications, including autonomous driving and robotics.
In this paper, we present VFMM3D, an innovative framework that leverages the capabilities of Vision Foundation Models (VFMs) to accurately transform single-view images into LiDAR point cloud representations.
arXiv Detail & Related papers (2024-04-15T03:12:12Z) - Sim2Real Bilevel Adaptation for Object Surface Classification using Vision-Based Tactile Sensors [14.835051543002164]
We train a Diffusion Model to bridge the Sim2Real gap in the field of vision-based tactile sensors for classifying object surfaces.
We employ a simulator to generate images by uniformly sampling the surface of objects from the YCB Model Set.
These simulated images are then translated into the real domain using the Diffusion Model and automatically labeled to train a classifier.
arXiv Detail & Related papers (2023-11-02T16:37:27Z) - Relightify: Relightable 3D Faces from a Single Image via Diffusion
Models [86.3927548091627]
We present the first approach to use diffusion models as a prior for highly accurate 3D facial BRDF reconstruction from a single image.
In contrast to existing methods, we directly acquire the observed texture from the input image, thus, resulting in more faithful and consistent estimation.
arXiv Detail & Related papers (2023-05-10T11:57:49Z) - Learning to Generate 3D Representations of Building Roofs Using
Single-View Aerial Imagery [68.3565370706598]
We present a novel pipeline for learning the conditional distribution of a building roof mesh given pixels from an aerial image.
Unlike alternative methods that require multiple images of the same object, our approach enables estimating 3D roof meshes using only a single image for predictions.
arXiv Detail & Related papers (2023-03-20T15:47:05Z) - Shape, Pose, and Appearance from a Single Image via Bootstrapped
Radiance Field Inversion [54.151979979158085]
We introduce a principled end-to-end reconstruction framework for natural images, where accurate ground-truth poses are not available.
We leverage an unconditional 3D-aware generator, to which we apply a hybrid inversion scheme where a model produces a first guess of the solution.
Our framework can de-render an image in as few as 10 steps, enabling its use in practical scenarios.
arXiv Detail & Related papers (2022-11-21T17:42:42Z) - Ground material classification and for UAV-based photogrammetric 3D data
A 2D-3D Hybrid Approach [1.3359609092684614]
In recent years, photogrammetry has been widely used in many areas to create 3D virtual data representing the physical environment.
These cutting-edge technologies have caught the US Army and Navy's attention for the purpose of rapid 3D battlefield reconstruction, virtual training, and simulations.
arXiv Detail & Related papers (2021-09-24T22:29:26Z) - FiG-NeRF: Figure-Ground Neural Radiance Fields for 3D Object Category
Modelling [11.432178728985956]
We use Neural Radiance Fields (NeRF) to learn high quality 3D object category models from collections of input images.
We show that this method can learn accurate 3D object category models using only photometric supervision and casually captured images.
arXiv Detail & Related papers (2021-04-17T01:38:54Z) - ShaRF: Shape-conditioned Radiance Fields from a Single View [54.39347002226309]
We present a method for estimating neural scenes representations of objects given only a single image.
The core of our method is the estimation of a geometric scaffold for the object.
We demonstrate in several experiments the effectiveness of our approach in both synthetic and real images.
arXiv Detail & Related papers (2021-02-17T16:40:28Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.