SharinGAN: Combining Synthetic and Real Data for Unsupervised Geometry
Estimation
- URL: http://arxiv.org/abs/2006.04026v1
- Date: Sun, 7 Jun 2020 02:45:33 GMT
- Title: SharinGAN: Combining Synthetic and Real Data for Unsupervised Geometry
Estimation
- Authors: Koutilya PNVR, Hao Zhou, David Jacobs
- Abstract summary: We propose a novel method for combining synthetic and real images when training networks.
We suggest a method for mapping both image types into a single, shared domain.
Our experiments demonstrate significant improvements over the state-of-the-art in two important domains.
- Score: 18.29202999419042
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a novel method for combining synthetic and real images when
training networks to determine geometric information from a single image. We
suggest a method for mapping both image types into a single, shared domain.
This is connected to a primary network for end-to-end training. Ideally, this
results in images from two domains that present shared information to the
primary network. Our experiments demonstrate significant improvements over the
state-of-the-art in two important domains, surface normal estimation of human
faces and monocular depth estimation for outdoor scenes, both in an
unsupervised setting.
Related papers
- Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - Self-Supervised Image Representation Learning with Geometric Set
Consistency [50.12720780102395]
We propose a method for self-supervised image representation learning under the guidance of 3D geometric consistency.
Specifically, we introduce 3D geometric consistency into a contrastive learning framework to enforce the feature consistency within image views.
arXiv Detail & Related papers (2022-03-29T08:57:33Z) - NeuralReshaper: Single-image Human-body Retouching with Deep Neural
Networks [50.40798258968408]
We present NeuralReshaper, a novel method for semantic reshaping of human bodies in single images using deep generative networks.
Our approach follows a fit-then-reshape pipeline, which first fits a parametric 3D human model to a source human image.
To deal with the lack-of-data problem that no paired data exist, we introduce a novel self-supervised strategy to train our network.
arXiv Detail & Related papers (2022-03-20T09:02:13Z) - Self-Supervised Learning of Domain Invariant Features for Depth
Estimation [35.74969527929284]
We tackle the problem of unsupervised synthetic-to-realistic domain adaptation for single image depth estimation.
An essential building block of single image depth estimation is an encoder-decoder task network that takes RGB images as input and produces depth maps as output.
We propose a novel training strategy to force the task network to learn domain invariant representations in a self-supervised manner.
arXiv Detail & Related papers (2021-06-04T16:45:48Z) - ShaRF: Shape-conditioned Radiance Fields from a Single View [54.39347002226309]
We present a method for estimating neural scenes representations of objects given only a single image.
The core of our method is the estimation of a geometric scaffold for the object.
We demonstrate in several experiments the effectiveness of our approach in both synthetic and real images.
arXiv Detail & Related papers (2021-02-17T16:40:28Z) - SOSD-Net: Joint Semantic Object Segmentation and Depth Estimation from
Monocular images [94.36401543589523]
We introduce the concept of semantic objectness to exploit the geometric relationship of these two tasks.
We then propose a Semantic Object and Depth Estimation Network (SOSD-Net) based on the objectness assumption.
To the best of our knowledge, SOSD-Net is the first network that exploits the geometry constraint for simultaneous monocular depth estimation and semantic segmentation.
arXiv Detail & Related papers (2021-01-19T02:41:03Z) - From Real to Synthetic and Back: Synthesizing Training Data for
Multi-Person Scene Understanding [0.7519872646378835]
We present a method for synthesizing naturally looking images of multiple people interacting in a specific scenario.
These images benefit from the advantages of synthetic data: being fully controllable and fully annotated with any type of standard or custom-defined ground truth.
To reduce the synthetic-to-real domain gap, we introduce a pipeline consisting of the following steps.
arXiv Detail & Related papers (2020-06-03T09:02:06Z) - Two-shot Spatially-varying BRDF and Shape Estimation [89.29020624201708]
We propose a novel deep learning architecture with a stage-wise estimation of shape and SVBRDF.
We create a large-scale synthetic training dataset with domain-randomized geometry and realistic materials.
Experiments on both synthetic and real-world datasets show that our network trained on a synthetic dataset can generalize well to real-world images.
arXiv Detail & Related papers (2020-04-01T12:56:13Z) - Introducing Pose Consistency and Warp-Alignment for Self-Supervised 6D
Object Pose Estimation in Color Images [38.9238085806793]
Most successful approaches to estimate the 6D pose of an object typically train a neural network by supervising the learning with annotated poses in real world images.
A two-stage 6D object pose estimator framework that can be applied on top of existing neural-network-based approaches is proposed.
arXiv Detail & Related papers (2020-03-27T11:53:38Z) - An End-to-End Network for Co-Saliency Detection in One Single Image [47.35448093528382]
Co-saliency detection within a single image is a common vision problem that has not yet been well addressed.
This study proposes a novel end-to-end trainable network comprising a backbone net and two branch nets.
We construct a new dataset of 2,019 natural images with co-saliency in each image to evaluate the proposed method.
arXiv Detail & Related papers (2019-10-25T16:00:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.