Domain Decluttering: Simplifying Images to Mitigate Synthetic-Real
Domain Shift and Improve Depth Estimation
- URL: http://arxiv.org/abs/2002.12114v2
- Date: Thu, 25 Jun 2020 07:37:13 GMT
- Title: Domain Decluttering: Simplifying Images to Mitigate Synthetic-Real
Domain Shift and Improve Depth Estimation
- Authors: Yunhan Zhao, Shu Kong, Daeyun Shin, Charless Fowlkes
- Abstract summary: We develop an attention module that learns to identify and remove difficult out-of-domain regions in real images.
Visualizing the removed regions provides interpretable insights into the synthetic-real domain gap.
- Score: 16.153683223016973
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Leveraging synthetically rendered data offers great potential to improve
monocular depth estimation and other geometric estimation tasks, but closing
the synthetic-real domain gap is a non-trivial and important task. While much
recent work has focused on unsupervised domain adaptation, we consider a more
realistic scenario where a large amount of synthetic training data is
supplemented by a small set of real images with ground-truth. In this setting,
we find that existing domain translation approaches are difficult to train and
offer little advantage over simple baselines that use a mix of real and
synthetic data. A key failure mode is that real-world images contain novel
objects and clutter not present in synthetic training. This high-level domain
shift isn't handled by existing image translation models.
Based on these observations, we develop an attention module that learns to
identify and remove difficult out-of-domain regions in real images in order to
improve depth prediction for a model trained primarily on synthetic data. We
carry out extensive experiments to validate our attend-remove-complete approach
(ARC) and find that it significantly outperforms state-of-the-art domain
adaptation methods for depth prediction. Visualizing the removed regions
provides interpretable insights into the synthetic-real domain gap.
Related papers
- Denoising as Adaptation: Noise-Space Domain Adaptation for Image Restoration [64.84134880709625]
We show that it is possible to perform domain adaptation via the noise space using diffusion models.
In particular, by leveraging the unique property of how auxiliary conditional inputs influence the multi-step denoising process, we derive a meaningful diffusion loss.
We present crucial strategies such as channel-shuffling layer and residual-swapping contrastive learning in the diffusion model.
arXiv Detail & Related papers (2024-06-26T17:40:30Z) - Deep Domain Adaptation: A Sim2Real Neural Approach for Improving Eye-Tracking Systems [80.62854148838359]
Eye image segmentation is a critical step in eye tracking that has great influence over the final gaze estimate.
We use dimensionality-reduction techniques to measure the overlap between the target eye images and synthetic training data.
Our methods result in robust, improved performance when tackling the discrepancy between simulation and real-world data samples.
arXiv Detail & Related papers (2024-03-23T22:32:06Z) - Domain-Adaptive Full-Face Gaze Estimation via Novel-View-Synthesis and Feature Disentanglement [12.857137513211866]
We propose an effective model training pipeline consisting of a training data synthesis and a gaze estimation model for unsupervised domain adaptation.
The proposed data synthesis leverages the single-image 3D reconstruction to expand the range of the head poses from the source domain without requiring a 3D facial shape dataset.
We propose a disentangling autoencoder network to separate gaze-related features and introduce background augmentation consistency loss to utilize the characteristics of the synthetic source domain.
arXiv Detail & Related papers (2023-05-25T15:15:03Z) - Domain Adaptation of Synthetic Driving Datasets for Real-World
Autonomous Driving [0.11470070927586014]
Network trained with synthetic data for certain computer vision tasks degrade significantly when tested on real world data.
In this paper, we propose and evaluate novel ways for the betterment of such approaches.
We propose a novel method to efficiently incorporate semantic supervision into this pair selection, which helps in boosting the performance of the model.
arXiv Detail & Related papers (2023-02-08T15:51:54Z) - One-Shot Domain Adaptive and Generalizable Semantic Segmentation with
Class-Aware Cross-Domain Transformers [96.51828911883456]
Unsupervised sim-to-real domain adaptation (UDA) for semantic segmentation aims to improve the real-world test performance of a model trained on simulated data.
Traditional UDA often assumes that there are abundant unlabeled real-world data samples available during training for the adaptation.
We explore the one-shot unsupervised sim-to-real domain adaptation (OSUDA) and generalization problem, where only one real-world data sample is available.
arXiv Detail & Related papers (2022-12-14T15:54:15Z) - Domain Adaptation for Underwater Image Enhancement [51.71570701102219]
We propose a novel Two-phase Underwater Domain Adaptation network (TUDA) to minimize the inter-domain and intra-domain gap.
In the first phase, a new dual-alignment network is designed, including a translation part for enhancing realism of input images, followed by an enhancement part.
In the second phase, we perform an easy-hard classification of real data according to the assessed quality of enhanced images, where a rank-based underwater quality assessment method is embedded.
arXiv Detail & Related papers (2021-08-22T06:38:19Z) - Occlusion-aware Unsupervised Learning of Depth from 4-D Light Fields [50.435129905215284]
We present an unsupervised learning-based depth estimation method for 4-D light field processing and analysis.
Based on the basic knowledge of the unique geometry structure of light field data, we explore the angular coherence among subsets of the light field views to estimate depth maps.
Our method can significantly shrink the performance gap between the previous unsupervised method and supervised ones, and produce depth maps with comparable accuracy to traditional methods with obviously reduced computational cost.
arXiv Detail & Related papers (2021-06-06T06:19:50Z) - Domain Adaptive Monocular Depth Estimation With Semantic Information [13.387521845596149]
We propose an adversarial training model that leverages semantic information to narrow the domain gap.
The proposed compact model achieves state-of-the-art performance comparable to complex latest models.
arXiv Detail & Related papers (2021-04-12T18:50:41Z) - Unsupervised Metric Relocalization Using Transform Consistency Loss [66.19479868638925]
Training networks to perform metric relocalization traditionally requires accurate image correspondences.
We propose a self-supervised solution, which exploits a key insight: localizing a query image within a map should yield the same absolute pose, regardless of the reference image used for registration.
We evaluate our framework on synthetic and real-world data, showing our approach outperforms other supervised methods when a limited amount of ground-truth information is available.
arXiv Detail & Related papers (2020-11-01T19:24:27Z) - Focus on defocus: bridging the synthetic to real domain gap for depth
estimation [9.023847175654602]
We tackle the issue of closing the synthetic-real domain gap by using domain invariant defocus blur as direct supervision.
We leverage defocus cues by using a permutation invariant convolutional neural network that encourages the network to learn from the differences between images with a different point of focus.
We are able to train our model completely on synthetic data and directly apply it to a wide range of real-world images.
arXiv Detail & Related papers (2020-05-19T17:52:37Z) - Focus on Semantic Consistency for Cross-domain Crowd Understanding [34.560447389853614]
Some domain adaptation algorithms try to liberate it by training models with synthetic data.
We found that a mass of estimation errors in the background areas impede the performance of the existing methods.
In this paper, we propose a domain adaptation method to eliminate it.
arXiv Detail & Related papers (2020-02-20T08:51:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.