Controllable Data Augmentation Through Deep Relighting
- URL: http://arxiv.org/abs/2110.13996v1
- Date: Tue, 26 Oct 2021 20:02:51 GMT
- Title: Controllable Data Augmentation Through Deep Relighting
- Authors: George Chogovadze and R\'emi Pautrat and Marc Pollefeys
- Abstract summary: We explore how to augment a varied set of image datasets through relighting so as to improve the ability of existing models to be invariant to illumination changes.
We develop a tool, based on an encoder-decoder network, that is able to quickly generate multiple variations of the illumination of various input scenes.
We demonstrate that by training models on datasets that have been augmented with our pipeline, it is possible to achieve higher performance on localization benchmarks.
- Score: 75.96144853354362
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: At the heart of the success of deep learning is the quality of the data.
Through data augmentation, one can train models with better generalization
capabilities and thus achieve greater results in their field of interest. In
this work, we explore how to augment a varied set of image datasets through
relighting so as to improve the ability of existing models to be invariant to
illumination changes, namely for learned descriptors. We develop a tool, based
on an encoder-decoder network, that is able to quickly generate multiple
variations of the illumination of various input scenes whilst also allowing the
user to define parameters such as the angle of incidence and intensity. We
demonstrate that by training models on datasets that have been augmented with
our pipeline, it is possible to achieve higher performance on localization
benchmarks.
Related papers
- Localized Gaussians as Self-Attention Weights for Point Clouds Correspondence [92.07601770031236]
We investigate semantically meaningful patterns in the attention heads of an encoder-only Transformer architecture.
We find that fixing the attention weights not only accelerates the training process but also enhances the stability of the optimization.
arXiv Detail & Related papers (2024-09-20T07:41:47Z) - A Simple Background Augmentation Method for Object Detection with Diffusion Model [53.32935683257045]
In computer vision, it is well-known that a lack of data diversity will impair model performance.
We propose a simple yet effective data augmentation approach by leveraging advancements in generative models.
Background augmentation, in particular, significantly improves the models' robustness and generalization capabilities.
arXiv Detail & Related papers (2024-08-01T07:40:00Z) - Generalization Gap in Data Augmentation: Insights from Illumination [3.470401787749558]
We investigate the differences in generalization between models trained with augmented data and those trained under real-world illumination conditions.
Results indicate that after applying various data augmentation methods, model performance has significantly improved.
Yet, a noticeable generalization gap still exists after utilizing various data augmentation methods.
arXiv Detail & Related papers (2024-04-11T07:11:43Z) - StableLLaVA: Enhanced Visual Instruction Tuning with Synthesized
Image-Dialogue Data [129.92449761766025]
We propose a novel data collection methodology that synchronously synthesizes images and dialogues for visual instruction tuning.
This approach harnesses the power of generative models, marrying the abilities of ChatGPT and text-to-image generative models.
Our research includes comprehensive experiments conducted on various datasets.
arXiv Detail & Related papers (2023-08-20T12:43:52Z) - Bilevel Fast Scene Adaptation for Low-Light Image Enhancement [50.639332885989255]
Enhancing images in low-light scenes is a challenging but widely concerned task in the computer vision.
Main obstacle lies in the modeling conundrum from distribution discrepancy across different scenes.
We introduce the bilevel paradigm to model the above latent correspondence.
A bilevel learning framework is constructed to endow the scene-irrelevant generality of the encoder towards diverse scenes.
arXiv Detail & Related papers (2023-06-02T08:16:21Z) - Instance-Conditioned GAN Data Augmentation for Representation Learning [29.36473147430433]
We introduce DA_IC-GAN, a learnable data augmentation module that can be used off-the-shelf in conjunction with most state-of-the-art training recipes.
We show that DA_IC-GAN can boost accuracy to between 1%p and 2%p with the highest capacity models.
We additionally couple DA_IC-GAN with a self-supervised training recipe and show that we can also achieve an improvement of 1%p in accuracy in some settings.
arXiv Detail & Related papers (2023-03-16T22:45:43Z) - Learning Representational Invariances for Data-Efficient Action
Recognition [52.23716087656834]
We show that our data augmentation strategy leads to promising performance on the Kinetics-100, UCF-101, and HMDB-51 datasets.
We also validate our data augmentation strategy in the fully supervised setting and demonstrate improved performance.
arXiv Detail & Related papers (2021-03-30T17:59:49Z) - Learned Camera Gain and Exposure Control for Improved Visual Feature
Detection and Matching [12.870196901446208]
We explore a data-driven approach to account for environmental lighting changes, improving the quality of images for use in visual odometry (VO) or visual simultaneous localization and mapping (SLAM)
We train a deep convolutional neural network model to predictively adjust camera gain and exposure time parameters.
We demonstrate through extensive real-world experiments that our network can anticipate and compensate for dramatic lighting changes.
arXiv Detail & Related papers (2021-02-08T16:46:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.