Learned Camera Gain and Exposure Control for Improved Visual Feature
Detection and Matching
- URL: http://arxiv.org/abs/2102.04341v1
- Date: Mon, 8 Feb 2021 16:46:09 GMT
- Title: Learned Camera Gain and Exposure Control for Improved Visual Feature
Detection and Matching
- Authors: Justin Tomasi, Brandon Wagstaff, Steven L. Waslander, Jonathan Kelly
- Abstract summary: We explore a data-driven approach to account for environmental lighting changes, improving the quality of images for use in visual odometry (VO) or visual simultaneous localization and mapping (SLAM)
We train a deep convolutional neural network model to predictively adjust camera gain and exposure time parameters.
We demonstrate through extensive real-world experiments that our network can anticipate and compensate for dramatic lighting changes.
- Score: 12.870196901446208
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Successful visual navigation depends upon capturing images that contain
sufficient useful information. In this paper, we explore a data-driven approach
to account for environmental lighting changes, improving the quality of images
for use in visual odometry (VO) or visual simultaneous localization and mapping
(SLAM). We train a deep convolutional neural network model to predictively
adjust camera gain and exposure time parameters such that consecutive images
contain a maximal number of matchable features. The training process is fully
self-supervised: our training signal is derived from an underlying VO or SLAM
pipeline and, as a result, the model is optimized to perform well with that
specific pipeline. We demonstrate through extensive real-world experiments that
our network can anticipate and compensate for dramatic lighting changes (e.g.,
transitions into and out of road tunnels), maintaining a substantially higher
number of inlier feature matches than competing camera parameter control
algorithms.
Related papers
- Efficient Camera Exposure Control for Visual Odometry via Deep Reinforcement Learning [10.886819238167286]
This study employs a deep reinforcement learning framework to train agents for exposure control.
A lightweight image simulator is developed to facilitate the training process.
Different levels of reward functions are crafted to enhance the VO systems.
arXiv Detail & Related papers (2024-08-30T04:37:52Z) - I$^2$-SLAM: Inverting Imaging Process for Robust Photorealistic Dense SLAM [10.464532720114052]
Casual video captures often suffer from motion blur and varying appearances, which degrade the final quality of coherent 3D visual representation.
We propose integrating the physical imaging into the SLAM system, which employs linear HDR radiance maps to collect measurements.
arXiv Detail & Related papers (2024-07-16T03:31:33Z) - Efficient Visual State Space Model for Image Deblurring [83.57239834238035]
Convolutional neural networks (CNNs) and Vision Transformers (ViTs) have achieved excellent performance in image restoration.
We propose a simple yet effective visual state space model (EVSSM) for image deblurring.
arXiv Detail & Related papers (2024-05-23T09:13:36Z) - A Non-Uniform Low-Light Image Enhancement Method with Multi-Scale
Attention Transformer and Luminance Consistency Loss [11.585269110131659]
Low-light image enhancement aims to improve the perception of images collected in dim environments.
Existing methods cannot adaptively extract the differentiated luminance information, which will easily cause over-exposure and under-exposure.
We propose a multi-scale attention Transformer named MSATr, which sufficiently extracts local and global features for light balance to improve the visual quality.
arXiv Detail & Related papers (2023-12-27T10:07:11Z) - SONIC: Sonar Image Correspondence using Pose Supervised Learning for Imaging Sonars [21.36164161799472]
We introduce SONIC, a pose-supervised network designed to yield robust feature correspondence capable of withstanding viewpoint variations.
Our method demonstrates significantly better performance in generating correspondences for sonar images.
arXiv Detail & Related papers (2023-10-23T15:21:46Z) - Improving Lens Flare Removal with General Purpose Pipeline and Multiple
Light Sources Recovery [69.71080926778413]
flare artifacts can affect image visual quality and downstream computer vision tasks.
Current methods do not consider automatic exposure and tone mapping in image signal processing pipeline.
We propose a solution to improve the performance of lens flare removal by revisiting the ISP and design a more reliable light sources recovery strategy.
arXiv Detail & Related papers (2023-08-31T04:58:17Z) - Ultra-High-Definition Low-Light Image Enhancement: A Benchmark and
Transformer-Based Method [51.30748775681917]
We consider the task of low-light image enhancement (LLIE) and introduce a large-scale database consisting of images at 4K and 8K resolution.
We conduct systematic benchmarking studies and provide a comparison of current LLIE algorithms.
As a second contribution, we introduce LLFormer, a transformer-based low-light enhancement method.
arXiv Detail & Related papers (2022-12-22T09:05:07Z) - Spatio-Temporal Outdoor Lighting Aggregation on Image Sequences using
Transformer Networks [23.6427456783115]
In this work, we focus on outdoor lighting estimation by aggregating individual noisy estimates from images.
Recent work based on deep neural networks has shown promising results for single image lighting estimation, but suffers from robustness.
We tackle this problem by combining lighting estimates from several image views sampled in the angular and temporal domain of an image sequence.
arXiv Detail & Related papers (2022-02-18T14:11:16Z) - Controllable Data Augmentation Through Deep Relighting [75.96144853354362]
We explore how to augment a varied set of image datasets through relighting so as to improve the ability of existing models to be invariant to illumination changes.
We develop a tool, based on an encoder-decoder network, that is able to quickly generate multiple variations of the illumination of various input scenes.
We demonstrate that by training models on datasets that have been augmented with our pipeline, it is possible to achieve higher performance on localization benchmarks.
arXiv Detail & Related papers (2021-10-26T20:02:51Z) - TransCamP: Graph Transformer for 6-DoF Camera Pose Estimation [77.09542018140823]
We propose a neural network approach with a graph transformer backbone, namely TransCamP, to address the camera relocalization problem.
TransCamP effectively fuses the image features, camera pose information and inter-frame relative camera motions into encoded graph attributes.
arXiv Detail & Related papers (2021-05-28T19:08:43Z) - Learning Monocular Dense Depth from Events [53.078665310545745]
Event cameras produce brightness changes in the form of a stream of asynchronous events instead of intensity frames.
Recent learning-based approaches have been applied to event-based data, such as monocular depth prediction.
We propose a recurrent architecture to solve this task and show significant improvement over standard feed-forward methods.
arXiv Detail & Related papers (2020-10-16T12:36:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.