IMU-Assisted Learning of Single-View Rolling Shutter Correction
- URL: http://arxiv.org/abs/2011.03106v2
- Date: Tue, 14 Sep 2021 14:56:02 GMT
- Title: IMU-Assisted Learning of Single-View Rolling Shutter Correction
- Authors: Jiawei Mo, Md Jahidul Islam, Junaed Sattar
- Abstract summary: Rolling shutter distortion is highly undesirable for photography and computer vision algorithms.
We propose a deep neural network to predict depth and row-wise pose from a single image for rolling shutter correction.
- Score: 16.242924916178282
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Rolling shutter distortion is highly undesirable for photography and computer
vision algorithms (e.g., visual SLAM) because pixels can be potentially
captured at different times and poses. In this paper, we propose a deep neural
network to predict depth and row-wise pose from a single image for rolling
shutter correction. Our contribution in this work is to incorporate inertial
measurement unit (IMU) data into the pose refinement process, which, compared
to the state-of-the-art, greatly enhances the pose prediction. The improved
accuracy and robustness make it possible for numerous vision algorithms to use
imagery captured by rolling shutter cameras and produce highly accurate
results. We also extend a dataset to have real rolling shutter images, IMU
data, depth maps, camera poses, and corresponding global shutter images for
rolling shutter correction training. We demonstrate the efficacy of the
proposed method by evaluating the performance of Direct Sparse Odometry (DSO)
algorithm on rolling shutter imagery corrected using the proposed approach.
Results show marked improvements of the DSO algorithm over using uncorrected
imagery, validating the proposed approach.
Related papers
- Cameras as Rays: Pose Estimation via Ray Diffusion [54.098613859015856]
Estimating camera poses is a fundamental task for 3D reconstruction and remains challenging given sparsely sampled views.
We propose a distributed representation of camera pose that treats a camera as a bundle of rays.
Our proposed methods, both regression- and diffusion-based, demonstrate state-of-the-art performance on camera pose estimation on CO3D.
arXiv Detail & Related papers (2024-02-22T18:59:56Z) - iComMa: Inverting 3D Gaussian Splatting for Camera Pose Estimation via Comparing and Matching [14.737266480464156]
We present a method named iComMa to address the 6D camera pose estimation problem in computer vision.
We propose an efficient method for accurate camera pose estimation by inverting 3D Gaussian Splatting (3DGS)
arXiv Detail & Related papers (2023-12-14T15:31:33Z) - Single Image Depth Prediction Made Better: A Multivariate Gaussian Take [163.14849753700682]
We introduce an approach that performs continuous modeling of per-pixel depth.
Our method's accuracy (named MG) is among the top on the KITTI depth-prediction benchmark leaderboard.
arXiv Detail & Related papers (2023-03-31T16:01:03Z) - Robust Camera Pose Refinement for Multi-Resolution Hash Encoding [19.655282240361835]
Multi-resolution hash encoding has been proposed to reduce the computational cost of neural renderings.
We propose a joint optimization algorithm to calibrate the camera pose and learn a geometric representation using efficient multi-resolution hash encoding.
arXiv Detail & Related papers (2023-02-03T06:49:27Z) - Learning Markerless Robot-Depth Camera Calibration and End-Effector Pose
Estimation [0.0]
We present a learning-based markerless extrinsic calibration system that uses a depth camera and does not rely on simulation data.
We learn models for end-effector (EE) segmentation, single-frame rotation prediction and keypoint detection, from automatically generated real-world data.
Our robustness with training data from multiple camera poses and test data from previously unseen poses give sub-centimeter evaluations and sub-deciradian average calibration and pose estimation errors.
arXiv Detail & Related papers (2022-12-15T00:53:42Z) - DeepRM: Deep Recurrent Matching for 6D Pose Refinement [77.34726150561087]
DeepRM is a novel recurrent network architecture for 6D pose refinement.
The architecture incorporates LSTM units to propagate information through each refinement step.
DeepRM achieves state-of-the-art performance on two widely accepted challenging datasets.
arXiv Detail & Related papers (2022-05-28T16:18:08Z) - Self-Supervised Camera Self-Calibration from Video [34.35533943247917]
We propose a learning algorithm to regress per-sequence calibration parameters using an efficient family of general camera models.
Our procedure achieves self-calibration results with sub-pixel reprojection error, outperforming other learning-based methods.
arXiv Detail & Related papers (2021-12-06T19:42:05Z) - An Adaptive Framework for Learning Unsupervised Depth Completion [59.17364202590475]
We present a method to infer a dense depth map from a color image and associated sparse depth measurements.
We show that regularization and co-visibility are related via the fitness of the model to data and can be unified into a single framework.
arXiv Detail & Related papers (2021-06-06T02:27:55Z) - Real-Time, Deep Synthetic Aperture Sonar (SAS) Autofocus [34.77467193499518]
Synthetic aperture sonar (SAS) requires precise time-of-flight measurements of the transmitted/received waveform to produce well-focused imagery.
To overcome this, an emphautofocus algorithm is employed as a post-processing step after image reconstruction to improve image focus.
We propose a deep learning technique to overcome these limitations and implicitly learn the weighting function in a data-driven manner.
arXiv Detail & Related papers (2021-03-18T15:16:29Z) - Robust Consistent Video Depth Estimation [65.53308117778361]
We present an algorithm for estimating consistent dense depth maps and camera poses from a monocular video.
Our algorithm combines two complementary techniques: (1) flexible deformation-splines for low-frequency large-scale alignment and (2) geometry-aware depth filtering for high-frequency alignment of fine depth details.
In contrast to prior approaches, our method does not require camera poses as input and achieves robust reconstruction for challenging hand-held cell phone captures containing a significant amount of noise, shake, motion blur, and rolling shutter deformations.
arXiv Detail & Related papers (2020-12-10T18:59:48Z) - Self-Supervised Linear Motion Deblurring [112.75317069916579]
Deep convolutional neural networks are state-of-the-art for image deblurring.
We present a differentiable reblur model for self-supervised motion deblurring.
Our experiments demonstrate that self-supervised single image deblurring is really feasible.
arXiv Detail & Related papers (2020-02-10T20:15:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.