Fisheye Distortion Rectification from Deep Straight Lines
- URL: http://arxiv.org/abs/2003.11386v1
- Date: Wed, 25 Mar 2020 13:20:00 GMT
- Title: Fisheye Distortion Rectification from Deep Straight Lines
- Authors: Zhu-Cun Xue, Nan Xue, Gui-Song Xia
- Abstract summary: We present a novel line-aware rectification network (LaRecNet) to address the problem of fisheye distortion rectification.
Our model achieves state-of-the-art performance in terms of both geometric accuracy and image quality.
In particular, the images rectified by LaRecNet achieve the highest peak signal-to-noise ratio (PSNR) and structure similarity index (SSIM) compared with the groundtruth.
- Score: 34.61402494687801
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper presents a novel line-aware rectification network (LaRecNet) to
address the problem of fisheye distortion rectification based on the classical
observation that straight lines in 3D space should be still straight in image
planes. Specifically, the proposed LaRecNet contains three sequential modules
to (1) learn the distorted straight lines from fisheye images; (2) estimate the
distortion parameters from the learned heatmaps and the image appearance; and
(3) rectify the input images via a proposed differentiable rectification layer.
To better train and evaluate the proposed model, we create a synthetic
line-rich fisheye (SLF) dataset that contains the distortion parameters and
well-annotated distorted straight lines of fisheye images. The proposed method
enables us to simultaneously calibrate the geometric distortion parameters and
rectify fisheye images. Extensive experiments demonstrate that our model
achieves state-of-the-art performance in terms of both geometric accuracy and
image quality on several evaluation metrics. In particular, the images
rectified by LaRecNet achieve an average reprojection error of 0.33 pixels on
the SLF dataset and produce the highest peak signal-to-noise ratio (PSNR) and
structure similarity index (SSIM) compared with the groundtruth.
Related papers
- Mode-GS: Monocular Depth Guided Anchored 3D Gaussian Splatting for Robust Ground-View Scene Rendering [47.879695094904015]
We present a novelview rendering algorithm, Mode-GS, for ground-robot trajectory datasets.
Our approach is based on using anchored Gaussian splats, which are designed to overcome the limitations of existing 3D Gaussian splatting algorithms.
Our method results in improved rendering performance, based on PSNR, SSIM, and LPIPS metrics, in ground scenes with free trajectory patterns.
arXiv Detail & Related papers (2024-10-06T23:01:57Z) - RoFIR: Robust Fisheye Image Rectification Framework Impervious to Optical Center Deviation [88.54817424560056]
We propose a distortion vector map (DVM) that measures the degree and direction of local distortion.
By learning the DVM, the model can independently identify local distortions at each pixel without relying on global distortion patterns.
In the pre-training stage, it predicts the distortion vector map and perceives the local distortion features of each pixel.
In the fine-tuning stage, it predicts a pixel-wise flow map for deviated fisheye image rectification.
arXiv Detail & Related papers (2024-06-27T06:38:56Z) - NeRFDeformer: NeRF Transformation from a Single View via 3D Scene Flows [60.291277312569285]
We present a method for automatically modifying a NeRF representation based on a single observation.
Our method defines the transformation as a 3D flow, specifically as a weighted linear blending of rigid transformations.
We also introduce a new dataset for exploring the problem of modifying a NeRF scene through a single observation.
arXiv Detail & Related papers (2024-06-15T07:58:08Z) - FishRecGAN: An End to End GAN Based Network for Fisheye Rectification
and Calibration [21.816020192280977]
We propose an end-to-end deep learning approach to rectify fisheye images and simultaneously calibrate camera and distortion parameters.
Our solution has achieved robust performance in high-resolution with a significant PSNR value of 22.343.
arXiv Detail & Related papers (2023-05-09T07:38:09Z) - 3D Human Pose and Shape Regression with Pyramidal Mesh Alignment
Feedback Loop [128.07841893637337]
Regression-based methods have recently shown promising results in reconstructing human meshes from monocular images.
Minor deviation in parameters may lead to noticeable misalignment between the estimated meshes and image evidences.
We propose a Pyramidal Mesh Alignment Feedback (PyMAF) loop to leverage a feature pyramid and rectify the predicted parameters.
arXiv Detail & Related papers (2021-03-30T17:07:49Z) - SIR: Self-supervised Image Rectification via Seeing the Same Scene from
Multiple Different Lenses [82.56853587380168]
We propose a novel self-supervised image rectification (SIR) method based on an important insight that the rectified results of distorted images of the same scene from different lens should be the same.
We leverage a differentiable warping module to generate the rectified images and re-distorted images from the distortion parameters.
Our method achieves comparable or even better performance than the supervised baseline method and representative state-of-the-art methods.
arXiv Detail & Related papers (2020-11-30T08:23:25Z) - A Deep Ordinal Distortion Estimation Approach for Distortion Rectification [62.72089758481803]
We propose a novel distortion rectification approach that can obtain more accurate parameters with higher efficiency.
We design a local-global associated estimation network that learns the ordinal distortion to approximate the realistic distortion distribution.
Considering the redundancy of distortion information, our approach only uses a part of distorted image for the ordinal distortion estimation.
arXiv Detail & Related papers (2020-07-21T10:03:42Z) - UnRectDepthNet: Self-Supervised Monocular Depth Estimation using a
Generic Framework for Handling Common Camera Distortion Models [8.484676769284578]
We propose a generic scale-aware self-supervised pipeline for estimating depth, euclidean distance, and visual odometry from unrectified monocular videos.
The proposed algorithm is evaluated further on the KITTI rectified dataset, and we achieve state-of-the-art results.
arXiv Detail & Related papers (2020-07-13T20:35:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.