Rethinking Generic Camera Models for Deep Single Image Camera
Calibration to Recover Rotation and Fisheye Distortion
- URL: http://arxiv.org/abs/2111.12927v1
- Date: Thu, 25 Nov 2021 05:58:23 GMT
- Title: Rethinking Generic Camera Models for Deep Single Image Camera
Calibration to Recover Rotation and Fisheye Distortion
- Authors: Nobuhiko Wakai, Satoshi Sato, Yasunori Ishii, Takayoshi Yamashita
- Abstract summary: We propose a generic camera model that has the potential to address various types of distortion.
Our proposed method outperformed conventional methods on two largescale datasets and images captured by off-the-shelf fisheye cameras.
- Score: 8.877834897951578
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although recent learning-based calibration methods can predict extrinsic and
intrinsic camera parameters from a single image, the accuracy of these methods
is degraded in fisheye images. This degradation is caused by mismatching
between the actual projection and expected projection. To address this problem,
we propose a generic camera model that has the potential to address various
types of distortion. Our generic camera model is utilized for learning-based
methods through a closed-form numerical calculation of the camera projection.
Simultaneously to recover rotation and fisheye distortion, we propose a
learning-based calibration method that uses the camera model. Furthermore, we
propose a loss function that alleviates the bias of the magnitude of errors for
four extrinsic and intrinsic camera parameters. Extensive experiments
demonstrated that our proposed method outperformed conventional methods on two
largescale datasets and images captured by off-the-shelf fisheye cameras.
Moreover, we are the first researchers to analyze the performance of
learning-based methods using various types of projection for off-the-shelf
cameras.
Related papers
- RoFIR: Robust Fisheye Image Rectification Framework Impervious to Optical Center Deviation [88.54817424560056]
We propose a distortion vector map (DVM) that measures the degree and direction of local distortion.
By learning the DVM, the model can independently identify local distortions at each pixel without relying on global distortion patterns.
In the pre-training stage, it predicts the distortion vector map and perceives the local distortion features of each pixel.
In the fine-tuning stage, it predicts a pixel-wise flow map for deviated fisheye image rectification.
arXiv Detail & Related papers (2024-06-27T06:38:56Z) - Learning Robust Multi-Scale Representation for Neural Radiance Fields
from Unposed Images [65.41966114373373]
We present an improved solution to the neural image-based rendering problem in computer vision.
The proposed approach could synthesize a realistic image of the scene from a novel viewpoint at test time.
arXiv Detail & Related papers (2023-11-08T08:18:23Z) - How to turn your camera into a perfect pinhole model [0.38233569758620056]
We propose a novel approach that involves a pre-processing step to remove distortions from images.
Our method does not need to assume any distortion model and can be applied to severely warped images.
This model allows for a serious upgrade of many algorithms and applications.
arXiv Detail & Related papers (2023-09-20T13:54:29Z) - Deep Learning for Camera Calibration and Beyond: A Survey [100.75060862015945]
Camera calibration involves estimating camera parameters to infer geometric features from captured sequences.
Recent efforts show that learning-based solutions have the potential to be used in place of the repeatability works of manual calibrations.
arXiv Detail & Related papers (2023-03-19T04:00:05Z) - Multi-task Learning for Camera Calibration [3.274290296343038]
We present a unique method for predicting intrinsic (principal point offset and focal length) and extrinsic (baseline, pitch, and translation) properties from a pair of images.
By reconstructing the 3D points using a camera model neural network and then using the loss in reconstruction to obtain the camera specifications, this innovative camera projection loss (CPL) method allows us that the desired parameters should be estimated.
arXiv Detail & Related papers (2022-11-22T17:39:31Z) - TartanCalib: Iterative Wide-Angle Lens Calibration using Adaptive
SubPixel Refinement of AprilTags [23.568127229446965]
Calibrating wide-angle lenses with current state-of-the-art techniques yields poor results due to extreme distortion at the edge.
We present our methodology for accurate wide-angle calibration.
arXiv Detail & Related papers (2022-10-05T18:57:07Z) - Self-Supervised Camera Self-Calibration from Video [34.35533943247917]
We propose a learning algorithm to regress per-sequence calibration parameters using an efficient family of general camera models.
Our procedure achieves self-calibration results with sub-pixel reprojection error, outperforming other learning-based methods.
arXiv Detail & Related papers (2021-12-06T19:42:05Z) - Self-Calibrating Neural Radiance Fields [68.64327335620708]
We jointly learn the geometry of the scene and the accurate camera parameters without any calibration objects.
Our camera model consists of a pinhole model, a fourth order radial distortion, and a generic noise model that can learn arbitrary non-linear camera distortions.
arXiv Detail & Related papers (2021-08-31T13:34:28Z) - Wide-angle Image Rectification: A Survey [86.36118799330802]
wide-angle images contain distortions that violate the assumptions underlying pinhole camera models.
Image rectification, which aims to correct these distortions, can solve these problems.
We present a detailed description and discussion of the camera models used in different approaches.
Next, we review both traditional geometry-based image rectification methods and deep learning-based methods.
arXiv Detail & Related papers (2020-10-30T17:28:40Z) - Neural Ray Surfaces for Self-Supervised Learning of Depth and Ego-motion [51.19260542887099]
We show that self-supervision can be used to learn accurate depth and ego-motion estimation without prior knowledge of the camera model.
Inspired by the geometric model of Grossberg and Nayar, we introduce Neural Ray Surfaces (NRS), convolutional networks that represent pixel-wise projection rays.
We demonstrate the use of NRS for self-supervised learning of visual odometry and depth estimation from raw videos obtained using a wide variety of camera systems.
arXiv Detail & Related papers (2020-08-15T02:29:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.