Tame a Wild Camera: In-the-Wild Monocular Camera Calibration
- URL: http://arxiv.org/abs/2306.10988v2
- Date: Wed, 22 Nov 2023 19:57:00 GMT
- Title: Tame a Wild Camera: In-the-Wild Monocular Camera Calibration
- Authors: Shengjie Zhu, Abhinav Kumar, Masa Hu and Xiaoming Liu
- Abstract summary: Previous methods for the monocular camera calibration rely on specific 3D objects or strong geometry prior.
Our method is assumption-free and calibrates the complete $4$ Degree-of-Freedom (DoF) intrinsic parameters.
We demonstrate downstream applications in image manipulation detection & restoration, uncalibrated two-view pose estimation, and 3D sensing.
- Score: 12.55056916519563
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: 3D sensing for monocular in-the-wild images, e.g., depth estimation and 3D
object detection, has become increasingly important. However, the unknown
intrinsic parameter hinders their development and deployment. Previous methods
for the monocular camera calibration rely on specific 3D objects or strong
geometry prior, such as using a checkerboard or imposing a Manhattan World
assumption. This work solves the problem from the other perspective by
exploiting the monocular 3D prior. Our method is assumption-free and calibrates
the complete $4$ Degree-of-Freedom (DoF) intrinsic parameters. First, we
demonstrate intrinsic is solved from two well-studied monocular priors, i.e.,
monocular depthmap, and surface normal map. However, this solution imposes a
low-bias and low-variance requirement for depth estimation. Alternatively, we
introduce a novel monocular 3D prior, the incidence field, defined as the
incidence rays between points in 3D space and pixels in the 2D imaging plane.
The incidence field is a pixel-wise parametrization of the intrinsic invariant
to image cropping and resizing. With the estimated incidence field, a robust
RANSAC algorithm recovers intrinsic. We demonstrate the effectiveness of our
method by showing superior performance on synthetic and zero-shot testing
datasets. Beyond calibration, we demonstrate downstream applications in image
manipulation detection & restoration, uncalibrated two-view pose estimation,
and 3D sensing. Codes, models, and data will be held in
https://github.com/ShngJZ/WildCamera.
Related papers
- Zolly: Zoom Focal Length Correctly for Perspective-Distorted Human Mesh
Reconstruction [66.10717041384625]
Zolly is the first 3DHMR method focusing on perspective-distorted images.
We propose a new camera model and a novel 2D representation, termed distortion image, which describes the 2D dense distortion scale of the human body.
We extend two real-world datasets tailored for this task, all containing perspective-distorted human images.
arXiv Detail & Related papers (2023-03-24T04:22:41Z) - Monocular 3D Object Detection with Depth from Motion [74.29588921594853]
We take advantage of camera ego-motion for accurate object depth estimation and detection.
Our framework, named Depth from Motion (DfM), then uses the established geometry to lift 2D image features to the 3D space and detects 3D objects thereon.
Our framework outperforms state-of-the-art methods by a large margin on the KITTI benchmark.
arXiv Detail & Related papers (2022-07-26T15:48:46Z) - Homography Loss for Monocular 3D Object Detection [54.04870007473932]
A differentiable loss function, termed as Homography Loss, is proposed to achieve the goal, which exploits both 2D and 3D information.
Our method yields the best performance compared with the other state-of-the-arts by a large margin on KITTI 3D datasets.
arXiv Detail & Related papers (2022-04-02T03:48:03Z) - Ray3D: ray-based 3D human pose estimation for monocular absolute 3D
localization [3.5379706873065917]
We propose a novel monocular ray-based 3D (Ray3D) absolute human pose estimation with calibrated camera.
Our method significantly outperforms existing state-of-the-art models.
arXiv Detail & Related papers (2022-03-22T05:42:31Z) - Probabilistic and Geometric Depth: Detecting Objects in Perspective [78.00922683083776]
3D object detection is an important capability needed in various practical applications such as driver assistance systems.
Monocular 3D detection, as an economical solution compared to conventional settings relying on binocular vision or LiDAR, has drawn increasing attention recently but still yields unsatisfactory results.
This paper first presents a systematic study on this problem and observes that the current monocular 3D detection problem can be simplified as an instance depth estimation problem.
arXiv Detail & Related papers (2021-07-29T16:30:33Z) - Learning Geometry-Guided Depth via Projective Modeling for Monocular 3D Object Detection [70.71934539556916]
We learn geometry-guided depth estimation with projective modeling to advance monocular 3D object detection.
Specifically, a principled geometry formula with projective modeling of 2D and 3D depth predictions in the monocular 3D object detection network is devised.
Our method remarkably improves the detection performance of the state-of-the-art monocular-based method without extra data by 2.80% on the moderate test setting.
arXiv Detail & Related papers (2021-07-29T12:30:39Z) - Probabilistic Monocular 3D Human Pose Estimation with Normalizing Flows [24.0966076588569]
We propose a normalizing flow based method that exploits the deterministic 3D-to-2D mapping to solve the ambiguous inverse 2D-to-3D problem.
We evaluate our approach on the two benchmark datasets Human3.6M and MPI-INF-3DHP, outperforming all comparable methods in most metrics.
arXiv Detail & Related papers (2021-07-29T07:33:14Z) - Neighbor-Vote: Improving Monocular 3D Object Detection through Neighbor
Distance Voting [12.611269919468999]
We present a novel neighbor-voting method that incorporates neighbor predictions to ameliorate object detection from severely deformed pseudo-LiDAR point clouds.
Our results on the bird's eye view detection outperform the state-of-the-art performance by a large margin, especially for the hard'' level detection.
arXiv Detail & Related papers (2021-07-06T09:18:33Z) - Gated3D: Monocular 3D Object Detection From Temporal Illumination Cues [28.806932489163888]
We propose a novel 3D object detection modality that exploits temporal illumination cues from a low-cost monocular gated imager.
We assess the proposed method on a novel 3D detection dataset that includes gated imagery captured in over 10,000 km of driving data.
arXiv Detail & Related papers (2021-02-06T16:06:51Z) - Single View Metrology in the Wild [94.7005246862618]
We present a novel approach to single view metrology that can recover the absolute scale of a scene represented by 3D heights of objects or camera height above the ground.
Our method relies on data-driven priors learned by a deep network specifically designed to imbibe weakly supervised constraints from the interplay of the unknown camera with 3D entities such as object heights.
We demonstrate state-of-the-art qualitative and quantitative results on several datasets as well as applications including virtual object insertion.
arXiv Detail & Related papers (2020-07-18T22:31:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.