3D Object Detection from a Single Fisheye Image Without a Single Fisheye
Training Image
- URL: http://arxiv.org/abs/2003.03759v3
- Date: Mon, 31 May 2021 05:56:56 GMT
- Title: 3D Object Detection from a Single Fisheye Image Without a Single Fisheye
Training Image
- Authors: Elad Plaut, Erez Ben Yaacov and Bat El Shlomo
- Abstract summary: We show how to use existing monocular 3D object detection models, trained only on rectilinear images, to detect 3D objects in images from fisheye cameras.
We outperform the only existing method for monocular 3D object detection in panoramas on a benchmark of synthetic data.
- Score: 7.86363825307044
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing monocular 3D object detection methods have been demonstrated on
rectilinear perspective images and fail in images with alternative projections
such as those acquired by fisheye cameras. Previous works on object detection
in fisheye images have focused on 2D object detection, partly due to the lack
of 3D datasets of such images. In this work, we show how to use existing
monocular 3D object detection models, trained only on rectilinear images, to
detect 3D objects in images from fisheye cameras, without using any fisheye
training data. We outperform the only existing method for monocular 3D object
detection in panoramas on a benchmark of synthetic data, despite the fact that
the existing method trains on the target non-rectilinear projection whereas we
train only on rectilinear images. We also experiment with an internal dataset
of real fisheye images.
Related papers
- VFMM3D: Releasing the Potential of Image by Vision Foundation Model for Monocular 3D Object Detection [80.62052650370416]
monocular 3D object detection holds significant importance across various applications, including autonomous driving and robotics.
In this paper, we present VFMM3D, an innovative framework that leverages the capabilities of Vision Foundation Models (VFMs) to accurately transform single-view images into LiDAR point cloud representations.
arXiv Detail & Related papers (2024-04-15T03:12:12Z) - 3D Object Aided Self-Supervised Monocular Depth Estimation [5.579605877061333]
We propose a new method to address dynamic object movements through monocular 3D object detection.
Specifically, we first detect 3D objects in the images and build the per-pixel correspondence of the dynamic pixels with the detected object pose.
In this way, the depth of every pixel can be learned via a meaningful geometry model.
arXiv Detail & Related papers (2022-12-04T08:52:33Z) - Self-supervised 3D Object Detection from Monocular Pseudo-LiDAR [9.361704310981196]
We propose a method for predicting absolute depth and detecting 3D objects using only monocular image sequences.
As a result, the proposed method surpasses other existing methods in performance on the KITTI 3D dataset.
arXiv Detail & Related papers (2022-09-20T05:55:49Z) - Learning Geometry-Guided Depth via Projective Modeling for Monocular 3D Object Detection [70.71934539556916]
We learn geometry-guided depth estimation with projective modeling to advance monocular 3D object detection.
Specifically, a principled geometry formula with projective modeling of 2D and 3D depth predictions in the monocular 3D object detection network is devised.
Our method remarkably improves the detection performance of the state-of-the-art monocular-based method without extra data by 2.80% on the moderate test setting.
arXiv Detail & Related papers (2021-07-29T12:30:39Z) - Disentangling and Vectorization: A 3D Visual Perception Approach for
Autonomous Driving Based on Surround-View Fisheye Cameras [3.485767750936058]
Multidimensional Vector is proposed to include the utilizable information generated in different dimensions and stages.
The experiments of real fisheye images demonstrate that our solution achieves state-of-the-art accuracy while being real-time in practice.
arXiv Detail & Related papers (2021-07-19T13:24:21Z) - Monocular Differentiable Rendering for Self-Supervised 3D Object
Detection [21.825158925459732]
3D object detection from monocular images is an ill-posed problem due to the projective entanglement of depth and scale.
We present a novel self-supervised method for textured 3D shape reconstruction and pose estimation of rigid objects.
Our method predicts the 3D location and meshes of each object in an image using differentiable rendering and a self-supervised objective.
arXiv Detail & Related papers (2020-09-30T09:21:43Z) - Kinematic 3D Object Detection in Monocular Video [123.7119180923524]
We propose a novel method for monocular video-based 3D object detection which carefully leverages kinematic motion to improve precision of 3D localization.
We achieve state-of-the-art performance on monocular 3D object detection and the Bird's Eye View tasks within the KITTI self-driving dataset.
arXiv Detail & Related papers (2020-07-19T01:15:12Z) - From Image Collections to Point Clouds with Self-supervised Shape and
Pose Networks [53.71440550507745]
Reconstructing 3D models from 2D images is one of the fundamental problems in computer vision.
We propose a deep learning technique for 3D object reconstruction from a single image.
We learn both 3D point cloud reconstruction and pose estimation networks in a self-supervised manner.
arXiv Detail & Related papers (2020-05-05T04:25:16Z) - DOPS: Learning to Detect 3D Objects and Predict their 3D Shapes [54.239416488865565]
We propose a fast single-stage 3D object detection method for LIDAR data.
The core novelty of our method is a fast, single-pass architecture that both detects objects in 3D and estimates their shapes.
We find that our proposed method achieves state-of-the-art results by 5% on object detection in ScanNet scenes, and it gets top results by 3.4% in the Open dataset.
arXiv Detail & Related papers (2020-04-02T17:48:50Z) - BirdNet+: End-to-End 3D Object Detection in LiDAR Bird's Eye View [117.44028458220427]
On-board 3D object detection in autonomous vehicles often relies on geometry information captured by LiDAR devices.
We present a fully end-to-end 3D object detection framework that can infer oriented 3D boxes solely from BEV images.
arXiv Detail & Related papers (2020-03-09T15:08:40Z) - SilhoNet-Fisheye: Adaptation of A ROI Based Object Pose Estimation
Network to Monocular Fisheye Images [15.573003283204958]
We present a novel framework for adapting a ROI-based 6D object pose estimation method to work on full fisheye images.
We also contribute a fisheye image dataset, called UWHandles, with 6D object pose and 2D bounding box annotations.
arXiv Detail & Related papers (2020-02-27T19:57:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.