3D Object Localization Using 2D Estimates for Computer Vision
Applications
- URL: http://arxiv.org/abs/2009.11446v2
- Date: Sat, 21 Aug 2021 09:37:25 GMT
- Title: 3D Object Localization Using 2D Estimates for Computer Vision
Applications
- Authors: Taha Hasan Masood Siddique and Muhammad Usman
- Abstract summary: A technique for object localization based on pose estimation and camera calibration is presented.
The 3-dimensional (3D) coordinates are estimated by collecting multiple 2-dimensional (2D) images of the object and are utilized for the calibration of the camera.
- Score: 0.9543667840503739
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A technique for object localization based on pose estimation and camera
calibration is presented. The 3-dimensional (3D) coordinates are estimated by
collecting multiple 2-dimensional (2D) images of the object and are utilized
for the calibration of the camera. The calibration steps involving a number of
parameter calculation including intrinsic and extrinsic parameters for the
removal of lens distortion, computation of object's size and camera's position
calculation are discussed. A transformation strategy to estimate the 3D pose
using the 2D images is presented. The proposed method is implemented on MATLAB
and validation experiments are carried out for both pose estimation and camera
calibration.
Related papers
- Multi-Modal Dataset Acquisition for Photometrically Challenging Object [56.30027922063559]
This paper addresses the limitations of current datasets for 3D vision tasks in terms of accuracy, size, realism, and suitable imaging modalities for photometrically challenging objects.
We propose a novel annotation and acquisition pipeline that enhances existing 3D perception and 6D object pose datasets.
arXiv Detail & Related papers (2023-08-21T10:38:32Z) - Monocular 3D Object Detection with Depth from Motion [74.29588921594853]
We take advantage of camera ego-motion for accurate object depth estimation and detection.
Our framework, named Depth from Motion (DfM), then uses the established geometry to lift 2D image features to the 3D space and detects 3D objects thereon.
Our framework outperforms state-of-the-art methods by a large margin on the KITTI benchmark.
arXiv Detail & Related papers (2022-07-26T15:48:46Z) - Level Set-Based Camera Pose Estimation From Multiple 2D/3D
Ellipse-Ellipsoid Correspondences [2.016317500787292]
We show that the definition of a cost function characterizing the projection of a 3D object onto a 2D object detection is not straightforward.
We develop an ellipse-ellipse cost based on level sets sampling, demonstrate its nice properties for handling partially visible objects and compare its performance with other common metrics.
arXiv Detail & Related papers (2022-07-16T14:09:54Z) - 3D object reconstruction and 6D-pose estimation from 2D shape for
robotic grasping of objects [2.330913682033217]
We propose a method for 3D object reconstruction and 6D-pose estimation from 2D images.
By computing transformation parameters directly from the 2D images, the number of free parameters required during the registration process is reduced.
In robot experiments, successful grasping of objects demonstrates its usability in real-world environments.
arXiv Detail & Related papers (2022-03-02T11:58:35Z) - A Critical Analysis of Image-based Camera Pose Estimation Techniques [18.566761146552537]
Camera localization could benefit many computer vision fields, such as autonomous driving, robot navigation, and augmented reality (AR)
In this survey, we first introduce specific application areas and the evaluation metrics for camera localization pose according to different sub-tasks.
arXiv Detail & Related papers (2022-01-15T09:57:45Z) - DETR3D: 3D Object Detection from Multi-view Images via 3D-to-2D Queries [43.02373021724797]
We introduce a framework for multi-camera 3D object detection.
Our method manipulates predictions directly in 3D space.
We achieve state-of-the-art performance on the nuScenes autonomous driving benchmark.
arXiv Detail & Related papers (2021-10-13T17:59:35Z) - Category-Level Metric Scale Object Shape and Pose Estimation [73.92460712829188]
We propose a framework that jointly estimates a metric scale shape and pose from a single RGB image.
We validated our method on both synthetic and real-world datasets to evaluate category-level object pose and shape.
arXiv Detail & Related papers (2021-09-01T12:16:46Z) - Calibrated and Partially Calibrated Semi-Generalized Homographies [65.29477277713205]
We propose the first minimal solutions for estimating the semi-generalized homography given a perspective and a generalized camera.
The proposed solvers are stable and efficient as demonstrated by a number of synthetic and real-world experiments.
arXiv Detail & Related papers (2021-03-11T08:56:24Z) - PLUME: Efficient 3D Object Detection from Stereo Images [95.31278688164646]
Existing methods tackle the problem in two steps: first depth estimation is performed, a pseudo LiDAR point cloud representation is computed from the depth estimates, and then object detection is performed in 3D space.
We propose a model that unifies these two tasks in the same metric space.
Our approach achieves state-of-the-art performance on the challenging KITTI benchmark, with significantly reduced inference time compared with existing methods.
arXiv Detail & Related papers (2021-01-17T05:11:38Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z) - Automatic Estimation of Sphere Centers from Images of Calibrated Cameras [11.816942730023886]
This paper deals with the automatic detection of ellipses in camera images, as well as to estimate the 3D position of the spheres corresponding to the detected 2D ellipses.
We propose two novel methods to (i) detect an ellipse in camera images and (ii) estimate the spatial location of the corresponding sphere if its size is known.
They are applied for calibrating the sensor system of autonomous cars equipped with digital cameras, depth sensors and LiDAR devices.
arXiv Detail & Related papers (2020-02-24T13:12:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.