Visually Guided Object Grasping
- URL: http://arxiv.org/abs/2311.12660v1
- Date: Tue, 21 Nov 2023 15:08:17 GMT
- Title: Visually Guided Object Grasping
- Authors: Radu Horaud, Fadi Dornaika and Bernard Espiau
- Abstract summary: We show how to represent a grasp or more generally, an alignment between two solids in 3-D projective space using an uncalibrated stereo rig.
We perform an analysis of the performances of the visual servoing algorithm and of the grasping precision that can be expected from this type of approach.
- Score: 19.71383212064634
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper we present a visual servoing approach to the problem of object
grasping and more generally, to the problem of aligning an end-effector with an
object. First we extend the method proposed by Espiau et al. [1] to the case of
a camera which is not mounted onto the robot being controlled and we stress the
importance of the real-time estimation of the image Jacobian. Second, we show
how to represent a grasp or more generally, an alignment between two solids in
3-D projective space using an uncalibrated stereo rig. Such a 3-D projective
representation is view-invariant in the sense that it can be easily mapped into
an image set-point without any knowledge about the camera parameters. Third, we
perform an analysis of the performances of the visual servoing algorithm and of
the grasping precision that can be expected from this type of approach.
Related papers
- Joint 3D Shape and Motion Estimation from Rolling Shutter Light-Field
Images [2.0277446818410994]
We propose an approach to address the problem of 3D reconstruction of scenes from a single image captured by a light-field camera equipped with a rolling shutter sensor.
Our method leverages the 3D information cues present in the light-field and the motion information provided by the rolling shutter effect.
We present a generic model for the imaging process of this sensor and a two-stage algorithm that minimizes the re-projection error.
arXiv Detail & Related papers (2023-11-02T15:08:18Z) - 3D-Aware Hypothesis & Verification for Generalizable Relative Object
Pose Estimation [69.73691477825079]
We present a new hypothesis-and-verification framework to tackle the problem of generalizable object pose estimation.
To measure reliability, we introduce a 3D-aware verification that explicitly applies 3D transformations to the 3D object representations learned from the two input images.
arXiv Detail & Related papers (2023-10-05T13:34:07Z) - Self-supervised Wide Baseline Visual Servoing via 3D Equivariance [35.93323183558956]
This paper presents a novel self-supervised visual servoing method for wide baseline images.
Existing approaches that regress absolute camera pose with respect to an object require 3D ground truth data of the object.
Ours yields more than 35% average distance error reduction and more than 90% success rate with 3cm error tolerance.
arXiv Detail & Related papers (2022-09-12T17:38:26Z) - Monocular 3D Object Detection with Depth from Motion [74.29588921594853]
We take advantage of camera ego-motion for accurate object depth estimation and detection.
Our framework, named Depth from Motion (DfM), then uses the established geometry to lift 2D image features to the 3D space and detects 3D objects thereon.
Our framework outperforms state-of-the-art methods by a large margin on the KITTI benchmark.
arXiv Detail & Related papers (2022-07-26T15:48:46Z) - Attentive and Contrastive Learning for Joint Depth and Motion Field
Estimation [76.58256020932312]
Estimating the motion of the camera together with the 3D structure of the scene from a monocular vision system is a complex task.
We present a self-supervised learning framework for 3D object motion field estimation from monocular videos.
arXiv Detail & Related papers (2021-10-13T16:45:01Z) - Object-Centric Multi-View Aggregation [86.94544275235454]
We present an approach for aggregating a sparse set of views of an object in order to compute a semi-implicit 3D representation in the form of a volumetric feature grid.
Key to our approach is an object-centric canonical 3D coordinate system into which views can be lifted, without explicit camera pose estimation.
We show that computing a symmetry-aware mapping from pixels to the canonical coordinate system allows us to better propagate information to unseen regions.
arXiv Detail & Related papers (2020-07-20T17:38:31Z) - Single View Metrology in the Wild [94.7005246862618]
We present a novel approach to single view metrology that can recover the absolute scale of a scene represented by 3D heights of objects or camera height above the ground.
Our method relies on data-driven priors learned by a deep network specifically designed to imbibe weakly supervised constraints from the interplay of the unknown camera with 3D entities such as object heights.
We demonstrate state-of-the-art qualitative and quantitative results on several datasets as well as applications including virtual object insertion.
arXiv Detail & Related papers (2020-07-18T22:31:33Z) - Lightweight Multi-View 3D Pose Estimation through Camera-Disentangled
Representation [57.11299763566534]
We present a solution to recover 3D pose from multi-view images captured with spatially calibrated cameras.
We exploit 3D geometry to fuse input images into a unified latent representation of pose, which is disentangled from camera view-points.
Our architecture then conditions the learned representation on camera projection operators to produce accurate per-view 2d detections.
arXiv Detail & Related papers (2020-04-05T12:52:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.