U-RED: Unsupervised 3D Shape Retrieval and Deformation for Partial Point
Clouds
- URL: http://arxiv.org/abs/2308.06383v1
- Date: Fri, 11 Aug 2023 20:56:05 GMT
- Title: U-RED: Unsupervised 3D Shape Retrieval and Deformation for Partial Point
Clouds
- Authors: Yan Di, Chenyangguang Zhang, Ruida Zhang, Fabian Manhardt, Yongzhi Su,
Jason Rambach, Didier Stricker, Xiangyang Ji and Federico Tombari
- Abstract summary: We propose U-RED, an Unsupervised shape REtrieval and Deformation pipeline.
It takes an arbitrary object observation as input, typically captured by RGB images or scans, and jointly retrieves and deforms the geometrically similar CAD models.
We show that U-RED surpasses existing state-of-the-art approaches by 47.3%, 16.7% and 31.6% respectively under Chamfer Distance.
- Score: 84.32525852378525
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this paper, we propose U-RED, an Unsupervised shape REtrieval and
Deformation pipeline that takes an arbitrary object observation as input,
typically captured by RGB images or scans, and jointly retrieves and deforms
the geometrically similar CAD models from a pre-established database to tightly
match the target. Considering existing methods typically fail to handle noisy
partial observations, U-RED is designed to address this issue from two aspects.
First, since one partial shape may correspond to multiple potential full
shapes, the retrieval method must allow such an ambiguous one-to-many
relationship. Thereby U-RED learns to project all possible full shapes of a
partial target onto the surface of a unit sphere. Then during inference, each
sampling on the sphere will yield a feasible retrieval. Second, since
real-world partial observations usually contain noticeable noise, a reliable
learned metric that measures the similarity between shapes is necessary for
stable retrieval. In U-RED, we design a novel point-wise residual-guided metric
that allows noise-robust comparison. Extensive experiments on the synthetic
datasets PartNet, ComplementMe and the real-world dataset Scan2CAD demonstrate
that U-RED surpasses existing state-of-the-art approaches by 47.3%, 16.7% and
31.6% respectively under Chamfer Distance.
Related papers
- Self-supervised 3D Point Cloud Completion via Multi-view Adversarial Learning [61.14132533712537]
We propose MAL-SPC, a framework that effectively leverages both object-level and category-specific geometric similarities to complete missing structures.
Our MAL-SPC does not require any 3D complete supervision and only necessitates a single partial point cloud for each object.
arXiv Detail & Related papers (2024-07-13T06:53:39Z) - Towards Human-Level 3D Relative Pose Estimation: Generalizable, Training-Free, with Single Reference [62.99706119370521]
Humans can easily deduce the relative pose of an unseen object, without label/training, given only a single query-reference image pair.
We propose a novel 3D generalizable relative pose estimation method by elaborating (i) with a 2.5D shape from an RGB-D reference, (ii) with an off-the-shelf differentiable, and (iii) with semantic cues from a pretrained model like DINOv2.
arXiv Detail & Related papers (2024-06-26T16:01:10Z) - KP-RED: Exploiting Semantic Keypoints for Joint 3D Shape Retrieval and Deformation [87.23575166061413]
KP-RED is a unified KeyPoint-driven REtrieval and Deformation framework.
It takes object scans as input and jointly retrieves and deforms the most geometrically similar CAD models.
arXiv Detail & Related papers (2024-03-15T08:44:56Z) - Deformation-Guided Unsupervised Non-Rigid Shape Matching [7.327850781641328]
We present an unsupervised data-driven approach for non-rigid shape matching.
Our approach is particularly robust when matching digitized shapes using 3D scanners.
arXiv Detail & Related papers (2023-11-27T09:55:55Z) - From Points to Multi-Object 3D Reconstruction [71.17445805257196]
We propose a method to detect and reconstruct multiple 3D objects from a single RGB image.
A keypoint detector localizes objects as center points and directly predicts all object properties, including 9-DoF bounding boxes and 3D shapes.
The presented approach performs lightweight reconstruction in a single-stage, it is real-time capable, fully differentiable and end-to-end trainable.
arXiv Detail & Related papers (2020-12-21T18:52:21Z) - Monocular 3D Detection with Geometric Constraints Embedding and
Semi-supervised Training [3.8073142980733]
We propose a novel framework for monocular 3D objects detection using only RGB images, called KM3D-Net.
We design a fully convolutional model to predict object keypoints, dimension, and orientation, and then combine these estimations with perspective geometry constraints to compute position attribute.
arXiv Detail & Related papers (2020-09-02T00:51:51Z) - Monocular Human Pose and Shape Reconstruction using Part Differentiable
Rendering [53.16864661460889]
Recent works succeed in regression-based methods which estimate parametric models directly through a deep neural network supervised by 3D ground truth.
In this paper, we introduce body segmentation as critical supervision.
To improve the reconstruction with part segmentation, we propose a part-level differentiable part that enables part-based models to be supervised by part segmentation.
arXiv Detail & Related papers (2020-03-24T14:25:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.