Learning 3D-3D Correspondences for One-shot Partial-to-partial
Registration
- URL: http://arxiv.org/abs/2006.04523v2
- Date: Tue, 16 Jun 2020 15:57:22 GMT
- Title: Learning 3D-3D Correspondences for One-shot Partial-to-partial
Registration
- Authors: Zheng Dang, Fei Wang and Mathieu Salzmann
- Abstract summary: We show that learning-based partial-to-partial registration can be achieved in a one-shot manner.
We propose an Optimal Transport layer able to account for occluded points thanks to the use of bins.
The resulting OPRNet framework outperforms the state of the art on standard benchmarks.
- Score: 66.41922513553367
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: While 3D-3D registration is traditionally tacked by optimization-based
methods, recent work has shown that learning-based techniques could achieve
faster and more robust results. In this context, however, only PRNet can handle
the partial-to-partial registration scenario. Unfortunately, this is achieved
at the cost of relying on an iterative procedure, with a complex network
architecture. Here, we show that learning-based partial-to-partial registration
can be achieved in a one-shot manner, jointly reducing network complexity and
increasing registration accuracy. To this end, we propose an Optimal Transport
layer able to account for occluded points thanks to the use of outlier bins.
The resulting OPRNet framework outperforms the state of the art on standard
benchmarks, demonstrating better robustness and generalization ability than
existing techniques.
Related papers
- Best Practices for a Handwritten Text Recognition System [8.334691351242753]
Handwritten text recognition has been developed rapidly in the recent years.
Non-trivial deviation in performance can be detected even when small pre-processing elements are changed.
This work highlights simple yet effective empirical practices that can further help training and provide well-performing handwritten text recognition systems.
arXiv Detail & Related papers (2024-04-17T13:00:05Z) - Implicit Shape and Appearance Priors for Few-Shot Full Head
Reconstruction [17.254539604491303]
In this paper, we address the problem of few-shot full 3D head reconstruction.
We accomplish this by incorporating a probabilistic shape and appearance prior into coordinate-based representations.
We extend the H3DS dataset, which now comprises 60 high-resolution 3D full head scans and their corresponding posed images and masks.
arXiv Detail & Related papers (2023-10-12T07:35:30Z) - Laplacian ICP for Progressive Registration of 3D Human Head Meshes [18.31701623205813]
We present a progressive 3D registration framework that is a highly-efficient variant of classical non-rigid Iterative Closest Points (N-ICP)
Since it uses the Laplace-Beltrami operator for deformation regularisation, we view the overall process as Laplacian ICP (L-ICP)
arXiv Detail & Related papers (2023-02-04T16:39:38Z) - 3D-QueryIS: A Query-based Framework for 3D Instance Segmentation [74.6998931386331]
Previous methods for 3D instance segmentation often maintain inter-task dependencies and the tendency towards a lack of robustness.
We propose a novel query-based method, termed as 3D-QueryIS, which is detector-free, semantic segmentation-free, and cluster-free.
Our 3D-QueryIS is free from the accumulated errors caused by the inter-task dependencies.
arXiv Detail & Related papers (2022-11-17T07:04:53Z) - Improving Point Cloud Based Place Recognition with Ranking-based Loss
and Large Batch Training [1.116812194101501]
The paper presents a simple and effective learning-based method for computing a discriminative 3D point cloud descriptor.
We employ recent advances in image retrieval and propose a modified version of a loss function based on a differentiable average precision approximation.
arXiv Detail & Related papers (2022-03-02T09:29:28Z) - Multi-initialization Optimization Network for Accurate 3D Human Pose and
Shape Estimation [75.44912541912252]
We propose a three-stage framework named Multi-Initialization Optimization Network (MION)
In the first stage, we strategically select different coarse 3D reconstruction candidates which are compatible with the 2D keypoints of input sample.
In the second stage, we design a mesh refinement transformer (MRT) to respectively refine each coarse reconstruction result via a self-attention mechanism.
Finally, a Consistency Estimation Network (CEN) is proposed to find the best result from mutiple candidates by evaluating if the visual evidence in RGB image matches a given 3D reconstruction.
arXiv Detail & Related papers (2021-12-24T02:43:58Z) - Dynamic Iterative Refinement for Efficient 3D Hand Pose Estimation [87.54604263202941]
We propose a tiny deep neural network of which partial layers are iteratively exploited for refining its previous estimations.
We employ learned gating criteria to decide whether to exit from the weight-sharing loop, allowing per-sample adaptation in our model.
Our method consistently outperforms state-of-the-art 2D/3D hand pose estimation approaches in terms of both accuracy and efficiency for widely used benchmarks.
arXiv Detail & Related papers (2021-11-11T23:31:34Z) - FlowStep3D: Model Unrolling for Self-Supervised Scene Flow Estimation [87.74617110803189]
Estimating the 3D motion of points in a scene, known as scene flow, is a core problem in computer vision.
We present a recurrent architecture that learns a single step of an unrolled iterative alignment procedure for refining scene flow predictions.
arXiv Detail & Related papers (2020-11-19T23:23:48Z) - Deep-3DAligner: Unsupervised 3D Point Set Registration Network With
Optimizable Latent Vector [15.900382629390297]
We propose to develop a novel model that integrates the optimization to learning, aiming to address the technical challenges in 3D registration.
In addition to the deep transformation decoding network, our framework introduce an optimizable deep underlineSpatial underlineCorrelation underlineRepresentation.
arXiv Detail & Related papers (2020-09-29T22:44:38Z) - Spatial-Spectral Residual Network for Hyperspectral Image
Super-Resolution [82.1739023587565]
We propose a novel spectral-spatial residual network for hyperspectral image super-resolution (SSRNet)
Our method can effectively explore spatial-spectral information by using 3D convolution instead of 2D convolution, which enables the network to better extract potential information.
In each unit, we employ spatial and temporal separable 3D convolution to extract spatial and spectral information, which not only reduces unaffordable memory usage and high computational cost, but also makes the network easier to train.
arXiv Detail & Related papers (2020-01-14T03:34:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.