Surface Reconstruction from Point Clouds via Grid-based Intersection Prediction
- URL: http://arxiv.org/abs/2403.14085v2
- Date: Tue, 9 Apr 2024 02:59:41 GMT
- Title: Surface Reconstruction from Point Clouds via Grid-based Intersection Prediction
- Authors: Hui Tian, Kai Xu,
- Abstract summary: We propose a novel approach that directly predicts the intersection points between line segment of point pairs and implicit surfaces.
Our approach demonstrates state-of-the-art performance on three datasets: ShapeNet, MGN, and ScanNet.
- Score: 12.329450385760051
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Surface reconstruction from point clouds is a crucial task in the fields of computer vision and computer graphics. SDF-based methods excel at reconstructing smooth meshes with minimal error and artefacts but struggle with representing open surfaces. On the other hand, UDF-based methods can effectively represent open surfaces but often introduce noise, leading to artefacts in the mesh. In this work, we propose a novel approach that directly predicts the intersection points between line segment of point pairs and implicit surfaces. To achieve it, we propose two modules named Relative Intersection Module and Sign Module respectively with the feature of point pair as input. To preserve the continuity of the surface, we also integrate symmetry into the two modules, which means the position of predicted intersection will not change even if the input order of the point pair changes. This method not only preserves the ability to represent open surfaces but also eliminates most artefacts on the mesh. Our approach demonstrates state-of-the-art performance on three datasets: ShapeNet, MGN, and ScanNet. The code will be made available upon acceptance.
Related papers
- SpaceMesh: A Continuous Representation for Learning Manifold Surface Meshes [61.110517195874074]
We present a scheme to directly generate manifold, polygonal meshes of complex connectivity as the output of a neural network.
Our key innovation is to define a continuous latent connectivity space at each mesh, which implies the discrete mesh.
In applications, this approach not only yields high-quality outputs from generative models, but also enables directly learning challenging geometry processing tasks such as mesh repair.
arXiv Detail & Related papers (2024-09-30T17:59:03Z) - Flattening-Net: Deep Regular 2D Representation for 3D Point Cloud
Analysis [66.49788145564004]
We present an unsupervised deep neural architecture called Flattening-Net to represent irregular 3D point clouds of arbitrary geometry and topology.
Our methods perform favorably against the current state-of-the-art competitors.
arXiv Detail & Related papers (2022-12-17T15:05:25Z) - RangeUDF: Semantic Surface Reconstruction from 3D Point Clouds [106.54285912111888]
We present RangeUDF, a new implicit representation based framework to recover the geometry and semantics of continuous 3D scene surfaces from point clouds.
We show that RangeUDF clearly surpasses state-of-the-art approaches for surface reconstruction on four point cloud datasets.
arXiv Detail & Related papers (2022-04-19T21:39:45Z) - MeshUDF: Fast and Differentiable Meshing of Unsigned Distance Field
Networks [68.82901764109685]
Recent work modelling 3D open surfaces train deep neural networks to approximate Unsigned Distance Fields (UDFs)
We propose to directly mesh deep UDFs as open surfaces with an extension of marching cubes, by locally detecting surface crossings.
Our method is order of magnitude faster than meshing a dense point cloud, and more accurate than inflating open surfaces.
arXiv Detail & Related papers (2021-11-29T14:24:02Z) - Deep Implicit Surface Point Prediction Networks [49.286550880464866]
Deep neural representations of 3D shapes as implicit functions have been shown to produce high fidelity models.
This paper presents a novel approach that models such surfaces using a new class of implicit representations called the closest surface-point (CSP) representation.
arXiv Detail & Related papers (2021-06-10T14:31:54Z) - DualConv: Dual Mesh Convolutional Networks for Shape Correspondence [44.94765770516059]
Convolutional neural networks have been extremely successful for 2D images and are readily extended to handle 3D voxel data.
In this paper we explore how these networks can be extended to the dual face-based representation of triangular meshes.
Our experiments demonstrate that building additionally convolutional models that explicitly leverage the neighborhood size regularity of dual meshes enables learning shape representations that perform on par or better than previous approaches.
arXiv Detail & Related papers (2021-03-23T11:22:47Z) - Primal-Dual Mesh Convolutional Neural Networks [62.165239866312334]
We propose a primal-dual framework drawn from the graph-neural-network literature to triangle meshes.
Our method takes features for both edges and faces of a 3D mesh as input and dynamically aggregates them.
We provide theoretical insights of our approach using tools from the mesh-simplification literature.
arXiv Detail & Related papers (2020-10-23T14:49:02Z) - DualConvMesh-Net: Joint Geodesic and Euclidean Convolutions on 3D Meshes [28.571946680616765]
We propose a family of deep hierarchical convolutional networks over 3D geometric data.
The first type, geodesic convolutions, defines the kernel weights over mesh surfaces or graphs.
The second type, Euclidean convolutions, is independent of any underlying mesh structure.
arXiv Detail & Related papers (2020-04-02T13:52:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.