WrappingNet: Mesh Autoencoder via Deep Sphere Deformation
- URL: http://arxiv.org/abs/2308.15413v1
- Date: Tue, 29 Aug 2023 16:13:04 GMT
- Title: WrappingNet: Mesh Autoencoder via Deep Sphere Deformation
- Authors: Eric Lei, Muhammad Asad Lodhi, Jiahao Pang, Junghyun Ahn, Dong Tian
- Abstract summary: WrappingNet is the first mesh autoencoder enabling general mesh unsupervised learning over heterogeneous objects.
It introduces a novel base graph in the bottleneck dedicated to representing mesh connectivity.
It is shown to facilitate learning a shared latent space representing object shape.
- Score: 10.934595072086324
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: There have been recent efforts to learn more meaningful representations via
fixed length codewords from mesh data, since a mesh serves as a complete model
of underlying 3D shape compared to a point cloud. However, the mesh
connectivity presents new difficulties when constructing a deep learning
pipeline for meshes. Previous mesh unsupervised learning approaches typically
assume category-specific templates, e.g., human face/body templates. It
restricts the learned latent codes to only be meaningful for objects in a
specific category, so the learned latent spaces are unable to be used across
different types of objects. In this work, we present WrappingNet, the first
mesh autoencoder enabling general mesh unsupervised learning over heterogeneous
objects. It introduces a novel base graph in the bottleneck dedicated to
representing mesh connectivity, which is shown to facilitate learning a shared
latent space representing object shape. The superiority of WrappingNet mesh
learning is further demonstrated via improved reconstruction quality and
competitive classification compared to point cloud learning, as well as latent
interpolation between meshes of different categories.
Related papers
- SpaceMesh: A Continuous Representation for Learning Manifold Surface Meshes [61.110517195874074]
We present a scheme to directly generate manifold, polygonal meshes of complex connectivity as the output of a neural network.
Our key innovation is to define a continuous latent connectivity space at each mesh, which implies the discrete mesh.
In applications, this approach not only yields high-quality outputs from generative models, but also enables directly learning challenging geometry processing tasks such as mesh repair.
arXiv Detail & Related papers (2024-09-30T17:59:03Z) - Segment Any 3D Object with Language [58.471327490684295]
We introduce Segment any 3D Object with LanguagE (SOLE), a semantic geometric and-aware visual-language learning framework with strong generalizability.
Specifically, we propose a multimodal fusion network to incorporate multimodal semantics in both backbone and decoder.
Our SOLE outperforms previous methods by a large margin on ScanNetv2, ScanNet200, and Replica benchmarks.
arXiv Detail & Related papers (2024-04-02T17:59:10Z) - Single-view 3D Mesh Reconstruction for Seen and Unseen Categories [69.29406107513621]
Single-view 3D Mesh Reconstruction is a fundamental computer vision task that aims at recovering 3D shapes from single-view RGB images.
This paper tackles Single-view 3D Mesh Reconstruction, to study the model generalization on unseen categories.
We propose an end-to-end two-stage network, GenMesh, to break the category boundaries in reconstruction.
arXiv Detail & Related papers (2022-08-04T14:13:35Z) - OMNet: Learning Overlapping Mask for Partial-to-Partial Point Cloud
Registration [31.108056345511976]
OMNet is a global feature based iterative network for partial-to-partial point cloud registration.
We learn masks in a coarse-to-fine manner to reject non-overlapping regions, which converting the partial-to-partial registration to the registration of the same shapes.
arXiv Detail & Related papers (2021-03-01T11:59:59Z) - PatchNets: Patch-Based Generalizable Deep Implicit 3D Shape
Representations [75.42959184226702]
We present a new mid-level patch-based surface representation for object-agnostic training.
We show several applications of our new representation, including shape and partial point cloud completion.
arXiv Detail & Related papers (2020-08-04T15:34:46Z) - Meshing Point Clouds with Predicted Intrinsic-Extrinsic Ratio Guidance [30.863194319818223]
We propose to leverage the input point cloud as much as possible, by only adding connectivity information to existing points.
Our key innovation is a surrogate of local connectivity, calculated by comparing the intrinsic/extrinsic metrics.
We demonstrate that our method can not only preserve details, handle ambiguous structures, but also possess strong generalizability to unseen categories.
arXiv Detail & Related papers (2020-07-17T22:36:00Z) - Deep Geometric Texture Synthesis [83.9404865744028]
We propose a novel framework for synthesizing geometric textures.
It learns texture statistics from local neighborhoods of a single reference 3D model.
Our network displaces mesh vertices in any direction, enabling synthesis of geometric textures.
arXiv Detail & Related papers (2020-06-30T19:36:38Z) - TearingNet: Point Cloud Autoencoder to Learn Topology-Friendly
Representations [20.318695890515613]
We propose an autoencoder, TearingNet, which tackles the challenging task of representing point clouds using a fixed-length descriptor.
Our TearingNet is characterized by a proposed Tearing network module and a Folding network module interacting with each other iteratively.
Experimentation shows the superiority of our proposal in terms of reconstructing point clouds as well as generating more topology-friendly representations than benchmarks.
arXiv Detail & Related papers (2020-06-17T22:42:43Z) - MeshWalker: Deep Mesh Understanding by Random Walks [19.594977587417247]
We look at the most popular representation of 3D shapes in computer graphics - a triangular mesh - and ask how it can be utilized within deep learning.
This paper proposes a very different approach, termed MeshWalker, to learn the shape directly from a given mesh.
We show that our approach achieves state-of-the-art results for two fundamental shape analysis tasks.
arXiv Detail & Related papers (2020-06-09T15:35:41Z) - Neural Subdivision [58.97214948753937]
This paper introduces Neural Subdivision, a novel framework for data-driven coarseto-fine geometry modeling.
We optimize for the same set of network weights across all local mesh patches, thus providing an architecture that is not constrained to a specific input mesh, fixed genus, or category.
We demonstrate that even when trained on a single high-resolution mesh our method generates reasonable subdivisions for novel shapes.
arXiv Detail & Related papers (2020-05-04T20:03:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.