Variational Autoencoding of Dental Point Clouds
- URL: http://arxiv.org/abs/2307.10895v3
- Date: Wed, 31 Jan 2024 12:40:51 GMT
- Title: Variational Autoencoding of Dental Point Clouds
- Authors: Johan Ziruo Ye, Thomas {\O}rkild, Peter Lempel S{\o}ndergaard,
S{\o}ren Hauberg
- Abstract summary: This paper introduces the FDI 16 dataset, an extensive collection of tooth meshes and point clouds.
We present a novel approach: Variational FoldingNet (VF-Net), a fully probabilistic variational autoencoder designed for point clouds.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Digital dentistry has made significant advancements, yet numerous challenges
remain. This paper introduces the FDI 16 dataset, an extensive collection of
tooth meshes and point clouds. Additionally, we present a novel approach:
Variational FoldingNet (VF-Net), a fully probabilistic variational autoencoder
designed for point clouds. Notably, prior latent variable models for point
clouds lack a one-to-one correspondence between input and output points.
Instead, they rely on optimizing Chamfer distances, a metric that lacks a
normalized distributional counterpart, rendering it unsuitable for
probabilistic modeling. We replace the explicit minimization of Chamfer
distances with a suitable encoder, increasing computational efficiency while
simplifying the probabilistic extension. This allows for straightforward
application in various tasks, including mesh generation, shape completion, and
representation learning. Empirically, we provide evidence of lower
reconstruction error in dental reconstruction and interpolation, showcasing
state-of-the-art performance in dental sample generation while identifying
valuable latent representations.
Related papers
- Variational Bayes image restoration with compressive autoencoders [5.514022022280169]
Regularization of inverse problems is of paramount importance in computational imaging.
State-of-the-art plug-and-play methods rely on an implicit regularization provided by neural denoisers.
We introduce the Variational Bayes Latent Estimation (VBLE) algorithm, which performs latent estimation.
arXiv Detail & Related papers (2023-11-29T15:49:31Z) - Complexity Matters: Rethinking the Latent Space for Generative Modeling [65.64763873078114]
In generative modeling, numerous successful approaches leverage a low-dimensional latent space, e.g., Stable Diffusion.
In this study, we aim to shed light on this under-explored topic by rethinking the latent space from the perspective of model complexity.
arXiv Detail & Related papers (2023-07-17T07:12:29Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - Controllable Mesh Generation Through Sparse Latent Point Diffusion
Models [105.83595545314334]
We design a novel sparse latent point diffusion model for mesh generation.
Our key insight is to regard point clouds as an intermediate representation of meshes, and model the distribution of point clouds instead.
Our proposed sparse latent point diffusion model achieves superior performance in terms of generation quality and controllability.
arXiv Detail & Related papers (2023-03-14T14:25:29Z) - Latent Autoregressive Source Separation [5.871054749661012]
This paper introduces vector-quantized Latent Autoregressive Source Separation (i.e., de-mixing an input signal into its constituent sources) without requiring additional gradient-based optimization or modifications of existing models.
Our separation method relies on the Bayesian formulation in which the autoregressive models are the priors, and a discrete (non-parametric) likelihood function is constructed by performing frequency counts over latent sums of addend tokens.
arXiv Detail & Related papers (2023-01-09T17:32:00Z) - Differentiable Convolution Search for Point Cloud Processing [114.66038862207118]
We propose a novel differential convolution search paradigm on point clouds.
It can work in a purely data-driven manner and thus is capable of auto-creating a group of suitable convolutions for geometric shape modeling.
We also propose a joint optimization framework for simultaneous search of internal convolution and external architecture, and introduce epsilon-greedy algorithm to alleviate the effect of discretization error.
arXiv Detail & Related papers (2021-08-29T14:42:03Z) - Low-rank Characteristic Tensor Density Estimation Part II: Compression
and Latent Density Estimation [31.631861197477185]
Learning generative probabilistic models is a core problem in machine learning.
This paper proposes a joint dimensionality reduction and non-parametric density estimation framework.
We demonstrate that the proposed model achieves very promising results on regression tasks, sampling, and anomaly detection.
arXiv Detail & Related papers (2021-06-20T00:38:56Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - Deterministic Decoding for Discrete Data in Variational Autoencoders [5.254093731341154]
We study a VAE model with a deterministic decoder (DD-VAE) for sequential data that selects the highest-scoring tokens instead of sampling.
We demonstrate the performance of DD-VAE on multiple datasets, including molecular generation and optimization problems.
arXiv Detail & Related papers (2020-03-04T16:36:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.