SALD: Sign Agnostic Learning with Derivatives
- URL: http://arxiv.org/abs/2006.05400v2
- Date: Sat, 3 Oct 2020 17:24:48 GMT
- Title: SALD: Sign Agnostic Learning with Derivatives
- Authors: Matan Atzmon and Yaron Lipman
- Abstract summary: We introduce SALD: a method for learning implicit neural representations of shapes directly from raw data.
We demonstrate the efficacy of SALD for shape space learning on two challenging datasets.
- Score: 42.43016094317574
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning 3D geometry directly from raw data, such as point clouds, triangle
soups, or unoriented meshes is still a challenging task that feeds many
downstream computer vision and graphics applications.
In this paper, we introduce SALD: a method for learning implicit neural
representations of shapes directly from raw data. We generalize sign agnostic
learning (SAL) to include derivatives: given an unsigned distance function to
the input raw data, we advocate a novel sign agnostic regression loss,
incorporating both pointwise values and gradients of the unsigned distance
function. Optimizing this loss leads to a signed implicit function solution,
the zero level set of which is a high quality and valid manifold approximation
to the input 3D data. The motivation behind SALD is that incorporating
derivatives in a regression loss leads to a lower sample complexity, and
consequently better fitting. In addition, we prove that SAL enjoys a minimal
length property in 2D, favoring minimal length solutions. More importantly, we
are able to show that this property still holds for SALD, i.e., with
derivatives included.
We demonstrate the efficacy of SALD for shape space learning on two
challenging datasets: ShapeNet that contains inconsistent orientation and
non-manifold meshes, and D-Faust that contains raw 3D scans (triangle soups).
On both these datasets, we present state-of-the-art results.
Related papers
- GaussianUDF: Inferring Unsigned Distance Functions through 3D Gaussian Splatting [49.60513072330759]
We propose a novel approach to bridge the gap between 3D Gaussians and UDFs.
Our key idea is to overfit thin and flat 2D Gaussian planes on surfaces, and then, leverage the self-supervision and gradient-based inference.
We show our advantages in terms of accuracy, efficiency, completeness, and sharpness of reconstructed open surfaces with boundaries.
arXiv Detail & Related papers (2025-03-25T08:46:55Z) - Rethinking End-to-End 2D to 3D Scene Segmentation in Gaussian Splatting [86.15347226865826]
We design a new end-to-end object-aware lifting approach, named Unified-Lift.
We augment each Gaussian point with an additional Gaussian-level feature learned using a contrastive loss to encode instance information.
We conduct experiments on three benchmarks: LERF-Masked, Replica, and Messy Rooms.
arXiv Detail & Related papers (2025-03-18T08:42:23Z) - Enhancing Generalizability of Representation Learning for Data-Efficient 3D Scene Understanding [50.448520056844885]
We propose a generative Bayesian network to produce diverse synthetic scenes with real-world patterns.
A series of experiments robustly display our method's consistent superiority over existing state-of-the-art pre-training approaches.
arXiv Detail & Related papers (2024-06-17T07:43:53Z) - Unsupervised Occupancy Learning from Sparse Point Cloud [8.732260277121547]
Implicit Neural Representations have gained prominence as a powerful framework for capturing complex data modalities.
In this paper, we propose a method to infer occupancy fields instead of Neural Signed Distance Functions.
We highlight its capacity to improve implicit shape inference with respect to baselines and the state-of-the-art using synthetic and real data.
arXiv Detail & Related papers (2024-04-03T14:05:39Z) - 3D Adversarial Augmentations for Robust Out-of-Domain Predictions [115.74319739738571]
We focus on improving the generalization to out-of-domain data.
We learn a set of vectors that deform the objects in an adversarial fashion.
We perform adversarial augmentation by applying the learned sample-independent vectors to the available objects when training a model.
arXiv Detail & Related papers (2023-08-29T17:58:55Z) - Self-supervised Human Mesh Recovery with Cross-Representation Alignment [20.69546341109787]
Self-supervised human mesh recovery methods have poor generalizability due to limited availability and diversity of 3D-annotated benchmark datasets.
We propose cross-representation alignment utilizing the complementary information from the robust but sparse representation (2D keypoints)
This adaptive cross-representation alignment explicitly learns from the deviations and captures complementary information: richness from sparse representation and robustness from dense representation.
arXiv Detail & Related papers (2022-09-10T04:47:20Z) - 3PSDF: Three-Pole Signed Distance Function for Learning Surfaces with
Arbitrary Topologies [18.609959464825636]
We present a novel learnable implicit representation called the three-pole signed distance function (3PSDF)
It can represent non-watertight 3D shapes with arbitrary topologies while supporting easy field-to-mesh conversion.
We propose a dedicated learning framework to effectively learn 3PSDF without worrying about the vanishing gradient due to the null labels.
arXiv Detail & Related papers (2022-05-31T07:24:04Z) - Learning-based Point Cloud Registration for 6D Object Pose Estimation in
the Real World [55.7340077183072]
We tackle the task of estimating the 6D pose of an object from point cloud data.
Recent learning-based approaches to addressing this task have shown great success on synthetic datasets.
We analyze the causes of these failures, which we trace back to the difference between the feature distributions of the source and target point clouds.
arXiv Detail & Related papers (2022-03-29T07:55:04Z) - Rel3D: A Minimally Contrastive Benchmark for Grounding Spatial Relations
in 3D [71.11034329713058]
Existing datasets lack large-scale, high-quality 3D ground truth information.
Rel3D is the first large-scale, human-annotated dataset for grounding spatial relations in 3D.
We propose minimally contrastive data collection -- a novel crowdsourcing method for reducing dataset bias.
arXiv Detail & Related papers (2020-12-03T01:51:56Z) - Implicit Functions in Feature Space for 3D Shape Reconstruction and
Completion [53.885984328273686]
Implicit Feature Networks (IF-Nets) deliver continuous outputs, can handle multiple topologies, and complete shapes for missing or sparse input data.
IF-Nets clearly outperform prior work in 3D object reconstruction in ShapeNet, and obtain significantly more accurate 3D human reconstructions.
arXiv Detail & Related papers (2020-03-03T11:14:29Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.