Staying in Shape: Learning Invariant Shape Representations using
Contrastive Learning
- URL: http://arxiv.org/abs/2107.03552v1
- Date: Thu, 8 Jul 2021 00:53:24 GMT
- Title: Staying in Shape: Learning Invariant Shape Representations using
Contrastive Learning
- Authors: Jeffrey Gu and Serena Yeung
- Abstract summary: Most existing invariant shape representations arehandcrafted, and previous work on learning shaperepresentations do not focus on producing invariants.
We show experimentally that our methodoutperforms previous unsupervised learning ap-proaches in both effectiveness and robustness.
- Score: 5.100152971410397
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Creating representations of shapes that are invari-ant to isometric or
almost-isometric transforma-tions has long been an area of interest in shape
anal-ysis, since enforcing invariance allows the learningof more effective and
robust shape representations.Most existing invariant shape representations
arehandcrafted, and previous work on learning shaperepresentations do not focus
on producing invariantrepresentations. To solve the problem of
learningunsupervised invariant shape representations, weuse contrastive
learning, which produces discrimi-native representations through learning
invarianceto user-specified data augmentations. To producerepresentations that
are specifically isometry andalmost-isometry invariant, we propose new
dataaugmentations that randomly sample these transfor-mations. We show
experimentally that our methodoutperforms previous unsupervised learning
ap-proaches in both effectiveness and robustness.
Related papers
- Revealing Multimodal Contrastive Representation Learning through Latent
Partial Causal Models [85.67870425656368]
We introduce a unified causal model specifically designed for multimodal data.
We show that multimodal contrastive representation learning excels at identifying latent coupled variables.
Experiments demonstrate the robustness of our findings, even when the assumptions are violated.
arXiv Detail & Related papers (2024-02-09T07:18:06Z) - Interpreting Equivariant Representations [5.325297567945828]
In this paper, we demonstrate that the inductive bias imposed on the by an equivariant model must also be taken into account when using latent representations.
We show how not accounting for the inductive biases leads to decreased performance on downstream tasks, and vice versa.
arXiv Detail & Related papers (2024-01-23T09:43:30Z) - Flow Factorized Representation Learning [109.51947536586677]
We introduce a generative model which specifies a distinct set of latent probability paths that define different input transformations.
We show that our model achieves higher likelihoods on standard representation learning benchmarks while simultaneously being closer to approximately equivariant models.
arXiv Detail & Related papers (2023-09-22T20:15:37Z) - Semantic-Aware Implicit Template Learning via Part Deformation
Consistency [18.63665468429503]
We propose a semantic-aware implicit template learning framework to enable semantically plausible deformation.
By leveraging semantic prior from a self-supervised feature extractor, we suggest local conditioning with novel semantic-aware deformation code.
Our experiments demonstrate the superiority of the proposed method over baselines in various tasks.
arXiv Detail & Related papers (2023-08-23T05:02:17Z) - Domain Generalization In Robust Invariant Representation [10.132611239890345]
In this paper, we investigate the generalization of invariant representations on out-of-distribution data.
We show that the invariant model learns unstructured latent representations that are robust to distribution shifts.
arXiv Detail & Related papers (2023-04-07T00:58:30Z) - Leveraging Equivariant Features for Absolute Pose Regression [9.30597356471664]
We show that a translation and rotation equivariant Convolutional Neural Network directly induces representations of camera motions into the feature space.
We then show that this geometric property allows for implicitly augmenting the training data under a whole group of image plane-preserving transformations.
arXiv Detail & Related papers (2022-04-05T12:44:20Z) - Topographic VAEs learn Equivariant Capsules [84.33745072274942]
We introduce the Topographic VAE: a novel method for efficiently training deep generative models with topographically organized latent variables.
We show that such a model indeed learns to organize its activations according to salient characteristics such as digit class, width, and style on MNIST.
We demonstrate approximate equivariance to complex transformations, expanding upon the capabilities of existing group equivariant neural networks.
arXiv Detail & Related papers (2021-09-03T09:25:57Z) - Augmenting Implicit Neural Shape Representations with Explicit
Deformation Fields [95.39603371087921]
Implicit neural representation is a recent approach to learn shape collections as zero level-sets of neural networks.
We advocate deformation-aware regularization for implicit neural representations, aiming at producing plausible deformations as latent code changes.
arXiv Detail & Related papers (2021-08-19T22:07:08Z) - SNARF: Differentiable Forward Skinning for Animating Non-Rigid Neural
Implicit Shapes [117.76767853430243]
We introduce SNARF, which combines the advantages of linear blend skinning for polygonal meshes with neural implicit surfaces.
We propose a forward skinning model that finds all canonical correspondences of any deformed point using iterative root finding.
Compared to state-of-the-art neural implicit representations, our approach generalizes better to unseen poses while preserving accuracy.
arXiv Detail & Related papers (2021-04-08T17:54:59Z) - Learning Disentangled Representations with Latent Variation
Predictability [102.4163768995288]
This paper defines the variation predictability of latent disentangled representations.
Within an adversarial generation process, we encourage variation predictability by maximizing the mutual information between latent variations and corresponding image pairs.
We develop an evaluation metric that does not rely on the ground-truth generative factors to measure the disentanglement of latent representations.
arXiv Detail & Related papers (2020-07-25T08:54:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.