Invariant Representations of Embedded Simplicial Complexes
- URL: http://arxiv.org/abs/2302.13565v1
- Date: Mon, 27 Feb 2023 07:49:05 GMT
- Title: Invariant Representations of Embedded Simplicial Complexes
- Authors: Taejin Paik
- Abstract summary: Analyzing embedded simplicial complexes, such as triangular meshes and graphs, is an important problem in many fields.
We propose a new approach for analyzing embedded simplicial complexes in a subdivision-invariant and isometry-invariant way using only topological and geometric information.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Analyzing embedded simplicial complexes, such as triangular meshes and
graphs, is an important problem in many fields. We propose a new approach for
analyzing embedded simplicial complexes in a subdivision-invariant and
isometry-invariant way using only topological and geometric information. Our
approach is based on creating and analyzing sufficient statistics and uses a
graph neural network. We demonstrate the effectiveness of our approach using a
synthetic mesh data set.
Related papers
- Simplicial Representation Learning with Neural $k$-Forms [14.566552361705499]
This paper focuses on leveraging geometric information from simplicial complexes embedded in $mathbbRn$ using node coordinates.
We use differential k-forms in mathbbRn to create representations of simplices, offering interpretability and geometric consistency without message passing.
Our method is efficient, versatile, and applicable to various input complexes, including graphs, simplicial complexes, and cell complexes.
arXiv Detail & Related papers (2023-12-13T21:03:39Z) - From Complexity to Clarity: Analytical Expressions of Deep Neural Network Weights via Clifford's Geometric Algebra and Convexity [54.01594785269913]
We show that optimal weights of deep ReLU neural networks are given by the wedge product of training samples when trained with standard regularized loss.
The training problem reduces to convex optimization over wedge product features, which encode the geometric structure of the training dataset.
arXiv Detail & Related papers (2023-09-28T15:19:30Z) - Towards a mathematical understanding of learning from few examples with
nonlinear feature maps [68.8204255655161]
We consider the problem of data classification where the training set consists of just a few data points.
We reveal key relationships between the geometry of an AI model's feature space, the structure of the underlying data distributions, and the model's generalisation capabilities.
arXiv Detail & Related papers (2022-11-07T14:52:58Z) - Dist2Cycle: A Simplicial Neural Network for Homology Localization [66.15805004725809]
Simplicial complexes can be viewed as high dimensional generalizations of graphs that explicitly encode multi-way ordered relations.
We propose a graph convolutional model for learning functions parametrized by the $k$-homological features of simplicial complexes.
arXiv Detail & Related papers (2021-10-28T14:59:41Z) - Probabilistic methods for approximate archetypal analysis [8.829245587252435]
Archetypal analysis is an unsupervised learning method for exploratory data analysis.
We introduce two preprocessing techniques to reduce the dimension and representation cardinality of the data.
We demonstrate the usefulness of our results by applying our method to summarize several moderately large-scale datasets.
arXiv Detail & Related papers (2021-08-12T14:27:11Z) - Analysis of Truncated Orthogonal Iteration for Sparse Eigenvector
Problems [78.95866278697777]
We propose two variants of the Truncated Orthogonal Iteration to compute multiple leading eigenvectors with sparsity constraints simultaneously.
We then apply our algorithms to solve the sparse principle component analysis problem for a wide range of test datasets.
arXiv Detail & Related papers (2021-03-24T23:11:32Z) - Joint Network Topology Inference via Structured Fusion Regularization [70.30364652829164]
Joint network topology inference represents a canonical problem of learning multiple graph Laplacian matrices from heterogeneous graph signals.
We propose a general graph estimator based on a novel structured fusion regularization.
We show that the proposed graph estimator enjoys both high computational efficiency and rigorous theoretical guarantee.
arXiv Detail & Related papers (2021-03-05T04:42:32Z) - Principled Simplicial Neural Networks for Trajectory Prediction [17.016397531234393]
We consider the construction of neural network architectures for data on simplicial complexes.
Based on these properties, we propose a simple convolutional architecture for the problem of trajectory prediction.
We show that it obeys all three of these properties when an odd, nonlinear activation function is used.
arXiv Detail & Related papers (2021-02-19T17:37:43Z) - Isometric Multi-Shape Matching [50.86135294068138]
Finding correspondences between shapes is a fundamental problem in computer vision and graphics.
While isometries are often studied in shape correspondence problems, they have not been considered explicitly in the multi-matching setting.
We present a suitable optimisation algorithm for solving our formulation and provide a convergence and complexity analysis.
arXiv Detail & Related papers (2020-12-04T15:58:34Z) - Isometric Transformation Invariant and Equivariant Graph Convolutional
Networks [5.249805590164902]
We propose a set of transformation invariant and equivariant models based on graph convolutional networks, called IsoGCNs.
We demonstrate that the proposed model has a competitive performance compared to state-of-the-art methods on tasks related to geometrical and physical simulation data.
arXiv Detail & Related papers (2020-05-13T13:44:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.