Simplicial Complex Representation Learning
- URL: http://arxiv.org/abs/2103.04046v2
- Date: Tue, 9 Mar 2021 07:18:25 GMT
- Title: Simplicial Complex Representation Learning
- Authors: Mustafa Hajij, Ghada Zamzmi, Xuanting Cai
- Abstract summary: Simplicial complexes form an important class of topological spaces that are frequently used in computer-aided design, computer graphics, and simulation.
In this work, we propose a method for simplicial complex-level representation learning that embeds a simplicial complex to a universal embedding space.
Our method utilizes a simplex-level embedding induced by a pre-trained simplicial autoencoder to learn an entire simplicial complex representation.
- Score: 0.7734726150561088
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Simplicial complexes form an important class of topological spaces that are
frequently used to in many applications areas such as computer-aided design,
computer graphics, and simulation. The representation learning on graphs, which
are just 1-d simplicial complexes, has witnessed a great attention and success
in the past few years. Due to the additional complexity higher dimensional
simplicial hold, there has not been enough effort to extend representation
learning to these objects especially when it comes to learn entire-simplicial
complex representation. In this work, we propose a method for simplicial
complex-level representation learning that embeds a simplicial complex to a
universal embedding space in a way that complex-to-complex proximity is
preserved. Our method utilizes a simplex-level embedding induced by a
pre-trained simplicial autoencoder to learn an entire simplicial complex
representation. To the best of our knowledge, this work presents the first
method for learning simplicial complex-level representation.
Related papers
- VisualPredicator: Learning Abstract World Models with Neuro-Symbolic Predicates for Robot Planning [86.59849798539312]
We present Neuro-Symbolic Predicates, a first-order abstraction language that combines the strengths of symbolic and neural knowledge representations.
We show that our approach offers better sample complexity, stronger out-of-distribution generalization, and improved interpretability.
arXiv Detail & Related papers (2024-10-30T16:11:05Z) - Understanding Visual Feature Reliance through the Lens of Complexity [14.282243225622093]
We introduce a new metric for quantifying feature complexity, based on $mathscrV$-information.
We analyze the complexities of 10,000 features, represented as directions in the penultimate layer, that were extracted from a standard ImageNet-trained vision model.
arXiv Detail & Related papers (2024-07-08T16:21:53Z) - Finding structure in logographic writing with library learning [55.63800121311418]
We develop a computational framework for discovering structure in a writing system.
Our framework discovers known linguistic structures in the Chinese writing system.
We demonstrate how a library learning approach may help reveal the fundamental computational principles that underlie the creation of structures in human cognition.
arXiv Detail & Related papers (2024-05-11T04:23:53Z) - Simplicity in Complexity : Explaining Visual Complexity using Deep Segmentation Models [6.324765782436764]
We propose to model complexity using segment-based representations of images.
We find that complexity is well-explained by a simple linear model with these two features across six diverse image-sets.
arXiv Detail & Related papers (2024-03-05T17:21:31Z) - SC-MAD: Mixtures of Higher-order Networks for Data Augmentation [36.33265644447091]
The simplicial complex has inspired generalizations of graph neural networks (GNNs) to simplicial complex-based models.
We propose data augmentation of simplicial complexes through both linear and nonlinear mixup mechanisms.
We theoretically demonstrate that the resultant synthetic simplicial complexes interpolate among existing data with respect to homomorphism densities.
arXiv Detail & Related papers (2023-09-14T06:25:39Z) - Spectral Convergence of Complexon Shift Operators [38.89310649097387]
We study the transferability of Topological Signal Processing via a generalized higher-order version of graphon, known as complexon.
Inspired by the graphon shift operator and message-passing neural network, we construct a marginal complexon and complexon shift operator.
We prove that when a simplicial complex signal sequence converges to a complexon signal, the eigenvalues, eigenspaces, and Fourier transform of the corresponding CSOs converge to that of the limit complexon signal.
arXiv Detail & Related papers (2023-09-12T08:40:20Z) - Generalized Simplicial Attention Neural Networks [22.171364354867723]
We introduce Generalized Simplicial Attention Neural Networks (GSANs)
GSANs process data living on simplicial complexes using masked self-attentional layers.
These schemes learn how to combine data associated with neighbor simplices of consecutive order in a task-oriented fashion.
arXiv Detail & Related papers (2023-09-05T11:29:25Z) - On the Complexity of Representation Learning in Contextual Linear
Bandits [110.84649234726442]
We show that representation learning is fundamentally more complex than linear bandits.
In particular, learning with a given set of representations is never simpler than learning with the worst realizable representation in the set.
arXiv Detail & Related papers (2022-12-19T13:08:58Z) - Structured information extraction from complex scientific text with
fine-tuned large language models [55.96705756327738]
We present a simple sequence-to-sequence approach to joint named entity recognition and relation extraction.
The approach leverages a pre-trained large language model (LLM), GPT-3, that is fine-tuned on approximately 500 pairs of prompts.
This approach represents a simple, accessible, and highly-flexible route to obtaining large databases of structured knowledge extracted from unstructured text.
arXiv Detail & Related papers (2022-12-10T07:51:52Z) - On the Complexity of Bayesian Generalization [141.21610899086392]
We consider concept generalization at a large scale in the diverse and natural visual spectrum.
We study two modes when the problem space scales up, and the $complexity$ of concepts becomes diverse.
arXiv Detail & Related papers (2022-11-20T17:21:37Z) - Dist2Cycle: A Simplicial Neural Network for Homology Localization [66.15805004725809]
Simplicial complexes can be viewed as high dimensional generalizations of graphs that explicitly encode multi-way ordered relations.
We propose a graph convolutional model for learning functions parametrized by the $k$-homological features of simplicial complexes.
arXiv Detail & Related papers (2021-10-28T14:59:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.