Metric Distribution to Vector: Constructing Data Representation via
Broad-Scale Discrepancies
- URL: http://arxiv.org/abs/2210.00415v1
- Date: Sun, 2 Oct 2022 03:18:30 GMT
- Title: Metric Distribution to Vector: Constructing Data Representation via
Broad-Scale Discrepancies
- Authors: Xue Liu, Dan Sun, Xiaobo Cao, Hao Ye, Wei Wei
- Abstract summary: We present a novel embedding strategy named $mathbfMetricDistribution2vec$ to extract distribution characteristics into the vectorial representation for each data.
We demonstrate the application and effectiveness of our representation method in the supervised prediction tasks on extensive real-world structural graph datasets.
- Score: 15.40538348604094
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph embedding provides a feasible methodology to conduct pattern
classification for graph-structured data by mapping each data into the
vectorial space. Various pioneering works are essentially coding method that
concentrates on a vectorial representation about the inner properties of a
graph in terms of the topological constitution, node attributions, link
relations, etc. However, the classification for each targeted data is a
qualitative issue based on understanding the overall discrepancies within the
dataset scale. From the statistical point of view, these discrepancies manifest
a metric distribution over the dataset scale if the distance metric is adopted
to measure the pairwise similarity or dissimilarity. Therefore, we present a
novel embedding strategy named $\mathbf{MetricDistribution2vec}$ to extract
such distribution characteristics into the vectorial representation for each
data. We demonstrate the application and effectiveness of our representation
method in the supervised prediction tasks on extensive real-world structural
graph datasets. The results have gained some unexpected increases compared with
a surge of baselines on all the datasets, even if we take the lightweight
models as classifiers. Moreover, the proposed methods also conducted
experiments in Few-Shot classification scenarios, and the results still show
attractive discrimination in rare training samples based inference.
Related papers
- Disentangled Hyperbolic Representation Learning for Heterogeneous Graphs [29.065531121422204]
We propose $textDis-H2textGCN$, a Disentangled Hyperbolic Heterogeneous Graph Convolutional Network.
We evaluate our proposed $textDis-H2textGCN$ on five real-world heterogeneous graph datasets.
arXiv Detail & Related papers (2024-06-14T18:50:47Z) - Improving embedding of graphs with missing data by soft manifolds [51.425411400683565]
The reliability of graph embeddings depends on how much the geometry of the continuous space matches the graph structure.
We introduce a new class of manifold, named soft manifold, that can solve this situation.
Using soft manifold for graph embedding, we can provide continuous spaces to pursue any task in data analysis over complex datasets.
arXiv Detail & Related papers (2023-11-29T12:48:33Z) - Graph Out-of-Distribution Generalization with Controllable Data
Augmentation [51.17476258673232]
Graph Neural Network (GNN) has demonstrated extraordinary performance in classifying graph properties.
Due to the selection bias of training and testing data, distribution deviation is widespread.
We propose OOD calibration to measure the distribution deviation of virtual samples.
arXiv Detail & Related papers (2023-08-16T13:10:27Z) - Bures-Wasserstein Means of Graphs [60.42414991820453]
We propose a novel framework for defining a graph mean via embeddings in the space of smooth graph signal distributions.
By finding a mean in this embedding space, we can recover a mean graph that preserves structural information.
We establish the existence and uniqueness of the novel graph mean, and provide an iterative algorithm for computing it.
arXiv Detail & Related papers (2023-05-31T11:04:53Z) - Fascinating Supervisory Signals and Where to Find Them: Deep Anomaly
Detection with Scale Learning [11.245813423781415]
We devise novel data-driven supervision for data by introducing a characteristic -- scale -- as data labels.
Scales serve as labels attached to transformed representations, thus offering ample labeled data for neural network training.
This paper further proposes a scale learning-based anomaly detection method.
arXiv Detail & Related papers (2023-05-25T14:48:00Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - Effective and Efficient Graph Learning for Multi-view Clustering [173.8313827799077]
We propose an effective and efficient graph learning model for multi-view clustering.
Our method exploits the view-similar between graphs of different views by the minimization of tensor Schatten p-norm.
Our proposed algorithm is time-economical and obtains the stable results and scales well with the data size.
arXiv Detail & Related papers (2021-08-15T13:14:28Z) - Size-Invariant Graph Representations for Graph Classification
Extrapolations [6.143735952091508]
In general, graph representation learning methods assume that the test and train data come from the same distribution.
Our work shows it is possible to use a causal model to learn approximately invariant representations that better extrapolate between train and test data.
arXiv Detail & Related papers (2021-03-08T20:01:59Z) - Learning non-Gaussian graphical models via Hessian scores and triangular
transport [6.308539010172309]
We propose an algorithm for learning the Markov structure of continuous and non-Gaussian distributions.
Our algorithm SING estimates the density using a deterministic coupling, induced by a triangular transport map, and iteratively exploits sparse structure in the map to reveal sparsity in the graph.
arXiv Detail & Related papers (2021-01-08T16:42:42Z) - Structured Graph Learning for Clustering and Semi-supervised
Classification [74.35376212789132]
We propose a graph learning framework to preserve both the local and global structure of data.
Our method uses the self-expressiveness of samples to capture the global structure and adaptive neighbor approach to respect the local structure.
Our model is equivalent to a combination of kernel k-means and k-means methods under certain condition.
arXiv Detail & Related papers (2020-08-31T08:41:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.