EGG-GAE: scalable graph neural networks for tabular data imputation
- URL: http://arxiv.org/abs/2210.10446v1
- Date: Wed, 19 Oct 2022 10:26:17 GMT
- Title: EGG-GAE: scalable graph neural networks for tabular data imputation
- Authors: Lev Telyatnikov and Simone Scardapane
- Abstract summary: We propose a novel EdGe Generation Graph AutoEncoder (EGG-GAE) for missing data imputation.
EGG-GAE works on randomly sampled mini-batches of the input data, and it automatically infers the best connectivity across the mini-batch for each architecture layer.
- Score: 8.775728170359024
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Missing data imputation (MDI) is crucial when dealing with tabular datasets
across various domains. Autoencoders can be trained to reconstruct missing
values, and graph autoencoders (GAE) can additionally consider similar patterns
in the dataset when imputing new values for a given instance. However,
previously proposed GAEs suffer from scalability issues, requiring the user to
define a similarity metric among patterns to build the graph connectivity
beforehand. In this paper, we leverage recent progress in latent graph
imputation to propose a novel EdGe Generation Graph AutoEncoder (EGG-GAE) for
missing data imputation that overcomes these two drawbacks. EGG-GAE works on
randomly sampled mini-batches of the input data (hence scaling to larger
datasets), and it automatically infers the best connectivity across the
mini-batch for each architecture layer. We also experiment with several
extensions, including an ensemble strategy for inference and the inclusion of
what we call prototype nodes, obtaining significant improvements, both in terms
of imputation error and final downstream accuracy, across multiple benchmarks
and baselines.
Related papers
- Scalable Weibull Graph Attention Autoencoder for Modeling Document Networks [50.42343781348247]
We develop a graph Poisson factor analysis (GPFA) which provides analytic conditional posteriors to improve the inference accuracy.
We also extend GPFA to a multi-stochastic-layer version named graph Poisson gamma belief network (GPGBN) to capture the hierarchical document relationships at multiple semantic levels.
Our models can extract high-quality hierarchical latent document representations and achieve promising performance on various graph analytic tasks.
arXiv Detail & Related papers (2024-10-13T02:22:14Z) - Revisiting, Benchmarking and Understanding Unsupervised Graph Domain Adaptation [31.106636947179005]
Unsupervised Graph Domain Adaptation involves the transfer of knowledge from a label-rich source graph to an unlabeled target graph.
We present the first comprehensive benchmark for unsupervised graph domain adaptation named GDABench.
We observe that the performance of current UGDA models varies significantly across different datasets and adaptation scenarios.
arXiv Detail & Related papers (2024-07-09T06:44:09Z) - DPGAN: A Dual-Path Generative Adversarial Network for Missing Data Imputation in Graphs [17.847551850315895]
This paper proposes a novel framework, called Dual-Pathrative Adversarial Network (DPGAN)
DPGAN can deal simultaneously with missing data and avoid over-smoothing problems.
Comprehensive experiments across various benchmark datasets substantiate that DPGAN consistently rivals, if not outperforms, existing state-of-the-art imputation algorithms.
arXiv Detail & Related papers (2024-04-26T05:26:10Z) - Novel Representation Learning Technique using Graphs for Performance
Analytics [0.0]
We propose a novel idea of transforming performance data into graphs to leverage the advancement of Graph Neural Network-based (GNN) techniques.
In contrast to other Machine Learning application domains, such as social networks, the graph is not given; instead, we need to build it.
We evaluate the effectiveness of the generated embeddings from GNNs based on how well they make even a simple feed-forward neural network perform for regression tasks.
arXiv Detail & Related papers (2024-01-19T16:34:37Z) - ADAMM: Anomaly Detection of Attributed Multi-graphs with Metadata: A
Unified Neural Network Approach [39.211176955683285]
We propose ADAMM, a novel graph neural network model that handles directed multi-graphs.
ADAMM fuses metadata and graph-level representation learning through an unsupervised anomaly detection objective.
arXiv Detail & Related papers (2023-11-13T14:19:36Z) - Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - Integrating Transformer and Autoencoder Techniques with Spectral Graph
Algorithms for the Prediction of Scarcely Labeled Molecular Data [2.8360662552057323]
This work introduces three graph-based models incorporating Merriman-Bence-Osher (MBO) techniques to tackle this challenge.
Specifically, graph-based modifications of the MBO scheme is integrated with state-of-the-art techniques, including a home-made transformer and an autoencoder.
The proposed models are validated using five benchmark data sets.
arXiv Detail & Related papers (2022-11-12T22:45:32Z) - Dynamic Graph Message Passing Networks for Visual Recognition [112.49513303433606]
Modelling long-range dependencies is critical for scene understanding tasks in computer vision.
A fully-connected graph is beneficial for such modelling, but its computational overhead is prohibitive.
We propose a dynamic graph message passing network, that significantly reduces the computational complexity.
arXiv Detail & Related papers (2022-09-20T14:41:37Z) - A Robust Stacking Framework for Training Deep Graph Models with
Multifaceted Node Features [61.92791503017341]
Graph Neural Networks (GNNs) with numerical node features and graph structure as inputs have demonstrated superior performance on various supervised learning tasks with graph data.
The best models for such data types in most standard supervised learning settings with IID (non-graph) data are not easily incorporated into a GNN.
Here we propose a robust stacking framework that fuses graph-aware propagation with arbitrary models intended for IID data.
arXiv Detail & Related papers (2022-06-16T22:46:33Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.