Convergent Boosted Smoothing for Modeling Graph Data with Tabular Node
Features
- URL: http://arxiv.org/abs/2110.13413v1
- Date: Tue, 26 Oct 2021 04:53:12 GMT
- Title: Convergent Boosted Smoothing for Modeling Graph Data with Tabular Node
Features
- Authors: Jiuhai Chen, Jonas Mueller, Vassilis N. Ioannidis, Soji Adeshina,
Yangkun Wang, Tom Goldstein, David Wipf
- Abstract summary: We propose a framework for iterating boosting with graph propagation steps.
Our approach is anchored in a principled meta loss function.
Across a variety of non-iid graph datasets, our method achieves comparable or superior performance.
- Score: 46.052312251801
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: For supervised learning with tabular data, decision tree ensembles produced
via boosting techniques generally dominate real-world applications involving
iid training/test sets. However for graph data where the iid assumption is
violated due to structured relations between samples, it remains unclear how to
best incorporate this structure within existing boosting pipelines. To this
end, we propose a generalized framework for iterating boosting with graph
propagation steps that share node/sample information across edges connecting
related samples. Unlike previous efforts to integrate graph-based models with
boosting, our approach is anchored in a principled meta loss function such that
provable convergence can be guaranteed under relatively mild assumptions.
Across a variety of non-iid graph datasets with tabular node features, our
method achieves comparable or superior performance than both tabular and graph
neural network models, as well as existing hybrid strategies that combine the
two. Beyond producing better predictive performance than recently proposed
graph models, our proposed techniques are easy to implement, computationally
more efficient, and enjoy stronger theoretical guarantees (which make our
results more reproducible).
Related papers
- Virtual Node Generation for Node Classification in Sparsely-Labeled Graphs [2.0060301665996016]
This paper presents a novel node generation method that infuses a small set of high-quality synthesized nodes into the graph as additional labeled nodes.
It is compatible with most popular graph pre-training (self-supervised learning), semi-supervised learning, and meta-learning methods.
Our Experiments demonstrate statistically significant performance improvements over 14 baselines on 10 publicly available datasets.
arXiv Detail & Related papers (2024-09-12T02:36:44Z) - Contrastive Learning for Non-Local Graphs with Multi-Resolution
Structural Views [1.4445779250002606]
We propose a novel multiview contrastive learning approach that integrates diffusion filters on graphs.
By incorporating multiple graph views as augmentations, our method captures the structural equivalence in heterophilic graphs.
arXiv Detail & Related papers (2023-08-19T17:42:02Z) - A Robust Stacking Framework for Training Deep Graph Models with
Multifaceted Node Features [61.92791503017341]
Graph Neural Networks (GNNs) with numerical node features and graph structure as inputs have demonstrated superior performance on various supervised learning tasks with graph data.
The best models for such data types in most standard supervised learning settings with IID (non-graph) data are not easily incorporated into a GNN.
Here we propose a robust stacking framework that fuses graph-aware propagation with arbitrary models intended for IID data.
arXiv Detail & Related papers (2022-06-16T22:46:33Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - Bayesian Graph Contrastive Learning [55.36652660268726]
We propose a novel perspective of graph contrastive learning methods showing random augmentations leads to encoders.
Our proposed method represents each node by a distribution in the latent space in contrast to existing techniques which embed each node to a deterministic vector.
We show a considerable improvement in performance compared to existing state-of-the-art methods on several benchmark datasets.
arXiv Detail & Related papers (2021-12-15T01:45:32Z) - A Robust and Generalized Framework for Adversarial Graph Embedding [73.37228022428663]
We propose a robust framework for adversarial graph embedding, named AGE.
AGE generates the fake neighbor nodes as the enhanced negative samples from the implicit distribution.
Based on this framework, we propose three models to handle three types of graph data.
arXiv Detail & Related papers (2021-05-22T07:05:48Z) - Heuristic Semi-Supervised Learning for Graph Generation Inspired by
Electoral College [80.67842220664231]
We propose a novel pre-processing technique, namely ELectoral COllege (ELCO), which automatically expands new nodes and edges to refine the label similarity within a dense subgraph.
In all setups tested, our method boosts the average score of base models by a large margin of 4.7 points, as well as consistently outperforms the state-of-the-art.
arXiv Detail & Related papers (2020-06-10T14:48:48Z) - Residual Correlation in Graph Neural Network Regression [39.54530450932135]
We show that conditional independence assumption severely limits predictive power.
We address this problem with an interpretable and efficient framework.
Our framework achieves substantially higher accuracy than competing baselines.
arXiv Detail & Related papers (2020-02-19T16:32:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.