Sub-Graph Learning for Spatiotemporal Forecasting via Knowledge
Distillation
- URL: http://arxiv.org/abs/2211.09740v1
- Date: Thu, 17 Nov 2022 18:02:55 GMT
- Title: Sub-Graph Learning for Spatiotemporal Forecasting via Knowledge
Distillation
- Authors: Mehrtash Mehrabi and Yingxue Zhang
- Abstract summary: We present a new framework called KD-SGL to effectively learn the sub-graphs.
We define one global model to learn the overall structure of the graph and multiple local models for each sub-graph.
- Score: 22.434970343698676
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One of the challenges in studying the interactions in large graphs is to
learn their diverse pattern and various interaction types. Hence, considering
only one distribution and model to study all nodes and ignoring their diversity
and local features in their neighborhoods, might severely affect the overall
performance. Based on the structural information of the nodes in the graph and
the interactions between them, the main graph can be divided into multiple
sub-graphs. This graph partitioning can tremendously affect the learning
process, however the overall performance is highly dependent on the clustering
method to avoid misleading the model. In this work, we present a new framework
called KD-SGL to effectively learn the sub-graphs, where we define one global
model to learn the overall structure of the graph and multiple local models for
each sub-graph. We assess the performance of the proposed framework and
evaluate it on public datasets. Based on the achieved results, it can improve
the performance of the state-of-the-arts spatiotemporal models with comparable
results compared to ensemble of models with less complexity.
Related papers
- Multi-View Stochastic Block Models [34.55723218769512]
We formalize a new family of models, called textitmulti-view block models that captures this setting.
For this model, we first study efficient algorithms that naively work on the union of multiple graphs.
Then, we introduce a new efficient algorithm that provably outperforms previous approaches by analyzing the structure of each graph separately.
arXiv Detail & Related papers (2024-06-07T11:45:31Z) - GraphGLOW: Universal and Generalizable Structure Learning for Graph
Neural Networks [72.01829954658889]
This paper introduces the mathematical definition of this novel problem setting.
We devise a general framework that coordinates a single graph-shared structure learner and multiple graph-specific GNNs.
The well-trained structure learner can directly produce adaptive structures for unseen target graphs without any fine-tuning.
arXiv Detail & Related papers (2023-06-20T03:33:22Z) - Globally Interpretable Graph Learning via Distribution Matching [12.885580925389352]
We aim to answer an important question that is not yet well studied: how to provide a global interpretation for the graph learning procedure?
We formulate this problem as globally interpretable graph learning, which targets on distilling high-level and human-intelligible patterns that dominate the learning procedure.
We propose a novel model fidelity metric, tailored for evaluating the fidelity of the resulting model trained on interpretations.
arXiv Detail & Related papers (2023-06-18T00:50:36Z) - Learning Strong Graph Neural Networks with Weak Information [64.64996100343602]
We develop a principled approach to the problem of graph learning with weak information (GLWI)
We propose D$2$PT, a dual-channel GNN framework that performs long-range information propagation on the input graph with incomplete structure, but also on a global graph that encodes global semantic similarities.
arXiv Detail & Related papers (2023-05-29T04:51:09Z) - Effective and Efficient Graph Learning for Multi-view Clustering [173.8313827799077]
We propose an effective and efficient graph learning model for multi-view clustering.
Our method exploits the view-similar between graphs of different views by the minimization of tensor Schatten p-norm.
Our proposed algorithm is time-economical and obtains the stable results and scales well with the data size.
arXiv Detail & Related papers (2021-08-15T13:14:28Z) - Graph Representation Learning by Ensemble Aggregating Subgraphs via
Mutual Information Maximization [5.419711903307341]
We introduce a self-supervised learning method to enhance the representations of graph-level learned by Graph Neural Networks.
To get a comprehensive understanding of the graph structure, we propose an ensemble-learning like subgraph method.
And to achieve efficient and effective contrasive learning, a Head-Tail contrastive samples construction method is proposed.
arXiv Detail & Related papers (2021-03-24T12:06:12Z) - Model-Agnostic Graph Regularization for Few-Shot Learning [60.64531995451357]
We present a comprehensive study on graph embedded few-shot learning.
We introduce a graph regularization approach that allows a deeper understanding of the impact of incorporating graph information between labels.
Our approach improves the performance of strong base learners by up to 2% on Mini-ImageNet and 6.7% on ImageNet-FS.
arXiv Detail & Related papers (2021-02-14T05:28:13Z) - Structured Graph Learning for Clustering and Semi-supervised
Classification [74.35376212789132]
We propose a graph learning framework to preserve both the local and global structure of data.
Our method uses the self-expressiveness of samples to capture the global structure and adaptive neighbor approach to respect the local structure.
Our model is equivalent to a combination of kernel k-means and k-means methods under certain condition.
arXiv Detail & Related papers (2020-08-31T08:41:20Z) - Structural Landmarking and Interaction Modelling: on Resolution Dilemmas
in Graph Classification [50.83222170524406]
We study the intrinsic difficulty in graph classification under the unified concept of resolution dilemmas''
We propose SLIM'', an inductive neural network model for Structural Landmarking and Interaction Modelling.
arXiv Detail & Related papers (2020-06-29T01:01:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.