Attentive Graph Neural Networks for Few-Shot Learning
- URL: http://arxiv.org/abs/2007.06878v2
- Date: Fri, 2 Oct 2020 13:08:56 GMT
- Title: Attentive Graph Neural Networks for Few-Shot Learning
- Authors: Hao Cheng, Joey Tianyi Zhou, Wee Peng Tay and Bihan Wen
- Abstract summary: Graph Neural Networks (GNN) has demonstrated the superior performance in many challenging applications, including the few-shot learning tasks.
Despite its powerful capacity to learn and generalize the model from few samples, GNN usually suffers from severe over-fitting and over-smoothing as the model becomes deep.
We propose a novel Attentive GNN to tackle these challenges, by incorporating a triple-attention mechanism.
- Score: 74.01069516079379
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Neural Networks (GNN) has demonstrated the superior performance in many
challenging applications, including the few-shot learning tasks. Despite its
powerful capacity to learn and generalize the model from few samples, GNN
usually suffers from severe over-fitting and over-smoothing as the model
becomes deep, which limit the scalability. In this work, we propose a novel
Attentive GNN to tackle these challenges, by incorporating a triple-attention
mechanism, i.e. node self-attention, neighborhood attention, and layer memory
attention. We explain why the proposed attentive modules can improve GNN for
few-shot learning with theoretical analysis and illustrations. Extensive
experiments show that the proposed Attentive GNN model achieves the promising
results, comparing to the state-of-the-art GNN- and CNN-based methods for
few-shot learning tasks, over the mini-ImageNet and tiered-ImageNet benchmarks,
under ConvNet-4 and ResNet-based backbone with both inductive and transductive
settings. The codes will be made publicly available.
Related papers
- CNN2GNN: How to Bridge CNN with GNN [59.42117676779735]
We propose a novel CNN2GNN framework to unify CNN and GNN together via distillation.
The performance of distilled boosted'' two-layer GNN on Mini-ImageNet is much higher than CNN containing dozens of layers such as ResNet152.
arXiv Detail & Related papers (2024-04-23T08:19:08Z) - Information Flow in Graph Neural Networks: A Clinical Triage Use Case [49.86931948849343]
Graph Neural Networks (GNNs) have gained popularity in healthcare and other domains due to their ability to process multi-modal and multi-relational graphs.
We investigate how the flow of embedding information within GNNs affects the prediction of links in Knowledge Graphs (KGs)
Our results demonstrate that incorporating domain knowledge into the GNN connectivity leads to better performance than using the same connectivity as the KG or allowing unconstrained embedding propagation.
arXiv Detail & Related papers (2023-09-12T09:18:12Z) - LazyGNN: Large-Scale Graph Neural Networks via Lazy Propagation [51.552170474958736]
We propose to capture long-distance dependency in graphs by shallower models instead of deeper models, which leads to a much more efficient model, LazyGNN, for graph representation learning.
LazyGNN is compatible with existing scalable approaches (such as sampling methods) for further accelerations through the development of mini-batch LazyGNN.
Comprehensive experiments demonstrate its superior prediction performance and scalability on large-scale benchmarks.
arXiv Detail & Related papers (2023-02-03T02:33:07Z) - Understanding and Improving Deep Graph Neural Networks: A Probabilistic
Graphical Model Perspective [22.82625446308785]
We propose a novel view for understanding graph neural networks (GNNs)
In this work, we focus on deep GNNs and propose a novel view for understanding them.
We design a more powerful GNN: coupling graph neural network (CoGNet)
arXiv Detail & Related papers (2023-01-25T12:02:12Z) - Addressing Over-Smoothing in Graph Neural Networks via Deep Supervision [13.180922099929765]
Deep graph neural networks (GNNs) suffer from over-smoothing when the number of layers increases.
We propose DSGNNs enhanced with deep supervision where representations learned at all layers are used for training.
We show that DSGNNs are resilient to over-smoothing and can outperform competitive benchmarks on node and graph property prediction problems.
arXiv Detail & Related papers (2022-02-25T06:05:55Z) - IGNNITION: Bridging the Gap Between Graph Neural Networks and Networking
Systems [4.1591055164123665]
We present IGNNITION, a novel open-source framework that enables fast prototyping of Graph Neural Networks (GNNs) for networking systems.
IGNNITION is based on an intuitive high-level abstraction that hides the complexity behind GNNs.
Our results show that the GNN models produced by IGNNITION are equivalent in terms of accuracy and performance to their native implementations.
arXiv Detail & Related papers (2021-09-14T14:28:21Z) - GNNLens: A Visual Analytics Approach for Prediction Error Diagnosis of
Graph Neural Networks [42.222552078920216]
Graph Neural Networks (GNNs) aim to extend deep learning techniques to graph data.
GNNs behave like a black box with their details hidden from model developers and users.
It is therefore difficult to diagnose possible errors of GNNs.
This paper fills the research gap with an interactive visual analysis tool, GNNLens, to assist model developers and users in understanding and analyzing GNNs.
arXiv Detail & Related papers (2020-11-22T16:09:08Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - A Survey on The Expressive Power of Graph Neural Networks [6.09170287691728]
Graph neural networks (GNNs) are effective machine learning models for various graph learning problems.
Despite their empirical successes, the theoretical limitations of GNNs have been revealed recently.
We provide a comprehensive overview of the expressive power of GNNs and provably powerful variants of GNNs.
arXiv Detail & Related papers (2020-03-09T12:37:40Z) - Efficient Probabilistic Logic Reasoning with Graph Neural Networks [63.099999467118245]
Markov Logic Networks (MLNs) can be used to address many knowledge graph problems.
Inference in MLN is computationally intensive, making the industrial-scale application of MLN very difficult.
We propose a graph neural network (GNN) variant, named ExpressGNN, which strikes a nice balance between the representation power and the simplicity of the model.
arXiv Detail & Related papers (2020-01-29T23:34:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.