Graph Convolutional Networks from the Perspective of Sheaves and the
Neural Tangent Kernel
- URL: http://arxiv.org/abs/2208.09309v1
- Date: Fri, 19 Aug 2022 12:46:49 GMT
- Title: Graph Convolutional Networks from the Perspective of Sheaves and the
Neural Tangent Kernel
- Authors: Thomas Gebhart
- Abstract summary: Graph convolutional networks are a popular class of deep neural network algorithms.
Despite their success, graph convolutional networks exhibit a number of peculiar features, including a bias towards learning oversmoothed and homophilic functions.
We propose to bridge this gap by studying the neural tangent kernel of sheaf convolutional networks.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph convolutional networks are a popular class of deep neural network
algorithms which have shown success in a number of relational learning tasks.
Despite their success, graph convolutional networks exhibit a number of
peculiar features, including a bias towards learning oversmoothed and
homophilic functions, which are not easily diagnosed due to the complex nature
of these algorithms. We propose to bridge this gap in understanding by studying
the neural tangent kernel of sheaf convolutional networks--a topological
generalization of graph convolutional networks. To this end, we derive a
parameterization of the neural tangent kernel for sheaf convolutional networks
which separates the function into two parts: one driven by a forward diffusion
process determined by the graph, and the other determined by the composite
effect of nodes' activations on the output layer. This geometrically-focused
derivation produces a number of immediate insights which we discuss in detail.
Related papers
- Emergence of Globally Attracting Fixed Points in Deep Neural Networks With Nonlinear Activations [24.052411316664017]
We introduce a theoretical framework for the evolution of the kernel sequence, which measures the similarity between the hidden representation for two different inputs.
For nonlinear activations, the kernel sequence converges globally to a unique fixed point, which can correspond to similar representations depending on the activation and network architecture.
This work provides new insights into the implicit biases of deep neural networks and how architectural choices influence the evolution of representations across layers.
arXiv Detail & Related papers (2024-10-26T07:10:47Z) - Graph Neural Networks Provably Benefit from Structural Information: A
Feature Learning Perspective [53.999128831324576]
Graph neural networks (GNNs) have pioneered advancements in graph representation learning.
This study investigates the role of graph convolution within the context of feature learning theory.
arXiv Detail & Related papers (2023-06-24T10:21:11Z) - Understanding Convolutional Neural Networks from Theoretical Perspective
via Volterra Convolution [22.058311878382142]
This study explores the relationship between convolutional neural networks and finite Volterra convolutions.
It provides a novel approach to explain and study the overall characteristics of neural networks without being disturbed by the complex network architectures.
arXiv Detail & Related papers (2021-10-19T12:07:46Z) - Deep Learning Chromatic and Clique Numbers of Graphs [0.0]
We develop deep learning models to predict the chromatic number and maximum clique size of graphs.
In particular, we show that deep neural networks, and in particular convolutional neural networks, obtain strong performance on this problem.
arXiv Detail & Related papers (2021-08-04T02:02:53Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - Learning Connectivity of Neural Networks from a Topological Perspective [80.35103711638548]
We propose a topological perspective to represent a network into a complete graph for analysis.
By assigning learnable parameters to the edges which reflect the magnitude of connections, the learning process can be performed in a differentiable manner.
This learning process is compatible with existing networks and owns adaptability to larger search spaces and different tasks.
arXiv Detail & Related papers (2020-08-19T04:53:31Z) - Towards Deeper Graph Neural Networks [63.46470695525957]
Graph convolutions perform neighborhood aggregation and represent one of the most important graph operations.
Several recent studies attribute this performance deterioration to the over-smoothing issue.
We propose Deep Adaptive Graph Neural Network (DAGNN) to adaptively incorporate information from large receptive fields.
arXiv Detail & Related papers (2020-07-18T01:11:14Z) - Graph Structure of Neural Networks [104.33754950606298]
We show how the graph structure of neural networks affect their predictive performance.
A "sweet spot" of relational graphs leads to neural networks with significantly improved predictive performance.
Top-performing neural networks have graph structure surprisingly similar to those of real biological neural networks.
arXiv Detail & Related papers (2020-07-13T17:59:31Z) - Geometrically Principled Connections in Graph Neural Networks [66.51286736506658]
We argue geometry should remain the primary driving force behind innovation in the emerging field of geometric deep learning.
We relate graph neural networks to widely successful computer graphics and data approximation models: radial basis functions (RBFs)
We introduce affine skip connections, a novel building block formed by combining a fully connected layer with any graph convolution operator.
arXiv Detail & Related papers (2020-04-06T13:25:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.