Structure-Aware DropEdge Towards Deep Graph Convolutional Networks
- URL: http://arxiv.org/abs/2306.12091v1
- Date: Wed, 21 Jun 2023 08:11:40 GMT
- Title: Structure-Aware DropEdge Towards Deep Graph Convolutional Networks
- Authors: Jiaqi Han, Wenbing Huang, Yu Rong, Tingyang Xu, Fuchun Sun, Junzhou
Huang
- Abstract summary: Graph Convolutional Networks (GCNs) encounter a remarkable drop in performance when multiple layers are piled up.
Over-smoothing isolates the network output from the input with the increase of network depth, weakening expressivity and trainability.
We investigate refined measures upon DropEdge -- an existing simple yet effective technique to relieve over-smoothing.
- Score: 83.38709956935095
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: It has been discovered that Graph Convolutional Networks (GCNs) encounter a
remarkable drop in performance when multiple layers are piled up. The main
factor that accounts for why deep GCNs fail lies in over-smoothing, which
isolates the network output from the input with the increase of network depth,
weakening expressivity and trainability. In this paper, we start by
investigating refined measures upon DropEdge -- an existing simple yet
effective technique to relieve over-smoothing. We term our method as DropEdge++
for its two structure-aware samplers in contrast to DropEdge: layer-dependent
sampler and feature-dependent sampler. Regarding the layer-dependent sampler,
we interestingly find that increasingly sampling edges from the bottom layer
yields superior performance than the decreasing counterpart as well as
DropEdge. We theoretically reveal this phenomenon with Mean-Edge-Number (MEN),
a metric closely related to over-smoothing. For the feature-dependent sampler,
we associate the edge sampling probability with the feature similarity of node
pairs, and prove that it further correlates the convergence subspace of the
output layer with the input features. Extensive experiments on several node
classification benchmarks, including both full- and semi- supervised tasks,
illustrate the efficacy of DropEdge++ and its compatibility with a variety of
backbones by achieving generally better performance over DropEdge and the
no-drop version.
Related papers
- ADEdgeDrop: Adversarial Edge Dropping for Robust Graph Neural Networks [53.41164429486268]
Graph Neural Networks (GNNs) have exhibited the powerful ability to gather graph-structured information from neighborhood nodes.
The performance of GNNs is limited by poor generalization and fragile robustness caused by noisy and redundant graph data.
We propose a novel adversarial edge-dropping method (ADEdgeDrop) that leverages an adversarial edge predictor guiding the removal of edges.
arXiv Detail & Related papers (2024-03-14T08:31:39Z) - Edge-Aware Mirror Network for Camouflaged Object Detection [5.032585246295627]
We propose a novel Edge-aware Mirror Network (EAMNet) to model edge detection and camouflaged object segmentation.
EAMNet has a two-branch architecture, where a segmentation-induced edge aggregation module and an edge-induced integrity aggregation module are designed to cross-guide the segmentation branch and edge detection branch.
Experiment results show that EAMNet outperforms existing cutting-edge baselines on three widely used COD datasets.
arXiv Detail & Related papers (2023-07-08T08:14:49Z) - Finding Heterophilic Neighbors via Confidence-based Subgraph Matching
for Semi-supervised Node Classification [1.3190581566723918]
Graph Neural Networks (GNNs) have proven to be powerful in many graph-based applications.
However, they fail to generalize well under heterophilic setups.
arXiv Detail & Related papers (2023-02-20T04:23:34Z) - Refined Edge Usage of Graph Neural Networks for Edge Prediction [51.06557652109059]
We propose a novel edge prediction paradigm named Edge-aware Message PassIng neuRal nEtworks (EMPIRE)
We first introduce an edge splitting technique to specify use of each edge where each edge is solely used as either the topology or the supervision.
In order to emphasize the differences between pairs connected by supervision edges and pairs unconnected, we further weight the messages to highlight the relative ones that can reflect the differences.
arXiv Detail & Related papers (2022-12-25T23:19:56Z) - Improved Convergence Guarantees for Shallow Neural Networks [91.3755431537592]
We prove convergence of depth 2 neural networks, trained via gradient descent, to a global minimum.
Our model has the following features: regression with quadratic loss function, fully connected feedforward architecture, RelU activations, Gaussian data instances, adversarial labels.
They strongly suggest that, at least in our model, the convergence phenomenon extends well beyond the NTK regime''
arXiv Detail & Related papers (2022-12-05T14:47:52Z) - Joint Multi-Dimension Pruning via Numerical Gradient Update [120.59697866489668]
We present joint multi-dimension pruning (abbreviated as JointPruning), an effective method of pruning a network on three crucial aspects: spatial, depth and channel simultaneously.
We show that our method is optimized collaboratively across the three dimensions in a single end-to-end training and it is more efficient than the previous exhaustive methods.
arXiv Detail & Related papers (2020-05-18T17:57:09Z) - DHP: Differentiable Meta Pruning via HyperNetworks [158.69345612783198]
This paper introduces a differentiable pruning method via hypernetworks for automatic network pruning.
Latent vectors control the output channels of the convolutional layers in the backbone network and act as a handle for the pruning of the layers.
Experiments are conducted on various networks for image classification, single image super-resolution, and denoising.
arXiv Detail & Related papers (2020-03-30T17:59:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.