Smoothness Sensor: Adaptive Smoothness-Transition Graph Convolutions for
Attributed Graph Clustering
- URL: http://arxiv.org/abs/2009.05743v1
- Date: Sat, 12 Sep 2020 08:12:27 GMT
- Title: Smoothness Sensor: Adaptive Smoothness-Transition Graph Convolutions for
Attributed Graph Clustering
- Authors: Chaojie Ji, Hongwei Chen, Ruxin Wang, Yunpeng Cai, Hongyan Wu
- Abstract summary: We propose a smoothness sensor for attributed graph clustering based on adaptive smoothness-transition graph convolutions.
As an alternative to graph-level smoothness, a novel fine-gained node-wise level assessment of smoothness is proposed.
Experiments show that the proposed methods significantly outperform 12 other state-of-the-art baselines in terms of three different metrics.
- Score: 10.905770964670191
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Clustering techniques attempt to group objects with similar properties into a
cluster. Clustering the nodes of an attributed graph, in which each node is
associated with a set of feature attributes, has attracted significant
attention. Graph convolutional networks (GCNs) represent an effective approach
for integrating the two complementary factors of node attributes and structural
information for attributed graph clustering. However, oversmoothing of GCNs
produces indistinguishable representations of nodes, such that the nodes in a
graph tend to be grouped into fewer clusters, and poses a challenge due to the
resulting performance drop. In this study, we propose a smoothness sensor for
attributed graph clustering based on adaptive smoothness-transition graph
convolutions, which senses the smoothness of a graph and adaptively terminates
the current convolution once the smoothness is saturated to prevent
oversmoothing. Furthermore, as an alternative to graph-level smoothness, a
novel fine-gained node-wise level assessment of smoothness is proposed, in
which smoothness is computed in accordance with the neighborhood conditions of
a given node at a certain order of graph convolution. In addition, a
self-supervision criterion is designed considering both the tightness within
clusters and the separation between clusters to guide the whole neural network
training process. Experiments show that the proposed methods significantly
outperform 12 other state-of-the-art baselines in terms of three different
metrics across four benchmark datasets. In addition, an extensive study reveals
the reasons for their effectiveness and efficiency.
Related papers
- Modularity aided consistent attributed graph clustering via coarsening [6.522020196906943]
Graph clustering is an important unsupervised learning technique for partitioning graphs with attributes and detecting communities.
We propose a loss function incorporating log-determinant, smoothness, and modularity components using a block majorization-minimization technique.
Our algorithm seamlessly integrates graph neural networks (GNNs) and variational graph autoencoders (VGAEs) to learn enhanced node features and deliver exceptional clustering performance.
arXiv Detail & Related papers (2024-07-09T10:42:19Z) - GraphRARE: Reinforcement Learning Enhanced Graph Neural Network with Relative Entropy [21.553180564868306]
GraphRARE is a framework built upon node relative entropy and deep reinforcement learning.
An innovative node relative entropy is used to measure mutual information between node pairs.
A deep reinforcement learning-based algorithm is developed to optimize the graph topology.
arXiv Detail & Related papers (2023-12-15T11:30:18Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - Data-heterogeneity-aware Mixing for Decentralized Learning [63.83913592085953]
We characterize the dependence of convergence on the relationship between the mixing weights of the graph and the data heterogeneity across nodes.
We propose a metric that quantifies the ability of a graph to mix the current gradients.
Motivated by our analysis, we propose an approach that periodically and efficiently optimize the metric.
arXiv Detail & Related papers (2022-04-13T15:54:35Z) - Graph Decipher: A transparent dual-attention graph neural network to
understand the message-passing mechanism for the node classification [2.0047096160313456]
We propose a new transparent network called Graph Decipher to investigate the message-passing mechanism.
Our algorithm achieves state-of-the-art performance while imposing a substantially lower burden under the node classification task.
arXiv Detail & Related papers (2022-01-04T23:24:00Z) - Effective and Efficient Graph Learning for Multi-view Clustering [173.8313827799077]
We propose an effective and efficient graph learning model for multi-view clustering.
Our method exploits the view-similar between graphs of different views by the minimization of tensor Schatten p-norm.
Our proposed algorithm is time-economical and obtains the stable results and scales well with the data size.
arXiv Detail & Related papers (2021-08-15T13:14:28Z) - Explicit Pairwise Factorized Graph Neural Network for Semi-Supervised
Node Classification [59.06717774425588]
We propose the Explicit Pairwise Factorized Graph Neural Network (EPFGNN), which models the whole graph as a partially observed Markov Random Field.
It contains explicit pairwise factors to model output-output relations and uses a GNN backbone to model input-output relations.
We conduct experiments on various datasets, which shows that our model can effectively improve the performance for semi-supervised node classification on graphs.
arXiv Detail & Related papers (2021-07-27T19:47:53Z) - Anisotropic Graph Convolutional Network for Semi-supervised Learning [7.843067454030999]
Graph convolutional networks learn effective node embeddings that have proven to be useful in achieving high-accuracy prediction results.
These networks suffer from the issue of over-smoothing and shrinking effect of the graph due in large part to the fact that they diffuse features across the edges of the graph using a linear Laplacian flow.
We propose an anisotropic graph convolutional network for semi-supervised node classification by introducing a nonlinear function that captures informative features from nodes, while preventing oversmoothing.
arXiv Detail & Related papers (2020-10-20T13:56:03Z) - Graph InfoClust: Leveraging cluster-level node information for
unsupervised graph representation learning [12.592903558338444]
We propose a graph representation learning method called Graph InfoClust.
It seeks to additionally capture cluster-level information content.
This optimization leads the node representations to capture richer information and nodal interactions, which improves their quality.
arXiv Detail & Related papers (2020-09-15T09:33:20Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z) - Graph Inference Learning for Semi-supervised Classification [50.55765399527556]
We propose a Graph Inference Learning framework to boost the performance of semi-supervised node classification.
For learning the inference process, we introduce meta-optimization on structure relations from training nodes to validation nodes.
Comprehensive evaluations on four benchmark datasets demonstrate the superiority of our proposed GIL when compared against state-of-the-art methods.
arXiv Detail & Related papers (2020-01-17T02:52:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.