Graph Convolutional Network for Recommendation with Low-pass
Collaborative Filters
- URL: http://arxiv.org/abs/2006.15516v3
- Date: Mon, 18 Jan 2021 15:31:25 GMT
- Title: Graph Convolutional Network for Recommendation with Low-pass
Collaborative Filters
- Authors: Wenhui Yu and Zheng Qin
- Abstract summary: We propose a textbfLow-pass textbfCollaborative textbfFilter (textbfLCF) to make it applicable to the large graph.
LCF is designed to remove the noise caused by exposure and quantization in the observed data.
Experiments show that LCF improves the effectiveness and efficiency of graph convolution and our GCN outperforms existing GCNs significantly.
- Score: 32.76804332450971
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: \textbf{G}raph \textbf{C}onvolutional \textbf{N}etwork (\textbf{GCN}) is
widely used in graph data learning tasks such as recommendation. However, when
facing a large graph, the graph convolution is very computationally expensive
thus is simplified in all existing GCNs, yet is seriously impaired due to the
oversimplification. To address this gap, we leverage the \textit{original graph
convolution} in GCN and propose a \textbf{L}ow-pass \textbf{C}ollaborative
\textbf{F}ilter (\textbf{LCF}) to make it applicable to the large graph. LCF is
designed to remove the noise caused by exposure and quantization in the
observed data, and it also reduces the complexity of graph convolution in an
unscathed way. Experiments show that LCF improves the effectiveness and
efficiency of graph convolution and our GCN outperforms existing GCNs
significantly. Codes are available on \url{https://github.com/Wenhui-Yu/LCFN}.
Related papers
- Mitigating Over-Squashing in Graph Neural Networks by Spectrum-Preserving Sparsification [81.06278257153835]
We propose a graph rewiring method that balances structural bottleneck reduction and graph property preservation.<n>Our method generates graphs with enhanced connectivity while maintaining sparsity and largely preserving the original graph spectrum.
arXiv Detail & Related papers (2025-06-19T08:01:00Z) - Schreier-Coset Graph Propagation [1.1126342180866644]
This work introduces Schrier-Coset Graph Propagation (SCGP), a group-theoretic augmentation method that enriches node features through Schreier-coset embeddings without altering the input graph topology.<n>SCGP embeds bottleneck-free connectivity patterns into a compact feature space, improving long-range message passing while maintaining computational efficiency.
arXiv Detail & Related papers (2025-05-15T15:14:02Z) - Efficient Graph Condensation via Gaussian Process [11.304327316816561]
Graph condensation reduces the size of large graphs while preserving performance.
Existing methods often rely on bi-level optimization, requiring extensive GNN training and limiting their scalability.
This paper proposes Graph Condensation via Gaussian Process (GCGP), a novel and computationally efficient approach to graph condensation.
arXiv Detail & Related papers (2025-01-05T14:43:07Z) - Cayley Graph Propagation [0.0]
Graph neural networks (GNNs) are notoriously vulnerable to over-squashing.<n>In this work, we show that truncation is detrimental to the coveted expansion properties.<n>We propose CGP, a method to propagate information over a complete Cayley graph structure.
arXiv Detail & Related papers (2024-10-04T13:32:34Z) - Scalable Graph Compressed Convolutions [68.85227170390864]
We propose a differentiable method that applies permutations to calibrate input graphs for Euclidean convolution.
Based on the graph calibration, we propose the Compressed Convolution Network (CoCN) for hierarchical graph representation learning.
arXiv Detail & Related papers (2024-07-26T03:14:13Z) - RobGC: Towards Robust Graph Condensation [61.259453496191696]
Graph neural networks (GNNs) have attracted widespread attention for their impressive capability of graph representation learning.<n>However, the increasing prevalence of large-scale graphs presents a significant challenge for GNN training due to their computational demands.<n>We propose graph condensation (GC) to generate an informative compact graph that enables efficient training of GNNs while retaining performance.
arXiv Detail & Related papers (2024-06-19T04:14:57Z) - Graph Data Condensation via Self-expressive Graph Structure Reconstruction [7.4525875528900665]
We introduce a novel framework named textbfGraph Data textbfCondensation via textbfSelf-expressive Graph Structure textbfReconstruction.
Our method explicitly incorporates the original graph structure into the condensing process and captures the nuanced interdependencies between the condensed nodes.
arXiv Detail & Related papers (2024-03-12T03:54:25Z) - Disentangled Condensation for Large-scale Graphs [31.781721873508978]
Graph condensation has emerged as an intriguing technique to save the expensive training costs of Graph Neural Networks (GNNs)
We propose to disentangle the condensation process into a two-stage GNN-free paradigm, independently condensing nodes and generating edges.
This simple yet effective approach achieves at least 10 times faster than state-of-the-art methods with comparable accuracy on medium-scale graphs.
arXiv Detail & Related papers (2024-01-18T09:59:00Z) - Fused Gromov-Wasserstein Graph Mixup for Graph-level Classifications [44.15102300886821]
We propose a novel graph mixup algorithm called FGWMixup, which seeks a midpoint of source graphs in the Fused Gromov-Wasserstein metric space.
Experiments conducted on five datasets demonstrate that FGWMixup effectively improves the generalizability and robustness of GNNs.
arXiv Detail & Related papers (2023-06-28T07:00:12Z) - Beyond Graph Convolutional Network: An Interpretable
Regularizer-centered Optimization Framework [12.116373546916078]
Graph convolutional networks (GCNs) have been attracting widespread attentions due to their encouraging performance and powerful generalizations.
In this paper, we induce an interpretable regularizer-centerd optimization framework, in which by building appropriate regularizers we can interpret most GCNs.
Under the proposed framework, we devise a dual-regularizer graph convolutional network (dubbed tsGCN) to capture topological and semantic structures from graph data.
arXiv Detail & Related papers (2023-01-11T05:51:33Z) - FoSR: First-order spectral rewiring for addressing oversquashing in GNNs [0.0]
Graph neural networks (GNNs) are able to leverage the structure of graph data by passing messages along the edges of the graph.
We propose a computationally efficient algorithm that prevents oversquashing by systematically adding edges to the graph.
We find experimentally that our algorithm outperforms existing graph rewiring methods in several graph classification tasks.
arXiv Detail & Related papers (2022-10-21T07:58:03Z) - Scaling R-GCN Training with Graph Summarization [71.06855946732296]
Training of Relation Graph Convolutional Networks (R-GCN) does not scale well with the size of the graph.
In this work, we experiment with the use of graph summarization techniques to compress the graph.
We obtain reasonable results on the AIFB, MUTAG and AM datasets.
arXiv Detail & Related papers (2022-03-05T00:28:43Z) - AnchorGAE: General Data Clustering via $O(n)$ Bipartite Graph
Convolution [79.44066256794187]
We show how to convert a non-graph dataset into a graph by introducing the generative graph model, which is used to build graph convolution networks (GCNs)
A bipartite graph constructed by anchors is updated dynamically to exploit the high-level information behind data.
We theoretically prove that the simple update will lead to degeneration and a specific strategy is accordingly designed.
arXiv Detail & Related papers (2021-11-12T07:08:13Z) - Node Feature Kernels Increase Graph Convolutional Network Robustness [19.076912727990326]
The robustness of Graph Convolutional Networks (GCNs) to perturbations of their input is becoming a topic of increasing importance.
In this paper, a random matrix theory analysis is possible.
It is observed that enhancing the message passing step in GCNs by adding the node feature kernel to the adjacency matrix of the graph structure solves this problem.
arXiv Detail & Related papers (2021-09-04T04:20:45Z) - Fast Graph Attention Networks Using Effective Resistance Based Graph
Sparsification [70.50751397870972]
FastGAT is a method to make attention based GNNs lightweight by using spectral sparsification to generate an optimal pruning of the input graph.
We experimentally evaluate FastGAT on several large real world graph datasets for node classification tasks.
arXiv Detail & Related papers (2020-06-15T22:07:54Z) - DeeperGCN: All You Need to Train Deeper GCNs [66.64739331859226]
Graph Convolutional Networks (GCNs) have been drawing significant attention with the power of representation learning on graphs.
Unlike Convolutional Neural Networks (CNNs), which are able to take advantage of stacking very deep layers, GCNs suffer from vanishing gradient, over-smoothing and over-fitting issues when going deeper.
This paper proposes DeeperGCN that is capable of successfully and reliably training very deep GCNs.
arXiv Detail & Related papers (2020-06-13T23:00:22Z) - Graph Highway Networks [77.38665506495553]
Graph Convolution Networks (GCN) are widely used in learning graph representations due to their effectiveness and efficiency.
They suffer from the notorious over-smoothing problem, in which the learned representations converge to alike vectors when many layers are stacked.
We propose Graph Highway Networks (GHNet) which utilize gating units to balance the trade-off between homogeneity and heterogeneity in the GCN learning process.
arXiv Detail & Related papers (2020-04-09T16:26:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.