DeepSN: A Sheaf Neural Framework for Influence Maximization
- URL: http://arxiv.org/abs/2412.12416v1
- Date: Mon, 16 Dec 2024 23:49:51 GMT
- Title: DeepSN: A Sheaf Neural Framework for Influence Maximization
- Authors: Asela Hevapathige, Qing Wang, Ahad N. Zehmakan,
- Abstract summary: Influence is key topic in data mining, with broad applications in social network analysis and viral marketing.
In recent years, researchers have increasingly turned to machine learning techniques to address this problem.
DeepSN employs sheaf neural diffusion to learn diverse influence patterns in a data-driven end-to-end manner.
- Score: 7.2716257100195385
- License:
- Abstract: Influence maximization is key topic in data mining, with broad applications in social network analysis and viral marketing. In recent years, researchers have increasingly turned to machine learning techniques to address this problem. They have developed methods to learn the underlying diffusion processes in a data-driven manner, which enhances the generalizability of the solution, and have designed optimization objectives to identify the optimal seed set. Nonetheless, two fundamental gaps remain unsolved: (1) Graph Neural Networks (GNNs) are increasingly used to learn diffusion models, but in their traditional form, they often fail to capture the complex dynamics of influence diffusion, (2) Designing optimization objectives is challenging due to combinatorial explosion when solving this problem. To address these challenges, we propose a novel framework, DeepSN. Our framework employs sheaf neural diffusion to learn diverse influence patterns in a data-driven, end-to-end manner, providing enhanced separability in capturing diffusion characteristics. We also propose an optimization technique that accounts for overlapping influence between vertices, which helps to reduce the search space and identify the optimal seed set effectively and efficiently. Finally, we conduct extensive experiments on both synthetic and real-world datasets to demonstrate the effectiveness of our framework.
Related papers
- Diffusion Models as Network Optimizers: Explorations and Analysis [71.69869025878856]
generative diffusion models (GDMs) have emerged as a promising new approach to network optimization.
In this study, we first explore the intrinsic characteristics of generative models.
We provide a concise theoretical and intuitive demonstration of the advantages of generative models over discriminative network optimization.
arXiv Detail & Related papers (2024-11-01T09:05:47Z) - Influence Maximization via Graph Neural Bandits [54.45552721334886]
We set the IM problem in a multi-round diffusion campaign, aiming to maximize the number of distinct users that are influenced.
We propose the framework IM-GNB (Influence Maximization with Graph Neural Bandits), where we provide an estimate of the users' probabilities of being influenced.
arXiv Detail & Related papers (2024-06-18T17:54:33Z) - DSCom: A Data-Driven Self-Adaptive Community-Based Framework for
Influence Maximization in Social Networks [3.97535858363999]
We reformulate the problem on the attributed network and leverage the node attributes to estimate the closeness between connected nodes.
Specifically, we propose a machine learning-based framework, named DSCom, to address this problem.
Compared to the previous theoretical works, we carefully designed empirical experiments with parameterized diffusion models based on real-world social networks.
arXiv Detail & Related papers (2023-11-18T14:03:43Z) - Deep Graph Representation Learning and Optimization for Influence
Maximization [10.90744025490539]
In Influence (IM) is formulated as selecting a set of initial users from a social network to maximize the expected number of influenced users.
We propose a novel framework DeepIM to generatively characterize the latent representation of seed sets.
We also design a novel objective function to infer optimal seed sets under flexible node-centrality-based budget constraints.
arXiv Detail & Related papers (2023-05-01T15:45:01Z) - ToupleGDD: A Fine-Designed Solution of Influence Maximization by Deep
Reinforcement Learning [4.266866385061998]
We propose a novel end-to-end DRL framework, ToupleGDD, to address the Influence Maximization (IM) problem.
Our model is trained on several small randomly generated graphs with a small budget, and tested on completely different networks under various large budgets.
arXiv Detail & Related papers (2022-10-14T03:56:53Z) - DRFLM: Distributionally Robust Federated Learning with Inter-client
Noise via Local Mixup [58.894901088797376]
federated learning has emerged as a promising approach for training a global model using data from multiple organizations without leaking their raw data.
We propose a general framework to solve the above two challenges simultaneously.
We provide comprehensive theoretical analysis including robustness analysis, convergence analysis, and generalization ability.
arXiv Detail & Related papers (2022-04-16T08:08:29Z) - Learning Neural Causal Models with Active Interventions [83.44636110899742]
We introduce an active intervention-targeting mechanism which enables a quick identification of the underlying causal structure of the data-generating process.
Our method significantly reduces the required number of interactions compared with random intervention targeting.
We demonstrate superior performance on multiple benchmarks from simulated to real-world data.
arXiv Detail & Related papers (2021-09-06T13:10:37Z) - Grain: Improving Data Efficiency of Graph Neural Networks via
Diversified Influence Maximization [24.25156825467544]
Graph Neural Networks (GNNs) go beyond the models existing data selection methods are designed for.
We present Grain, an efficient framework that opens up a new perspective through connecting data selection in GNNs with social influence.
Empirical studies on public datasets demonstrate that Grain significantly improves both the performance and efficiency of data selection for GNNs.
arXiv Detail & Related papers (2021-07-31T11:39:00Z) - Influence Estimation and Maximization via Neural Mean-Field Dynamics [60.91291234832546]
We propose a novel learning framework using neural mean-field (NMF) dynamics for inference and estimation problems.
Our framework can simultaneously learn the structure of the diffusion network and the evolution of node infection probabilities.
arXiv Detail & Related papers (2021-06-03T00:02:05Z) - Network Diffusions via Neural Mean-Field Dynamics [52.091487866968286]
We propose a novel learning framework for inference and estimation problems of diffusion on networks.
Our framework is derived from the Mori-Zwanzig formalism to obtain an exact evolution of the node infection probabilities.
Our approach is versatile and robust to variations of the underlying diffusion network models.
arXiv Detail & Related papers (2020-06-16T18:45:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.