Identifying Key Nodes for the Influence Spread using a Machine Learning Approach
- URL: http://arxiv.org/abs/2412.01949v1
- Date: Mon, 02 Dec 2024 20:17:44 GMT
- Title: Identifying Key Nodes for the Influence Spread using a Machine Learning Approach
- Authors: Mateusz Stolarski, Adam Piróg, Piotr Bródka,
- Abstract summary: We present an enhanced machine learning-based framework for the influence spread problem.
We focus on identifying key nodes for the Independent Cascade model, which is a popular reference method.
Next, we show that our methodology allows ML models to not only predict the influence of a given node, but to also determine other characteristics of the spreading process.
- Score: 0.8192907805418583
- License:
- Abstract: The identification of key nodes in complex networks is an important topic in many network science areas. It is vital to a variety of real-world applications, including viral marketing, epidemic spreading and influence maximization. In recent years, machine learning algorithms have proven to outperform the conventional, centrality-based methods in accuracy and consistency, but this approach still requires further refinement. What information about the influencers can be extracted from the network? How can we precisely obtain the labels required for training? Can these models generalize well? In this paper, we answer these questions by presenting an enhanced machine learning-based framework for the influence spread problem. We focus on identifying key nodes for the Independent Cascade model, which is a popular reference method. Our main contribution is an improved process of obtaining the labels required for training by introducing 'Smart Bins' and proving their advantage over known methods. Next, we show that our methodology allows ML models to not only predict the influence of a given node, but to also determine other characteristics of the spreading process-which is another novelty to the relevant literature. Finally, we extensively test our framework and its ability to generalize beyond complex networks of different types and sizes, gaining important insight into the properties of these methods.
Related papers
- Multi-perspective Memory Enhanced Network for Identifying Key Nodes in Social Networks [51.54002032659713]
We propose a novel Multi-perspective Memory Enhanced Network (MMEN) for identifying key nodes in social networks.
MMEN mines key nodes from multiple perspectives and utilizes memory networks to store historical information.
Our method significantly outperforms previous methods.
arXiv Detail & Related papers (2024-03-22T14:29:03Z) - Impact of network topology on the performance of Decentralized Federated
Learning [4.618221836001186]
Decentralized machine learning is gaining momentum, addressing infrastructure challenges and privacy concerns.
This study investigates the interplay between network structure and learning performance using three network topologies and six data distribution methods.
We highlight the challenges in transferring knowledge from peripheral to central nodes, attributed to a dilution effect during model aggregation.
arXiv Detail & Related papers (2024-02-28T11:13:53Z) - Deep Internal Learning: Deep Learning from a Single Input [88.59966585422914]
In many cases there is value in training a network just from the input at hand.
This is particularly relevant in many signal and image processing problems where training data is scarce and diversity is large.
This survey paper aims at covering deep internal-learning techniques that have been proposed in the past few years for these two important directions.
arXiv Detail & Related papers (2023-12-12T16:48:53Z) - Unsupervised Learning via Network-Aware Embeddings [0.0]
We show how to create network-aware embeddings by estimating the network distance between numeric node attributes.
Our method is fully open source and data and code are available to reproduce all results in the paper.
arXiv Detail & Related papers (2023-09-19T08:17:48Z) - The effect of network topologies on fully decentralized learning: a
preliminary investigation [2.9592782993171918]
In a decentralized machine learning system, data is partitioned among multiple devices or nodes, each of which trains a local model using its own data.
We investigate how different types of topologies impact the "spreading of knowledge"
Specifically, we highlight the different roles in this process of more or less connected nodes (hubs and leaves)
arXiv Detail & Related papers (2023-07-29T09:39:17Z) - Deep networks for system identification: a Survey [56.34005280792013]
System identification learns mathematical descriptions of dynamic systems from input-output data.
Main aim of the identified model is to predict new data from previous observations.
We discuss architectures commonly adopted in the literature, like feedforward, convolutional, and recurrent networks.
arXiv Detail & Related papers (2023-01-30T12:38:31Z) - Inducing Gaussian Process Networks [80.40892394020797]
We propose inducing Gaussian process networks (IGN), a simple framework for simultaneously learning the feature space as well as the inducing points.
The inducing points, in particular, are learned directly in the feature space, enabling a seamless representation of complex structured domains.
We report on experimental results for real-world data sets showing that IGNs provide significant advances over state-of-the-art methods.
arXiv Detail & Related papers (2022-04-21T05:27:09Z) - Vital Node Identification in Complex Networks Using a Machine
Learning-Based Approach [12.898094758070474]
We propose a machine learning-based, data driven approach for vital node identification.
The main idea is to train the model with a small portion of the graph, say 0.5% of the nodes, and do the prediction on the rest of the nodes.
Several machine learning models are trained on the node representations, but the best results are achieved by a Support Vector Regression machine with RBF kernel.
arXiv Detail & Related papers (2022-02-13T06:54:18Z) - A neural anisotropic view of underspecification in deep learning [60.119023683371736]
We show that the way neural networks handle the underspecification of problems is highly dependent on the data representation.
Our results highlight that understanding the architectural inductive bias in deep learning is fundamental to address the fairness, robustness, and generalization of these systems.
arXiv Detail & Related papers (2021-04-29T14:31:09Z) - Model-Based Deep Learning [155.063817656602]
Signal processing, communications, and control have traditionally relied on classical statistical modeling techniques.
Deep neural networks (DNNs) use generic architectures which learn to operate from data, and demonstrate excellent performance.
We are interested in hybrid techniques that combine principled mathematical models with data-driven systems to benefit from the advantages of both approaches.
arXiv Detail & Related papers (2020-12-15T16:29:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.