Understanding the network formation pattern for better link prediction
- URL: http://arxiv.org/abs/2110.08850v1
- Date: Sun, 17 Oct 2021 15:30:04 GMT
- Title: Understanding the network formation pattern for better link prediction
- Authors: Jiating Yu and Ling-Yun Wu
- Abstract summary: We propose a novel method named Link prediction using Multiple Order Local Information (MOLI)
MOLI exploits the local information from the neighbors of different distances, with parameters that can be a prior-driven based on prior knowledge.
We show that MOLI outperforms the other 11 widely used link prediction algorithms on 11 different types of simulated and real-world networks.
- Score: 4.8334761517444855
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: As a classical problem in the field of complex networks, link prediction has
attracted much attention from researchers, which is of great significance to
help us understand the evolution and dynamic development mechanisms of
networks. Although various network type-specific algorithms have been proposed
to tackle the link prediction problem, most of them suppose that the network
structure is dominated by the Triadic Closure Principle. We still lack an
adaptive and comprehensive understanding of network formation patterns for
predicting potential links. In addition, it is valuable to investigate how
network local information can be better utilized. To this end, we proposed a
novel method named Link prediction using Multiple Order Local Information
(MOLI) that exploits the local information from the neighbors of different
distances, with parameters that can be a prior-driven based on prior knowledge,
or data-driven by solving an optimization problem on observed networks. MOLI
defined a local network diffusion process via random walks on the graph,
resulting in better use of network information. We show that MOLI outperforms
the other 11 widely used link prediction algorithms on 11 different types of
simulated and real-world networks. We also conclude that there are different
patterns of local information utilization for different networks, including
social networks, communication networks, biological networks, etc. In
particular, the classical common neighbor-based algorithm is not as adaptable
to all social networks as it is perceived to be; instead, some of the social
networks obey the Quadrilateral Closure Principle which preferentially connects
paths of length three.
Related papers
- Generalization emerges from local optimization in a self-organized learning network [0.0]
We design and analyze a new paradigm for building supervised learning networks, driven only by local optimization rules without relying on a global error function.
Our network stores new knowledge in the nodes accurately and instantaneously, in the form of a lookup table.
We show on numerous examples of classification tasks that the networks generated by our algorithm systematically reach such a state of perfect generalization when the number of learned examples becomes sufficiently large.
We report on the dynamics of the change of state and show that it is abrupt and has the distinctive characteristics of a first order phase transition, a phenomenon already observed for traditional learning networks and known as grokking.
arXiv Detail & Related papers (2024-10-03T15:32:08Z) - The Principles of Deep Learning Theory [19.33681537640272]
This book develops an effective theory approach to understanding deep neural networks of practical relevance.
We explain how these effectively-deep networks learn nontrivial representations from training.
We show that the depth-to-width ratio governs the effective model complexity of the ensemble of trained networks.
arXiv Detail & Related papers (2021-06-18T15:00:00Z) - Learning Autonomy in Management of Wireless Random Networks [102.02142856863563]
This paper presents a machine learning strategy that tackles a distributed optimization task in a wireless network with an arbitrary number of randomly interconnected nodes.
We develop a flexible deep neural network formalism termed distributed message-passing neural network (DMPNN) with forward and backward computations independent of the network topology.
arXiv Detail & Related papers (2021-06-15T09:03:28Z) - Learning Structures for Deep Neural Networks [99.8331363309895]
We propose to adopt the efficient coding principle, rooted in information theory and developed in computational neuroscience.
We show that sparse coding can effectively maximize the entropy of the output signals.
Our experiments on a public image classification dataset demonstrate that using the structure learned from scratch by our proposed algorithm, one can achieve a classification accuracy comparable to the best expert-designed structure.
arXiv Detail & Related papers (2021-05-27T12:27:24Z) - Anomaly Detection on Attributed Networks via Contrastive Self-Supervised
Learning [50.24174211654775]
We present a novel contrastive self-supervised learning framework for anomaly detection on attributed networks.
Our framework fully exploits the local information from network data by sampling a novel type of contrastive instance pair.
A graph neural network-based contrastive learning model is proposed to learn informative embedding from high-dimensional attributes and local structure.
arXiv Detail & Related papers (2021-02-27T03:17:20Z) - Firefly Neural Architecture Descent: a General Approach for Growing
Neural Networks [50.684661759340145]
Firefly neural architecture descent is a general framework for progressively and dynamically growing neural networks.
We show that firefly descent can flexibly grow networks both wider and deeper, and can be applied to learn accurate but resource-efficient neural architectures.
In particular, it learns networks that are smaller in size but have higher average accuracy than those learned by the state-of-the-art methods.
arXiv Detail & Related papers (2021-02-17T04:47:18Z) - Learning Connectivity of Neural Networks from a Topological Perspective [80.35103711638548]
We propose a topological perspective to represent a network into a complete graph for analysis.
By assigning learnable parameters to the edges which reflect the magnitude of connections, the learning process can be performed in a differentiable manner.
This learning process is compatible with existing networks and owns adaptability to larger search spaces and different tasks.
arXiv Detail & Related papers (2020-08-19T04:53:31Z) - MODEL: Motif-based Deep Feature Learning for Link Prediction [23.12527010960999]
We propose a novel embedding algorithm that incorporates network motifs to capture higher-order structures in the network.
Experiments were conducted on three types of networks: social networks, biological networks, and academic networks.
Our algorithm outperforms both the traditional similarity-based algorithms by 20% and the state-of-the-art embedding-based algorithms by 19%.
arXiv Detail & Related papers (2020-08-09T03:39:28Z) - Heuristics for Link Prediction in Multiplex Networks [15.637077180633735]
We propose a novel framework and three families of families for multiplex network link prediction.
We show that the proposeds show increased performance with the richness of connection type correlation structure and significantly outperform their baselines for ordinary networks with a single connection type.
arXiv Detail & Related papers (2020-04-09T17:36:18Z) - Detecting Communities in Heterogeneous Multi-Relational Networks:A
Message Passing based Approach [89.19237792558687]
Community is a common characteristic of networks including social networks, biological networks, computer and information networks.
We propose an efficient message passing based algorithm to simultaneously detect communities for all homogeneous networks.
arXiv Detail & Related papers (2020-04-06T17:36:24Z) - Modeling Dynamic Heterogeneous Network for Link Prediction using
Hierarchical Attention with Temporal RNN [16.362525151483084]
We propose a novel dynamic heterogeneous network embedding method, termed as DyHATR.
It uses hierarchical attention to learn heterogeneous information and incorporates recurrent neural networks with temporal attention to capture evolutionary patterns.
We benchmark our method on four real-world datasets for the task of link prediction.
arXiv Detail & Related papers (2020-04-01T17:16:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.