DeHIN: A Decentralized Framework for Embedding Large-scale Heterogeneous
Information Networks
- URL: http://arxiv.org/abs/2201.02757v1
- Date: Sat, 8 Jan 2022 04:08:36 GMT
- Title: DeHIN: A Decentralized Framework for Embedding Large-scale Heterogeneous
Information Networks
- Authors: Mubashir Imran, Hongzhi Yin, Tong Chen, Zi Huang, Kai Zheng
- Abstract summary: We present textitDecentralized Embedding Framework for Heterogeneous Information Network (DeHIN) in this paper.
DeHIN presents a context preserving partition mechanism that innovatively formulates a large HIN as a hypergraph.
Our framework then adopts a decentralized strategy to efficiently partition HINs by adopting a tree-like pipeline.
- Score: 64.62314068155997
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Modeling heterogeneity by extraction and exploitation of high-order
information from heterogeneous information networks (HINs) has been attracting
immense research attention in recent times. Such heterogeneous network
embedding (HNE) methods effectively harness the heterogeneity of small-scale
HINs. However, in the real world, the size of HINs grow exponentially with the
continuous introduction of new nodes and different types of links, making it a
billion-scale network. Learning node embeddings on such HINs creates a
performance bottleneck for existing HNE methods that are commonly centralized,
i.e., complete data and the model are both on a single machine. To address
large-scale HNE tasks with strong efficiency and effectiveness guarantee, we
present \textit{Decentralized Embedding Framework for Heterogeneous Information
Network} (DeHIN) in this paper. In DeHIN, we generate a distributed parallel
pipeline that utilizes hypergraphs in order to infuse parallelization into the
HNE task. DeHIN presents a context preserving partition mechanism that
innovatively formulates a large HIN as a hypergraph, whose hyperedges connect
semantically similar nodes. Our framework then adopts a decentralized strategy
to efficiently partition HINs by adopting a tree-like pipeline. Then, each
resulting subnetwork is assigned to a distributed worker, which employs the
deep information maximization theorem to locally learn node embeddings from the
partition it receives. We further devise a novel embedding alignment scheme to
precisely project independently learned node embeddings from all subnetworks
onto a common vector space, thus allowing for downstream tasks like link
prediction and node classification.
Related papers
- Heterogenous Memory Augmented Neural Networks [84.29338268789684]
We introduce a novel heterogeneous memory augmentation approach for neural networks.
By introducing learnable memory tokens with attention mechanism, we can effectively boost performance without huge computational overhead.
We show our approach on various image and graph-based tasks under both in-distribution (ID) and out-of-distribution (OOD) conditions.
arXiv Detail & Related papers (2023-10-17T01:05:28Z) - Network Alignment with Transferable Graph Autoencoders [79.89704126746204]
We propose a novel graph autoencoder architecture designed to extract powerful and robust node embeddings.
We prove that the generated embeddings are associated with the eigenvalues and eigenvectors of the graphs.
Our proposed framework also leverages transfer learning and data augmentation to achieve efficient network alignment at a very large scale without retraining.
arXiv Detail & Related papers (2023-10-05T02:58:29Z) - Self-Ensembling GAN for Cross-Domain Semantic Segmentation [107.27377745720243]
This paper proposes a self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain data for semantic segmentation.
In SE-GAN, a teacher network and a student network constitute a self-ensembling model for generating semantic segmentation maps, which together with a discriminator, forms a GAN.
Despite its simplicity, we find SE-GAN can significantly boost the performance of adversarial training and enhance the stability of the model.
arXiv Detail & Related papers (2021-12-15T09:50:25Z) - Attention-driven Graph Clustering Network [49.040136530379094]
We propose a novel deep clustering method named Attention-driven Graph Clustering Network (AGCN)
AGCN exploits a heterogeneous-wise fusion module to dynamically fuse the node attribute feature and the topological graph feature.
AGCN can jointly perform feature learning and cluster assignment in an unsupervised fashion.
arXiv Detail & Related papers (2021-08-12T02:30:38Z) - A Framework for Joint Unsupervised Learning of Cluster-Aware Embedding
for Heterogeneous Networks [6.900303913555705]
Heterogeneous Information Network (HIN) embedding refers to the low-dimensional projections of the HIN nodes that preserve the HIN structure and semantics.
We propose ours for joint learning of cluster embeddings as well as cluster-aware HIN embedding.
arXiv Detail & Related papers (2021-08-09T11:36:36Z) - GAHNE: Graph-Aggregated Heterogeneous Network Embedding [32.44836376873812]
Heterogeneous network embedding aims to embed nodes into low-dimensional vectors which capture rich intrinsic information of heterogeneous networks.
Existing models either depend on manually designing meta-paths, ignore mutual effects between different semantics, or omit some aspects of information from global networks.
In GAHNE model, we develop several mechanisms that can aggregate semantic representations from different single-type sub-networks as well as fuse the global information into final embeddings.
arXiv Detail & Related papers (2020-12-23T07:11:30Z) - Deep Kernel Supervised Hashing for Node Classification in Structural
Networks [39.459721876872266]
We propose a novel Deep Kernel Supervised Hashing (DKSH) method to learn the hashing representations of nodes for node classification.
The proposed method significantly outperforms the state-of-the-art baselines over three real world benchmark datasets.
arXiv Detail & Related papers (2020-10-26T13:38:23Z) - DINE: A Framework for Deep Incomplete Network Embedding [33.97952453310253]
We propose a Deep Incomplete Network Embedding method, namely DINE.
We first complete the missing part including both nodes and edges in a partially observable network by using the expectation-maximization framework.
We evaluate DINE over three networks on multi-label classification and link prediction tasks.
arXiv Detail & Related papers (2020-08-09T04:59:35Z) - EPINE: Enhanced Proximity Information Network Embedding [2.257737378757467]
In this work, we devote to mining valuable information in adjacency matrices at a deeper level.
Under the same objective, many NE methods calculate high-order proximity by the powers of adjacency matrices.
We propose to redefine high-order proximity in a more intuitive manner.
arXiv Detail & Related papers (2020-03-04T15:57:17Z) - Dense Residual Network: Enhancing Global Dense Feature Flow for
Character Recognition [75.4027660840568]
This paper explores how to enhance the local and global dense feature flow by exploiting hierarchical features fully from all the convolution layers.
Technically, we propose an efficient and effective CNN framework, i.e., Fast Dense Residual Network (FDRN) for text recognition.
arXiv Detail & Related papers (2020-01-23T06:55:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.