Boosting the Speed of Entity Alignment 10*: Dual Attention Matching
Network with Normalized Hard Sample Mining
- URL: http://arxiv.org/abs/2103.15452v1
- Date: Mon, 29 Mar 2021 09:35:07 GMT
- Title: Boosting the Speed of Entity Alignment 10*: Dual Attention Matching
Network with Normalized Hard Sample Mining
- Authors: Xin Mao, Wenting Wang, Yuanbin Wu, Man Lan
- Abstract summary: We propose a novel KG encoder -- Dual Attention Matching Network (Dual-AMN)
The experimental results on widely used public datasets indicate that our method achieves both high accuracy and high efficiency.
- Score: 26.04006507181558
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Seeking the equivalent entities among multi-source Knowledge Graphs (KGs) is
the pivotal step to KGs integration, also known as \emph{entity alignment}
(EA). However, most existing EA methods are inefficient and poor in
scalability. A recent summary points out that some of them even require several
days to deal with a dataset containing 200,000 nodes (DWY100K). We believe
over-complex graph encoder and inefficient negative sampling strategy are the
two main reasons. In this paper, we propose a novel KG encoder -- Dual
Attention Matching Network (Dual-AMN), which not only models both intra-graph
and cross-graph information smartly, but also greatly reduces computational
complexity. Furthermore, we propose the Normalized Hard Sample Mining Loss to
smoothly select hard negative samples with reduced loss shift. The experimental
results on widely used public datasets indicate that our method achieves both
high accuracy and high efficiency. On DWY100K, the whole running process of our
method could be finished in 1,100 seconds, at least 10* faster than previous
work. The performances of our method also outperform previous works across all
datasets, where Hits@1 and MRR have been improved from 6% to 13%.
Related papers
- Scalable and Efficient Temporal Graph Representation Learning via Forward Recent Sampling [7.803172953008744]
Temporal graph representation learning (TGRL) is essential for modeling dynamic systems in real-world networks.
Traditional TGRL methods often face significant computational challenges and inference delays due to the inefficient sampling of temporal neighbors.
We propose a novel TGRL framework, No-Looking-Back (NLB), which overcomes these challenges by introducing a forward recent sampling strategy.
arXiv Detail & Related papers (2024-02-03T00:12:36Z) - Efficient Heterogeneous Graph Learning via Random Projection [58.4138636866903]
Heterogeneous Graph Neural Networks (HGNNs) are powerful tools for deep learning on heterogeneous graphs.
Recent pre-computation-based HGNNs use one-time message passing to transform a heterogeneous graph into regular-shaped tensors.
We propose a hybrid pre-computation-based HGNN, named Random Projection Heterogeneous Graph Neural Network (RpHGNN)
arXiv Detail & Related papers (2023-10-23T01:25:44Z) - Improving Dual-Encoder Training through Dynamic Indexes for Negative
Mining [61.09807522366773]
We introduce an algorithm that approximates the softmax with provable bounds and that dynamically maintains the tree.
In our study on datasets with over twenty million targets, our approach cuts error by half in relation to oracle brute-force negative mining.
arXiv Detail & Related papers (2023-03-27T15:18:32Z) - Simplex Autoencoders [1.3960152426268768]
We propose a new approach that models the latent space of an Autoencoder as a simplex, allowing for a novel for determining the number of components in the mixture model.
We evaluate our approaches on a synthetic dataset and demonstrate their performance on three benchmark datasets.
arXiv Detail & Related papers (2023-01-16T15:57:03Z) - RSC: Accelerating Graph Neural Networks Training via Randomized Sparse
Computations [56.59168541623729]
Training graph neural networks (GNNs) is time consuming because sparse graph-based operations are hard to be accelerated by hardware.
We explore trading off the computational precision to reduce the time complexity via sampling-based approximation.
We propose Randomized Sparse Computation, which for the first time demonstrate the potential of training GNNs with approximated operations.
arXiv Detail & Related papers (2022-10-19T17:25:33Z) - Communication-Efficient Adam-Type Algorithms for Distributed Data Mining [93.50424502011626]
We propose a class of novel distributed Adam-type algorithms (emphi.e., SketchedAMSGrad) utilizing sketching.
Our new algorithm achieves a fast convergence rate of $O(frac1sqrtnT + frac1(k/d)2 T)$ with the communication cost of $O(k log(d))$ at each iteration.
arXiv Detail & Related papers (2022-10-14T01:42:05Z) - SCARA: Scalable Graph Neural Networks with Feature-Oriented Optimization [23.609017952951454]
We propose SCARA, a scalable Graph Neural Network (GNN) with feature-oriented optimization for graph computation.
SCARA efficiently computes graph embedding from node features, and further selects and reuses feature results to reduce overhead.
It is efficient to process precomputation on the largest available billion-scale GNN dataset Papers100M (111M nodes, 1.6B edges) in 100 seconds.
arXiv Detail & Related papers (2022-07-19T10:32:11Z) - Condensing Graphs via One-Step Gradient Matching [50.07587238142548]
We propose a one-step gradient matching scheme, which performs gradient matching for only one single step without training the network weights.
Our theoretical analysis shows this strategy can generate synthetic graphs that lead to lower classification loss on real graphs.
In particular, we are able to reduce the dataset size by 90% while approximating up to 98% of the original performance.
arXiv Detail & Related papers (2022-06-15T18:20:01Z) - Are Negative Samples Necessary in Entity Alignment? An Approach with
High Performance, Scalability and Robustness [26.04006507181558]
We propose a novel EA method with three new components to enable high Performance, high Scalability, and high Robustness.
We conduct detailed experiments on several public datasets to examine the effectiveness and efficiency of our proposed method.
arXiv Detail & Related papers (2021-08-11T15:20:41Z) - Augmentation Strategies for Learning with Noisy Labels [3.698228929379249]
We evaluate different augmentation strategies for algorithms tackling the "learning with noisy labels" problem.
We find that using one set of augmentations for loss modeling tasks and another set for learning is the most effective.
We introduce this augmentation strategy to the state-of-the-art technique and demonstrate that we can improve performance across all evaluated noise levels.
arXiv Detail & Related papers (2021-03-03T02:19:35Z) - Combining Label Propagation and Simple Models Out-performs Graph Neural
Networks [52.121819834353865]
We show that for many standard transductive node classification benchmarks, we can exceed or match the performance of state-of-the-art GNNs.
We call this overall procedure Correct and Smooth (C&S)
Our approach exceeds or nearly matches the performance of state-of-the-art GNNs on a wide variety of benchmarks.
arXiv Detail & Related papers (2020-10-27T02:10:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.