Are Negative Samples Necessary in Entity Alignment? An Approach with
High Performance, Scalability and Robustness
- URL: http://arxiv.org/abs/2108.05278v2
- Date: Thu, 12 Aug 2021 03:06:55 GMT
- Title: Are Negative Samples Necessary in Entity Alignment? An Approach with
High Performance, Scalability and Robustness
- Authors: Xin Mao, Wenting Wang, Yuanbin Wu, Man Lan
- Abstract summary: We propose a novel EA method with three new components to enable high Performance, high Scalability, and high Robustness.
We conduct detailed experiments on several public datasets to examine the effectiveness and efficiency of our proposed method.
- Score: 26.04006507181558
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Entity alignment (EA) aims to find the equivalent entities in different KGs,
which is a crucial step in integrating multiple KGs. However, most existing EA
methods have poor scalability and are unable to cope with large-scale datasets.
We summarize three issues leading to such high time-space complexity in
existing EA methods: (1) Inefficient graph encoders, (2) Dilemma of negative
sampling, and (3) "Catastrophic forgetting" in semi-supervised learning. To
address these challenges, we propose a novel EA method with three new
components to enable high Performance, high Scalability, and high Robustness
(PSR): (1) Simplified graph encoder with relational graph sampling, (2)
Symmetric negative-free alignment loss, and (3) Incremental semi-supervised
learning. Furthermore, we conduct detailed experiments on several public
datasets to examine the effectiveness and efficiency of our proposed method.
The experimental results show that PSR not only surpasses the previous SOTA in
performance but also has impressive scalability and robustness.
Related papers
- Decoupled and Interactive Regression Modeling for High-performance One-stage 3D Object Detection [8.531052087985097]
Inadequate bounding box modeling in regression tasks constrains the performance of one-stage 3D object detection.
We propose Decoupled and Interactive Regression Modeling (DIRM) for one-stage detection.
arXiv Detail & Related papers (2024-09-01T10:47:22Z) - When hard negative sampling meets supervised contrastive learning [17.173114048398947]
We introduce a new supervised contrastive learning objective, SCHaNe, which incorporates hard negative sampling during the fine-tuning phase.
SchaNe outperforms the strong baseline BEiT-3 in Top-1 accuracy across various benchmarks.
Our proposed objective sets a new state-of-the-art for base models on ImageNet-1k, achieving an 86.14% accuracy.
arXiv Detail & Related papers (2023-08-28T20:30:10Z) - ContrasInver: Ultra-Sparse Label Semi-supervised Regression for
Multi-dimensional Seismic Inversion [7.356328937024184]
ContrasInver is a method that achieves seismic inversion using as few as two or three well logs.
In experiments, ContrasInver achieved state-of-the-art performance in the synthetic data SEAM I.
It's the first data-driven approach yielding reliable results on the Netherlands F3 and Delft, using only three and two well logs respectively.
arXiv Detail & Related papers (2023-02-13T15:19:51Z) - Open-Set Semi-Supervised Learning for 3D Point Cloud Understanding [62.17020485045456]
It is commonly assumed in semi-supervised learning (SSL) that the unlabeled data are drawn from the same distribution as that of the labeled ones.
We propose to selectively utilize unlabeled data through sample weighting, so that only conducive unlabeled data would be prioritized.
arXiv Detail & Related papers (2022-05-02T16:09:17Z) - Scale-Equivalent Distillation for Semi-Supervised Object Detection [57.59525453301374]
Recent Semi-Supervised Object Detection (SS-OD) methods are mainly based on self-training, generating hard pseudo-labels by a teacher model on unlabeled data as supervisory signals.
We analyze the challenges these methods meet with the empirical experiment results.
We introduce a novel approach, Scale-Equivalent Distillation (SED), which is a simple yet effective end-to-end knowledge distillation framework robust to large object size variance and class imbalance.
arXiv Detail & Related papers (2022-03-23T07:33:37Z) - Generalization of Neural Combinatorial Solvers Through the Lens of
Adversarial Robustness [68.97830259849086]
Most datasets only capture a simpler subproblem and likely suffer from spurious features.
We study adversarial robustness - a local generalization property - to reveal hard, model-specific instances and spurious features.
Unlike in other applications, where perturbation models are designed around subjective notions of imperceptibility, our perturbation models are efficient and sound.
Surprisingly, with such perturbations, a sufficiently expressive neural solver does not suffer from the limitations of the accuracy-robustness trade-off common in supervised learning.
arXiv Detail & Related papers (2021-10-21T07:28:11Z) - Guided Point Contrastive Learning for Semi-supervised Point Cloud
Semantic Segmentation [90.2445084743881]
We present a method for semi-supervised point cloud semantic segmentation to adopt unlabeled point clouds in training to boost the model performance.
Inspired by the recent contrastive loss in self-supervised tasks, we propose the guided point contrastive loss to enhance the feature representation and model generalization ability.
arXiv Detail & Related papers (2021-10-15T16:38:54Z) - Structure-Aware Hard Negative Mining for Heterogeneous Graph Contrastive
Learning [21.702342154458623]
This work investigates Contrastive Learning (CL) on Graph Neural Networks (GNNs)
We first generate multiple semantic views according to metapaths and network schemas.
We then push node embeddings corresponding to different semantic views close to each other (positives) and pulling other embeddings apart (negatives)
Considering the complex graph structure and the smoothing nature of GNNs, we propose a structure-aware hard negative mining scheme.
arXiv Detail & Related papers (2021-08-31T14:44:49Z) - WSSOD: A New Pipeline for Weakly- and Semi-Supervised Object Detection [75.80075054706079]
We propose a weakly- and semi-supervised object detection framework (WSSOD)
An agent detector is first trained on a joint dataset and then used to predict pseudo bounding boxes on weakly-annotated images.
The proposed framework demonstrates remarkable performance on PASCAL-VOC and MSCOCO benchmark, achieving a high performance comparable to those obtained in fully-supervised settings.
arXiv Detail & Related papers (2021-05-21T11:58:50Z) - Enhanced Principal Component Analysis under A Collaborative-Robust
Framework [89.28334359066258]
We introduce a general collaborative-robust weight learning framework that combines weight learning and robust loss in a non-trivial way.
Under the proposed framework, only a part of well-fitting samples are activated which indicates more importance during training, and others, whose errors are large, will not be ignored.
In particular, the negative effects of inactivated samples are alleviated by the robust loss function.
arXiv Detail & Related papers (2021-03-22T15:17:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.