Dependency-aware Self-training for Entity Alignment
- URL: http://arxiv.org/abs/2211.16101v1
- Date: Tue, 29 Nov 2022 11:24:14 GMT
- Title: Dependency-aware Self-training for Entity Alignment
- Authors: Bing Liu, Tiancheng Lan, Wen Hua, Guido Zuccon
- Abstract summary: Entity Alignment (EA) aims to detect entity mappings in different Knowledge Graphs (KGs)
EA methods dominate current EA research but still suffer from their reliance on labelled mappings.
We propose exploiting the dependencies between entities, a particularity of EA, to suppress the noise without hurting the recall of True Positive mappings.
- Score: 28.158354625969668
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Entity Alignment (EA), which aims to detect entity mappings (i.e. equivalent
entity pairs) in different Knowledge Graphs (KGs), is critical for KG fusion.
Neural EA methods dominate current EA research but still suffer from their
reliance on labelled mappings. To solve this problem, a few works have explored
boosting the training of EA models with self-training, which adds confidently
predicted mappings into the training data iteratively. Though the effectiveness
of self-training can be glimpsed in some specific settings, we still have very
limited knowledge about it. One reason is the existing works concentrate on
devising EA models and only treat self-training as an auxiliary tool. To fill
this knowledge gap, we change the perspective to self-training to shed light on
it. In addition, the existing self-training strategies have limited impact
because they introduce either much False Positive noise or a low quantity of
True Positive pseudo mappings. To improve self-training for EA, we propose
exploiting the dependencies between entities, a particularity of EA, to
suppress the noise without hurting the recall of True Positive mappings.
Through extensive experiments, we show that the introduction of dependency
makes the self-training strategy for EA reach a new level. The value of
self-training in alleviating the reliance on annotation is actually much higher
than what has been realised. Furthermore, we suggest future study on smart data
annotation to break the ceiling of EA performance.
Related papers
- KBAlign: Efficient Self Adaptation on Specific Knowledge Bases [75.78948575957081]
Large language models (LLMs) usually rely on retrieval-augmented generation to exploit knowledge materials in an instant manner.
We propose KBAlign, an approach designed for efficient adaptation to downstream tasks involving knowledge bases.
Our method utilizes iterative training with self-annotated data such as Q&A pairs and revision suggestions, enabling the model to grasp the knowledge content efficiently.
arXiv Detail & Related papers (2024-11-22T08:21:03Z) - Incremental Self-training for Semi-supervised Learning [56.57057576885672]
IST is simple yet effective and fits existing self-training-based semi-supervised learning methods.
We verify the proposed IST on five datasets and two types of backbone, effectively improving the recognition accuracy and learning speed.
arXiv Detail & Related papers (2024-04-14T05:02:00Z) - Self-Alignment for Factuality: Mitigating Hallucinations in LLMs via Self-Evaluation [71.91287418249688]
Large language models (LLMs) often struggle with factual inaccuracies, even when they hold relevant knowledge.
We leverage the self-evaluation capability of an LLM to provide training signals that steer the model towards factuality.
We show that the proposed self-alignment approach substantially enhances factual accuracy over Llama family models across three key knowledge-intensive tasks.
arXiv Detail & Related papers (2024-02-14T15:52:42Z) - Augmenting Unsupervised Reinforcement Learning with Self-Reference [63.68018737038331]
Humans possess the ability to draw on past experiences explicitly when learning new tasks.
We propose the Self-Reference (SR) approach, an add-on module explicitly designed to leverage historical information.
Our approach achieves state-of-the-art results in terms of Interquartile Mean (IQM) performance and Optimality Gap reduction on the Unsupervised Reinforcement Learning Benchmark.
arXiv Detail & Related papers (2023-11-16T09:07:34Z) - Making Users Indistinguishable: Attribute-wise Unlearning in Recommender
Systems [28.566330708233824]
We find that attackers can extract private information, i.e., gender, race, and age, from a trained model even if it has not been explicitly encountered during training.
To protect the sensitive attribute of users, Attribute Unlearning (AU) aims to degrade attacking performance and make target attributes indistinguishable.
arXiv Detail & Related papers (2023-10-06T09:36:44Z) - Guiding Neural Entity Alignment with Compatibility [32.22210683891481]
We argue that different entities within one Knowledge Graph should have compatible counterparts in the other KG.
Making compatible predictions should be one of the goals of training an EA model along with fitting the labelled data.
We devise a training framework by addressing three problems: (1) how to measure the compatibility of an EA model; (2) how to inject the property of being compatible into an EA model; and (3) how to optimise parameters of the compatibility model.
arXiv Detail & Related papers (2022-11-29T00:05:08Z) - Learning Domain Adaptive Object Detection with Probabilistic Teacher [93.76128726257946]
We present a simple yet effective framework, termed as Probabilistic Teacher (PT)
PT aims to capture the uncertainty of unlabeled target data from a gradually evolving teacher and guides the learning of a student in a mutually beneficial manner.
We also present a novel Entropy Focal Loss (EFL) to further facilitate the uncertainty-guided self-training.
arXiv Detail & Related papers (2022-06-13T16:24:22Z) - ICLEA: Interactive Contrastive Learning for Self-supervised Entity
Alignment [27.449414854756913]
Self-supervised entity alignment (EA) aims to link equivalent entities across different knowledge graphs (KGs) without seed alignments.
The current SOTA self-supervised EA method draws inspiration from contrastive learning, originally designed in computer vision.
We propose an interactive contrastive learning model for self-supervised EA.
arXiv Detail & Related papers (2022-01-17T06:04:00Z) - ActiveEA: Active Learning for Neural Entity Alignment [31.212894129845093]
Entity alignment (EA) aims to match equivalent entities across different Knowledge Graphs (KGs)
Current mainstream methods -- neural EA models -- rely on training with seed alignment, i.e., a set of pre-aligned entity pairs.
We devise a novel Active Learning (AL) framework for neural EA, aiming to create highly informative seed alignment.
arXiv Detail & Related papers (2021-10-13T03:38:04Z) - Reducing Label Effort: Self-Supervised meets Active Learning [32.4747118398236]
Recent developments in self-training have achieved very impressive results rivaling supervised learning on some datasets.
Our experiments reveal that self-training is remarkably more efficient than active learning at reducing the labeling effort.
The performance gap between active learning trained either with self-training or from scratch diminishes as we approach to the point where almost half of the dataset is labeled.
arXiv Detail & Related papers (2021-08-25T20:04:44Z) - Two-phase Pseudo Label Densification for Self-training based Domain
Adaptation [93.03265290594278]
We propose a novel Two-phase Pseudo Label Densification framework, referred to as TPLD.
In the first phase, we use sliding window voting to propagate the confident predictions, utilizing intrinsic spatial-correlations in the images.
In the second phase, we perform a confidence-based easy-hard classification.
To ease the training process and avoid noisy predictions, we introduce the bootstrapping mechanism to the original self-training loss.
arXiv Detail & Related papers (2020-12-09T02:35:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.