Open Knowledge Base Canonicalization with Multi-task Unlearning
- URL: http://arxiv.org/abs/2310.16419v1
- Date: Wed, 25 Oct 2023 07:13:06 GMT
- Title: Open Knowledge Base Canonicalization with Multi-task Unlearning
- Authors: Bingchen Liu, Shihao Hou, Weixin Zeng, Xiang Zhao, Shijun Liu, Li Pan
- Abstract summary: MulCanon is a multi-task unlearning framework to tackle machine unlearning problem in OKB canonicalization.
A thorough experimental study on popular OKB canonicalization datasets validates that MulCanon achieves advanced machine unlearning effects.
- Score: 19.130159457887
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The construction of large open knowledge bases (OKBs) is integral to many
applications in the field of mobile computing. Noun phrases and relational
phrases in OKBs often suffer from redundancy and ambiguity, which calls for the
investigation on OKB canonicalization. However, in order to meet the
requirements of some privacy protection regulations and to ensure the
timeliness of the data, the canonicalized OKB often needs to remove some
sensitive information or outdated data. The machine unlearning in OKB
canonicalization is an excellent solution to the above problem. Current
solutions address OKB canonicalization by devising advanced clustering
algorithms and using knowledge graph embedding (KGE) to further facilitate the
canonicalization process. Effective schemes are urgently needed to fully
synergise machine unlearning with clustering and KGE learning. To this end, we
put forward a multi-task unlearning framework, namely MulCanon, to tackle
machine unlearning problem in OKB canonicalization. Specifically, the noise
characteristics in the diffusion model are utilized to achieve the effect of
machine unlearning for data in OKB. MulCanon unifies the learning objectives of
diffusion model, KGE and clustering algorithms, and adopts a two-step
multi-task learning paradigm for training. A thorough experimental study on
popular OKB canonicalization datasets validates that MulCanon achieves advanced
machine unlearning effects.
Related papers
- Automated Knowledge Concept Annotation and Question Representation Learning for Knowledge Tracing [59.480951050911436]
We present KCQRL, a framework for automated knowledge concept annotation and question representation learning.
We demonstrate the effectiveness of KCQRL across 15 KT algorithms on two large real-world Math learning datasets.
arXiv Detail & Related papers (2024-10-02T16:37:19Z) - Open Knowledge Base Canonicalization with Multi-task Learning [18.053863554106307]
Large open knowledge bases (OKBs) are integral to many knowledge-driven applications on the world wide web such as web search.
noun phrases and relational phrases in OKBs often suffer from redundancy and ambiguity, which calls for the investigation on OKB canonicalization.
Current solutions address OKB canonicalization by devising advanced clustering algorithms and using knowledge graph embedding (KGE) to further facilitate the canonicalization process.
We put forward a multi-task learning framework, namely MulCanon, to tackle OKB canonicalization.
arXiv Detail & Related papers (2024-03-21T08:03:46Z) - KGA: A General Machine Unlearning Framework Based on Knowledge Gap
Alignment [51.15802100354848]
We propose a general unlearning framework called KGA to induce forgetfulness.
Experiments on large-scale datasets show that KGA yields comprehensive improvements over baselines.
arXiv Detail & Related papers (2023-05-11T02:44:29Z) - Federated Gradient Matching Pursuit [17.695717854068715]
Traditional machine learning techniques require centralizing all training data on one server or data hub.
In particular, federated learning (FL) provides such a solution to learn a shared model while keeping training data at local clients.
We propose a novel algorithmic framework, federated gradient matching pursuit (FedGradMP), to solve the sparsity constrained minimization problem in the FL setting.
arXiv Detail & Related papers (2023-02-20T16:26:29Z) - Towards Unbounded Machine Unlearning [13.31957848633701]
We study unlearning for different applications (RB, RC, UP), with the view that each has its own desiderata, definitions for forgetting' and associated metrics for forget quality.
For UP, we propose a novel adaptation of a strong Membership Inference Attack for unlearning.
We also propose SCRUB, a novel unlearning algorithm, which is consistently a top performer for forget quality across the different application-dependent metrics for RB, RC, and UP.
arXiv Detail & Related papers (2023-02-20T10:15:36Z) - Multi-View Clustering for Open Knowledge Base Canonicalization [9.976636206355394]
Noun phrases and relation phrases in large open knowledge bases (OKBs) are not canonicalized.
We propose CMVC, a novel unsupervised framework that leverages two views of knowledge jointly for canonicalizing OKBs.
We demonstrate the superiority of our framework through extensive experiments on multiple real-world OKB data sets against state-of-the-art methods.
arXiv Detail & Related papers (2022-06-22T14:23:16Z) - Learning to Detect Critical Nodes in Sparse Graphs via Feature Importance Awareness [53.351863569314794]
The critical node problem (CNP) aims to find a set of critical nodes from a network whose deletion maximally degrades the pairwise connectivity of the residual network.
This work proposes a feature importance-aware graph attention network for node representation.
It combines it with dueling double deep Q-network to create an end-to-end algorithm to solve CNP for the first time.
arXiv Detail & Related papers (2021-12-03T14:23:05Z) - Boosting Weakly Supervised Object Detection via Learning Bounding Box
Adjusters [76.36104006511684]
Weakly-supervised object detection (WSOD) has emerged as an inspiring recent topic to avoid expensive instance-level object annotations.
We defend the problem setting for improving localization performance by leveraging the bounding box regression knowledge from a well-annotated auxiliary dataset.
Our method performs favorably against state-of-the-art WSOD methods and knowledge transfer model with similar problem setting.
arXiv Detail & Related papers (2021-08-03T13:38:20Z) - Network Support for High-performance Distributed Machine Learning [17.919773898228716]
We propose a system model that captures both learning nodes (that perform computations) and information nodes (that provide data)
We then formulate the problem of selecting (i) which learning and information nodes should cooperate to complete the learning task, and (ii) the number of iterations to perform.
We devise an algorithm, named DoubleClimb, that can find a 1+1/|I|-competitive solution with cubic worst-case complexity.
arXiv Detail & Related papers (2021-02-05T19:38:57Z) - Semi-Supervised Learning with Meta-Gradient [123.26748223837802]
We propose a simple yet effective meta-learning algorithm in semi-supervised learning.
We find that the proposed algorithm performs favorably against state-of-the-art methods.
arXiv Detail & Related papers (2020-07-08T08:48:56Z) - Belief Propagation Reloaded: Learning BP-Layers for Labeling Problems [83.98774574197613]
We take one of the simplest inference methods, a truncated max-product Belief propagation, and add what is necessary to make it a proper component of a deep learning model.
This BP-Layer can be used as the final or an intermediate block in convolutional neural networks (CNNs)
The model is applicable to a range of dense prediction problems, is well-trainable and provides parameter-efficient and robust solutions in stereo, optical flow and semantic segmentation.
arXiv Detail & Related papers (2020-03-13T13:11:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.