Propagation and Pitfalls: Reasoning-based Assessment of Knowledge
Editing through Counterfactual Tasks
- URL: http://arxiv.org/abs/2401.17585v1
- Date: Wed, 31 Jan 2024 04:12:59 GMT
- Title: Propagation and Pitfalls: Reasoning-based Assessment of Knowledge
Editing through Counterfactual Tasks
- Authors: Wenyue Hua, Jiang Guo, Mingwen Dong, Henghui Zhu, Patrick Ng, Zhiguo
Wang
- Abstract summary: We introduce a novel reasoning-based benchmark -- ReCoE (Reasoning-based Counterfactual Editing dataset)
We conduct a thorough analysis of existing knowledge editing techniques, including input augmentation, finetuning, and locate-and-edit.
All model editing methods show notably low performance on this dataset, especially in certain reasoning schemes.
- Score: 36.292901021210575
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Current approaches of knowledge editing struggle to effectively propagate
updates to interconnected facts. In this work, we delve into the barriers that
hinder the appropriate propagation of updated knowledge within these models for
accurate reasoning. To support our analysis, we introduce a novel
reasoning-based benchmark -- ReCoE (Reasoning-based Counterfactual Editing
dataset) -- which covers six common reasoning schemes in real world. We conduct
a thorough analysis of existing knowledge editing techniques, including input
augmentation, finetuning, and locate-and-edit. We found that all model editing
methods show notably low performance on this dataset, especially in certain
reasoning schemes. Our analysis over the chain-of-thought generation of edited
models further uncover key reasons behind the inadequacy of existing knowledge
editing methods from a reasoning standpoint, involving aspects on fact-wise
editing, fact recall ability, and coherence in generation. We will make our
benchmark publicly available.
Related papers
- Outdated Issue Aware Decoding for Reasoning Questions on Edited Knowledge [93.54427119091174]
We propose outDated ISsue aware deCOding to enhance the performance of edited models on reasoning questions.
We capture the difference in the probability distribution between the original and edited models.
We amplify the difference of the token prediction in the edited model to alleviate the outdated issue.
arXiv Detail & Related papers (2024-06-05T03:00:15Z) - Editing the Mind of Giants: An In-Depth Exploration of Pitfalls of Knowledge Editing in Large Language Models [26.516571783335824]
Recent studies have identified concerning side effects, such as knowledge distortion and the deterioration of general abilities, that have emerged after editing.
This survey presents a comprehensive study of these side effects, providing a unified view of the challenges associated with knowledge editing in Large Language Models.
arXiv Detail & Related papers (2024-06-03T15:28:21Z) - Updating Language Models with Unstructured Facts: Towards Practical
Knowledge Editing [87.35944788684958]
We propose a new benchmark, Unstructured Knowledge Editing (UKE)
UKE evaluates editing performance directly using unstructured texts as knowledge updates, termed unstructured facts.
We conduct extensive experiments on newly built datasets and demonstrate that UKE poses a significant challenge to state-of-the-art knowledge editing methods.
arXiv Detail & Related papers (2024-02-29T07:08:34Z) - EVEDIT: Event-based Knowledge Editing with Deductive Editing Boundaries [69.72012539060731]
We introduce a theoretical framework for efficient knowledge editing (KE) in large language models (LLMs)
We propose a novel task of event-based knowledge editing that pairs facts with event descriptions.
We empirically demonstrate the superiority of event-based editing over the existing setting on resolving uncertainty in edited models.
arXiv Detail & Related papers (2024-02-17T16:34:50Z) - Model Editing Harms General Abilities of Large Language Models: Regularization to the Rescue [122.20016030723043]
Model editing is a technique that edits large language models (LLMs) with updated knowledge to alleviate hallucinations without resource-intensive retraining.
Current model editing methods can effectively modify a model's behavior within a specific area of interest.
They often overlook the potential unintended side effects on the general abilities of LLMs.
arXiv Detail & Related papers (2024-01-09T18:03:15Z) - A Comprehensive Study of Knowledge Editing for Large Language Models [82.65729336401027]
Large Language Models (LLMs) have shown extraordinary capabilities in understanding and generating text that closely mirrors human communication.
This paper defines the knowledge editing problem and provides a comprehensive review of cutting-edge approaches.
We introduce a new benchmark, KnowEdit, for a comprehensive empirical evaluation of representative knowledge editing approaches.
arXiv Detail & Related papers (2024-01-02T16:54:58Z) - Assessing Knowledge Editing in Language Models via Relation Perspective [21.64869056276927]
This paper constructs a new benchmark named RaKE, which focuses on relation-based knowledge editing.
We establish a suite of innovative metrics for evaluation and conduct comprehensive experiments involving various knowledge editing baselines.
Our research results confirm that knowledge related to relations is not only stored in the FFN network but also in the attention layers.
arXiv Detail & Related papers (2023-11-15T15:44:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.