Does the Generator Mind its Contexts? An Analysis of Generative Model
Faithfulness under Context Transfer
- URL: http://arxiv.org/abs/2402.14488v1
- Date: Thu, 22 Feb 2024 12:26:07 GMT
- Title: Does the Generator Mind its Contexts? An Analysis of Generative Model
Faithfulness under Context Transfer
- Authors: Xinshuo Hu and Baotian Hu and Dongfang Li and Xiaoguang Li and Lifeng
Shang
- Abstract summary: The present study introduces the knowledge-augmented generator, which is specifically designed to produce information that remains grounded in contextual knowledge.
Our objective is to explore the existence of hallucinations arising from parametric memory when contextual knowledge undergoes changes.
- Score: 42.081311699224585
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: The present study introduces the knowledge-augmented generator, which is
specifically designed to produce information that remains grounded in
contextual knowledge, regardless of alterations in the context. Previous
research has predominantly focused on examining hallucinations stemming from
static input, such as in the domains of summarization or machine translation.
However, our investigation delves into the faithfulness of generative question
answering in the presence of dynamic knowledge. Our objective is to explore the
existence of hallucinations arising from parametric memory when contextual
knowledge undergoes changes, while also analyzing the underlying causes for
their occurrence. In order to efficiently address this issue, we propose a
straightforward yet effective measure for detecting such hallucinations.
Intriguingly, our investigation uncovers that all models exhibit a tendency to
generate previous answers as hallucinations. To gain deeper insights into the
underlying causes of this phenomenon, we conduct a series of experiments that
verify the critical role played by context in hallucination, both during
training and testing, from various perspectives.
Related papers
- Why and How LLMs Hallucinate: Connecting the Dots with Subsequence Associations [82.42811602081692]
This paper introduces a subsequence association framework to systematically trace and understand hallucinations.
Key insight is hallucinations that arise when dominant hallucinatory associations outweigh faithful ones.
We propose a tracing algorithm that identifies causal subsequences by analyzing hallucination probabilities across randomized input contexts.
arXiv Detail & Related papers (2025-04-17T06:34:45Z) - KSHSeek: Data-Driven Approaches to Mitigating and Detecting Knowledge-Shortcut Hallucinations in Generative Models [17.435794516702256]
Large language models (LLMs) have significantly advanced the development of natural language processing (NLP)
Model hallucinations remain a major challenge in natural language generation (NLG) tasks due to their complex causes.
This work introduces a new paradigm for mitigating specific hallucination issues in generative models, enhancing their robustness and reliability in real-world applications.
arXiv Detail & Related papers (2025-03-25T09:18:27Z) - What are Models Thinking about? Understanding Large Language Model Hallucinations "Psychology" through Model Inner State Analysis [5.8289838136060155]
Internal states of large language models' inference have been widely used in numerous research works.
In this paper, we analyze different internal states' revealing features during inference forward and evaluate their ability in hallucination detection.
arXiv Detail & Related papers (2025-02-19T07:23:18Z) - Can Your Uncertainty Scores Detect Hallucinated Entity? [14.432545893757677]
We propose a new data set, HalluEntity, which annotates hallucination at the entity level.
Based on the dataset, we evaluate uncertainty-based hallucination detection approaches across 17 modern LLMs.
Our experimental results show that uncertainty estimation approaches focusing on individual token probabilities tend to over-predict hallucinations.
arXiv Detail & Related papers (2025-02-17T16:01:41Z) - Who Brings the Frisbee: Probing Hidden Hallucination Factors in Large Vision-Language Model via Causality Analysis [14.033320167387194]
A major challenge in their real-world application is hallucination, where LVLMs generate non-existent visual elements, eroding user trust.
We hypothesize that hidden factors, such as objects, contexts, and semantic foreground-background structures, induce hallucination.
By analyzing the causality between images, text prompts, and network saliency, we systematically explore interventions to block these factors.
arXiv Detail & Related papers (2024-12-04T01:23:57Z) - Mitigating Knowledge Conflicts in Language Model-Driven Question Answering [15.29366851382021]
In this work, we argue that hallucination could be mitigated via explicit correlation between input source and generated content.
We focus on a typical example of hallucination, entity-based knowledge conflicts in question answering, where correlation of entities and their description at training time hinders model behaviour during inference.
arXiv Detail & Related papers (2024-11-18T07:33:10Z) - A Cause-Effect Look at Alleviating Hallucination of Knowledge-grounded Dialogue Generation [51.53917938874146]
We propose a possible solution for alleviating the hallucination in KGD by exploiting the dialogue-knowledge interaction.
Experimental results of our example implementation show that this method can reduce hallucination without disrupting other dialogue performance.
arXiv Detail & Related papers (2024-04-04T14:45:26Z) - On Large Language Models' Hallucination with Regard to Known Facts [74.96789694959894]
Large language models are successful in answering factoid questions but are also prone to hallucination.
We investigate the phenomenon of LLMs possessing correct answer knowledge yet still hallucinating from the perspective of inference dynamics.
Our study shed light on understanding the reasons for LLMs' hallucinations on their known facts, and more importantly, on accurately predicting when they are hallucinating.
arXiv Detail & Related papers (2024-03-29T06:48:30Z) - In-Context Sharpness as Alerts: An Inner Representation Perspective for
Hallucination Mitigation [36.31646727970656]
Large language models (LLMs) frequently hallucinate and produce factual errors.
correct generations tend to have sharper context activations in the hidden states of the in-context tokens, compared to the incorrect ones.
We propose an entropy-based metric to quantify the sharpness'' among the in-context hidden states and incorporate it into the decoding process.
arXiv Detail & Related papers (2024-03-03T15:53:41Z) - Towards Mitigating Hallucination in Large Language Models via
Self-Reflection [63.2543947174318]
Large language models (LLMs) have shown promise for generative and knowledge-intensive tasks including question-answering (QA) tasks.
This paper analyses the phenomenon of hallucination in medical generative QA systems using widely adopted LLMs and datasets.
arXiv Detail & Related papers (2023-10-10T03:05:44Z) - AutoHall: Automated Hallucination Dataset Generation for Large Language Models [56.92068213969036]
This paper introduces a method for automatically constructing model-specific hallucination datasets based on existing fact-checking datasets called AutoHall.
We also propose a zero-resource and black-box hallucination detection method based on self-contradiction.
arXiv Detail & Related papers (2023-09-30T05:20:02Z) - On Hallucination and Predictive Uncertainty in Conditional Language
Generation [76.18783678114325]
Higher predictive uncertainty corresponds to a higher chance of hallucination.
Epistemic uncertainty is more indicative of hallucination than aleatoric or total uncertainties.
It helps to achieve better results of trading performance in standard metric for less hallucination with the proposed beam search variant.
arXiv Detail & Related papers (2021-03-28T00:32:27Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.