Counterfactual Debiasing for Generating Factually Consistent Text
Summaries
- URL: http://arxiv.org/abs/2305.10736v1
- Date: Thu, 18 May 2023 06:15:45 GMT
- Title: Counterfactual Debiasing for Generating Factually Consistent Text
Summaries
- Authors: Chenhe Dong, Yuexiang Xie, Yaliang Li, Ying Shen
- Abstract summary: We construct causal graphs for abstractive text summarization and identify the intrinsic causes of the factual inconsistency.
We propose a debiasing framework, named CoFactSum, to alleviate the causal effects of these biases by counterfactual estimation.
Experiments on two widely-used summarization datasets demonstrate the effectiveness of CoFactSum.
- Score: 46.88138136539263
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Despite substantial progress in abstractive text summarization to generate
fluent and informative texts, the factual inconsistency in the generated
summaries remains an important yet challenging problem to be solved. In this
paper, we construct causal graphs for abstractive text summarization and
identify the intrinsic causes of the factual inconsistency, i.e., the language
bias and irrelevancy bias, and further propose a debiasing framework, named
CoFactSum, to alleviate the causal effects of these biases by counterfactual
estimation. Specifically, the proposed CoFactSum provides two counterfactual
estimation strategies, i.e., Explicit Counterfactual Masking with an explicit
dynamic masking strategy, and Implicit Counterfactual Training with an implicit
discriminative cross-attention mechanism. Meanwhile, we design a Debiasing
Degree Adjustment mechanism to dynamically adapt the debiasing degree at each
decoding step. Extensive experiments on two widely-used summarization datasets
demonstrate the effectiveness of CoFactSum in enhancing the factual consistency
of generated summaries compared with several baselines.
Related papers
- AMRFact: Enhancing Summarization Factuality Evaluation with AMR-Driven Negative Samples Generation [57.8363998797433]
We propose AMRFact, a framework that generates perturbed summaries using Abstract Meaning Representations (AMRs)
Our approach parses factually consistent summaries into AMR graphs and injects controlled factual inconsistencies to create negative examples, allowing for coherent factually inconsistent summaries to be generated with high error-type coverage.
arXiv Detail & Related papers (2023-11-16T02:56:29Z) - Disentangled Representation Learning with Transmitted Information Bottleneck [57.22757813140418]
We present textbfDisTIB (textbfTransmitted textbfInformation textbfBottleneck for textbfDisd representation learning), a novel objective that navigates the balance between information compression and preservation.
arXiv Detail & Related papers (2023-11-03T03:18:40Z) - Interpretable Automatic Fine-grained Inconsistency Detection in Text
Summarization [56.94741578760294]
We propose the task of fine-grained inconsistency detection, the goal of which is to predict the fine-grained types of factual errors in a summary.
Motivated by how humans inspect factual inconsistency in summaries, we propose an interpretable fine-grained inconsistency detection model, FineGrainFact.
arXiv Detail & Related papers (2023-05-23T22:11:47Z) - Questioning the Validity of Summarization Datasets and Improving Their
Factual Consistency [14.974996886744083]
We release SummFC, a filtered summarization dataset with improved factual consistency.
We argue that our dataset should become a valid benchmark for developing and evaluating summarization systems.
arXiv Detail & Related papers (2022-10-31T15:04:20Z) - Factual Consistency Evaluation for Text Summarization via Counterfactual
Estimation [42.63902468258758]
We propose a novel metric to evaluate the factual consistency in text summarization via counterfactual estimation.
We conduct a series of experiments on three public abstractive text summarization datasets.
arXiv Detail & Related papers (2021-08-30T11:48:41Z) - The Factual Inconsistency Problem in Abstractive Text Summarization: A
Survey [25.59111855107199]
neural encoder-decoder models pioneered by Seq2Seq framework have been proposed to achieve the goal of generating more abstractive summaries.
At a high level, such neural models can freely generate summaries without any constraint on the words or phrases used.
However, the neural model's abstraction ability is a double-edged sword.
arXiv Detail & Related papers (2021-04-30T08:46:13Z) - Multi-Fact Correction in Abstractive Text Summarization [98.27031108197944]
Span-Fact is a suite of two factual correction models that leverages knowledge learned from question answering models to make corrections in system-generated summaries via span selection.
Our models employ single or multi-masking strategies to either iteratively or auto-regressively replace entities in order to ensure semantic consistency w.r.t. the source text.
Experiments show that our models significantly boost the factual consistency of system-generated summaries without sacrificing summary quality in terms of both automatic metrics and human evaluation.
arXiv Detail & Related papers (2020-10-06T02:51:02Z) - Enhancing Factual Consistency of Abstractive Summarization [57.67609672082137]
We propose a fact-aware summarization model FASum to extract and integrate factual relations into the summary generation process.
We then design a factual corrector model FC to automatically correct factual errors from summaries generated by existing systems.
arXiv Detail & Related papers (2020-03-19T07:36:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.