Causal Intervention for Abstractive Related Work Generation
- URL: http://arxiv.org/abs/2305.13685v1
- Date: Tue, 23 May 2023 04:48:30 GMT
- Title: Causal Intervention for Abstractive Related Work Generation
- Authors: Jiachang Liu, Qi Zhang, Chongyang Shi, Usman Naseem, Shoujin Wang,
Ivor Tsang
- Abstract summary: We propose a novel Causal Intervention Module for Related Work Generation (CaM)
We first model the relations among sentence order, document relation, and transitional content in related work generation using a causal graph.
We then use do-calculus to derive ordinary conditional probabilities and identify causal effects through CaM.
- Score: 16.7515135295061
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Abstractive related work generation has attracted increasing attention in
generating coherent related work that better helps readers grasp the background
in the current research. However, most existing abstractive models ignore the
inherent causality of related work generation, leading to low quality of
generated related work and spurious correlations that affect the models'
generalizability. In this study, we argue that causal intervention can address
these limitations and improve the quality and coherence of the generated
related works. To this end, we propose a novel Causal Intervention Module for
Related Work Generation (CaM) to effectively capture causalities in the
generation process and improve the quality and coherence of the generated
related works. Specifically, we first model the relations among sentence order,
document relation, and transitional content in related work generation using a
causal graph. Then, to implement the causal intervention and mitigate the
negative impact of spurious correlations, we use do-calculus to derive ordinary
conditional probabilities and identify causal effects through CaM. Finally, we
subtly fuse CaM with Transformer to obtain an end-to-end generation model.
Extensive experiments on two real-world datasets show that causal interventions
in CaM can effectively promote the model to learn causal relations and produce
related work of higher quality and coherence.
Related papers
- Deep Causal Generative Models with Property Control [11.604321459670315]
We propose a novel deep generative framework called the Correlation-aware Causal Variational Auto-encoder (C2VAE)
C2VAE simultaneously recovers the correlation and causal relationships between properties using disentangled latent vectors.
arXiv Detail & Related papers (2024-05-25T13:07:27Z) - Rethinking Radiology Report Generation via Causal Inspired Counterfactual Augmentation [11.266364967223556]
Radiology Report Generation (RRG) draws attention as a vision-and-language interaction of biomedical fields.
Previous works inherited the ideology of traditional language generation tasks, aiming to generate paragraphs with high readability as reports.
Despite significant progress, the independence between diseases-a specific property of RRG-was neglected, yielding the models being confused by the co-occurrence of diseases brought on by the biased data distribution.
arXiv Detail & Related papers (2023-11-22T10:55:36Z) - Identifiable Latent Polynomial Causal Models Through the Lens of Change [82.14087963690561]
Causal representation learning aims to unveil latent high-level causal representations from observed low-level data.
One of its primary tasks is to provide reliable assurance of identifying these latent causal models, known as identifiability.
arXiv Detail & Related papers (2023-10-24T07:46:10Z) - From Identifiable Causal Representations to Controllable Counterfactual Generation: A Survey on Causal Generative Modeling [17.074858228123706]
We focus on fundamental theory, methodology, drawbacks, datasets, and metrics.
We cover applications of causal generative models in fairness, privacy, out-of-distribution generalization, precision medicine, and biological sciences.
arXiv Detail & Related papers (2023-10-17T05:45:32Z) - Inducing Causal Structure for Abstractive Text Summarization [76.1000380429553]
We introduce a Structural Causal Model (SCM) to induce the underlying causal structure of the summarization data.
We propose a Causality Inspired Sequence-to-Sequence model (CI-Seq2Seq) to learn the causal representations that can mimic the causal factors.
Experimental results on two widely used text summarization datasets demonstrate the advantages of our approach.
arXiv Detail & Related papers (2023-08-24T16:06:36Z) - Less is More: Mitigate Spurious Correlations for Open-Domain Dialogue
Response Generation Models by Causal Discovery [52.95935278819512]
We conduct the first study on spurious correlations for open-domain response generation models based on a corpus CGDIALOG curated in our work.
Inspired by causal discovery algorithms, we propose a novel model-agnostic method for training and inference of response generation model.
arXiv Detail & Related papers (2023-03-02T06:33:48Z) - CausalDialogue: Modeling Utterance-level Causality in Conversations [83.03604651485327]
We have compiled and expanded upon a new dataset called CausalDialogue through crowd-sourcing.
This dataset includes multiple cause-effect pairs within a directed acyclic graph (DAG) structure.
We propose a causality-enhanced method called Exponential Average Treatment Effect (ExMATE) to enhance the impact of causality at the utterance level in training neural conversation models.
arXiv Detail & Related papers (2022-12-20T18:31:50Z) - De-Biasing Generative Models using Counterfactual Methods [0.0]
We propose a new decoder based framework named the Causal Counterfactual Generative Model (CCGM)
Our proposed method combines a causal latent space VAE model with specific modification to emphasize causal fidelity.
We explore how better disentanglement of causal learning and encoding/decoding generates higher causal intervention quality.
arXiv Detail & Related papers (2022-07-04T16:53:20Z) - Active Bayesian Causal Inference [72.70593653185078]
We propose Active Bayesian Causal Inference (ABCI), a fully-Bayesian active learning framework for integrated causal discovery and reasoning.
ABCI jointly infers a posterior over causal models and queries of interest.
We show that our approach is more data-efficient than several baselines that only focus on learning the full causal graph.
arXiv Detail & Related papers (2022-06-04T22:38:57Z) - Amortized Causal Discovery: Learning to Infer Causal Graphs from
Time-Series Data [63.15776078733762]
We propose Amortized Causal Discovery, a novel framework to learn to infer causal relations from time-series data.
We demonstrate experimentally that this approach, implemented as a variational model, leads to significant improvements in causal discovery performance.
arXiv Detail & Related papers (2020-06-18T19:59:12Z) - Resolving Spurious Correlations in Causal Models of Environments via
Interventions [2.836066255205732]
We consider the problem of inferring a causal model of a reinforcement learning environment.
Our method designs a reward function that incentivizes an agent to do an intervention to find errors in the causal model.
The experimental results in a grid-world environment show that our approach leads to better causal models compared to baselines.
arXiv Detail & Related papers (2020-02-12T20:20:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.