SSL Framework for Causal Inconsistency between Structures and
Representations
- URL: http://arxiv.org/abs/2310.18634v1
- Date: Sat, 28 Oct 2023 08:29:49 GMT
- Title: SSL Framework for Causal Inconsistency between Structures and
Representations
- Authors: Hang Chen and Xinyu Yang and Keqing Du
- Abstract summary: Cross-pollination of deep learning and causal discovery has catalyzed a burgeoning field of research seeking to elucidate causal relationships within non-statistical data forms like images, videos, and text.
We theoretically develop intervention strategies suitable for indefinite data and derive causal consistency condition (CCC)
CCC could potentially play an influential role in various fields.
- Score: 23.035761299444953
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The cross-pollination of deep learning and causal discovery has catalyzed a
burgeoning field of research seeking to elucidate causal relationships within
non-statistical data forms like images, videos, and text. Such data, often
being named `indefinite data', exhibit unique challenges-inconsistency between
causal structure and representation, which are not common in conventional data
forms. To tackle this issue, we theoretically develop intervention strategies
suitable for indefinite data and derive causal consistency condition (CCC).
Moreover, we design a self-supervised learning (SSL) framework that considers
interventions as `views' and CCC as a `philosophy' with two implement examples
on Supervised Specialized Models (SSMs) and Large Language Models (LLMs),
respectively. To evaluate pure inconsistency manifestations, we have prepared
the first high-quality causal dialogue dataset-Causalogue. Evaluations are also
performed on three other downstream tasks. Extensive experimentation has
substantiated the efficacy of our methodology, illuminating how CCC could
potentially play an influential role in various fields.
Related papers
- Towards Causal Relationship in Indefinite Data: Baseline Model and New
Datasets [23.035761299444953]
"Indefinite Data" is characterized by multi-structure data and multi-value representations.
We release two high-quality datasets - Causalogue and Causaction.
We propose a probabilistic framework as a baseline, incorporating three designed highlights for this gap.
arXiv Detail & Related papers (2024-01-16T09:15:43Z) - Causal Representation Learning Made Identifiable by Grouping of Observational Variables [8.157856010838382]
Causal Representation Learning aims to learn a causal model for hidden features in a data-driven manner.
Here, we show identifiability based on novel, weak constraints.
We also propose a novel self-supervised estimation framework consistent with the model.
arXiv Detail & Related papers (2023-10-24T10:38:02Z) - Towards Causal Foundation Model: on Duality between Causal Inference and Attention [18.046388712804042]
We take a first step towards building causally-aware foundation models for treatment effect estimations.
We propose a novel, theoretically justified method called Causal Inference with Attention (CInA)
arXiv Detail & Related papers (2023-10-01T22:28:34Z) - Inducing Causal Structure for Abstractive Text Summarization [76.1000380429553]
We introduce a Structural Causal Model (SCM) to induce the underlying causal structure of the summarization data.
We propose a Causality Inspired Sequence-to-Sequence model (CI-Seq2Seq) to learn the causal representations that can mimic the causal factors.
Experimental results on two widely used text summarization datasets demonstrate the advantages of our approach.
arXiv Detail & Related papers (2023-08-24T16:06:36Z) - Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Learning a Structural Causal Model for Intuition Reasoning in
Conversation [20.243323155177766]
Reasoning, a crucial aspect of NLP research, has not been adequately addressed by prevailing models.
We develop a conversation cognitive model ( CCM) that explains how each utterance receives and activates channels of information.
By leveraging variational inference, it explores substitutes for implicit causes, addresses the issue of their unobservability, and reconstructs the causal representations of utterances through the evidence lower bounds.
arXiv Detail & Related papers (2023-05-28T13:54:09Z) - Causal Triplet: An Open Challenge for Intervention-centric Causal
Representation Learning [98.78136504619539]
Causal Triplet is a causal representation learning benchmark featuring visually more complex scenes.
We show that models built with the knowledge of disentangled or object-centric representations significantly outperform their distributed counterparts.
arXiv Detail & Related papers (2023-01-12T17:43:38Z) - Effect Identification in Cluster Causal Diagrams [51.42809552422494]
We introduce a new type of graphical model called cluster causal diagrams (for short, C-DAGs)
C-DAGs allow for the partial specification of relationships among variables based on limited prior knowledge.
We develop the foundations and machinery for valid causal inferences over C-DAGs.
arXiv Detail & Related papers (2022-02-22T21:27:31Z) - Towards Robust and Adaptive Motion Forecasting: A Causal Representation
Perspective [72.55093886515824]
We introduce a causal formalism of motion forecasting, which casts the problem as a dynamic process with three groups of latent variables.
We devise a modular architecture that factorizes the representations of invariant mechanisms and style confounders to approximate a causal graph.
Experiment results on synthetic and real datasets show that our three proposed components significantly improve the robustness and reusability of the learned motion representations.
arXiv Detail & Related papers (2021-11-29T18:59:09Z) - Uncovering Main Causalities for Long-tailed Information Extraction [14.39860866665021]
Long-tailed distributions caused by the selection bias of a dataset may lead to incorrect correlations.
This motivates us to propose counterfactual IE (CFIE), a novel framework that aims to uncover the main causalities behind data.
arXiv Detail & Related papers (2021-09-11T08:08:24Z) - On Disentangled Representations Learned From Correlated Data [59.41587388303554]
We bridge the gap to real-world scenarios by analyzing the behavior of the most prominent disentanglement approaches on correlated data.
We show that systematically induced correlations in the dataset are being learned and reflected in the latent representations.
We also demonstrate how to resolve these latent correlations, either using weak supervision during training or by post-hoc correcting a pre-trained model with a small number of labels.
arXiv Detail & Related papers (2020-06-14T12:47:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.