CREST: A Joint Framework for Rationalization and Counterfactual Text
Generation
- URL: http://arxiv.org/abs/2305.17075v1
- Date: Fri, 26 May 2023 16:34:58 GMT
- Title: CREST: A Joint Framework for Rationalization and Counterfactual Text
Generation
- Authors: Marcos Treviso, Alexis Ross, Nuno M. Guerreiro, Andr\'e F. T. Martins
- Abstract summary: We introduce CREST (ContRastive Edits with Sparse raTionalization), a framework for selective rationalization and counterfactual text generation.
CREST generates valid counterfactuals that are more natural than those produced by previous methods.
New loss function that leverages CREST counterfactuals to regularize selective rationales improves both model robustness and rationale quality.
- Score: 5.606679908174783
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Selective rationales and counterfactual examples have emerged as two
effective, complementary classes of interpretability methods for analyzing and
training NLP models. However, prior work has not explored how these methods can
be integrated to combine their complementary advantages. We overcome this
limitation by introducing CREST (ContRastive Edits with Sparse
raTionalization), a joint framework for selective rationalization and
counterfactual text generation, and show that this framework leads to
improvements in counterfactual quality, model robustness, and interpretability.
First, CREST generates valid counterfactuals that are more natural than those
produced by previous methods, and subsequently can be used for data
augmentation at scale, reducing the need for human-generated examples. Second,
we introduce a new loss function that leverages CREST counterfactuals to
regularize selective rationales and show that this regularization improves both
model robustness and rationale quality, compared to methods that do not
leverage CREST counterfactuals. Our results demonstrate that CREST successfully
bridges the gap between selective rationales and counterfactual examples,
addressing the limitations of existing methods and providing a more
comprehensive view of a model's predictions.
Related papers
- A Survey of Low-shot Vision-Language Model Adaptation via Representer Theorem [38.84662767814454]
Key challenge to address under the condition of limited training data is how to fine-tune pre-trained vision-language models in a parameter-efficient manner.
This paper proposes a unified computational framework to integrate existing methods together, identify their nature and support in-depth comparison.
As a demonstration, we extend existing methods by modeling inter-class correlation between representers in reproducing kernel Hilbert space (RKHS)
arXiv Detail & Related papers (2024-10-15T15:22:30Z) - Advancing Semantic Textual Similarity Modeling: A Regression Framework with Translated ReLU and Smooth K2 Loss [3.435381469869212]
This paper presents an innovative regression framework for Sentence-BERT STS tasks.
It proposes two simple yet effective loss functions: Translated ReLU and Smooth K2 Loss.
Experimental results demonstrate that our method achieves convincing performance across seven established STS benchmarks.
arXiv Detail & Related papers (2024-06-08T02:52:43Z) - Make Your Decision Convincing! A Unified Two-Stage Framework:
Self-Attribution and Decision-Making [24.906886146275127]
We propose a unified two-stage framework known as Self-Attribution and Decision-Making (SADM)
We demonstrate that our framework not only establishes a more reliable link between the generated rationale and model decision but also achieves competitive results in task performance and the quality of rationale.
arXiv Detail & Related papers (2023-10-20T15:59:57Z) - REX: Rapid Exploration and eXploitation for AI Agents [103.68453326880456]
We propose an enhanced approach for Rapid Exploration and eXploitation for AI Agents called REX.
REX introduces an additional layer of rewards and integrates concepts similar to Upper Confidence Bound (UCB) scores, leading to more robust and efficient AI agent performance.
arXiv Detail & Related papers (2023-07-18T04:26:33Z) - Advancing Counterfactual Inference through Nonlinear Quantile Regression [77.28323341329461]
We propose a framework for efficient and effective counterfactual inference implemented with neural networks.
The proposed approach enhances the capacity to generalize estimated counterfactual outcomes to unseen data.
Empirical results conducted on multiple datasets offer compelling support for our theoretical assertions.
arXiv Detail & Related papers (2023-06-09T08:30:51Z) - Less is More: Mitigate Spurious Correlations for Open-Domain Dialogue
Response Generation Models by Causal Discovery [52.95935278819512]
We conduct the first study on spurious correlations for open-domain response generation models based on a corpus CGDIALOG curated in our work.
Inspired by causal discovery algorithms, we propose a novel model-agnostic method for training and inference of response generation model.
arXiv Detail & Related papers (2023-03-02T06:33:48Z) - When to Update Your Model: Constrained Model-based Reinforcement
Learning [50.74369835934703]
We propose a novel and general theoretical scheme for a non-decreasing performance guarantee of model-based RL (MBRL)
Our follow-up derived bounds reveal the relationship between model shifts and performance improvement.
A further example demonstrates that learning models from a dynamically-varying number of explorations benefit the eventual returns.
arXiv Detail & Related papers (2022-10-15T17:57:43Z) - Towards Robust and Reliable Algorithmic Recourse [11.887537452826624]
We propose a novel framework, RObust Algorithmic Recourse (ROAR), that leverages adversarial training for finding recourses that are robust to model shifts.
We also carry out detailed theoretical analysis which underscores the importance of constructing recourses that are robust to model shifts.
arXiv Detail & Related papers (2021-02-26T17:38:52Z) - Control as Hybrid Inference [62.997667081978825]
We present an implementation of CHI which naturally mediates the balance between iterative and amortised inference.
We verify the scalability of our algorithm on a continuous control benchmark, demonstrating that it outperforms strong model-free and model-based baselines.
arXiv Detail & Related papers (2020-07-11T19:44:09Z) - Decomposed Adversarial Learned Inference [118.27187231452852]
We propose a novel approach, Decomposed Adversarial Learned Inference (DALI)
DALI explicitly matches prior and conditional distributions in both data and code spaces.
We validate the effectiveness of DALI on the MNIST, CIFAR-10, and CelebA datasets.
arXiv Detail & Related papers (2020-04-21T20:00:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.