REFER: An End-to-end Rationale Extraction Framework for Explanation
Regularization
- URL: http://arxiv.org/abs/2310.14418v1
- Date: Sun, 22 Oct 2023 21:20:52 GMT
- Title: REFER: An End-to-end Rationale Extraction Framework for Explanation
Regularization
- Authors: Mohammad Reza Ghasemi Madani, Pasquale Minervini
- Abstract summary: We propose REFER, a framework that employs a differentiable rationale extractor that allows to back-propagate through the rationale extraction process.
We analyze the impact of using human highlights during training by jointly training the task model and the rationale extractor.
- Score: 12.409398096527829
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Human-annotated textual explanations are becoming increasingly important in
Explainable Natural Language Processing. Rationale extraction aims to provide
faithful (i.e., reflective of the behavior of the model) and plausible (i.e.,
convincing to humans) explanations by highlighting the inputs that had the
largest impact on the prediction without compromising the performance of the
task model. In recent works, the focus of training rationale extractors was
primarily on optimizing for plausibility using human highlights, while the task
model was trained on jointly optimizing for task predictive accuracy and
faithfulness. We propose REFER, a framework that employs a differentiable
rationale extractor that allows to back-propagate through the rationale
extraction process. We analyze the impact of using human highlights during
training by jointly training the task model and the rationale extractor. In our
experiments, REFER yields significantly better results in terms of
faithfulness, plausibility, and downstream task accuracy on both
in-distribution and out-of-distribution data. On both e-SNLI and CoS-E, our
best setting produces better results in terms of composite normalized relative
gain than the previous baselines by 11% and 3%, respectively.
Related papers
- Discussion: Effective and Interpretable Outcome Prediction by Training Sparse Mixtures of Linear Experts [4.178382980763478]
We propose to train a sparse Mixture-of-Experts where both the gate'' and expert'' sub-nets are Logistic Regressors.
This ensemble-like model is trained end-to-end while automatically selecting a subset of input features in each sub-net.
arXiv Detail & Related papers (2024-07-18T13:59:10Z) - Plausible Extractive Rationalization through Semi-Supervised Entailment Signal [29.67884478799914]
We take a semi-supervised approach to optimize for the plausibility of extracted rationales.
We adopt a pre-trained natural language inference (NLI) model and further fine-tune it on a small set of supervised rationales.
We show that, by enforcing the alignment agreement between the explanation and answer in a question-answering task, the performance can be improved without access to ground truth labels.
arXiv Detail & Related papers (2024-02-13T14:12:32Z) - From Prediction to Action: Critical Role of Performance Estimation for
Machine-Learning-Driven Materials Discovery [2.3243389656894595]
We argue that the lack of proper performance estimation methods from pre-computed data collections is a fundamental problem for improving data-driven materials discovery.
We propose a novel such estimator that, in contrast to na"ive reward estimation, successfully predicts Gaussian processes with the "expected improvement" acquisition function.
arXiv Detail & Related papers (2023-11-27T05:29:43Z) - Target Variable Engineering [0.0]
We compare the predictive performance of regression models trained to predict numeric targets vs. classifiers trained to predict their binarized counterparts.
We find that regression requires significantly more computational effort to converge upon the optimal performance.
arXiv Detail & Related papers (2023-10-13T23:12:21Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Improved Fine-tuning by Leveraging Pre-training Data: Theory and
Practice [52.11183787786718]
Fine-tuning a pre-trained model on the target data is widely used in many deep learning applications.
Recent studies have empirically shown that training from scratch has the final performance that is no worse than this pre-training strategy.
We propose a novel selection strategy to select a subset from pre-training data to help improve the generalization on the target task.
arXiv Detail & Related papers (2021-11-24T06:18:32Z) - Dynamic Iterative Refinement for Efficient 3D Hand Pose Estimation [87.54604263202941]
We propose a tiny deep neural network of which partial layers are iteratively exploited for refining its previous estimations.
We employ learned gating criteria to decide whether to exit from the weight-sharing loop, allowing per-sample adaptation in our model.
Our method consistently outperforms state-of-the-art 2D/3D hand pose estimation approaches in terms of both accuracy and efficiency for widely used benchmarks.
arXiv Detail & Related papers (2021-11-11T23:31:34Z) - SAIS: Supervising and Augmenting Intermediate Steps for Document-Level
Relation Extraction [51.27558374091491]
We propose to explicitly teach the model to capture relevant contexts and entity types by supervising and augmenting intermediate steps (SAIS) for relation extraction.
Based on a broad spectrum of carefully designed tasks, our proposed SAIS method not only extracts relations of better quality due to more effective supervision, but also retrieves the corresponding supporting evidence more accurately.
arXiv Detail & Related papers (2021-09-24T17:37:35Z) - Towards More Fine-grained and Reliable NLP Performance Prediction [85.78131503006193]
We make two contributions to improving performance prediction for NLP tasks.
First, we examine performance predictors for holistic measures of accuracy like F1 or BLEU.
Second, we propose methods to understand the reliability of a performance prediction model from two angles: confidence intervals and calibration.
arXiv Detail & Related papers (2021-02-10T15:23:20Z) - An Information Bottleneck Approach for Controlling Conciseness in
Rationale Extraction [84.49035467829819]
We show that it is possible to better manage this trade-off by optimizing a bound on the Information Bottleneck (IB) objective.
Our fully unsupervised approach jointly learns an explainer that predicts sparse binary masks over sentences, and an end-task predictor that considers only the extracted rationale.
arXiv Detail & Related papers (2020-05-01T23:26:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.