Towards Interpretable Natural Language Understanding with Explanations
as Latent Variables
- URL: http://arxiv.org/abs/2011.05268v3
- Date: Thu, 26 May 2022 22:12:42 GMT
- Title: Towards Interpretable Natural Language Understanding with Explanations
as Latent Variables
- Authors: Wangchunshu Zhou, Jinyi Hu, Hanlin Zhang, Xiaodan Liang, Maosong Sun,
Chenyan Xiong, Jian Tang
- Abstract summary: We develop a framework for interpretable natural language understanding that requires only a small set of human annotated explanations for training.
Our framework treats natural language explanations as latent variables that model the underlying reasoning process of a neural model.
- Score: 146.83882632854485
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently generating natural language explanations has shown very promising
results in not only offering interpretable explanations but also providing
additional information and supervision for prediction. However, existing
approaches usually require a large set of human annotated explanations for
training while collecting a large set of explanations is not only time
consuming but also expensive. In this paper, we develop a general framework for
interpretable natural language understanding that requires only a small set of
human annotated explanations for training. Our framework treats natural
language explanations as latent variables that model the underlying reasoning
process of a neural model. We develop a variational EM framework for
optimization where an explanation generation module and an
explanation-augmented prediction module are alternatively optimized and
mutually enhance each other. Moreover, we further propose an explanation-based
self-training method under this framework for semi-supervised learning. It
alternates between assigning pseudo-labels to unlabeled data and generating new
explanations to iteratively improve each other. Experiments on two natural
language understanding tasks demonstrate that our framework can not only make
effective predictions in both supervised and semi-supervised settings, but also
generate good natural language explanation.
Related papers
- Explaining Text Similarity in Transformer Models [52.571158418102584]
Recent advances in explainable AI have made it possible to mitigate limitations by leveraging improved explanations for Transformers.
We use BiLRP, an extension developed for computing second-order explanations in bilinear similarity models, to investigate which feature interactions drive similarity in NLP models.
Our findings contribute to a deeper understanding of different semantic similarity tasks and models, highlighting how novel explainable AI methods enable in-depth analyses and corpus-level insights.
arXiv Detail & Related papers (2024-05-10T17:11:31Z) - Explanation-aware Soft Ensemble Empowers Large Language Model In-context
Learning [50.00090601424348]
Large language models (LLMs) have shown remarkable capabilities in various natural language understanding tasks.
We propose EASE, an Explanation-Aware Soft Ensemble framework to empower in-context learning with LLMs.
arXiv Detail & Related papers (2023-11-13T06:13:38Z) - Explainability for Large Language Models: A Survey [59.67574757137078]
Large language models (LLMs) have demonstrated impressive capabilities in natural language processing.
This paper introduces a taxonomy of explainability techniques and provides a structured overview of methods for explaining Transformer-based language models.
arXiv Detail & Related papers (2023-09-02T22:14:26Z) - MaNtLE: Model-agnostic Natural Language Explainer [9.43206883360088]
We introduce MaNtLE, a model-agnostic natural language explainer that analyzes multiple classifier predictions.
MaNtLE uses multi-task training on thousands of synthetic classification tasks to generate faithful explanations.
Simulated user studies indicate that, on average, MaNtLE-generated explanations are at least 11% more faithful compared to LIME and Anchors explanations.
arXiv Detail & Related papers (2023-05-22T12:58:06Z) - Parallel Sentence-Level Explanation Generation for Real-World
Low-Resource Scenarios [18.5713713816771]
This paper is the first to explore the problem smoothly from weak-supervised learning to unsupervised learning.
We propose a non-autoregressive interpretable model to facilitate parallel explanation generation and simultaneous prediction.
arXiv Detail & Related papers (2023-02-21T14:52:21Z) - Explanation Regeneration via Information Bottleneck [29.92996769997743]
We develop an information bottleneck method EIB to produce refined explanations that are sufficient and concise.
Our approach regenerates the free-text explanation by polishing the single-pass output from the pretrained language model.
arXiv Detail & Related papers (2022-12-19T16:41:19Z) - Explanations from Large Language Models Make Small Reasoners Better [61.991772773700006]
We show that our method can consistently and significantly outperform finetuning baselines across different settings.
As a side benefit, human evaluation shows that our method can generate high-quality explanations to justify its predictions.
arXiv Detail & Related papers (2022-10-13T04:50:02Z) - Leap-Of-Thought: Teaching Pre-Trained Models to Systematically Reason
Over Implicit Knowledge [96.92252296244233]
Large pre-trained language models (LMs) acquire some reasoning capacity, but this ability is difficult to control.
We show that LMs can be trained to reliably perform systematic reasoning combining both implicit, pre-trained knowledge and explicit natural language statements.
Our work paves a path towards open-domain systems that constantly improve by interacting with users who can instantly correct a model by adding simple natural language statements.
arXiv Detail & Related papers (2020-06-11T17:02:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.