e-SNLI-VE: Corrected Visual-Textual Entailment with Natural Language
Explanations
- URL: http://arxiv.org/abs/2004.03744v3
- Date: Thu, 19 Aug 2021 09:26:21 GMT
- Title: e-SNLI-VE: Corrected Visual-Textual Entailment with Natural Language
Explanations
- Authors: Virginie Do, Oana-Maria Camburu, Zeynep Akata and Thomas Lukasiewicz
- Abstract summary: We present a data collection effort to correct the class with the highest error rate in SNLI-VE.
Thirdly, we introduce e-SNLI-VE, which appends human-written natural language explanations to SNLI-VE.
We train models that learn from these explanations at training time, and output such explanations at testing time.
- Score: 87.71914254873857
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recently proposed SNLI-VE corpus for recognising visual-textual
entailment is a large, real-world dataset for fine-grained multimodal
reasoning. However, the automatic way in which SNLI-VE has been assembled (via
combining parts of two related datasets) gives rise to a large number of errors
in the labels of this corpus. In this paper, we first present a data collection
effort to correct the class with the highest error rate in SNLI-VE. Secondly,
we re-evaluate an existing model on the corrected corpus, which we call
SNLI-VE-2.0, and provide a quantitative comparison with its performance on the
non-corrected corpus. Thirdly, we introduce e-SNLI-VE, which appends
human-written natural language explanations to SNLI-VE-2.0. Finally, we train
models that learn from these explanations at training time, and output such
explanations at testing time.
Related papers
- A Novel Cartography-Based Curriculum Learning Method Applied on RoNLI: The First Romanian Natural Language Inference Corpus [71.77214818319054]
Natural language inference is a proxy for natural language understanding.
There is no publicly available NLI corpus for the Romanian language.
We introduce the first Romanian NLI corpus (RoNLI) comprising 58K training sentence pairs.
arXiv Detail & Related papers (2024-05-20T08:41:15Z) - Improving Domain-Specific Retrieval by NLI Fine-Tuning [64.79760042717822]
This article investigates the fine-tuning potential of natural language inference (NLI) data to improve information retrieval and ranking.
We employ both monolingual and multilingual sentence encoders fine-tuned by a supervised method utilizing contrastive loss and NLI data.
Our results point to the fact that NLI fine-tuning increases the performance of the models in both tasks and both languages, with the potential to improve mono- and multilingual models.
arXiv Detail & Related papers (2023-08-06T12:40:58Z) - Can current NLI systems handle German word order? Investigating language
model performance on a new German challenge set of minimal pairs [3.0534660670547873]
Compared to English, German word order is freer and therefore poses additional challenges for natural language inference (NLI)
We create WOGLI (Word Order in German Language Inference), the first adversarial NLI dataset for German word order.
We show that current German autoencoding models fine-tuned on translated NLI data can struggle on this challenge set.
arXiv Detail & Related papers (2023-06-07T15:33:07Z) - Embracing Ambiguity: Shifting the Training Target of NLI Models [21.153381668435145]
We explore the option of training on the estimated label distribution of the annotators in the Natural Language Inference (NLI) task.
We show it is possible to reduce ChaosNLI divergence scores when finetuning on this data, a promising first step towards learning how to capture linguistic ambiguity.
arXiv Detail & Related papers (2021-06-06T03:18:53Z) - LIREx: Augmenting Language Inference with Relevant Explanation [1.4780878458667916]
Natural language explanations (NLEs) are a form of data annotation in which annotators identify rationales when assigning labels to data instances.
NLEs have been shown to capture human reasoning better, but not as beneficial for natural language inference.
We propose a novel framework, LIREx, that incorporates both a rationale-enabled explanation generator and an instance selector to select only relevant NLEs.
arXiv Detail & Related papers (2020-12-16T18:49:29Z) - Evaluating Semantic Accuracy of Data-to-Text Generation with Natural
Language Inference [3.42658286826597]
A major challenge in evaluating data-to-text (D2T) generation is measuring the semantic accuracy of the generated text.
We propose a new metric for evaluating the semantic accuracy of D2T generation based on a neural model pretrained for natural language inference (NLI)
Our experiments on two recent D2T datasets show that our metric can achieve high accuracy in identifying erroneous system outputs.
arXiv Detail & Related papers (2020-11-21T16:37:28Z) - Word Shape Matters: Robust Machine Translation with Visual Embedding [78.96234298075389]
We introduce a new encoding of the input symbols for character-level NLP models.
It encodes the shape of each character through the images depicting the letters when printed.
We name this new strategy visual embedding and it is expected to improve the robustness of NLP models.
arXiv Detail & Related papers (2020-10-20T04:08:03Z) - Improving Text Generation with Student-Forcing Optimal Transport [122.11881937642401]
We propose using optimal transport (OT) to match the sequences generated in training and testing modes.
An extension is also proposed to improve the OT learning, based on the structural and contextual information of the text sequences.
The effectiveness of the proposed method is validated on machine translation, text summarization, and text generation tasks.
arXiv Detail & Related papers (2020-10-12T19:42:25Z) - Mining Knowledge for Natural Language Inference from Wikipedia
Categories [53.26072815839198]
We introduce WikiNLI: a resource for improving model performance on NLI and LE tasks.
It contains 428,899 pairs of phrases constructed from naturally annotated category hierarchies in Wikipedia.
We show that we can improve strong baselines such as BERT and RoBERTa by pretraining them on WikiNLI and transferring the models on downstream tasks.
arXiv Detail & Related papers (2020-10-03T00:45:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.