Universal Natural Language Processing with Limited Annotations: Try
Few-shot Textual Entailment as a Start
- URL: http://arxiv.org/abs/2010.02584v1
- Date: Tue, 6 Oct 2020 09:50:25 GMT
- Title: Universal Natural Language Processing with Limited Annotations: Try
Few-shot Textual Entailment as a Start
- Authors: Wenpeng Yin, Nazneen Fatema Rajani, Dragomir Radev, Richard Socher,
Caiming Xiong
- Abstract summary: Universal Few-shot textual Entailment (UFO-Entail)
We introduce Universal Few-shot textual Entailment (UFO-Entail)
We demonstrate that this framework enables a pretrained entailment model to work well on new entailment domains in a few-shot setting.
- Score: 125.23550801424328
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A standard way to address different NLP problems is by first constructing a
problem-specific dataset, then building a model to fit this dataset. To build
the ultimate artificial intelligence, we desire a single machine that can
handle diverse new problems, for which task-specific annotations are limited.
We bring up textual entailment as a unified solver for such NLP problems.
However, current research of textual entailment has not spilled much ink on the
following questions: (i) How well does a pretrained textual entailment system
generalize across domains with only a handful of domain-specific examples? and
(ii) When is it worth transforming an NLP task into textual entailment? We
argue that the transforming is unnecessary if we can obtain rich annotations
for this task. Textual entailment really matters particularly when the target
NLP task has insufficient annotations.
Universal NLP can be probably achieved through different routines. In this
work, we introduce Universal Few-shot textual Entailment (UFO-Entail). We
demonstrate that this framework enables a pretrained entailment model to work
well on new entailment domains in a few-shot setting, and show its
effectiveness as a unified solver for several downstream NLP tasks such as
question answering and coreference resolution when the end-task annotations are
limited. Code: https://github.com/salesforce/UniversalFewShotNLP
Related papers
- Scalable and Domain-General Abstractive Proposition Segmentation [20.532804009152255]
We focus on the task of abstractive proposition segmentation (APS): transforming text into simple, self-contained, well-formed sentences.
We first introduce evaluation metrics for the task to measure several dimensions of quality.
We then propose a scalable, yet accurate, proposition segmentation model.
arXiv Detail & Related papers (2024-06-28T10:24:31Z) - SeqGPT: An Out-of-the-box Large Language Model for Open Domain Sequence
Understanding [103.34092301324425]
Large language models (LLMs) have shown impressive ability for open-domain NLP tasks.
We present SeqGPT, a bilingual (i.e., English and Chinese) open-source autoregressive model specially enhanced for open-domain natural language understanding.
arXiv Detail & Related papers (2023-08-21T07:31:19Z) - EXnet: Efficient In-context Learning for Data-less Text classification [0.0]
We present EXnet, a model specifically designed to perform in-context learning without limitations on the number of examples.
We argue that in-context learning is an effective method to increase task accuracy, and providing examples facilitates cross-task generalization.
With extensive experiments, we show that even our smallest model (15M parameters) generalizes to several unseen classification tasks and domains.
arXiv Detail & Related papers (2023-05-24T01:40:57Z) - Binding Language Models in Symbolic Languages [146.3027328556881]
Binder is a training-free neural-symbolic framework that maps the task input to a program.
In the parsing stage, Codex is able to identify the part of the task input that cannot be answerable by the original programming language.
In the execution stage, Codex can perform versatile functionalities given proper prompts in the API calls.
arXiv Detail & Related papers (2022-10-06T12:55:17Z) - Evaluating NLP Systems On a Novel Cloze Task: Judging the Plausibility
of Possible Fillers in Instructional Texts [2.3449131636069898]
Cloze task is a widely used task to evaluate an NLP system's language understanding ability.
New task is proposed: predicting if a filler word in a cloze task is a good, neutral, or bad candidate.
arXiv Detail & Related papers (2021-12-03T12:02:52Z) - DocNLI: A Large-scale Dataset for Document-level Natural Language
Inference [55.868482696821815]
Natural language inference (NLI) is formulated as a unified framework for solving various NLP problems.
This work presents DocNLI -- a newly-constructed large-scale dataset for document-level NLI.
arXiv Detail & Related papers (2021-06-17T13:02:26Z) - Language Models are Few-Shot Learners [61.36677350504291]
We show that scaling up language models greatly improves task-agnostic, few-shot performance.
We train GPT-3, an autoregressive language model with 175 billion parameters, and test its performance in the few-shot setting.
GPT-3 achieves strong performance on many NLP datasets, including translation, question-answering, and cloze tasks.
arXiv Detail & Related papers (2020-05-28T17:29:03Z) - Enabling Language Models to Fill in the Blanks [81.59381915581892]
We present a simple approach for text infilling, the task of predicting missing spans of text at any position in a document.
We train (or fine-tune) off-the-shelf language models on sequences containing the concatenation of artificially-masked text and the text which was masked.
We show that this approach, which we call infilling by language modeling, can enable LMs to infill entire sentences effectively on three different domains: short stories, scientific abstracts, and lyrics.
arXiv Detail & Related papers (2020-05-11T18:00:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.