SemEval-2020 Task 6: Definition extraction from free text with the DEFT
corpus
- URL: http://arxiv.org/abs/2008.13694v1
- Date: Mon, 31 Aug 2020 15:55:24 GMT
- Title: SemEval-2020 Task 6: Definition extraction from free text with the DEFT
corpus
- Authors: Sasha Spala, Nicholas A Miller, Franck Dernoncourt, Carl Dockhorn
- Abstract summary: We present DeftEval, a SemEval shared task in which participants extract definitions from free text.
DeftEval involved 3 distinct subtasks:Sentence classification, sequence labeling, and relation extraction.
- Score: 28.67911239741097
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Research on definition extraction has been conducted for well over a decade,
largely with significant constraints on the type of definitions considered. In
this work, we present DeftEval, a SemEval shared task in which participants
must extract definitions from free text using a term-definition pair corpus
that reflects the complex reality of definitions in natural language.
Definitions and glosses in free text often appear without explicit indicators,
across sentences boundaries, or in an otherwise complex linguistic manner.
DeftEval involved 3 distinct subtasks: 1)Sentence classification, 2) sequence
labeling, and 3) relation extraction.
Related papers
- Boundary-Driven Table-Filling with Cross-Granularity Contrastive Learning for Aspect Sentiment Triplet Extraction [8.011606196420757]
The Aspect Sentiment Triplet Extraction task is one of the most prominent subtasks in fine-grained sentiment analysis.
Most existing approaches frame triplet extraction as a 2D table-filling process in an end-to-end manner.
We propose boundary-driven table-filling with cross-granularity contrastive learning (BTF-CCL) to enhance the semantic consistency between sentence-level representations and word-level representations.
arXiv Detail & Related papers (2025-02-04T02:23:45Z) - Natural Language Decompositions of Implicit Content Enable Better Text
Representations [56.85319224208865]
We introduce a method for the analysis of text that takes implicitly communicated content explicitly into account.
We use a large language model to produce sets of propositions that are inferentially related to the text that has been observed.
Our results suggest that modeling the meanings behind observed language, rather than the literal text alone, is a valuable direction for NLP.
arXiv Detail & Related papers (2023-05-23T23:45:20Z) - PropSegmEnt: A Large-Scale Corpus for Proposition-Level Segmentation and
Entailment Recognition [63.51569687229681]
We argue for the need to recognize the textual entailment relation of each proposition in a sentence individually.
We propose PropSegmEnt, a corpus of over 45K propositions annotated by expert human raters.
Our dataset structure resembles the tasks of (1) segmenting sentences within a document to the set of propositions, and (2) classifying the entailment relation of each proposition with respect to a different yet topically-aligned document.
arXiv Detail & Related papers (2022-12-21T04:03:33Z) - Textual Entailment Recognition with Semantic Features from Empirical
Text Representation [60.31047947815282]
A text entails a hypothesis if and only if the true value of the hypothesis follows the text.
In this paper, we propose a novel approach to identifying the textual entailment relationship between text and hypothesis.
We employ an element-wise Manhattan distance vector-based feature that can identify the semantic entailment relationship between the text-hypothesis pair.
arXiv Detail & Related papers (2022-10-18T10:03:51Z) - Semeval-2022 Task 1: CODWOE -- Comparing Dictionaries and Word
Embeddings [1.5293427903448025]
We focus on relating opaque word vectors with human-readable definitions.
This problem naturally divides into two subtasks: converting definitions into embeddings, and converting embeddings into definitions.
This task was conducted in a multilingual setting, using comparable sets of embeddings trained homogeneously.
arXiv Detail & Related papers (2022-05-27T09:40:33Z) - CDM: Combining Extraction and Generation for Definition Modeling [8.487707405248242]
We propose to combine extraction and generation for definition modeling.
First extract self- and correlative definitional information of target terms from the Web.
Then generate the final definitions by incorporating the extracted definitional information.
arXiv Detail & Related papers (2021-11-14T08:03:18Z) - Connect-the-Dots: Bridging Semantics between Words and Definitions via
Aligning Word Sense Inventories [47.03271152494389]
Word Sense Disambiguation aims to automatically identify the exact meaning of one word according to its context.
Existing supervised models struggle to make correct predictions on rare word senses due to limited training data.
We propose a gloss alignment algorithm that can align definition sentences with the same meaning from different sense inventories to collect rich lexical knowledge.
arXiv Detail & Related papers (2021-10-27T00:04:33Z) - Unsupervised Technical Domain Terms Extraction using Term Extractor [9.23545668304066]
The goal of terminology extraction is to extract relevant words or phrases from a given corpus automatically.
This paper focuses on the unsupervised automated domain term extraction method that considers chunking, preprocessing, and ranking domain-specific terms.
arXiv Detail & Related papers (2021-01-22T09:24:09Z) - Measuring Association Between Labels and Free-Text Rationales [60.58672852655487]
In interpretable NLP, we require faithful rationales that reflect the model's decision-making process for an explained instance.
We demonstrate that pipelines, existing models for faithful extractive rationalization on information-extraction style tasks, do not extend as reliably to "reasoning" tasks requiring free-text rationales.
We turn to models that jointly predict and rationalize, a class of widely used high-performance models for free-text rationalization whose faithfulness is not yet established.
arXiv Detail & Related papers (2020-10-24T03:40:56Z) - Constrained Abstractive Summarization: Preserving Factual Consistency
with Constrained Generation [93.87095877617968]
We propose Constrained Abstractive Summarization (CAS), a general setup that preserves the factual consistency of abstractive summarization.
We adopt lexically constrained decoding, a technique generally applicable to autoregressive generative models, to fulfill CAS.
We observe up to 13.8 ROUGE-2 gains when only one manual constraint is used in interactive summarization.
arXiv Detail & Related papers (2020-10-24T00:27:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.