Nominal Compound Chain Extraction: A New Task for Semantic-enriched
Lexical Chain
- URL: http://arxiv.org/abs/2009.09173v1
- Date: Sat, 19 Sep 2020 06:20:37 GMT
- Title: Nominal Compound Chain Extraction: A New Task for Semantic-enriched
Lexical Chain
- Authors: Bobo Li and Hao Fei and Yafeng Ren and Donghong Ji
- Abstract summary: We introduce a novel task, Nominal Compound Chain Extraction (NCCE), extracting and clustering all the nominal compounds that share identical semantic topics.
In addition, we model the task as a two-stage prediction (i.e., compound extraction and chain detection), which is handled via a proposed joint framework.
The experiments are based on our manually annotated corpus, and the results prove the necessity of the NCCE task.
- Score: 34.352862428120126
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Lexical chain consists of cohesion words in a document, which implies the
underlying structure of a text, and thus facilitates downstream NLP tasks.
Nevertheless, existing work focuses on detecting the simple surface lexicons
with shallow syntax associations, ignoring the semantic-aware lexical compounds
as well as the latent semantic frames, (e.g., topic), which can be much more
crucial for real-world NLP applications. In this paper, we introduce a novel
task, Nominal Compound Chain Extraction (NCCE), extracting and clustering all
the nominal compounds that share identical semantic topics. In addition, we
model the task as a two-stage prediction (i.e., compound extraction and chain
detection), which is handled via a proposed joint framework. The model employs
the BERT encoder to yield contextualized document representation. Also, HowNet
is exploited as external resources for offering rich sememe information. The
experiments are based on our manually annotated corpus, and the results prove
the necessity of the NCCE task as well as the effectiveness of our joint
approach.
Related papers
- DepNeCTI: Dependency-based Nested Compound Type Identification for
Sanskrit [7.04795623262177]
This work introduces the novel task of nested compound type identification (NeCTI)
It aims to identify nested spans of a multi-component compound and decode the implicit semantic relations between them.
To the best of our knowledge, this is the first attempt in the field of lexical semantics to propose this task.
arXiv Detail & Related papers (2023-10-14T06:11:53Z) - Large Language Model Prompt Chaining for Long Legal Document
Classification [2.3148470932285665]
Chaining is a strategy used to decompose complex tasks into smaller, manageable components.
We demonstrate that through prompt chaining, we can not only enhance the performance over zero-shot, but also surpass the micro-F1 score achieved by larger models.
arXiv Detail & Related papers (2023-08-08T08:57:01Z) - Enhanced word embeddings using multi-semantic representation through
lexical chains [1.8199326045904998]
We propose two novel algorithms, called Flexible Lexical Chain II and Fixed Lexical Chain II.
These algorithms combine the semantic relations derived from lexical chains, prior knowledge from lexical databases, and the robustness of the distributional hypothesis in word embeddings as building blocks forming a single system.
Our results show the integration between lexical chains and word embeddings representations sustain state-of-the-art results, even against more complex systems.
arXiv Detail & Related papers (2021-01-22T09:43:33Z) - Structured Prediction as Translation between Augmented Natural Languages [109.50236248762877]
We propose a new framework, Translation between Augmented Natural Languages (TANL), to solve many structured prediction language tasks.
Instead of tackling the problem by training task-specific discriminatives, we frame it as a translation task between augmented natural languages.
Our approach can match or outperform task-specific models on all tasks, and in particular, achieves new state-of-the-art results on joint entity and relation extraction.
arXiv Detail & Related papers (2021-01-14T18:32:21Z) - Relation Clustering in Narrative Knowledge Graphs [71.98234178455398]
relational sentences in the original text are embedded (with SBERT) and clustered in order to merge together semantically similar relations.
Preliminary tests show that such clustering might successfully detect similar relations, and provide a valuable preprocessing for semi-supervised approaches.
arXiv Detail & Related papers (2020-11-27T10:43:04Z) - Keyphrase Extraction with Dynamic Graph Convolutional Networks and
Diversified Inference [50.768682650658384]
Keyphrase extraction (KE) aims to summarize a set of phrases that accurately express a concept or a topic covered in a given document.
Recent Sequence-to-Sequence (Seq2Seq) based generative framework is widely used in KE task, and it has obtained competitive performance on various benchmarks.
In this paper, we propose to adopt the Dynamic Graph Convolutional Networks (DGCN) to solve the above two problems simultaneously.
arXiv Detail & Related papers (2020-10-24T08:11:23Z) - Pairwise Representation Learning for Event Coreference [73.10563168692667]
We develop a Pairwise Representation Learning (PairwiseRL) scheme for the event mention pairs.
Our representation supports a finer, structured representation of the text snippet to facilitate encoding events and their arguments.
We show that PairwiseRL, despite its simplicity, outperforms the prior state-of-the-art event coreference systems on both cross-document and within-document event coreference benchmarks.
arXiv Detail & Related papers (2020-10-24T06:55:52Z) - Cross-Supervised Joint-Event-Extraction with Heterogeneous Information
Networks [61.950353376870154]
Joint-event-extraction is a sequence-to-sequence labeling task with a tag set composed of tags of triggers and entities.
We propose a Cross-Supervised Mechanism (CSM) to alternately supervise the extraction of triggers or entities.
Our approach outperforms the state-of-the-art methods in both entity and trigger extraction.
arXiv Detail & Related papers (2020-10-13T11:51:17Z) - Joint Semantic Analysis with Document-Level Cross-Task Coherence Rewards [13.753240692520098]
We present a neural network architecture for joint coreference resolution and semantic role labeling for English.
We use reinforcement learning to encourage global coherence over the document and between semantic annotations.
This leads to improvements on both tasks in multiple datasets from different domains.
arXiv Detail & Related papers (2020-10-12T09:36:24Z) - Extractive Summarization as Text Matching [123.09816729675838]
This paper creates a paradigm shift with regard to the way we build neural extractive summarization systems.
We formulate the extractive summarization task as a semantic text matching problem.
We have driven the state-of-the-art extractive result on CNN/DailyMail to a new level (44.41 in ROUGE-1)
arXiv Detail & Related papers (2020-04-19T08:27:57Z) - Selective Attention Encoders by Syntactic Graph Convolutional Networks
for Document Summarization [21.351111598564987]
We propose a graph to connect the parsing trees from the sentences in a document and utilize the stacked graph convolutional networks (GCNs) to learn the syntactic representation for a document.
The proposed GCNs based selective attention approach outperforms the baselines and achieves the state-of-the-art performance on the dataset.
arXiv Detail & Related papers (2020-03-18T01:30:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.