CASE: Context-Aware Semantic Expansion
- URL: http://arxiv.org/abs/1912.13194v1
- Date: Tue, 31 Dec 2019 06:38:57 GMT
- Title: CASE: Context-Aware Semantic Expansion
- Authors: Jialong Han, Aixin Sun, Haisong Zhang, Chenliang Li, Shuming Shi
- Abstract summary: This paper defines and studies a new task called Context-Aware Semantic Expansion (CASE)
Given a seed term in a sentential context, we aim to suggest other terms that well fit the context as the seed.
We show that annotations for this task can be harvested at scale from existing corpora, in a fully automatic manner.
- Score: 68.30244980290742
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we define and study a new task called Context-Aware Semantic
Expansion (CASE). Given a seed term in a sentential context, we aim to suggest
other terms that well fit the context as the seed. CASE has many interesting
applications such as query suggestion, computer-assisted writing, and word
sense disambiguation, to name a few. Previous explorations, if any, only
involve some similar tasks, and all require human annotations for evaluation.
In this study, we demonstrate that annotations for this task can be harvested
at scale from existing corpora, in a fully automatic manner. On a dataset of
1.8 million sentences thus derived, we propose a network architecture that
encodes the context and seed term separately before suggesting alternative
terms. The context encoder in this architecture can be easily extended by
incorporating seed-aware attention. Our experiments demonstrate that
competitive results are achieved with appropriate choices of context encoder
and attention scoring function.
Related papers
- Syntax and Semantics Meet in the "Middle": Probing the Syntax-Semantics
Interface of LMs Through Agentivity [68.8204255655161]
We present the semantic notion of agentivity as a case study for probing such interactions.
This suggests LMs may potentially serve as more useful tools for linguistic annotation, theory testing, and discovery.
arXiv Detail & Related papers (2023-05-29T16:24:01Z) - In-context Learning as Maintaining Coherency: A Study of On-the-fly
Machine Translation Using Large Language Models [15.309754694595322]
We present a perspective of in-context learning as the desired generation task maintaining coherency with its context.
We first investigate randomly sampled prompts across 4 domains, and find that translation performance improves when shown in-domain prompts.
In doing so, we demonstrate the efficacy of In-context Machine Translation for on-the-fly adaptation.
arXiv Detail & Related papers (2023-05-05T14:30:20Z) - Dense Paraphrasing for Textual Enrichment [7.6233489924270765]
We define the process of rewriting a textual expression (lexeme or phrase) such that it reduces ambiguity while also making explicit the underlying semantics that is not (necessarily) expressed in the economy of sentence structure as Dense Paraphrasing (DP)
We build the first complete DP dataset, provide the scope and design of the annotation task, and present results demonstrating how this DP process can enrich a source text to improve inferencing and QA task performance.
arXiv Detail & Related papers (2022-10-20T19:58:31Z) - ezCoref: Towards Unifying Annotation Guidelines for Coreference
Resolution [28.878540389202367]
We develop a crowdsourcing-friendly coreference annotation methodology, ezCoref, consisting of an annotation tool and an interactive tutorial.
We use ezCoref to re-annotate 240 passages from seven existing English coreference datasets (spanning fiction, news, and multiple other domains) while teaching annotators only cases that are treated similarly across these datasets.
Surprisingly, we find that reasonable quality annotations were already achievable (>90% agreement between the crowd and expert annotations) even without extensive training.
arXiv Detail & Related papers (2022-10-13T17:09:59Z) - Out of Context: A New Clue for Context Modeling of Aspect-based
Sentiment Analysis [54.735400754548635]
ABSA aims to predict the sentiment expressed in a review with respect to a given aspect.
The given aspect should be considered as a new clue out of context in the context modeling process.
We design several aspect-aware context encoders based on different backbones.
arXiv Detail & Related papers (2021-06-21T02:26:03Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - Context Dependent Semantic Parsing: A Survey [56.69006903481575]
semantic parsing is the task of translating natural language utterances into machine-readable meaning representations.
Currently, most semantic parsing methods are not able to utilize contextual information.
To address this issue, context dependent semantic parsing has recently drawn a lot of attention.
arXiv Detail & Related papers (2020-11-02T07:51:05Z) - Video captioning with stacked attention and semantic hard pull [0.0]
The task of generating a semantically accurate description of a video is quite complex.
This paper proposes a novel architecture, namely Semantically Sensible Video Captioning (SSVC)
The paper reports that the use of the novelties improves the performance of state-of-the-art architectures.
arXiv Detail & Related papers (2020-09-15T19:34:37Z) - Empower Entity Set Expansion via Language Model Probing [58.78909391545238]
Existing set expansion methods bootstrap the seed entity set by adaptively selecting context features and extracting new entities.
A key challenge for entity set expansion is to avoid selecting ambiguous context features which will shift the class semantics and lead to accumulative errors in later iterations.
We propose a novel iterative set expansion framework that leverages automatically generated class names to address the semantic drift issue.
arXiv Detail & Related papers (2020-04-29T00:09:43Z) - Finding Black Cat in a Coal Cellar -- Keyphrase Extraction &
Keyphrase-Rubric Relationship Classification from Complex Assignments [5.067828201066184]
This paper aims to quantify the effectiveness of supervised and unsupervised approaches for the task for keyphrase extraction.
We find that (i) unsupervised MultiPartiteRank produces the best result for keyphrase extraction.
We also present a comprehensive analysis and derive useful observations for those interested in these tasks for the future.
arXiv Detail & Related papers (2020-04-03T13:18:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.