Verb Sense Clustering using Contextualized Word Representations for
Semantic Frame Induction
- URL: http://arxiv.org/abs/2105.13465v1
- Date: Thu, 27 May 2021 21:53:40 GMT
- Title: Verb Sense Clustering using Contextualized Word Representations for
Semantic Frame Induction
- Authors: Kosuke Yamada, Ryohei Sasano, Koichi Takeda
- Abstract summary: Contextualized word representations have proven useful for various natural language processing tasks.
In this paper, we focus on verbs that evoke different frames depending on the context.
We investigate how well contextualized word representations can recognize the difference of frames that the same verb evokes.
- Score: 9.93359829907774
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Contextualized word representations have proven useful for various natural
language processing tasks. However, it remains unclear to what extent these
representations can cover hand-coded semantic information such as semantic
frames, which specify the semantic role of the arguments associated with a
predicate. In this paper, we focus on verbs that evoke different frames
depending on the context, and we investigate how well contextualized word
representations can recognize the difference of frames that the same verb
evokes. We also explore which types of representation are suitable for semantic
frame induction. In our experiments, we compare seven different contextualized
word representations for two English frame-semantic resources, FrameNet and
PropBank. We demonstrate that several contextualized word representations,
especially BERT and its variants, are considerably informative for semantic
frame induction. Furthermore, we examine the extent to which the contextualized
representation of a verb can estimate the number of frames that the verb can
evoke.
Related papers
- The Impact of Word Splitting on the Semantic Content of Contextualized
Word Representations [3.4668147567693453]
The quality of representations of words that are split is often, but not always, worse than that of the embeddings of known words.
Our analysis reveals, among other interesting findings, that the quality of representations of words that are split is often, but not always, worse than that of the embeddings of known words.
arXiv Detail & Related papers (2024-02-22T15:04:24Z) - Unsupervised Mapping of Arguments of Deverbal Nouns to Their
Corresponding Verbal Labels [52.940886615390106]
Deverbal nouns are verbs commonly used in written English texts to describe events or actions, as well as their arguments.
The solutions that do exist for handling arguments of nominalized constructions are based on semantic annotation.
We propose to adopt a more syntactic approach, which maps the arguments of deverbal nouns to the corresponding verbal construction.
arXiv Detail & Related papers (2023-06-24T10:07:01Z) - Interpretable Word Sense Representations via Definition Generation: The
Case of Semantic Change Analysis [3.515619810213763]
We propose using automatically generated natural language definitions of contextualised word usages as interpretable word and word sense representations.
We demonstrate how the resulting sense labels can make existing approaches to semantic change analysis more interpretable.
arXiv Detail & Related papers (2023-05-19T20:36:21Z) - PropSegmEnt: A Large-Scale Corpus for Proposition-Level Segmentation and
Entailment Recognition [63.51569687229681]
We argue for the need to recognize the textual entailment relation of each proposition in a sentence individually.
We propose PropSegmEnt, a corpus of over 45K propositions annotated by expert human raters.
Our dataset structure resembles the tasks of (1) segmenting sentences within a document to the set of propositions, and (2) classifying the entailment relation of each proposition with respect to a different yet topically-aligned document.
arXiv Detail & Related papers (2022-12-21T04:03:33Z) - Lost in Context? On the Sense-wise Variance of Contextualized Word
Embeddings [11.475144702935568]
We quantify how much the contextualized embeddings of each word sense vary across contexts in typical pre-trained models.
We find that word representations are position-biased, where the first words in different contexts tend to be more similar.
arXiv Detail & Related papers (2022-08-20T12:27:25Z) - GSRFormer: Grounded Situation Recognition Transformer with Alternate
Semantic Attention Refinement [73.73599110214828]
Grounded Situation Recognition (GSR) aims to generate structured semantic summaries of images for human-like'' event understanding.
Inspired by object detection and image captioning tasks, existing methods typically employ a two-stage framework.
We propose a novel two-stage framework that focuses on utilizing such bidirectional relations within verbs and roles.
arXiv Detail & Related papers (2022-08-18T17:13:59Z) - Cross-Modal Discrete Representation Learning [73.68393416984618]
We present a self-supervised learning framework that learns a representation that captures finer levels of granularity across different modalities.
Our framework relies on a discretized embedding space created via vector quantization that is shared across different modalities.
arXiv Detail & Related papers (2021-06-10T00:23:33Z) - Accurate Word Representations with Universal Visual Guidance [55.71425503859685]
This paper proposes a visual representation method to explicitly enhance conventional word embedding with multiple-aspect senses from visual guidance.
We build a small-scale word-image dictionary from a multimodal seed dataset where each word corresponds to diverse related images.
Experiments on 12 natural language understanding and machine translation tasks further verify the effectiveness and the generalization capability of the proposed approach.
arXiv Detail & Related papers (2020-12-30T09:11:50Z) - Unsupervised Distillation of Syntactic Information from Contextualized
Word Representations [62.230491683411536]
We tackle the task of unsupervised disentanglement between semantics and structure in neural language representations.
To this end, we automatically generate groups of sentences which are structurally similar but semantically different.
We demonstrate that our transformation clusters vectors in space by structural properties, rather than by lexical semantics.
arXiv Detail & Related papers (2020-10-11T15:13:18Z) - Context-theoretic Semantics for Natural Language: an Algebraic Framework [0.0]
We present a framework for natural language semantics in which words, phrases and sentences are all represented as vectors.
We show that the vector representations of words can be considered as elements of an algebra over a field.
arXiv Detail & Related papers (2020-09-22T13:31:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.