Semantic Frame Induction with Deep Metric Learning
- URL: http://arxiv.org/abs/2304.14286v1
- Date: Thu, 27 Apr 2023 15:46:09 GMT
- Title: Semantic Frame Induction with Deep Metric Learning
- Authors: Kosuke Yamada, Ryohei Sasano, Koichi Takeda
- Abstract summary: We propose a model that uses deep metric learning to fine-tune a contextualized embedding model.
We apply the fine-tuned contextualized embeddings to perform semantic frame induction.
- Score: 24.486546938073907
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent studies have demonstrated the usefulness of contextualized word
embeddings in unsupervised semantic frame induction. However, they have also
revealed that generic contextualized embeddings are not always consistent with
human intuitions about semantic frames, which causes unsatisfactory performance
for frame induction based on contextualized embeddings. In this paper, we
address supervised semantic frame induction, which assumes the existence of
frame-annotated data for a subset of predicates in a corpus and aims to build a
frame induction model that leverages the annotated data. We propose a model
that uses deep metric learning to fine-tune a contextualized embedding model,
and we apply the fine-tuned contextualized embeddings to perform semantic frame
induction. Our experiments on FrameNet show that fine-tuning with deep metric
learning considerably improves the clustering evaluation scores, namely, the
B-cubed F-score and Purity F-score, by about 8 points or more. We also
demonstrate that our approach is effective even when the number of training
instances is small.
Related papers
- Manual Verbalizer Enrichment for Few-Shot Text Classification [1.860409237919611]
acrshortmave is an approach for verbalizer construction by enrichment of class labels.
Our model achieves state-of-the-art results while using significantly fewer resources.
arXiv Detail & Related papers (2024-10-08T16:16:47Z) - Point Contrastive Prediction with Semantic Clustering for
Self-Supervised Learning on Point Cloud Videos [71.20376514273367]
We propose a unified point cloud video self-supervised learning framework for object-centric and scene-centric data.
Our method outperforms supervised counterparts on a wide range of downstream tasks.
arXiv Detail & Related papers (2023-08-18T02:17:47Z) - Scalable Learning of Latent Language Structure With Logical Offline
Cycle Consistency [71.42261918225773]
Conceptually, LOCCO can be viewed as a form of self-learning where the semantic being trained is used to generate annotations for unlabeled text.
As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model.
arXiv Detail & Related papers (2023-05-31T16:47:20Z) - Interpretable Automatic Fine-grained Inconsistency Detection in Text
Summarization [56.94741578760294]
We propose the task of fine-grained inconsistency detection, the goal of which is to predict the fine-grained types of factual errors in a summary.
Motivated by how humans inspect factual inconsistency in summaries, we propose an interpretable fine-grained inconsistency detection model, FineGrainFact.
arXiv Detail & Related papers (2023-05-23T22:11:47Z) - Acquiring Frame Element Knowledge with Deep Metric Learning for Semantic
Frame Induction [24.486546938073907]
We propose a method that applies deep metric learning to semantic frame induction tasks.
A pre-trained language model is fine-tuned to be suitable for distinguishing frame element roles.
Experimental results on FrameNet demonstrate that our method achieves substantially better performance than existing methods.
arXiv Detail & Related papers (2023-05-23T11:02:28Z) - FineDiving: A Fine-grained Dataset for Procedure-aware Action Quality
Assessment [93.09267863425492]
We argue that understanding both high-level semantics and internal temporal structures of actions in competitive sports videos is the key to making predictions accurate and interpretable.
We construct a new fine-grained dataset, called FineDiving, developed on diverse diving events with detailed annotations on action procedures.
arXiv Detail & Related papers (2022-04-07T17:59:32Z) - Convex Polytope Modelling for Unsupervised Derivation of Semantic
Structure for Data-efficient Natural Language Understanding [31.888489552069146]
A Convex-Polytopic-Model-based framework shows great potential in automatically extracting semantic patterns by exploiting the raw dialog corpus.
We show that this framework can exploit semantic-frame-related features in the corpus, reveal the underlying semantic structure of the utterances, and boost the performance of the state-of-the-art NLU model with minimal supervision.
arXiv Detail & Related papers (2022-01-25T19:12:44Z) - Probing as Quantifying the Inductive Bias of Pre-trained Representations [99.93552997506438]
We present a novel framework for probing where the goal is to evaluate the inductive bias of representations for a particular task.
We apply our framework to a series of token-, arc-, and sentence-level tasks.
arXiv Detail & Related papers (2021-10-15T22:01:16Z) - Comprehensive Studies for Arbitrary-shape Scene Text Detection [78.50639779134944]
We propose a unified framework for the bottom-up based scene text detection methods.
Under the unified framework, we ensure the consistent settings for non-core modules.
With the comprehensive investigations and elaborate analyses, it reveals the advantages and disadvantages of previous models.
arXiv Detail & Related papers (2021-07-25T13:18:55Z) - Learning structure-aware semantic segmentation with image-level
supervision [36.40302533324508]
We argue that the lost structure information in CAM limits its application in downstream semantic segmentation.
We introduce an auxiliary semantic boundary detection module, which penalizes the deteriorated predictions.
Experimental results on the PASCAL-VOC dataset illustrate the effectiveness of the proposed solution.
arXiv Detail & Related papers (2021-04-15T03:33:20Z) - BERT as a Teacher: Contextual Embeddings for Sequence-Level Reward [23.176481887478634]
We show that the underlying operations, counting words and comparing counts, can be lifted to embedding words and comparing embeddings.
An in-depth analysis of BERT embeddings shows empirically that contextual embeddings can be employed to capture the required dependencies.
We cast unconditional generation as a reinforcement learning problem and show that our reward function indeed provides a more effective learning signal than n-gram reward in this challenging setting.
arXiv Detail & Related papers (2020-03-05T16:06:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.