FrameEOL: Semantic Frame Induction using Causal Language Models
- URL: http://arxiv.org/abs/2510.09097v1
- Date: Fri, 10 Oct 2025 07:52:07 GMT
- Title: FrameEOL: Semantic Frame Induction using Causal Language Models
- Authors: Chihiro Yano, Kosuke Yamada, Hayato Tsukagoshi, Ryohei Sasano, Koichi Takeda,
- Abstract summary: We propose a new method for semantic frame induction based on causal language models (CLMs)<n>We leverage in-context learning (ICL) and deep metric learning (DML) to obtain embeddings more suitable for frame induction.<n> Experimental results on the English and Japanese FrameNet demonstrate that the proposed methods outperform existing frame induction methods.
- Score: 18.542847631796725
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Semantic frame induction is the task of clustering frame-evoking words according to the semantic frames they evoke. In recent years, leveraging embeddings of frame-evoking words that are obtained using masked language models (MLMs) such as BERT has led to high-performance semantic frame induction. Although causal language models (CLMs) such as the GPT and Llama series succeed in a wide range of language comprehension tasks and can engage in dialogue as if they understood frames, they have not yet been applied to semantic frame induction. We propose a new method for semantic frame induction based on CLMs. Specifically, we introduce FrameEOL, a prompt-based method for obtaining Frame Embeddings that outputs One frame-name as a Label representing the given situation. To obtain embeddings more suitable for frame induction, we leverage in-context learning (ICL) and deep metric learning (DML). Frame induction is then performed by clustering the resulting embeddings. Experimental results on the English and Japanese FrameNet datasets demonstrate that the proposed methods outperform existing frame induction methods. In particular, for Japanese, which lacks extensive frame resources, the CLM-based method using only 5 ICL examples achieved comparable performance to the MLM-based method fine-tuned with DML.
Related papers
- Event-Anchored Frame Selection for Effective Long-Video Understanding [67.56884568828508]
Event-Anchored Frame Selection (EFS) is a hierarchical, event-aware pipeline.<n>As a training-free, plug-and-play module, EFS can be seamlessly integrated into off-the-shelf LVLMs.
arXiv Detail & Related papers (2026-03-01T08:25:37Z) - FrameMind: Frame-Interleaved Video Reasoning via Reinforcement Learning [65.42201665046505]
Current video understanding models rely on fixed frame sampling strategies, processing predetermined visual inputs regardless of the specific reasoning requirements of each question.<n>This static approach limits their ability to adaptively gather visual evidence, leading to suboptimal performance on tasks that require broad temporal coverage or fine-grained spatial detail.<n>We introduce FrameMind, an end-to-end framework trained with reinforcement learning that enables models to dynamically request visual information during reasoning through Frame-Interleaved Chain-of-Thought (FiCOT)<n>Unlike traditional approaches, FrameMind operates in multiple turns where the model alternates between textual reasoning and active visual perception, using tools to extract
arXiv Detail & Related papers (2025-09-28T17:59:43Z) - Do LLMs Encode Frame Semantics? Evidence from Frame Identification [4.508786802660182]
We investigate whether large language models encode latent knowledge of frame semantics, focusing on frame identification.<n>We evaluate models under prompt-based inference and observe that they can perform frame identification effectively even without explicit supervision.
arXiv Detail & Related papers (2025-09-23T20:09:32Z) - Exploring In-Context Learning for Frame-Semantic Parsing [0.8192907805418581]
We propose a method that automatically generates task-specific prompts for the Frame Identification (FI) and Frame Semantic Role Labeling (FSRL) subtasks.<n>The method achieves competitive results, with F1 scores of 94.3% for FI and 77.4% for FSRL.<n>The findings suggest that ICL offers a practical and effective alternative to traditional fine-tuning for domain-specific FSP tasks.
arXiv Detail & Related papers (2025-07-30T20:29:17Z) - ULLME: A Unified Framework for Large Language Model Embeddings with Generation-Augmented Learning [72.90823351726374]
We introduce the Unified framework for Large Language Model Embedding (ULLME), a flexible, plug-and-play implementation that enables bidirectional attention across various LLMs.
We also propose Generation-augmented Representation Learning (GRL), a novel fine-tuning method to boost LLMs for text embedding tasks.
To showcase our framework's flexibility and effectiveness, we release three pre-trained models from ULLME with different backbone architectures.
arXiv Detail & Related papers (2024-08-06T18:53:54Z) - RAR: Retrieving And Ranking Augmented MLLMs for Visual Recognition [78.97487780589574]
Multimodal Large Language Models (MLLMs) excel at classifying fine-grained categories.
This paper introduces a Retrieving And Ranking augmented method for MLLMs.
Our proposed approach not only addresses the inherent limitations in fine-grained recognition but also preserves the model's comprehensive knowledge base.
arXiv Detail & Related papers (2024-03-20T17:59:55Z) - Learning Referring Video Object Segmentation from Weak Annotation [78.45828085350936]
Referring video object segmentation (RVOS) is a task that aims to segment the target object in all video frames based on a sentence describing the object.
We propose a new annotation scheme that reduces the annotation effort by 8 times, while providing sufficient supervision for RVOS.
Our scheme only requires a mask for the frame where the object first appears and bounding boxes for the rest of the frames.
arXiv Detail & Related papers (2023-08-04T06:50:52Z) - Acquiring Frame Element Knowledge with Deep Metric Learning for Semantic
Frame Induction [24.486546938073907]
We propose a method that applies deep metric learning to semantic frame induction tasks.
A pre-trained language model is fine-tuned to be suitable for distinguishing frame element roles.
Experimental results on FrameNet demonstrate that our method achieves substantially better performance than existing methods.
arXiv Detail & Related papers (2023-05-23T11:02:28Z) - Semantic Frame Induction with Deep Metric Learning [24.486546938073907]
We propose a model that uses deep metric learning to fine-tune a contextualized embedding model.
We apply the fine-tuned contextualized embeddings to perform semantic frame induction.
arXiv Detail & Related papers (2023-04-27T15:46:09Z) - Knowledge-augmented Frame Semantic Parsing with Hybrid Prompt-tuning [17.6573121083417]
We propose a Knowledge-Augmented Frame Semantic Parsing Architecture (KAF-SPA) to enhance semantic representation.
A Memory-based Knowledge Extraction Module (MKEM) is devised to select accurate frame knowledge and construct the continuous templates.
We also design a Task-oriented Knowledge Probing Module (TKPM) using hybrid prompts to incorporate the selected knowledge into the PLMs and adapt PLMs to the tasks of frame and argument identification.
arXiv Detail & Related papers (2023-03-25T06:41:19Z) - Linguistic Structure Guided Context Modeling for Referring Image
Segmentation [61.701577239317785]
We propose a "gather-propagate-distribute" scheme to model multimodal context by cross-modal interaction.
Our LSCM module builds a Dependency Parsing Tree Word Graph (DPT-WG) which guides all the words to include valid multimodal context of the sentence.
arXiv Detail & Related papers (2020-10-01T16:03:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.