DiCoRe: Enhancing Zero-shot Event Detection via Divergent-Convergent LLM Reasoning
- URL: http://arxiv.org/abs/2506.05128v1
- Date: Thu, 05 Jun 2025 15:16:14 GMT
- Title: DiCoRe: Enhancing Zero-shot Event Detection via Divergent-Convergent LLM Reasoning
- Authors: Tanmay Parekh, Kartik Mehta, Ninareh Mehrabi, Kai-Wei Chang, Nanyun Peng,
- Abstract summary: DiCoRe is a divergent-convergent reasoning framework that decouples the task of ED using Dreamer and Grounder.<n>We demonstrate how DiCoRe consistently outperforms prior zero-shot, transfer-learning, and reasoning baselines.
- Score: 89.44918500792394
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Zero-shot Event Detection (ED), the task of identifying event mentions in natural language text without any training data, is critical for document understanding in specialized domains. Understanding the complex event ontology, extracting domain-specific triggers from the passage, and structuring them appropriately overloads and limits the utility of Large Language Models (LLMs) for zero-shot ED. To this end, we propose DiCoRe, a divergent-convergent reasoning framework that decouples the task of ED using Dreamer and Grounder. Dreamer encourages divergent reasoning through open-ended event discovery, which helps to boost event coverage. Conversely, Grounder introduces convergent reasoning to align the free-form predictions with the task-specific instructions using finite-state machine guided constrained decoding. Additionally, an LLM-Judge verifies the final outputs to ensure high precision. Through extensive experiments on six datasets across five domains and nine LLMs, we demonstrate how DiCoRe consistently outperforms prior zero-shot, transfer-learning, and reasoning baselines, achieving 4-7% average F1 gains over the best baseline -- establishing DiCoRe as a strong zero-shot ED framework.
Related papers
- PrismRAG: Boosting RAG Factuality with Distractor Resilience and Strategized Reasoning [57.89188317734747]
PrismRAG trains the model with distractor-aware QA pairs mixing gold evidence with subtle distractor passages.<n>It instills reasoning-centric habits that make the LLM plan, rationalize, and synthesize without relying on extensive human engineered instructions.
arXiv Detail & Related papers (2025-07-25T00:15:31Z) - ESTR-CoT: Towards Explainable and Accurate Event Stream based Scene Text Recognition with Chain-of-Thought Reasoning [57.767536707234036]
We propose a novel chain-of-thought reasoning based event stream scene text recognition framework, termed ESTR-CoT.<n>Specifically, we first adopt the vision encoder EVA-CLIP to transform the input event stream into tokens and utilize a Llama tokenizer to encode the given generation prompt.<n>A Q-former is used to align the vision token to the pre-trained large language model Vicuna-7B and output both the answer and chain-of-thought (CoT) reasoning process simultaneously.
arXiv Detail & Related papers (2025-07-02T23:41:31Z) - The First Few Tokens Are All You Need: An Efficient and Effective Unsupervised Prefix Fine-Tuning Method for Reasoning Models [69.798277882245]
We introduce Unsupervised Prefix Fine-Tuning (UPFT) to enhance large language models' reasoning efficiency.<n>UPFT removes the need for labeled data or exhaustive sampling.<n> Experiments show that UPFT matches the performance of supervised methods.
arXiv Detail & Related papers (2025-03-04T18:56:03Z) - From Objects to Events: Unlocking Complex Visual Understanding in Object Detectors via LLM-guided Symbolic Reasoning [71.41062111470414]
The proposed plug-and-play framework interfaces with any open-vocabulary detector.<n>At its core, our approach combines (i) a symbolic regression mechanism exploring relationship patterns among detected entities.<n>We compared our training-free framework against specialized event recognition systems across diverse application domains.
arXiv Detail & Related papers (2025-02-09T10:30:54Z) - GIVE: Structured Reasoning of Large Language Models with Knowledge Graph Inspired Veracity Extrapolation [108.2008975785364]
Graph Inspired Veracity Extrapolation (GIVE) is a novel reasoning method that merges parametric and non-parametric memories to improve accurate reasoning with minimal external input.<n>GIVE guides the LLM agent to select the most pertinent expert data (observe), engage in query-specific divergent thinking (reflect), and then synthesize this information to produce the final output (speak)
arXiv Detail & Related papers (2024-10-11T03:05:06Z) - Enhancing Event Reasoning in Large Language Models through Instruction Fine-Tuning with Semantic Causal Graphs [7.9482258228279195]
We propose a novel approach for instruction fine-tuning LLMs for event detection.
Our method introduces Semantic Causal Graphs (SCGs) to capture both causal relationships and contextual information within text.
Our evaluations demonstrate that training LLMs with SCG Instructions outperforms standard instruction fine-tuning by an average of 35.69% on Event Trigger Classification.
arXiv Detail & Related papers (2024-08-30T18:56:06Z) - EVIT: Event-Oriented Instruction Tuning for Event Reasoning [18.012724531672813]
Event reasoning aims to infer events according to certain relations and predict future events.
Large language models (LLMs) have made significant advancements in event reasoning owing to their wealth of knowledge and reasoning capabilities.
However, smaller instruction-tuned models currently in use do not consistently demonstrate exceptional proficiency in managing these tasks.
arXiv Detail & Related papers (2024-04-18T08:14:53Z) - Identifying Factual Inconsistencies in Summaries: Grounding LLM Inference via Task Taxonomy [48.29181662640212]
Factual inconsistencies pose a significant hurdle for the faithful summarization by generative models.
We consolidate key error types of inconsistent facts in summaries, and incorporate them to facilitate both the zero-shot and supervised paradigms of LLMs.
arXiv Detail & Related papers (2024-02-20T08:41:23Z) - MsPrompt: Multi-step Prompt Learning for Debiasing Few-shot Event
Detection [16.98619925632727]
Event detection (ED) aims to identify the key trigger words in unstructured text and predict the event types accordingly.
Traditional ED models are too data-hungry to accommodate real applications with scarce labeled data.
We propose a multi-step prompt learning model (MsPrompt) for debiasing few-shot event detection.
arXiv Detail & Related papers (2023-05-16T10:19:12Z) - Trigger-free Event Detection via Derangement Reading Comprehension [4.728684358207039]
Event detection aims to detect events from texts and categorize them.
We propose a novel trigger-free ED method via Derangement mechanism on a machine Reading (DRC) framework.
We show that our proposed trigger-free ED model is remarkably competitive to mainstream trigger-based models.
arXiv Detail & Related papers (2022-08-20T11:01:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.