PASTA: A Dataset for Modeling Participant States in Narratives
- URL: http://arxiv.org/abs/2208.00329v2
- Date: Sat, 1 Jul 2023 22:34:52 GMT
- Title: PASTA: A Dataset for Modeling Participant States in Narratives
- Authors: Sayontan Ghosh, Mahnaz Koupaee, Isabella Chen, Francis Ferraro,
Nathanael Chambers, Niranjan Balasubramanian
- Abstract summary: We introduce a new crowdsourced English-language, Participant States dataset, PASTA.
This dataset contains inferable participant states; a counterfactual perturbation to each state; and the changes to the story that would be necessary if the counterfactual were true.
We introduce three state-based reasoning tasks that test for the ability to infer when a state is entailed by a story, to revise a story conditioned on a counterfactual state, and to explain the most likely state change given a revised story.
- Score: 24.982857364049664
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The events in a narrative are understood as a coherent whole via the
underlying states of their participants. Often, these participant states are
not explicitly mentioned, instead left to be inferred by the reader. A model
that understands narratives should likewise infer these implicit states, and
even reason about the impact of changes to these states on the narrative. To
facilitate this goal, we introduce a new crowdsourced English-language,
Participant States dataset, PASTA. This dataset contains inferable participant
states; a counterfactual perturbation to each state; and the changes to the
story that would be necessary if the counterfactual were true. We introduce
three state-based reasoning tasks that test for the ability to infer when a
state is entailed by a story, to revise a story conditioned on a counterfactual
state, and to explain the most likely state change given a revised story.
Experiments show that today's LLMs can reason about states to some degree, but
there is large room for improvement, especially in problems requiring access
and ability to reason with diverse types of knowledge (e.g. physical,
numerical, factual).
Related papers
- Learning Multiple Object States from Actions via Large Language Models [15.053419817253145]
We formulate object state recognition as a multi-label classification task that explicitly handles multiple states.
We generate pseudo-labels from the transcribed narrations, capturing the influence of past states.
Our model trained on LLM-generated pseudo-labels significantly outperforms strong vision-language models.
arXiv Detail & Related papers (2024-05-02T08:43:16Z) - SCHEMA: State CHangEs MAtter for Procedure Planning in Instructional
Videos [54.01116513202433]
We study the problem of procedure planning in instructional videos, which aims to make a goal-oriented sequence of action steps given partial visual state observations.
Recent works succeeded in sequence modeling of steps with only sequence-level annotations accessible during training, which overlooked the roles of states in the procedures.
We aim to establish a more structured state space by investigating the causal relations between steps and states in procedures.
arXiv Detail & Related papers (2024-03-03T19:53:06Z) - OSCaR: Object State Captioning and State Change Representation [52.13461424520107]
This paper introduces the Object State Captioning and State Change Representation (OSCaR) dataset and benchmark.
OSCaR consists of 14,084 annotated video segments with nearly 1,000 unique objects from various egocentric video collections.
It sets a new testbed for evaluating multimodal large language models (MLLMs)
arXiv Detail & Related papers (2024-02-27T01:48:19Z) - Language Modeling with Latent Situations [46.38670628102201]
SituationSupervision is a family of approaches for improving coherence in language models.
It trains models to construct and condition on explicit representations of entities and their states.
It produces major coherence improvements between 4-11%.
arXiv Detail & Related papers (2022-12-20T05:59:42Z) - MoNET: Tackle State Momentum via Noise-Enhanced Training for Dialogue
State Tracking [42.70799541159301]
Dialogue state tracking (DST) aims to convert the dialogue history into dialogue states which consist of slot-value pairs.
The dialogue state in the last turn is typically adopted as the input for predicting the current state by DST models.
We propose MoNET to tackle state momentum via noise-enhanced training.
arXiv Detail & Related papers (2022-11-10T11:55:25Z) - Generating Coherent Narratives by Learning Dynamic and Discrete Entity
States with a Contrastive Framework [68.1678127433077]
We extend the Transformer model to dynamically conduct entity state updates and sentence realization for narrative generation.
Experiments on two narrative datasets show that our model can generate more coherent and diverse narratives than strong baselines.
arXiv Detail & Related papers (2022-08-08T09:02:19Z) - SOCCER: An Information-Sparse Discourse State Tracking Collection in the
Sports Commentary Domain [7.119677737397071]
In pursuit of natural language understanding, there has been a long standing interest in tracking state changes throughout narratives.
This paper proposes to turn to simplified, fully observable systems that show some of these properties: Sports events.
We propose a new task formulation where, given paragraphs of commentary of a game at different timestamps, the system is asked to recognize the occurrence of in-game events.
arXiv Detail & Related papers (2021-06-03T16:21:13Z) - A New Bandit Setting Balancing Information from State Evolution and
Corrupted Context [52.67844649650687]
We propose a new sequential decision-making setting combining key aspects of two established online learning problems with bandit feedback.
The optimal action to play at any given moment is contingent on an underlying changing state which is not directly observable by the agent.
We present an algorithm that uses a referee to dynamically combine the policies of a contextual bandit and a multi-armed bandit.
arXiv Detail & Related papers (2020-11-16T14:35:37Z) - Paragraph-level Commonsense Transformers with Recurrent Memory [77.4133779538797]
We train a discourse-aware model that incorporates paragraph-level information to generate coherent commonsense inferences from narratives.
Our results show that PARA-COMET outperforms the sentence-level baselines, particularly in generating inferences that are both coherent and novel.
arXiv Detail & Related papers (2020-10-04T05:24:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.