Multiview Contextual Commonsense Inference: A New Dataset and Task
- URL: http://arxiv.org/abs/2210.02890v1
- Date: Thu, 6 Oct 2022 13:08:41 GMT
- Title: Multiview Contextual Commonsense Inference: A New Dataset and Task
- Authors: Siqi Shen, Deepanway Ghosal, Navonil Majumder, Henry Lim, Rada
Mihalcea, Soujanya Poria
- Abstract summary: CICEROv2 is a dataset consisting of 8,351 instances from 2,379 dialogues.
It contains multiple human-written answers for each contextual commonsense inference question.
We show that the inferences in CICEROv2 are more semantically diverse than other contextual commonsense inference datasets.
- Score: 40.566530682082714
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Contextual commonsense inference is the task of generating various types of
explanations around the events in a dyadic dialogue, including cause,
motivation, emotional reaction, and others. Producing a coherent and
non-trivial explanation requires awareness of the dialogue's structure and of
how an event is grounded in the context. In this work, we create CICEROv2, a
dataset consisting of 8,351 instances from 2,379 dialogues, containing multiple
human-written answers for each contextual commonsense inference question,
representing a type of explanation on cause, subsequent event, motivation, and
emotional reaction. We show that the inferences in CICEROv2 are more
semantically diverse than other contextual commonsense inference datasets. To
solve the inference task, we propose a collection of pre-training objectives,
including concept denoising and utterance sorting to prepare a pre-trained
model for the downstream contextual commonsense inference task. Our results
show that the proposed pre-training objectives are effective at adapting the
pre-trained T5-Large model for the contextual commonsense inference task.
Related papers
- Can Language Models Take A Hint? Prompting for Controllable Contextualized Commonsense Inference [12.941933077524919]
We introduce "hinting," a data augmentation technique that enhances contextualized commonsense inference.
"Hinting" employs a prefix prompting strategy using both hard and soft prompts to guide the inference process.
Our results show that "hinting" does not compromise the performance of contextual commonsense inference while offering improved controllability.
arXiv Detail & Related papers (2024-10-03T04:32:46Z) - DiPlomat: A Dialogue Dataset for Situated Pragmatic Reasoning [89.92601337474954]
Pragmatic reasoning plays a pivotal role in deciphering implicit meanings that frequently arise in real-life conversations.
We introduce a novel challenge, DiPlomat, aiming at benchmarking machines' capabilities on pragmatic reasoning and situated conversational understanding.
arXiv Detail & Related papers (2023-06-15T10:41:23Z) - Pre-training Multi-party Dialogue Models with Latent Discourse Inference [85.9683181507206]
We pre-train a model that understands the discourse structure of multi-party dialogues, namely, to whom each utterance is replying.
To fully utilize the unlabeled data, we propose to treat the discourse structures as latent variables, then jointly infer them and pre-train the discourse-aware model.
arXiv Detail & Related papers (2023-05-24T14:06:27Z) - COMET-M: Reasoning about Multiple Events in Complex Sentences [14.644677930985816]
We propose COMET-M (Multi-Event), an event-centric commonsense model capable of generating commonsense inferences for a target event within a complex sentence.
COMET-M builds upon COMET, which excels at generating event-centric inferences for simple sentences, but struggles with the complexity of multi-event sentences prevalent in natural text.
arXiv Detail & Related papers (2023-05-24T01:35:01Z) - UniDU: Towards A Unified Generative Dialogue Understanding Framework [62.8474841241855]
We investigate a unified generative dialogue understanding framework, namely UniDU, to achieve information exchange among DU tasks.
We conduct experiments on ten dialogue understanding datasets, which span five fundamental tasks.
The proposed UniDU framework outperforms task-specific well-designed methods on all 5 tasks.
arXiv Detail & Related papers (2022-04-10T09:32:34Z) - Topic-Aware Contrastive Learning for Abstractive Dialogue Summarization [41.75442239197745]
This work proposes two topic-aware contrastive learning objectives, namely coherence detection and sub-summary generation objectives.
Experiments on benchmark datasets demonstrate that the proposed simple method significantly outperforms strong baselines.
arXiv Detail & Related papers (2021-09-10T17:03:25Z) - Probing Task-Oriented Dialogue Representation from Language Models [106.02947285212132]
This paper investigates pre-trained language models to find out which model intrinsically carries the most informative representation for task-oriented dialogue tasks.
We fine-tune a feed-forward layer as the classifier probe on top of a fixed pre-trained language model with annotated labels in a supervised way.
arXiv Detail & Related papers (2020-10-26T21:34:39Z) - Fact-based Dialogue Generation with Convergent and Divergent Decoding [2.28438857884398]
This paper proposes an end-to-end fact-based dialogue system augmented with the ability of convergent and divergent thinking.
Our model incorporates a novel convergent and divergent decoding that can generate informative and diverse responses.
arXiv Detail & Related papers (2020-05-06T23:49:35Z) - Masking Orchestration: Multi-task Pretraining for Multi-role Dialogue
Representation Learning [50.5572111079898]
Multi-role dialogue understanding comprises a wide range of diverse tasks such as question answering, act classification, dialogue summarization etc.
While dialogue corpora are abundantly available, labeled data, for specific learning tasks, can be highly scarce and expensive.
In this work, we investigate dialogue context representation learning with various types unsupervised pretraining tasks.
arXiv Detail & Related papers (2020-02-27T04:36:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.