Construction and Evaluation of a Self-Attention Model for Semantic
Understanding of Sentence-Final Particles
- URL: http://arxiv.org/abs/2210.00282v1
- Date: Sat, 1 Oct 2022 13:54:54 GMT
- Title: Construction and Evaluation of a Self-Attention Model for Semantic
Understanding of Sentence-Final Particles
- Authors: Shuhei Mandokoro, Natsuki Oka, Akane Matsushima, Chie Fukada, Yuko
Yoshimura, Koji Kawahara and Kazuaki Tanaka
- Abstract summary: Sentence-final particles serve an essential role in spoken Japanese because they express the speaker's mental attitudes toward a proposition and/or an interlocutor.
This paper proposes a self-attention model that takes various subjective senses in addition to language and images as input and learns the relationship between words and subjective senses.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sentence-final particles serve an essential role in spoken Japanese because
they express the speaker's mental attitudes toward a proposition and/or an
interlocutor. They are acquired at early ages and occur very frequently in
everyday conversation. However, there has been little proposal for a
computational model of acquiring sentence-final particles. This paper proposes
Subjective BERT, a self-attention model that takes various subjective senses in
addition to language and images as input and learns the relationship between
words and subjective senses. An evaluation experiment revealed that the model
understands the usage of "yo", which expresses the speaker's intention to
communicate new information, and that of "ne", which denotes the speaker's
desire to confirm that some information is shared.
Related papers
- To Drop or Not to Drop? Predicting Argument Ellipsis Judgments: A Case Study in Japanese [26.659122101710068]
We study whether and why a particular argument should be omitted across over 2,000 data points in the balanced corpus of Japanese.
The data indicate that native speakers overall share common criteria for such judgments.
The gap between the systems' prediction and human judgments in specific linguistic aspects is revealed.
arXiv Detail & Related papers (2024-04-17T12:26:52Z) - MindDial: Belief Dynamics Tracking with Theory-of-Mind Modeling for Situated Neural Dialogue Generation [62.44907105496227]
MindDial is a novel conversational framework that can generate situated free-form responses with theory-of-mind modeling.
We introduce an explicit mind module that can track the speaker's belief and the speaker's prediction of the listener's belief.
Our framework is applied to both prompting and fine-tuning-based models, and is evaluated across scenarios involving both common ground alignment and negotiation.
arXiv Detail & Related papers (2023-06-27T07:24:32Z) - NOPE: A Corpus of Naturally-Occurring Presuppositions in English [33.69537711677911]
We introduce the Naturally-Occurring Presuppositions in English (NOPE) Corpus.
We investigate the context-sensitivity of 10 different types of presupposition triggers.
We evaluate machine learning models' ability to predict human inferences.
arXiv Detail & Related papers (2021-09-14T22:03:23Z) - Extending rational models of communication from beliefs to actions [10.169856458866088]
Speakers communicate to influence their partner's beliefs and shape their actions.
We develop three speaker models: a belief-oriented speaker with a purely informative objective; an action-oriented speaker with an instrumental objective; and a combined speaker which integrates the two.
We show that grounding production choices in future listener actions results in relevance effects and flexible uses of nonliteral language.
arXiv Detail & Related papers (2021-05-25T13:58:01Z) - Who Responded to Whom: The Joint Effects of Latent Topics and Discourse
in Conversation Structure [53.77234444565652]
We identify the responding relations in the conversation discourse, which link response utterances to their initiations.
We propose a model to learn latent topics and discourse in word distributions, and predict pairwise initiation-response links.
Experimental results on both English and Chinese conversations show that our model significantly outperforms the previous state of the arts.
arXiv Detail & Related papers (2021-04-17T17:46:00Z) - Did they answer? Subjective acts and intents in conversational discourse [48.63528550837949]
We present the first discourse dataset with multiple and subjective interpretations of English conversation.
We show disagreements are nuanced and require a deeper understanding of the different contextual factors.
arXiv Detail & Related papers (2021-04-09T16:34:19Z) - I Beg to Differ: A study of constructive disagreement in online
conversations [15.581515781839656]
We construct a corpus of 7 425 Wikipedia Talk page conversations that contain content disputes.
We define the task of predicting whether disagreements will be escalated to mediation by a moderator.
We develop a variety of neural models and show that taking into account the structure of the conversation improves predictive accuracy.
arXiv Detail & Related papers (2021-01-26T16:36:43Z) - Filling the Gap of Utterance-aware and Speaker-aware Representation for
Multi-turn Dialogue [76.88174667929665]
A multi-turn dialogue is composed of multiple utterances from two or more different speaker roles.
In the existing retrieval-based multi-turn dialogue modeling, the pre-trained language models (PrLMs) as encoder represent the dialogues coarsely.
We propose a novel model to fill such a gap by modeling the effective utterance-aware and speaker-aware representations entailed in a dialogue history.
arXiv Detail & Related papers (2020-09-14T15:07:19Z) - Improving Factual Consistency Between a Response and Persona Facts [64.30785349238619]
Neural models for response generation produce responses that are semantically plausible but not necessarily factually consistent with facts describing the speaker's persona.
We propose to fine-tune these models by reinforcement learning and an efficient reward function that explicitly captures the consistency between a response and persona facts as well as semantic plausibility.
arXiv Detail & Related papers (2020-04-30T18:08:22Z) - You Impress Me: Dialogue Generation via Mutual Persona Perception [62.89449096369027]
The research in cognitive science suggests that understanding is an essential signal for a high-quality chit-chat conversation.
Motivated by this, we propose P2 Bot, a transmitter-receiver based framework with the aim of explicitly modeling understanding.
arXiv Detail & Related papers (2020-04-11T12:51:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.