NormMark: A Weakly Supervised Markov Model for Socio-cultural Norm
Discovery
- URL: http://arxiv.org/abs/2305.16598v1
- Date: Fri, 26 May 2023 03:03:37 GMT
- Title: NormMark: A Weakly Supervised Markov Model for Socio-cultural Norm
Discovery
- Authors: Farhad Moghimifar, Shilin Qu, Tongtong Wu, Yuan-Fang Li, Gholamreza
Haffari
- Abstract summary: Existing methods for norm recognition tend to focus only on surface-level features of dialogues.
We propose NormMark, a probabilistic generative Markov model to carry the latent features throughout a dialogue.
We show that our approach achieves higher F1 score, outperforming current state-of-the-art methods, including GPT3.
- Score: 46.16583206206433
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Norms, which are culturally accepted guidelines for behaviours, can be
integrated into conversational models to generate utterances that are
appropriate for the socio-cultural context. Existing methods for norm
recognition tend to focus only on surface-level features of dialogues and do
not take into account the interactions within a conversation. To address this
issue, we propose NormMark, a probabilistic generative Markov model to carry
the latent features throughout a dialogue. These features are captured by
discrete and continuous latent variables conditioned on the conversation
history, and improve the model's ability in norm recognition. The model is
trainable on weakly annotated data using the variational technique. On a
dataset with limited norm annotations, we show that our approach achieves
higher F1 score, outperforming current state-of-the-art methods, including
GPT3.
Related papers
- MITA: Bridging the Gap between Model and Data for Test-time Adaptation [68.62509948690698]
Test-Time Adaptation (TTA) has emerged as a promising paradigm for enhancing the generalizability of models.
We propose Meet-In-The-Middle based MITA, which introduces energy-based optimization to encourage mutual adaptation of the model and data from opposing directions.
arXiv Detail & Related papers (2024-10-12T07:02:33Z) - Question Generation in Knowledge-Driven Dialog: Explainability and Evaluation [5.598219914856067]
We present a model which instead of directly generating a question, sequentially predicts first a fact then a question.
We evaluate our approach on 37k test dialogs adapted from the KGConv dataset.
arXiv Detail & Related papers (2024-04-11T15:24:50Z) - Quantifying the Plausibility of Context Reliance in Neural Machine
Translation [25.29330352252055]
We introduce Plausibility Evaluation of Context Reliance (PECoRe)
PECoRe is an end-to-end interpretability framework designed to quantify context usage in language models' generations.
We use pecore to quantify the plausibility of context-aware machine translation models.
arXiv Detail & Related papers (2023-10-02T13:26:43Z) - Uncertainty Estimation of Transformers' Predictions via Topological Analysis of the Attention Matrices [3.1466086042810884]
Transformer-based language models have set new benchmarks across a wide range of NLP tasks.
reliably estimating the uncertainty of their predictions remains a significant challenge.
We tackle these limitations by harnessing the geometry of attention maps across multiple heads and layers to assess model confidence.
Our method significantly outperforms existing uncertainty estimation techniques on benchmarks for acceptability judgments and artificial text detection.
arXiv Detail & Related papers (2023-08-22T09:17:45Z) - NormSAGE: Multi-Lingual Multi-Cultural Norm Discovery from Conversations
On-the-Fly [61.77957329364812]
We introduce a framework for addressing the novel task of conversation-grounded multi-lingual, multi-cultural norm discovery.
NormSAGE elicits knowledge about norms through directed questions representing the norm discovery task and conversation context.
It further addresses the risk of language model hallucination with a self-verification mechanism ensuring that the norms discovered are correct.
arXiv Detail & Related papers (2022-10-16T18:30:05Z) - Advanced Conditional Variational Autoencoders (A-CVAE): Towards
interpreting open-domain conversation generation via disentangling latent
feature representation [15.742077523458995]
This paper proposes to harness the generative model with a priori knowledge through a cognitive approach involving mesoscopic scale feature disentanglement.
We propose a new metric for open-domain dialogues, which can objectively evaluate the interpretability of the latent space distribution.
arXiv Detail & Related papers (2022-07-26T07:39:36Z) - Stylized Knowledge-Grounded Dialogue Generation via Disentangled
Template Rewriting [55.10977824136768]
We study a new problem: Stylized Knowledge-Grounded Dialogue Generation.
It presents two challenges: How to train a SKDG model where no context, knowledge, stylized response> triples are available.
We propose a novel disentangled template rewriting (DTR) method which generates responses via combing disentangled style templates and content templates.
arXiv Detail & Related papers (2022-04-12T08:17:21Z) - On the Use of Linguistic Features for the Evaluation of Generative
Dialogue Systems [17.749995931459136]
We propose that a metric based on linguistic features may be able to maintain good correlation with human judgment and be interpretable.
To support this proposition, we measure and analyze various linguistic features on dialogues produced by multiple dialogue models.
We find that the features' behaviour is consistent with the known properties of the models tested, and is similar across domains.
arXiv Detail & Related papers (2021-04-13T16:28:00Z) - I like fish, especially dolphins: Addressing Contradictions in Dialogue
Modeling [104.09033240889106]
We introduce the DialoguE COntradiction DEtection task (DECODE) and a new conversational dataset containing both human-human and human-bot contradictory dialogues.
We then compare a structured utterance-based approach of using pre-trained Transformer models for contradiction detection with the typical unstructured approach.
arXiv Detail & Related papers (2020-12-24T18:47:49Z) - Modeling Topical Relevance for Multi-Turn Dialogue Generation [61.87165077442267]
We propose a new model, named STAR-BTM, to tackle the problem of topic drift in multi-turn dialogue.
The Biterm Topic Model is pre-trained on the whole training dataset. Then, the topic level attention weights are computed based on the topic representation of each context.
Experimental results on both Chinese customer services data and English Ubuntu dialogue data show that STAR-BTM significantly outperforms several state-of-the-art methods.
arXiv Detail & Related papers (2020-09-27T03:33:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.