Investigating Robustness of Dialog Models to Popular Figurative Language
Constructs
- URL: http://arxiv.org/abs/2110.00687v1
- Date: Fri, 1 Oct 2021 23:55:16 GMT
- Title: Investigating Robustness of Dialog Models to Popular Figurative Language
Constructs
- Authors: Harsh Jhamtani, Varun Gangal, Eduard Hovy and Taylor Berg-Kirkpatrick
- Abstract summary: We analyze the performance of existing dialog models in situations where the input dialog context exhibits use of figurative language.
We propose lightweight solutions to help existing models become more robust to figurative language.
- Score: 30.841109045790862
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Humans often employ figurative language use in communication, including
during interactions with dialog systems. Thus, it is important for real-world
dialog systems to be able to handle popular figurative language constructs like
metaphor and simile. In this work, we analyze the performance of existing
dialog models in situations where the input dialog context exhibits use of
figurative language. We observe large gaps in handling of figurative language
when evaluating the models on two open domain dialog datasets. When faced with
dialog contexts consisting of figurative language, some models show very large
drops in performance compared to contexts without figurative language. We
encourage future research in dialog modeling to separately analyze and report
results on figurative language in order to better test model capabilities
relevant to real-world use. Finally, we propose lightweight solutions to help
existing models become more robust to figurative language by simply using an
external resource to translate figurative language to literal (non-figurative)
forms while preserving the meaning to the best extent possible.
Related papers
- SPECTRUM: Speaker-Enhanced Pre-Training for Long Dialogue Summarization [48.284512017469524]
Multi-turn dialogues are characterized by their extended length and the presence of turn-taking conversations.
Traditional language models often overlook the distinct features of these dialogues by treating them as regular text.
We propose a speaker-enhanced pre-training method for long dialogue summarization.
arXiv Detail & Related papers (2024-01-31T04:50:00Z) - DiPlomat: A Dialogue Dataset for Situated Pragmatic Reasoning [89.92601337474954]
Pragmatic reasoning plays a pivotal role in deciphering implicit meanings that frequently arise in real-life conversations.
We introduce a novel challenge, DiPlomat, aiming at benchmarking machines' capabilities on pragmatic reasoning and situated conversational understanding.
arXiv Detail & Related papers (2023-06-15T10:41:23Z) - LMs stand their Ground: Investigating the Effect of Embodiment in
Figurative Language Interpretation by Language Models [0.0]
Figurative language is a challenge for language models since its interpretation deviates from their conventional order and meaning.
Yet, humans can easily understand and interpret metaphors as they can be derived from embodied metaphors.
This study shows how larger language models perform better at interpreting metaphoric sentences when the action of the metaphorical sentence is more embodied.
arXiv Detail & Related papers (2023-05-05T11:44:12Z) - IRFL: Image Recognition of Figurative Language [20.472997304393413]
Figurative forms are often conveyed through multiple modalities (e.g., both text and images)
We develop the Image Recognition of Figurative Language dataset.
We introduce two novel tasks as a benchmark for multimodal figurative language understanding.
arXiv Detail & Related papers (2023-03-27T17:59:55Z) - VD-PCR: Improving Visual Dialog with Pronoun Coreference Resolution [79.05412803762528]
The visual dialog task requires an AI agent to interact with humans in multi-round dialogs based on a visual environment.
We propose VD-PCR, a novel framework to improve Visual Dialog understanding with Pronoun Coreference Resolution.
With the proposed implicit and explicit methods, VD-PCR achieves state-of-the-art experimental results on the VisDial dataset.
arXiv Detail & Related papers (2022-05-29T15:29:50Z) - Testing the Ability of Language Models to Interpret Figurative Language [69.59943454934799]
Figurative and metaphorical language are commonplace in discourse.
It remains an open question to what extent modern language models can interpret nonliteral phrases.
We introduce Fig-QA, a Winograd-style nonliteral language understanding task.
arXiv Detail & Related papers (2022-04-26T23:42:22Z) - Modeling Coreference Relations in Visual Dialog [18.926582410644375]
The occurrences of coreference relations in the dialog makes it a more challenging task than visual question-answering.
We propose two soft constraints that can improve the model's ability of resolving coreferences in dialog in an unsupervised way.
arXiv Detail & Related papers (2022-03-06T15:22:24Z) - It's not Rocket Science : Interpreting Figurative Language in Narratives [48.84507467131819]
We study the interpretation of two non-compositional figurative languages (idioms and similes)
Our experiments show that models based solely on pre-trained language models perform substantially worse than humans on these tasks.
We additionally propose knowledge-enhanced models, adopting human strategies for interpreting figurative language.
arXiv Detail & Related papers (2021-08-31T21:46:35Z) - Probing Contextual Language Models for Common Ground with Visual
Representations [76.05769268286038]
We design a probing model that evaluates how effective are text-only representations in distinguishing between matching and non-matching visual representations.
Our findings show that language representations alone provide a strong signal for retrieving image patches from the correct object categories.
Visually grounded language models slightly outperform text-only language models in instance retrieval, but greatly under-perform humans.
arXiv Detail & Related papers (2020-05-01T21:28:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.