M2H2: A Multimodal Multiparty Hindi Dataset For Humor Recognition in
Conversations
- URL: http://arxiv.org/abs/2108.01260v1
- Date: Tue, 3 Aug 2021 02:54:09 GMT
- Title: M2H2: A Multimodal Multiparty Hindi Dataset For Humor Recognition in
Conversations
- Authors: Dushyant Singh Chauhan, Gopendra Vikram Singh, Navonil Majumder, Amir
Zadeh, Asif Ekbal, Pushpak Bhattacharyya, Louis-philippe Morency, and
Soujanya Poria
- Abstract summary: We propose a dataset for Multimodal Multiparty Hindi Humor (M2H2) recognition in conversations containing 6,191 utterances from 13 episodes of a very popular TV series "Shrimaan Shrimati Phir Se"
Each utterance is annotated with humor/non-humor labels and encompasses acoustic, visual, and textual modalities.
The empirical results on M2H2 dataset demonstrate that multimodal information complements unimodal information for humor recognition.
- Score: 72.81164101048181
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Humor recognition in conversations is a challenging task that has recently
gained popularity due to its importance in dialogue understanding, including in
multimodal settings (i.e., text, acoustics, and visual). The few existing
datasets for humor are mostly in English. However, due to the tremendous growth
in multilingual content, there is a great demand to build models and systems
that support multilingual information access. To this end, we propose a dataset
for Multimodal Multiparty Hindi Humor (M2H2) recognition in conversations
containing 6,191 utterances from 13 episodes of a very popular TV series
"Shrimaan Shrimati Phir Se". Each utterance is annotated with humor/non-humor
labels and encompasses acoustic, visual, and textual modalities. We propose
several strong multimodal baselines and show the importance of contextual and
multimodal information for humor recognition in conversations. The empirical
results on M2H2 dataset demonstrate that multimodal information complements
unimodal information for humor recognition. The dataset and the baselines are
available at http://www.iitp.ac.in/~ai-nlp-ml/resources.html and
https://github.com/declare-lab/M2H2-dataset.
Related papers
- M2DS: Multilingual Dataset for Multi-document Summarisation [0.5071800070021028]
Multi-document Summarisation (MDS) has resulted in diverse datasets covering customer reviews, academic papers, medical and legal documents, and news articles.
However, the English-centric nature of these datasets has created a conspicuous void for multilingual datasets in today's globalised digital landscape.
This paper introduces M2DS, emphasising its unique multilingual aspect, and includes baseline scores from state-of-the-art MDS models evaluated on our dataset.
arXiv Detail & Related papers (2024-07-17T06:25:51Z) - Emotion and Intent Joint Understanding in Multimodal Conversation: A Benchmarking Dataset [74.74686464187474]
Emotion and Intent Joint Understanding in Multimodal Conversation (MC-EIU) aims to decode the semantic information manifested in a multimodal conversational history.
MC-EIU is enabling technology for many human-computer interfaces.
We propose an MC-EIU dataset, which features 7 emotion categories, 9 intent categories, 3 modalities, i.e., textual, acoustic, and visual content, and two languages, English and Mandarin.
arXiv Detail & Related papers (2024-07-03T01:56:00Z) - Multi3WOZ: A Multilingual, Multi-Domain, Multi-Parallel Dataset for
Training and Evaluating Culturally Adapted Task-Oriented Dialog Systems [64.40789703661987]
Multi3WOZ is a novel multilingual, multi-domain, multi-parallel ToD dataset.
It is large-scale and offers culturally adapted dialogs in 4 languages.
We describe a complex bottom-up data collection process that yielded the final dataset.
arXiv Detail & Related papers (2023-07-26T08:29:42Z) - MMT: A Multilingual and Multi-Topic Indian Social Media Dataset [1.0413233169366503]
Social media plays a significant role in cross-cultural communication.
A vast amount of this occurs in code-mixed and multilingual form.
We introduce a large-scale multilingual, and multi-topic dataset (MMT) collected from Twitter.
arXiv Detail & Related papers (2023-04-02T21:39:00Z) - TextMI: Textualize Multimodal Information for Integrating Non-verbal
Cues in Pre-trained Language Models [5.668457303716451]
We propose TextMI as a general, competitive baseline for multimodal behavioral analysis tasks.
Our approach significantly reduces model complexity, adds interpretability to the model's decision, and can be applied for a diverse set of tasks.
arXiv Detail & Related papers (2023-03-27T17:54:32Z) - TikTalk: A Video-Based Dialogue Dataset for Multi-Modal Chitchat in Real
World [97.58623810402563]
We introduce a new video-based multi-modal dialogue dataset, called TikTalk.
We collect 38K videos from a popular video-sharing platform, along with 367K conversations posted by users beneath them.
Users engage in spontaneous conversations based on their multi-modal experiences from watching videos, which helps recreate real-world chitchat context.
arXiv Detail & Related papers (2023-01-14T10:18:22Z) - LVP-M3: Language-aware Visual Prompt for Multilingual Multimodal Machine
Translation [94.33019040320507]
Multimodal Machine Translation (MMT) focuses on enhancing text-only translation with visual features.
Recent advances still struggle to train a separate model for each language pair, which is costly and unaffordable when the number of languages increases.
We propose the Multilingual MMT task by establishing two new Multilingual MMT benchmark datasets covering seven languages.
arXiv Detail & Related papers (2022-10-19T12:21:39Z) - M3ED: Multi-modal Multi-scene Multi-label Emotional Dialogue Database [139.08528216461502]
We propose a Multi-modal Multi-scene Multi-label Emotional Dialogue dataset, M3ED.
M3ED contains 990 dyadic emotional dialogues from 56 different TV series, a total of 9,082 turns and 24,449 utterances.
To the best of our knowledge, M3ED is the first multimodal emotional dialogue dataset in Chinese.
arXiv Detail & Related papers (2022-05-09T06:52:51Z) - Multi-modal Sarcasm Detection and Humor Classification in Code-mixed
Conversations [14.852199996061287]
We develop a Hindi-English code-mixed dataset, MaSaC, for the multi-modal sarcasm detection and humor classification in conversational dialog.
We propose MSH-COMICS, a novel attention-rich neural architecture for the utterance classification.
arXiv Detail & Related papers (2021-05-20T18:33:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.