"Judge me by my size (noun), do you?'' YodaLib: A Demographic-Aware
Humor Generation Framework
- URL: http://arxiv.org/abs/2006.00578v1
- Date: Sun, 31 May 2020 18:11:52 GMT
- Title: "Judge me by my size (noun), do you?'' YodaLib: A Demographic-Aware
Humor Generation Framework
- Authors: Aparna Garimella, Carmen Banea, Nabil Hossain, Rada Mihalcea
- Abstract summary: We propose an automatic humor generation framework for Mad Libs stories, while accounting for the demographic backgrounds of the desired audience.
We build upon the BERT platform to predict location-biased word fillings in incomplete sentences, and we fine tune BERT to classify location-specific humor in a sentence.
We leverage these components to produce YodaLib, a fully-automated Mad Libs style humor generation framework, which selects and ranks appropriate candidate words and sentences.
- Score: 31.115389392654492
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The subjective nature of humor makes computerized humor generation a
challenging task. We propose an automatic humor generation framework for
filling the blanks in Mad Libs stories, while accounting for the demographic
backgrounds of the desired audience. We collect a dataset consisting of such
stories, which are filled in and judged by carefully selected workers on Amazon
Mechanical Turk. We build upon the BERT platform to predict location-biased
word fillings in incomplete sentences, and we fine tune BERT to classify
location-specific humor in a sentence. We leverage these components to produce
YodaLib, a fully-automated Mad Libs style humor generation framework, which
selects and ranks appropriate candidate words and sentences in order to
generate a coherent and funny story tailored to certain demographics. Our
experimental results indicate that YodaLib outperforms a previous
semi-automated approach proposed for this task, while also surpassing human
annotators in both qualitative and quantitative analyses.
Related papers
- Can Pre-trained Language Models Understand Chinese Humor? [74.96509580592004]
This paper is the first work that systematically investigates the humor understanding ability of pre-trained language models (PLMs)
We construct a comprehensive Chinese humor dataset, which can fully meet all the data requirements of the proposed evaluation framework.
Our empirical study on the Chinese humor dataset yields some valuable observations, which are of great guiding value for future optimization of PLMs in humor understanding and generation.
arXiv Detail & Related papers (2024-07-04T18:13:38Z) - Getting Serious about Humor: Crafting Humor Datasets with Unfunny Large Language Models [27.936545041302377]
Large language models (LLMs) can generate synthetic data for humor detection via editing texts.
We benchmark LLMs on an existing human dataset and show that current LLMs display an impressive ability to 'unfun' jokes.
We extend our approach to a code-mixed English-Hindi humor dataset, where we find that GPT-4's synthetic data is highly rated by bilingual annotators.
arXiv Detail & Related papers (2024-02-23T02:58:12Z) - ExPUNations: Augmenting Puns with Keywords and Explanations [88.58174386894913]
We augment an existing dataset of puns with detailed crowdsourced annotations of keywords.
This is the first humor dataset with such extensive and fine-grained annotations specifically for puns.
We propose two tasks: explanation generation to aid with pun classification and keyword-conditioned pun generation.
arXiv Detail & Related papers (2022-10-24T18:12:02Z) - Towards Multimodal Prediction of Spontaneous Humour: A Novel Dataset and First Results [84.37263300062597]
Humor is a substantial element of human social behavior, affect, and cognition.
Current methods of humor detection have been exclusively based on staged data, making them inadequate for "real-world" applications.
We contribute to addressing this deficiency by introducing the novel Passau-Spontaneous Football Coach Humor dataset, comprising about 11 hours of recordings.
arXiv Detail & Related papers (2022-09-28T17:36:47Z) - DeHumor: Visual Analytics for Decomposing Humor [36.300283476950796]
We develop DeHumor, a visual system for analyzing humorous behaviors in public speaking.
To intuitively reveal the building blocks of each concrete example, DeHumor decomposes each humorous video into multimodal features.
We show that DeHumor is able to highlight various building blocks of humor examples.
arXiv Detail & Related papers (2021-07-18T04:01:07Z) - Predicting the Humorousness of Tweets Using Gaussian Process Preference
Learning [56.18809963342249]
We present a probabilistic approach that learns to rank and rate the humorousness of short texts by exploiting human preference judgments and automatically sourced linguistic annotations.
We report system performance for the campaign's two subtasks, humour detection and funniness score prediction, and discuss some issues arising from the conversion between the numeric scores used in the HAHA@IberLEF 2019 data and the pairwise judgment annotations required for our method.
arXiv Detail & Related papers (2020-08-03T13:05:42Z) - Enabling Language Models to Fill in the Blanks [81.59381915581892]
We present a simple approach for text infilling, the task of predicting missing spans of text at any position in a document.
We train (or fine-tune) off-the-shelf language models on sequences containing the concatenation of artificially-masked text and the text which was masked.
We show that this approach, which we call infilling by language modeling, can enable LMs to infill entire sentences effectively on three different domains: short stories, scientific abstracts, and lyrics.
arXiv Detail & Related papers (2020-05-11T18:00:03Z) - ColBERT: Using BERT Sentence Embedding in Parallel Neural Networks for
Computational Humor [0.0]
We propose a novel approach for detecting and rating humor in short texts based on a popular linguistic theory of humor.
The proposed technical method initiates by separating sentences of the given text and utilizing the BERT model to generate embeddings for each one.
We accompany the paper with a novel dataset for humor detection consisting of 200,000 formal short texts.
The proposed model obtained F1 scores of 0.982 and 0.869 in the humor detection experiments which outperform general and state-of-the-art models.
arXiv Detail & Related papers (2020-04-27T13:10:11Z) - Stimulating Creativity with FunLines: A Case Study of Humor Generation
in Headlines [9.367224590861913]
We introduce FunLines, a competitive game where players edit news headlines to make them funny.
FunLines makes the humor generation process fun, interactive, collaborative, rewarding and educational.
We show the effectiveness of this data by training humor classification models that outperform a previous benchmark.
arXiv Detail & Related papers (2020-02-05T22:56:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.