Integrating extracted information from bert and multiple embedding
methods with the deep neural network for humour detection
- URL: http://arxiv.org/abs/2105.05112v1
- Date: Tue, 11 May 2021 15:09:19 GMT
- Title: Integrating extracted information from bert and multiple embedding
methods with the deep neural network for humour detection
- Authors: Rida Miraj, Masaki Aono
- Abstract summary: We propose a framework for humour detection in short texts taken from news headlines.
Our proposed framework (IBEN) attempts to extract information from written text via the use of different layers of BERT.
The extracted information was then sent to a Bi-GRU neural network as an embedding matrix.
- Score: 3.612189440297043
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Humour detection from sentences has been an interesting and challenging task
in the last few years. In attempts to highlight humour detection, most research
was conducted using traditional approaches of embedding, e.g., Word2Vec or
Glove. Recently BERT sentence embedding has also been used for this task. In
this paper, we propose a framework for humour detection in short texts taken
from news headlines. Our proposed framework (IBEN) attempts to extract
information from written text via the use of different layers of BERT. After
several trials, weights were assigned to different layers of the BERT model.
The extracted information was then sent to a Bi-GRU neural network as an
embedding matrix. We utilized the properties of some external embedding models.
A multi-kernel convolution in our neural network was also employed to extract
higher-level sentence representations. This framework performed very well on
the task of humour detection.
Related papers
- Assaying on the Robustness of Zero-Shot Machine-Generated Text Detectors [57.7003399760813]
We explore advanced Large Language Models (LLMs) and their specialized variants, contributing to this field in several ways.
We uncover a significant correlation between topics and detection performance.
These investigations shed light on the adaptability and robustness of these detection methods across diverse topics.
arXiv Detail & Related papers (2023-12-20T10:53:53Z) - Word Sense Induction with Knowledge Distillation from BERT [6.88247391730482]
This paper proposes a method to distill multiple word senses from a pre-trained language model (BERT) by using attention over the senses of a word in a context.
Experiments on the contextual word similarity and sense induction tasks show that this method is superior to or competitive with state-of-the-art multi-sense embeddings.
arXiv Detail & Related papers (2023-04-20T21:05:35Z) - Can BERT Refrain from Forgetting on Sequential Tasks? A Probing Study [68.75670223005716]
We find that pre-trained language models like BERT have a potential ability to learn sequentially, even without any sparse memory replay.
Our experiments reveal that BERT can actually generate high quality representations for previously learned tasks in a long term, under extremely sparse replay or even no replay.
arXiv Detail & Related papers (2023-03-02T09:03:43Z) - Be More with Less: Hypergraph Attention Networks for Inductive Text
Classification [56.98218530073927]
Graph neural networks (GNNs) have received increasing attention in the research community and demonstrated their promising results on this canonical task.
Despite the success, their performance could be largely jeopardized in practice since they are unable to capture high-order interaction between words.
We propose a principled model -- hypergraph attention networks (HyperGAT) which can obtain more expressive power with less computational consumption for text representation learning.
arXiv Detail & Related papers (2020-11-01T00:21:59Z) - Intrinsic Probing through Dimension Selection [69.52439198455438]
Most modern NLP systems make use of pre-trained contextual representations that attain astonishingly high performance on a variety of tasks.
Such high performance should not be possible unless some form of linguistic structure inheres in these representations, and a wealth of research has sprung up on probing for it.
In this paper, we draw a distinction between intrinsic probing, which examines how linguistic information is structured within a representation, and the extrinsic probing popular in prior work, which only argues for the presence of such information by showing that it can be successfully extracted.
arXiv Detail & Related papers (2020-10-06T15:21:08Z) - A Comparative Study on Structural and Semantic Properties of Sentence
Embeddings [77.34726150561087]
We propose a set of experiments using a widely-used large-scale data set for relation extraction.
We show that different embedding spaces have different degrees of strength for the structural and semantic properties.
These results provide useful information for developing embedding-based relation extraction methods.
arXiv Detail & Related papers (2020-09-23T15:45:32Z) - A novel sentence embedding based topic detection method for micro-blog [5.821169298644354]
We present a novel approach based on neural network to detect topics in the micro-blogging dataset.
We use an unsupervised neural sentence embedding model to map the blogs to an embedding space.
In addition, we propose an improved clustering algorithm referred as relationship-aware DBSCAN (RADBSCAN)
arXiv Detail & Related papers (2020-06-10T09:58:57Z) - ColBERT: Using BERT Sentence Embedding in Parallel Neural Networks for
Computational Humor [0.0]
We propose a novel approach for detecting and rating humor in short texts based on a popular linguistic theory of humor.
The proposed technical method initiates by separating sentences of the given text and utilizing the BERT model to generate embeddings for each one.
We accompany the paper with a novel dataset for humor detection consisting of 200,000 formal short texts.
The proposed model obtained F1 scores of 0.982 and 0.869 in the humor detection experiments which outperform general and state-of-the-art models.
arXiv Detail & Related papers (2020-04-27T13:10:11Z) - Hybrid Attention-Based Transformer Block Model for Distant Supervision
Relation Extraction [20.644215991166902]
We propose a new framework using hybrid attention-based Transformer block with multi-instance learning to perform the DSRE task.
The proposed approach can outperform the state-of-the-art algorithms on the evaluation dataset.
arXiv Detail & Related papers (2020-03-10T13:05:52Z) - Incorporating BERT into Neural Machine Translation [251.54280200353674]
We propose a new algorithm named BERT-fused model, in which we first use BERT to extract representations for an input sequence.
We conduct experiments on supervised (including sentence-level and document-level translations), semi-supervised and unsupervised machine translation, and achieve state-of-the-art results on seven benchmark datasets.
arXiv Detail & Related papers (2020-02-17T08:13:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.