Zero-shot Entity and Tweet Characterization with Designed Conditional
Prompts and Contexts
- URL: http://arxiv.org/abs/2204.08405v1
- Date: Mon, 18 Apr 2022 17:01:49 GMT
- Title: Zero-shot Entity and Tweet Characterization with Designed Conditional
Prompts and Contexts
- Authors: Sharath Srivatsa, Tushar Mohan, Kumari Neha, Nishchay Malakar,
Ponnurangam Kumaraguru, and Srinath Srinivasa
- Abstract summary: We evaluate the zero-shot language model capabilities of Generative Pretrained Transformer 2 (GPT-2) to characterize and Tweets subjectively.
We fine-tune GPT-2 with a Tweets corpus from a few popular hashtags and evaluate characterizing tweets by priming the language model with prefixes, questions, and contextual synopsis prompts.
- Score: 6.38674533060275
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Online news and social media have been the de facto mediums to disseminate
information globally from the beginning of the last decade. However, bias in
content and purpose of intentions are not regulated, and managing bias is the
responsibility of content consumers. In this regard, understanding the stances
and biases of news sources towards specific entities becomes important. To
address this problem, we use pretrained language models, which have been shown
to bring about good results with no task-specific training or few-shot
training. In this work, we approach the problem of characterizing Named
Entities and Tweets as an open-ended text classification and open-ended fact
probing problem.We evaluate the zero-shot language model capabilities of
Generative Pretrained Transformer 2 (GPT-2) to characterize Entities and Tweets
subjectively with human psychology-inspired and logical conditional prefixes
and contexts. First, we fine-tune the GPT-2 model on a sufficiently large news
corpus and evaluate subjective characterization of popular entities in the
corpus by priming with prefixes. Second, we fine-tune GPT-2 with a Tweets
corpus from a few popular hashtags and evaluate characterizing tweets by
priming the language model with prefixes, questions, and contextual synopsis
prompts. Entity characterization results were positive across measures and
human evaluation.
Related papers
- Fine-tuning Language Models for Factuality [96.5203774943198]
Large pre-trained language models (LLMs) have led to their widespread use, sometimes even as a replacement for traditional search engines.
Yet language models are prone to making convincing but factually inaccurate claims, often referred to as 'hallucinations'
In this work, we fine-tune language models to be more factual, without human labeling.
arXiv Detail & Related papers (2023-11-14T18:59:15Z) - SOUL: Towards Sentiment and Opinion Understanding of Language [96.74878032417054]
We propose a new task called Sentiment and Opinion Understanding of Language (SOUL)
SOUL aims to evaluate sentiment understanding through two subtasks: Review (RC) and Justification Generation (JG)
arXiv Detail & Related papers (2023-10-27T06:48:48Z) - Characterizing Latent Perspectives of Media Houses Towards Public
Figures [0.0]
This work proposes a zero-shot approach for non-extractive or generative characterizations of person entities from a corpus using GPT-2.
We use well-articulated articles from several well-known news media houses as a corpus to build a sound argument for this approach.
arXiv Detail & Related papers (2023-09-12T10:27:39Z) - Analyzing Vietnamese Legal Questions Using Deep Neural Networks with
Biaffine Classifiers [3.116035935327534]
We propose using deep neural networks to extract important information from Vietnamese legal questions.
Given a legal question in natural language, the goal is to extract all the segments that contain the needed information to answer the question.
arXiv Detail & Related papers (2023-04-27T18:19:24Z) - Bias at a Second Glance: A Deep Dive into Bias for German Educational
Peer-Review Data Modeling [10.080007569933331]
We analyze bias across text and through multiple architectures on a corpus of 9,165 German peer- reviews over five years.
Our collected corpus does not reveal many biases in the co-occurrence analysis or in the GloVe embeddings.
Pre-trained German language models find substantial conceptual, racial, and gender bias.
arXiv Detail & Related papers (2022-09-21T13:08:16Z) - Leveraging Pre-trained Language Model for Speech Sentiment Analysis [58.78839114092951]
We explore the use of pre-trained language models to learn sentiment information of written texts for speech sentiment analysis.
We propose a pseudo label-based semi-supervised training strategy using a language model on an end-to-end speech sentiment approach.
arXiv Detail & Related papers (2021-06-11T20:15:21Z) - Sentiment analysis in tweets: an assessment study from classical to
modern text representation models [59.107260266206445]
Short texts published on Twitter have earned significant attention as a rich source of information.
Their inherent characteristics, such as the informal, and noisy linguistic style, remain challenging to many natural language processing (NLP) tasks.
This study fulfils an assessment of existing language models in distinguishing the sentiment expressed in tweets by using a rich collection of 22 datasets.
arXiv Detail & Related papers (2021-05-29T21:05:28Z) - TweetEval: Unified Benchmark and Comparative Evaluation for Tweet
Classification [22.265865542786084]
We propose a new evaluation framework (TweetEval) consisting of seven heterogeneous Twitter-specific classification tasks.
Our initial experiments show the effectiveness of starting off with existing pre-trained generic language models.
arXiv Detail & Related papers (2020-10-23T14:11:04Z) - Weakly-Supervised Aspect-Based Sentiment Analysis via Joint
Aspect-Sentiment Topic Embedding [71.2260967797055]
We propose a weakly-supervised approach for aspect-based sentiment analysis.
We learn sentiment, aspect> joint topic embeddings in the word embedding space.
We then use neural models to generalize the word-level discriminative information.
arXiv Detail & Related papers (2020-10-13T21:33:24Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.