BERT-Assisted Semantic Annotation Correction for Emotion-Related
Questions
- URL: http://arxiv.org/abs/2204.00916v1
- Date: Sat, 2 Apr 2022 18:00:49 GMT
- Title: BERT-Assisted Semantic Annotation Correction for Emotion-Related
Questions
- Authors: Abe Kazemzadeh
- Abstract summary: We use the BERT neural language model to feed information back into an annotation task in a question-asking game called Emotion Twenty Questions (EMO20Q)
We show this method to be an effective way to assess and revise annotations of textual user data with complex, utterance-level semantic labels.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Annotated data have traditionally been used to provide the input for training
a supervised machine learning (ML) model. However, current pre-trained ML
models for natural language processing (NLP) contain embedded linguistic
information that can be used to inform the annotation process. We use the BERT
neural language model to feed information back into an annotation task that
involves semantic labelling of dialog behavior in a question-asking game called
Emotion Twenty Questions (EMO20Q). First we describe the background of BERT,
the EMO20Q data, and assisted annotation tasks. Then we describe the methods
for fine-tuning BERT for the purpose of checking the annotated labels. To do
this, we use the paraphrase task as a way to check that all utterances with the
same annotation label are classified as paraphrases of each other. We show this
method to be an effective way to assess and revise annotations of textual user
data with complex, utterance-level semantic labels.
Related papers
- Zero-shot prompt-based classification: topic labeling in times of foundation models in German Tweets [1.734165485480267]
We propose a new tool for automatically annotating text using written guidelines without providing training samples.
Our results show that the prompt-based approach is comparable with the fine-tuned BERT but without any annotated training data.
Our findings emphasize the ongoing paradigm shift in the NLP landscape, i.e., the unification of downstream tasks and elimination of the need for pre-labeled training data.
arXiv Detail & Related papers (2024-06-26T10:44:02Z) - Eliciting Human Preferences with Language Models [56.68637202313052]
Language models (LMs) can be directed to perform target tasks by using labeled examples or natural language prompts.
We propose to use *LMs themselves* to guide the task specification process.
We study GATE in three domains: email validation, content recommendation, and moral reasoning.
arXiv Detail & Related papers (2023-10-17T21:11:21Z) - Description-Enhanced Label Embedding Contrastive Learning for Text
Classification [65.01077813330559]
Self-Supervised Learning (SSL) in model learning process and design a novel self-supervised Relation of Relation (R2) classification task.
Relation of Relation Learning Network (R2-Net) for text classification, in which text classification and R2 classification are treated as optimization targets.
external knowledge from WordNet to obtain multi-aspect descriptions for label semantic learning.
arXiv Detail & Related papers (2023-06-15T02:19:34Z) - Exploring Structured Semantic Prior for Multi Label Recognition with
Incomplete Labels [60.675714333081466]
Multi-label recognition (MLR) with incomplete labels is very challenging.
Recent works strive to explore the image-to-label correspondence in the vision-language model, ie, CLIP, to compensate for insufficient annotations.
We advocate remedying the deficiency of label supervision for the MLR with incomplete labels by deriving a structured semantic prior.
arXiv Detail & Related papers (2023-03-23T12:39:20Z) - Active Learning and Multi-label Classification for Ellipsis and
Coreference Detection in Conversational Question-Answering [5.984693203400407]
ellipsis and coreferences are commonly occurring linguistic phenomena.
We propose to use a multi-label classifier based on DistilBERT.
We show that these methods greatly enhance the performance of the classifier for detecting these phenomena on a manually labeled dataset.
arXiv Detail & Related papers (2022-07-07T08:14:54Z) - Enhanced Language Representation with Label Knowledge for Span
Extraction [2.4909170697740963]
We introduce a new paradigm to integrate label knowledge and propose a novel model to explicitly and efficiently integrate label knowledge into text representations.
Specifically, it encodes texts and label annotations independently and then integrates label knowledge into text representation with an elaborate-designed semantics fusion module.
We conduct extensive experiments on three typical span extraction tasks: flat NER, nested NER, and event detection.
Our method achieves state-of-the-art performance on four benchmarks, and 2) reduces training time and inference time by 76% and 77% on average, respectively, compared with the QA Formalization paradigm.
arXiv Detail & Related papers (2021-11-01T12:21:05Z) - Annotation Curricula to Implicitly Train Non-Expert Annotators [56.67768938052715]
voluntary studies often require annotators to familiarize themselves with the task, its annotation scheme, and the data domain.
This can be overwhelming in the beginning, mentally taxing, and induce errors into the resulting annotations.
We propose annotation curricula, a novel approach to implicitly train annotators.
arXiv Detail & Related papers (2021-06-04T09:48:28Z) - R$^2$-Net: Relation of Relation Learning Network for Sentence Semantic
Matching [58.72111690643359]
We propose a Relation of Relation Learning Network (R2-Net) for sentence semantic matching.
We first employ BERT to encode the input sentences from a global perspective.
Then a CNN-based encoder is designed to capture keywords and phrase information from a local perspective.
To fully leverage labels for better relation information extraction, we introduce a self-supervised relation of relation classification task.
arXiv Detail & Related papers (2020-12-16T13:11:30Z) - Understanding Pre-trained BERT for Aspect-based Sentiment Analysis [71.40586258509394]
This paper analyzes the pre-trained hidden representations learned from reviews on BERT for tasks in aspect-based sentiment analysis (ABSA)
It is not clear how the general proxy task of (masked) language model trained on unlabeled corpus without annotations of aspects or opinions can provide important features for downstream tasks in ABSA.
arXiv Detail & Related papers (2020-10-31T02:21:43Z) - Automatically Identifying Words That Can Serve as Labels for Few-Shot
Text Classification [12.418532541734193]
A recent approach for few-shot text classification is to convert textual inputs to cloze questions that contain some form of task description, process them with a pretrained language model and map the predicted words to labels.
To mitigate this issue, we devise an approach that automatically finds such a mapping given small amounts of training data.
For a number of tasks, the mapping found by our approach performs almost as well as hand-crafted label-to-word mappings.
arXiv Detail & Related papers (2020-10-26T14:56:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.