Sentiment-Aware Word and Sentence Level Pre-training for Sentiment
Analysis
- URL: http://arxiv.org/abs/2210.09803v2
- Date: Wed, 19 Oct 2022 12:22:05 GMT
- Title: Sentiment-Aware Word and Sentence Level Pre-training for Sentiment
Analysis
- Authors: Shuai Fan, Chen Lin, Haonan Li, Zhenghao Lin, Jinsong Su, Hang Zhang,
Yeyun Gong, Jian Guo, Nan Duan
- Abstract summary: SentiWSP is a Sentiment-aware pre-trained language model with combined Word-level and Sentence-level Pre-training tasks.
SentiWSP achieves new state-of-the-art performance on various sentence-level and aspect-level sentiment classification benchmarks.
- Score: 64.70116276295609
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most existing pre-trained language representation models (PLMs) are
sub-optimal in sentiment analysis tasks, as they capture the sentiment
information from word-level while under-considering sentence-level information.
In this paper, we propose SentiWSP, a novel Sentiment-aware pre-trained
language model with combined Word-level and Sentence-level Pre-training tasks.
The word level pre-training task detects replaced sentiment words, via a
generator-discriminator framework, to enhance the PLM's knowledge about
sentiment words. The sentence level pre-training task further strengthens the
discriminator via a contrastive learning framework, with similar sentences as
negative samples, to encode sentiments in a sentence. Extensive experimental
results show that SentiWSP achieves new state-of-the-art performance on various
sentence-level and aspect-level sentiment classification benchmarks. We have
made our code and model publicly available at
https://github.com/XMUDM/SentiWSP.
Related papers
- FRACTAL: Fine-Grained Scoring from Aggregate Text Labels [17.052047103156372]
Large language models (LLMs) are increasingly tuned to power complex generation tasks such as writing, fact-seeking, querying and reasoning.
Traditionally, human or model feedback for evaluating and tuning LLM performance has been provided at the response level.
Recent works indicate that sentence-level labels may provide more accurate and interpretable feedback for LLM optimization.
arXiv Detail & Related papers (2024-04-07T05:54:28Z) - SentiCSE: A Sentiment-aware Contrastive Sentence Embedding Framework with Sentiment-guided Textual Similarity [12.954271451359222]
Sentiment-aware pre-trained language models (PLMs) demonstrate impressive results in downstream sentiment analysis tasks.
We propose Sentiment-guided Textual Similarity (SgTS), a novel metric for evaluating the quality of sentiment representations.
We then propose SentiCSE, a novel Sentiment-aware Contrastive Sentence Embedding framework for constructing sentiment representations.
arXiv Detail & Related papers (2024-04-01T13:24:20Z) - Sentence Representation Learning with Generative Objective rather than
Contrastive Objective [86.01683892956144]
We propose a novel generative self-supervised learning objective based on phrase reconstruction.
Our generative learning achieves powerful enough performance improvement and outperforms the current state-of-the-art contrastive methods.
arXiv Detail & Related papers (2022-10-16T07:47:46Z) - Zero-shot Aspect-level Sentiment Classification via Explicit Utilization
of Aspect-to-Document Sentiment Composition [13.955534334487618]
In this work, we achieve zero-shot aspect-level sentiment classification by only using document-level reviews.
Our key intuition is that the sentiment representation of a document is composed of the sentiment representations of all the aspects of that document.
AF-DSC first learns sentiment representations for all potential aspects and then aggregates aspect-level sentiments into a document-level one to perform document-level sentiment classification.
arXiv Detail & Related papers (2022-09-06T08:02:55Z) - KESA: A Knowledge Enhanced Approach For Sentiment Analysis [13.937274761540925]
We study sentence-level sentiment analysis and propose two sentiment-aware auxiliary tasks named sentiment word cloze and conditional sentiment prediction.
The experimental results demonstrate that our approach consistently outperforms pre-trained models.
arXiv Detail & Related papers (2022-02-24T13:21:27Z) - To Augment or Not to Augment? A Comparative Study on Text Augmentation
Techniques for Low-Resource NLP [0.0]
We investigate three categories of text augmentation methodologies which perform changes on the syntax.
We compare them on part-of-speech tagging, dependency parsing and semantic role labeling for a diverse set of language families.
Our results suggest that the augmentation techniques can further improve over strong baselines based on mBERT.
arXiv Detail & Related papers (2021-11-18T10:52:48Z) - UCPhrase: Unsupervised Context-aware Quality Phrase Tagging [63.86606855524567]
UCPhrase is a novel unsupervised context-aware quality phrase tagger.
We induce high-quality phrase spans as silver labels from consistently co-occurring word sequences.
We show that our design is superior to state-of-the-art pre-trained, unsupervised, and distantly supervised methods.
arXiv Detail & Related papers (2021-05-28T19:44:24Z) - Narrative Incoherence Detection [76.43894977558811]
We propose the task of narrative incoherence detection as a new arena for inter-sentential semantic understanding.
Given a multi-sentence narrative, decide whether there exist any semantic discrepancies in the narrative flow.
arXiv Detail & Related papers (2020-12-21T07:18:08Z) - Grounded Compositional Outputs for Adaptive Language Modeling [59.02706635250856]
A language model's vocabulary$-$typically selected before training and permanently fixed later$-$affects its size.
We propose a fully compositional output embedding layer for language models.
To our knowledge, the result is the first word-level language model with a size that does not depend on the training vocabulary.
arXiv Detail & Related papers (2020-09-24T07:21:14Z) - SKEP: Sentiment Knowledge Enhanced Pre-training for Sentiment Analysis [69.80296394461149]
We introduce Sentiment Knowledge Enhanced Pre-training (SKEP) in order to learn a unified sentiment representation for multiple sentiment analysis tasks.
With the help of automatically-mined knowledge, SKEP conducts sentiment masking and constructs three sentiment knowledge prediction objectives.
Experiments on three kinds of sentiment tasks show that SKEP significantly outperforms strong pre-training baseline.
arXiv Detail & Related papers (2020-05-12T09:23:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.