Ensemble Creation via Anchored Regularization for Unsupervised Aspect
Extraction
- URL: http://arxiv.org/abs/2210.06829v1
- Date: Thu, 13 Oct 2022 08:23:56 GMT
- Title: Ensemble Creation via Anchored Regularization for Unsupervised Aspect
Extraction
- Authors: Pulah Dhandekar and Manu Joseph
- Abstract summary: Unsupervised aspect-based sentiment analysis allows us to generate insights without investing time or money in generating labels.
One of the models that we improve upon is ABAE that reconstructs the sentences as a linear combination of aspect terms present in it.
In this research we explore how we can use information from another unsupervised model to regularize ABAE, leading to better performance.
- Score: 1.8591803874887636
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Aspect Based Sentiment Analysis is the most granular form of sentiment
analysis that can be performed on the documents / sentences. Besides delivering
the most insights at a finer grain, it also poses equally daunting challenges.
One of them being the shortage of labelled data. To bring in value right out of
the box for the text data being generated at a very fast pace in today's world,
unsupervised aspect-based sentiment analysis allows us to generate insights
without investing time or money in generating labels. From topic modelling
approaches to recent deep learning-based aspect extraction models, this domain
has seen a lot of development. One of the models that we improve upon is ABAE
that reconstructs the sentences as a linear combination of aspect terms present
in it, In this research we explore how we can use information from another
unsupervised model to regularize ABAE, leading to better performance. We
contrast it with baseline rule based ensemble and show that the ensemble
methods work better than the individual models and the regularization based
ensemble performs better than the rule-based one.
Related papers
- Enhancement of Approximation Spaces by the Use of Primals and Neighborhood [0.0]
We introduce four new generalized rough set models that draw inspiration from "neighborhoods and primals"
We claim that the current models can preserve nearly all significant aspects associated with the rough set model.
We also demonstrate that the new strategy we define for our everyday health-related problem yields more accurate findings.
arXiv Detail & Related papers (2024-10-23T18:49:13Z) - Instruct-DeBERTa: A Hybrid Approach for Aspect-based Sentiment Analysis on Textual Reviews [2.0143010051030417]
Aspect-based Sentiment Analysis (ABSA) is a critical task in Natural Language Processing (NLP)
Traditional sentiment analysis methods, while useful for determining overall sentiment, often miss the implicit opinions about particular product or service features.
This paper presents a comprehensive review of the evolution of ABSA methodologies, from lexicon-based approaches to machine learning.
arXiv Detail & Related papers (2024-08-23T16:31:07Z) - Long-Span Question-Answering: Automatic Question Generation and QA-System Ranking via Side-by-Side Evaluation [65.16137964758612]
We explore the use of long-context capabilities in large language models to create synthetic reading comprehension data from entire books.
Our objective is to test the capabilities of LLMs to analyze, understand, and reason over problems that require a detailed comprehension of long spans of text.
arXiv Detail & Related papers (2024-05-31T20:15:10Z) - Corpus Considerations for Annotator Modeling and Scaling [9.263562546969695]
We show that the commonly used user token model consistently outperforms more complex models.
Our findings shed light on the relationship between corpus statistics and annotator modeling performance.
arXiv Detail & Related papers (2024-04-02T22:27:24Z) - General Greedy De-bias Learning [163.65789778416172]
We propose a General Greedy De-bias learning framework (GGD), which greedily trains the biased models and the base model like gradient descent in functional space.
GGD can learn a more robust base model under the settings of both task-specific biased models with prior knowledge and self-ensemble biased model without prior knowledge.
arXiv Detail & Related papers (2021-12-20T14:47:32Z) - SentimentArcs: A Novel Method for Self-Supervised Sentiment Analysis of
Time Series Shows SOTA Transformers Can Struggle Finding Narrative Arcs [0.0]
This paper introduces SentimentArcs, a new self-supervised time series sentiment analysis methodology.
A large ensemble of diverse models provides a synthetic ground truth for self-supervised learning.
Simple visualizations exploit the temporal structure in narratives so domain experts can quickly spot trends.
arXiv Detail & Related papers (2021-10-18T16:45:31Z) - Aspect-Controllable Opinion Summarization [58.5308638148329]
We propose an approach that allows the generation of customized summaries based on aspect queries.
Using a review corpus, we create a synthetic training dataset of (review, summary) pairs enriched with aspect controllers.
We fine-tune a pretrained model using our synthetic dataset and generate aspect-specific summaries by modifying the aspect controllers.
arXiv Detail & Related papers (2021-09-07T16:09:17Z) - WikiAsp: A Dataset for Multi-domain Aspect-based Summarization [69.13865812754058]
We propose WikiAsp, a large-scale dataset for multi-domain aspect-based summarization.
Specifically, we build the dataset using Wikipedia articles from 20 different domains, using the section titles and boundaries of each article as a proxy for aspect annotation.
Results highlight key challenges that existing summarization models face in this setting, such as proper pronoun handling of quoted sources and consistent explanation of time-sensitive events.
arXiv Detail & Related papers (2020-11-16T10:02:52Z) - Learning from Context or Names? An Empirical Study on Neural Relation
Extraction [112.06614505580501]
We study the effect of two main information sources in text: textual context and entity mentions (names)
We propose an entity-masked contrastive pre-training framework for relation extraction (RE)
Our framework can improve the effectiveness and robustness of neural models in different RE scenarios.
arXiv Detail & Related papers (2020-10-05T11:21:59Z) - Knowledge Graph-Augmented Abstractive Summarization with Semantic-Driven
Cloze Reward [42.925345819778656]
We present ASGARD, a novel framework for Abstractive Summarization with Graph-Augmentation and semantic-driven RewarD.
We propose the use of dual encoders---a sequential document encoder and a graph-structured encoder---to maintain the global context and local characteristics of entities.
Results show that our models produce significantly higher ROUGE scores than a variant without knowledge graph as input on both New York Times and CNN/Daily Mail datasets.
arXiv Detail & Related papers (2020-05-03T18:23:06Z) - Few-Shot Learning for Opinion Summarization [117.70510762845338]
Opinion summarization is the automatic creation of text reflecting subjective information expressed in multiple documents.
In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text.
Our approach substantially outperforms previous extractive and abstractive methods in automatic and human evaluation.
arXiv Detail & Related papers (2020-04-30T15:37:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.