Automatic Product Ontology Extraction from Textual Reviews
- URL: http://arxiv.org/abs/2105.10966v1
- Date: Sun, 23 May 2021 16:06:38 GMT
- Title: Automatic Product Ontology Extraction from Textual Reviews
- Authors: Joel Oksanen, Oana Cocarascu, Francesca Toni
- Abstract summary: We show that the generated by our method outperform hand-crafted (NetWord) and extracted by existing methods (Text2Onto and COMET) in several, diverse settings.
Our method is better able to determine recommended products based on their reviews, in alternative to using Amazon's standard score aggregations.
- Score: 12.235907063179278
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Ontologies have proven beneficial in different settings that make use of
textual reviews. However, manually constructing ontologies is a laborious and
time-consuming process in need of automation. We propose a novel methodology
for automatically extracting ontologies, in the form of meronomies, from
product reviews, using a very limited amount of hand-annotated training data.
We show that the ontologies generated by our method outperform hand-crafted
ontologies (WordNet) and ontologies extracted by existing methods (Text2Onto
and COMET) in several, diverse settings. Specifically, our generated ontologies
outperform the others when evaluated by human annotators as well as on an
existing Q&A dataset from Amazon. Moreover, our method is better able to
generalise, in capturing knowledge about unseen products. Finally, we consider
a real-world setting, showing that our method is better able to determine
recommended products based on their reviews, in alternative to using Amazon's
standard score aggregations.
Related papers
- Meronymic Ontology Extraction via Large Language Models [16.08771514313186]
Onologies have become essential in today's digital age as a way of organising the vast amount of readily available unstructured text.<n>In this paper, we develop a fully-automated method of extracting meronymies from texts.
arXiv Detail & Related papers (2025-10-11T11:54:38Z) - A Survey of Automatic Evaluation Methods on Text, Visual and Speech Generations [58.105900601078595]
We present a comprehensive review and a unified taxonomy of automatic evaluation methods for generated content across all three modalities.<n>Our analysis begins by examining evaluation methods for text generation, where techniques are most mature.<n>We then extend this framework to image and audio generation, demonstrating its broad applicability.
arXiv Detail & Related papers (2025-06-06T11:09:46Z) - End-to-End Ontology Learning with Large Language Models [11.755755139228219]
Large language models (LLMs) have been applied to solve various subtasks of ontology learning.
We address this gap by OLLM, a general and scalable method for building the taxonomic backbone of an ontology from scratch.
In contrast to standard metrics, our metrics use deep learning techniques to define more robust structural distance measures between graphs.
Our model can be effectively adapted to new domains, like arXiv, needing only a small number of training examples.
arXiv Detail & Related papers (2024-10-31T02:52:39Z) - Towards Better Open-Ended Text Generation: A Multicriteria Evaluation Framework [0.1979158763744267]
Open-ended text generation has become a prominent task in natural language processing.
Decoding methods often excel in some metrics while underperforming in others.
We present novel ranking strategies within this multicriteria framework.
arXiv Detail & Related papers (2024-10-24T11:32:01Z) - Automatic Generation of Behavioral Test Cases For Natural Language Processing Using Clustering and Prompting [6.938766764201549]
This paper introduces an automated approach to develop test cases by exploiting the power of large language models and statistical techniques.
We analyze the behavioral test profiles across four different classification algorithms and discuss the limitations and strengths of those models.
arXiv Detail & Related papers (2024-07-31T21:12:21Z) - CovScore: Evaluation of Multi-Document Abstractive Title Set Generation [16.516381474175986]
CovScore is an automatic reference-less methodology for evaluating thematic title sets.
We propose a novel methodology that decomposes quality into five main metrics along different aspects of evaluation.
arXiv Detail & Related papers (2024-07-24T16:14:15Z) - Long-Span Question-Answering: Automatic Question Generation and QA-System Ranking via Side-by-Side Evaluation [65.16137964758612]
We explore the use of long-context capabilities in large language models to create synthetic reading comprehension data from entire books.
Our objective is to test the capabilities of LLMs to analyze, understand, and reason over problems that require a detailed comprehension of long spans of text.
arXiv Detail & Related papers (2024-05-31T20:15:10Z) - Generative AI-Based Text Generation Methods Using Pre-Trained GPT-2 Model [2.6320841968362645]
This work delved into the realm of automatic text generation, exploring a variety of techniques ranging from traditional deterministic approaches to more modern methods.
Through analysis of greedy search, beam search, top-k sampling, top-p sampling, contrastive searching, and locally typical searching, this work has provided valuable insights into the strengths, weaknesses, and potential applications of each method.
arXiv Detail & Related papers (2024-04-02T09:49:53Z) - Automated Grading and Feedback Tools for Programming Education: A
Systematic Review [7.776434991976473]
Most papers assess the correctness of assignments in object-oriented languages.
Few tools assess the maintainability, readability or documentation of the source code.
Most tools offered fully automated assessment to allow for near-instantaneous feedback.
arXiv Detail & Related papers (2023-06-20T17:54:50Z) - Human-in-the-Loop Disinformation Detection: Stance, Sentiment, or
Something Else? [93.91375268580806]
Both politics and pandemics have recently provided ample motivation for the development of machine learning-enabled disinformation (a.k.a. fake news) detection algorithms.
Existing literature has focused primarily on the fully-automated case, but the resulting techniques cannot reliably detect disinformation on the varied topics, sources, and time scales required for military applications.
By leveraging an already-available analyst as a human-in-the-loop, canonical machine learning techniques of sentiment analysis, aspect-based sentiment analysis, and stance detection become plausible methods to use for a partially-automated disinformation detection system.
arXiv Detail & Related papers (2021-11-09T13:30:34Z) - Hierarchical Bi-Directional Self-Attention Networks for Paper Review
Rating Recommendation [81.55533657694016]
We propose a Hierarchical bi-directional self-attention Network framework (HabNet) for paper review rating prediction and recommendation.
Specifically, we leverage the hierarchical structure of the paper reviews with three levels of encoders: sentence encoder (level one), intra-review encoder (level two) and inter-review encoder (level three)
We are able to identify useful predictors to make the final acceptance decision, as well as to help discover the inconsistency between numerical review ratings and text sentiment conveyed by reviewers.
arXiv Detail & Related papers (2020-11-02T08:07:50Z) - On Learning Text Style Transfer with Direct Rewards [101.97136885111037]
Lack of parallel corpora makes it impossible to directly train supervised models for the text style transfer task.
We leverage semantic similarity metrics originally used for fine-tuning neural machine translation models.
Our model provides significant gains in both automatic and human evaluation over strong baselines.
arXiv Detail & Related papers (2020-10-24T04:30:02Z) - Few-Shot Learning for Opinion Summarization [117.70510762845338]
Opinion summarization is the automatic creation of text reflecting subjective information expressed in multiple documents.
In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text.
Our approach substantially outperforms previous extractive and abstractive methods in automatic and human evaluation.
arXiv Detail & Related papers (2020-04-30T15:37:38Z) - PONE: A Novel Automatic Evaluation Metric for Open-Domain Generative
Dialogue Systems [48.99561874529323]
There are three kinds of automatic methods to evaluate the open-domain generative dialogue systems.
Due to the lack of systematic comparison, it is not clear which kind of metrics are more effective.
We propose a novel and feasible learning-based metric that can significantly improve the correlation with human judgments.
arXiv Detail & Related papers (2020-04-06T04:36:33Z) - Mining customer product reviews for product development: A summarization
process [0.7742297876120561]
This research set out to identify and structure from online reviews the words and expressions related to customers' likes and dislikes to guide product development.
The authors propose a summarization model containing multiples aspects of user preference, such as product affordances, emotions, usage conditions.
A case study demonstrates that with the proposed model and the annotation guidelines, human annotators can structure the online reviews with high inter-agreement.
arXiv Detail & Related papers (2020-01-13T13:01:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.