Subsentence Extraction from Text Using Coverage-Based Deep Learning
Language Models
- URL: http://arxiv.org/abs/2104.09777v1
- Date: Tue, 20 Apr 2021 06:24:49 GMT
- Title: Subsentence Extraction from Text Using Coverage-Based Deep Learning
Language Models
- Authors: JongYoon Lim, Inkyu Sa, Ho Seok Ahn, Norina Gasteiger, Sanghyub John
Lee, Bruce MacDonald
- Abstract summary: We propose a coverage-based sentiment and subsentence extraction system.
The predicted subsentence consists of auxiliary information expressing a sentiment.
Our approach outperforms the state-of-the-art approaches by a large margin in subsentence prediction.
- Score: 3.3461339691835277
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Sentiment prediction remains a challenging and unresolved task in various
research fields, including psychology, neuroscience, and computer science. This
stems from its high degree of subjectivity and limited input sources that can
effectively capture the actual sentiment. This can be even more challenging
with only text-based input. Meanwhile, the rise of deep learning and an
unprecedented large volume of data have paved the way for artificial
intelligence to perform impressively accurate predictions or even human-level
reasoning. Drawing inspiration from this, we propose a coverage-based sentiment
and subsentence extraction system that estimates a span of input text and
recursively feeds this information back to the networks. The predicted
subsentence consists of auxiliary information expressing a sentiment. This is
an important building block for enabling vivid and epic sentiment delivery
(within the scope of this paper) and for other natural language processing
tasks such as text summarisation and Q&A. Our approach outperforms the
state-of-the-art approaches by a large margin in subsentence prediction (i.e.,
Average Jaccard scores from 0.72 to 0.89). For the evaluation, we designed
rigorous experiments consisting of 24 ablation studies. Finally, our learned
lessons are returned to the community by sharing software packages and a public
dataset that can reproduce the results presented in this paper.
Related papers
- On the effect of curriculum learning with developmental data for grammar
acquisition [4.4044968357361745]
This work explores the degree to which grammar acquisition is driven by language simplicity' and the source modality (speech vs. text) of data.
We find that grammar acquisition is largely driven by exposure to speech data, and in particular through exposure to two of the BabyLM training corpora: AO-Childes and Open Subtitles.
arXiv Detail & Related papers (2023-10-31T20:05:30Z) - Harnessing the Power of Text-image Contrastive Models for Automatic
Detection of Online Misinformation [50.46219766161111]
We develop a self-learning model to explore the constrastive learning in the domain of misinformation identification.
Our model shows the superior performance of non-matched image-text pair detection when the training data is insufficient.
arXiv Detail & Related papers (2023-04-19T02:53:59Z) - Lexical Complexity Prediction: An Overview [13.224233182417636]
The occurrence of unknown words in texts significantly hinders reading comprehension.
computational modelling has been applied to identify complex words in texts and substitute them for simpler alternatives.
We present an overview of computational approaches to lexical complexity prediction focusing on the work carried out on English data.
arXiv Detail & Related papers (2023-03-08T19:35:08Z) - Fine-Grained Visual Entailment [51.66881737644983]
We propose an extension of this task, where the goal is to predict the logical relationship of fine-grained knowledge elements within a piece of text to an image.
Unlike prior work, our method is inherently explainable and makes logical predictions at different levels of granularity.
We evaluate our method on a new dataset of manually annotated knowledge elements and show that our method achieves 68.18% accuracy at this challenging task.
arXiv Detail & Related papers (2022-03-29T16:09:38Z) - A Case Study to Reveal if an Area of Interest has a Trend in Ongoing
Tweets Using Word and Sentence Embeddings [0.0]
We have proposed an easily applicable automated methodology in which the Daily Mean Similarity Scores show the similarity between the daily tweet corpus and the target words.
The Daily Mean Similarity Scores have mainly based on cosine similarity and word/sentence embeddings.
We have also compared the effectiveness of using word versus sentence embeddings while applying our methodology and realized that both give almost the same results.
arXiv Detail & Related papers (2021-10-02T18:44:55Z) - The Challenges of Persian User-generated Textual Content: A Machine
Learning-Based Approach [0.0]
This research applies machine learning-based approaches to tackle the hurdles that come with Persian user-generated textual content.
The presented approach uses a machine-translated datasets to conduct sentiment analysis for the Persian language.
The results of the experiments have shown promising state-of-the-art performance in contrast to the previous efforts.
arXiv Detail & Related papers (2021-01-20T11:57:59Z) - Deep Learning for Text Style Transfer: A Survey [71.8870854396927]
Text style transfer is an important task in natural language generation, which aims to control certain attributes in the generated text.
We present a systematic survey of the research on neural text style transfer, spanning over 100 representative articles since the first neural text style transfer work in 2017.
We discuss the task formulation, existing datasets and subtasks, evaluation, as well as the rich methodologies in the presence of parallel and non-parallel data.
arXiv Detail & Related papers (2020-11-01T04:04:43Z) - Be More with Less: Hypergraph Attention Networks for Inductive Text
Classification [56.98218530073927]
Graph neural networks (GNNs) have received increasing attention in the research community and demonstrated their promising results on this canonical task.
Despite the success, their performance could be largely jeopardized in practice since they are unable to capture high-order interaction between words.
We propose a principled model -- hypergraph attention networks (HyperGAT) which can obtain more expressive power with less computational consumption for text representation learning.
arXiv Detail & Related papers (2020-11-01T00:21:59Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z) - Improving unsupervised neural aspect extraction for online discussions
using out-of-domain classification [11.746330029375745]
We introduce a simple approach based on sentence filtering to improve topical aspects learned from newsgroups-based content.
The positive effect of sentence filtering on topic coherence is demonstrated in comparison to aspect extraction models trained on unfiltered texts.
arXiv Detail & Related papers (2020-06-17T10:34:16Z) - ORB: An Open Reading Benchmark for Comprehensive Evaluation of Machine
Reading Comprehension [53.037401638264235]
We present an evaluation server, ORB, that reports performance on seven diverse reading comprehension datasets.
The evaluation server places no restrictions on how models are trained, so it is a suitable test bed for exploring training paradigms and representation learning.
arXiv Detail & Related papers (2019-12-29T07:27:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.