Intweetive Text Summarization
- URL: http://arxiv.org/abs/2001.11382v1
- Date: Thu, 16 Jan 2020 08:38:40 GMT
- Title: Intweetive Text Summarization
- Authors: Jean Val\`ere Cossu, Juan-Manuel Torres-Moreno, Eric SanJuan, Marc
El-B\`eze
- Abstract summary: We propose to automatically generated summaries of Micro-Blogs conversations dealing with public figures E-Reputation.
These summaries are generated using key-word queries or sample tweet and offer a focused view of the whole Micro-Blog network.
- Score: 1.1565654851982567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The amount of user generated contents from various social medias allows
analyst to handle a wide view of conversations on several topics related to
their business. Nevertheless keeping up-to-date with this amount of information
is not humanly feasible. Automatic Summarization then provides an interesting
mean to digest the dynamics and the mass volume of contents. In this paper, we
address the issue of tweets summarization which remains scarcely explored. We
propose to automatically generated summaries of Micro-Blogs conversations
dealing with public figures E-Reputation. These summaries are generated using
key-word queries or sample tweet and offer a focused view of the whole
Micro-Blog network. Since state-of-the-art is lacking on this point we conduct
and evaluate our experiments over the multilingual CLEF RepLab Topic-Detection
dataset according to an experimental evaluation process.
Related papers
- ProCIS: A Benchmark for Proactive Retrieval in Conversations [21.23826888841565]
We introduce a large-scale dataset for proactive document retrieval that consists of over 2.8 million conversations.
We conduct crowdsourcing experiments to obtain high-quality and relatively complete relevance judgments.
We also collect annotations related to the parts of the conversation that are related to each document, enabling us to evaluate proactive retrieval systems.
arXiv Detail & Related papers (2024-05-10T13:11:07Z) - AnswerSumm: A Manually-Curated Dataset and Pipeline for Answer
Summarization [73.91543616777064]
Community Question Answering (CQA) fora such as Stack Overflow and Yahoo! Answers contain a rich resource of answers to a wide range of community-based questions.
One goal of answer summarization is to produce a summary that reflects the range of answer perspectives.
This work introduces a novel dataset of 4,631 CQA threads for answer summarization, curated by professional linguists.
arXiv Detail & Related papers (2021-11-11T21:48:02Z) - EmailSum: Abstractive Email Thread Summarization [105.46012304024312]
We develop an abstractive Email Thread Summarization (EmailSum) dataset.
This dataset contains human-annotated short (30 words) and long (100 words) summaries of 2549 email threads.
Our results reveal the key challenges of current abstractive summarization models in this task.
arXiv Detail & Related papers (2021-07-30T15:13:14Z) - ConvoSumm: Conversation Summarization Benchmark and Improved Abstractive
Summarization with Argument Mining [61.82562838486632]
We crowdsource four new datasets on diverse online conversation forms of news comments, discussion forums, community question answering forums, and email threads.
We benchmark state-of-the-art models on our datasets and analyze characteristics associated with the data.
arXiv Detail & Related papers (2021-06-01T22:17:13Z) - Unsupervised Summarization for Chat Logs with Topic-Oriented Ranking and
Context-Aware Auto-Encoders [59.038157066874255]
We propose a novel framework called RankAE to perform chat summarization without employing manually labeled data.
RankAE consists of a topic-oriented ranking strategy that selects topic utterances according to centrality and diversity simultaneously.
A denoising auto-encoder is designed to generate succinct but context-informative summaries based on the selected utterances.
arXiv Detail & Related papers (2020-12-14T07:31:17Z) - Multi-View Sequence-to-Sequence Models with Conversational Structure for
Abstractive Dialogue Summarization [72.54873655114844]
Text summarization is one of the most challenging and interesting problems in NLP.
This work proposes a multi-view sequence-to-sequence model by first extracting conversational structures of unstructured daily chats from different views to represent conversations.
Experiments on a large-scale dialogue summarization corpus demonstrated that our methods significantly outperformed previous state-of-the-art models via both automatic evaluations and human judgment.
arXiv Detail & Related papers (2020-10-04T20:12:44Z) - Abstractive Summarization of Spoken and Written Instructions with BERT [66.14755043607776]
We present the first application of the BERTSum model to conversational language.
We generate abstractive summaries of narrated instructional videos across a wide variety of topics.
We envision this integrated as a feature in intelligent virtual assistants, enabling them to summarize both written and spoken instructional content upon request.
arXiv Detail & Related papers (2020-08-21T20:59:34Z) - Extending Text Informativeness Measures to Passage Interestingness
Evaluation (Language Model vs. Word Embedding) [1.2998637003026272]
This paper defines the concept of Interestingness as a generalization of Informativeness.
We then study the ability of state of the art Informativeness measures to cope with this generalization.
We prove that the CLEF-INEX Tweet Contextualization 2012 Logarithm Similarity measure provides best results.
arXiv Detail & Related papers (2020-04-14T18:22:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.