On the State of German (Abstractive) Text Summarization
- URL: http://arxiv.org/abs/2301.07095v1
- Date: Tue, 17 Jan 2023 18:59:20 GMT
- Title: On the State of German (Abstractive) Text Summarization
- Authors: Dennis Aumiller and Jing Fan and Michael Gertz
- Abstract summary: We assess the landscape of German abstractive text summarization.
We investigate why practically useful solutions for abstractive text summarization are still absent in industry.
- Score: 3.1776833268555134
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: With recent advancements in the area of Natural Language Processing, the
focus is slowly shifting from a purely English-centric view towards more
language-specific solutions, including German. Especially practical for
businesses to analyze their growing amount of textual data are text
summarization systems, which transform long input documents into compressed and
more digestible summary texts. In this work, we assess the particular landscape
of German abstractive text summarization and investigate the reasons why
practically useful solutions for abstractive text summarization are still
absent in industry. Our focus is two-fold, analyzing a) training resources, and
b) publicly available summarization systems. We are able to show that popular
existing datasets exhibit crucial flaws in their assumptions about the original
sources, which frequently leads to detrimental effects on system generalization
and evaluation biases. We confirm that for the most popular training dataset,
MLSUM, over 50% of the training set is unsuitable for abstractive summarization
purposes. Furthermore, available systems frequently fail to compare to simple
baselines, and ignore more effective and efficient extractive summarization
approaches. We attribute poor evaluation quality to a variety of different
factors, which are investigated in more detail in this work: A lack of
qualitative (and diverse) gold data considered for training, understudied (and
untreated) positional biases in some of the existing datasets, and the lack of
easily accessible and streamlined pre-processing strategies or analysis tools.
We provide a comprehensive assessment of available models on the cleaned
datasets, and find that this can lead to a reduction of more than 20 ROUGE-1
points during evaluation. The code for dataset filtering and reproducing
results can be found online at https://github.com/dennlinger/summaries
Related papers
- ACLSum: A New Dataset for Aspect-based Summarization of Scientific
Publications [10.529898520273063]
ACLSum is a novel summarization dataset carefully crafted and evaluated by domain experts.
In contrast to previous datasets, ACLSum facilitates multi-aspect summarization of scientific papers.
arXiv Detail & Related papers (2024-03-08T13:32:01Z) - On Context Utilization in Summarization with Large Language Models [83.84459732796302]
Large language models (LLMs) excel in abstractive summarization tasks, delivering fluent and pertinent summaries.
Recent advancements have extended their capabilities to handle long-input contexts, exceeding 100k tokens.
We conduct the first comprehensive study on context utilization and position bias in summarization.
arXiv Detail & Related papers (2023-10-16T16:45:12Z) - Lay Text Summarisation Using Natural Language Processing: A Narrative
Literature Review [1.8899300124593648]
The aim of this literature review is to describe and compare the different text summarisation approaches used to generate lay summaries.
We screened 82 articles and included eight relevant papers published between 2020 and 2021, using the same dataset.
A combination of extractive and abstractive summarisation methods in a hybrid approach was found to be most effective.
arXiv Detail & Related papers (2023-03-24T18:30:50Z) - Ensemble Transfer Learning for Multilingual Coreference Resolution [60.409789753164944]
A problem that frequently occurs when working with a non-English language is the scarcity of annotated training data.
We design a simple but effective ensemble-based framework that combines various transfer learning techniques.
We also propose a low-cost TL method that bootstraps coreference resolution models by utilizing Wikipedia anchor texts.
arXiv Detail & Related papers (2023-01-22T18:22:55Z) - OpineSum: Entailment-based self-training for abstractive opinion
summarization [6.584115526134759]
We present a novel self-training approach, OpineSum, for abstractive opinion summarization.
The summaries in this approach are built using a novel application of textual entailment.
OpineSum achieves state-of-the-art performance in both settings.
arXiv Detail & Related papers (2022-12-21T06:20:28Z) - mFACE: Multilingual Summarization with Factual Consistency Evaluation [79.60172087719356]
Abstractive summarization has enjoyed renewed interest in recent years, thanks to pre-trained language models and the availability of large-scale datasets.
Despite promising results, current models still suffer from generating factually inconsistent summaries.
We leverage factual consistency evaluation models to improve multilingual summarization.
arXiv Detail & Related papers (2022-12-20T19:52:41Z) - A Closer Look at Debiased Temporal Sentence Grounding in Videos:
Dataset, Metric, and Approach [53.727460222955266]
Temporal Sentence Grounding in Videos (TSGV) aims to ground a natural language sentence in an untrimmed video.
Recent studies have found that current benchmark datasets may have obvious moment annotation biases.
We introduce a new evaluation metric "dR@n,IoU@m" that discounts the basic recall scores to alleviate the inflating evaluation caused by biased datasets.
arXiv Detail & Related papers (2022-03-10T08:58:18Z) - Few-Shot Learning for Opinion Summarization [117.70510762845338]
Opinion summarization is the automatic creation of text reflecting subjective information expressed in multiple documents.
In this work, we show that even a handful of summaries is sufficient to bootstrap generation of the summary text.
Our approach substantially outperforms previous extractive and abstractive methods in automatic and human evaluation.
arXiv Detail & Related papers (2020-04-30T15:37:38Z) - Unsupervised Opinion Summarization with Noising and Denoising [85.49169453434554]
We create a synthetic dataset from a corpus of user reviews by sampling a review, pretending it is a summary, and generating noisy versions thereof.
At test time, the model accepts genuine reviews and generates a summary containing salient opinions, treating those that do not reach consensus as noise.
arXiv Detail & Related papers (2020-04-21T16:54:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.