Towards Unifying Multi-Lingual and Cross-Lingual Summarization
- URL: http://arxiv.org/abs/2305.09220v1
- Date: Tue, 16 May 2023 06:53:21 GMT
- Title: Towards Unifying Multi-Lingual and Cross-Lingual Summarization
- Authors: Jiaan Wang, Fandong Meng, Duo Zheng, Yunlong Liang, Zhixu Li, Jianfeng
Qu and Jie Zhou
- Abstract summary: We aim to unify multilingual summarization (MLS) and cross-lingual summarization ( CLS) into a more general setting, i.e., many-to-many summarization (M2MS)
As the first step towards M2MS, we conduct preliminary studies to show that M2MS can better transfer task knowledge across different languages than MLS and CLS.
We propose Pisces, a pre-trained M2MS model that learns language modeling, cross-lingual ability and summarization ability via three-stage pre-training.
- Score: 43.89340385650822
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: To adapt text summarization to the multilingual world, previous work proposes
multi-lingual summarization (MLS) and cross-lingual summarization (CLS).
However, these two tasks have been studied separately due to the different
definitions, which limits the compatible and systematic research on both of
them. In this paper, we aim to unify MLS and CLS into a more general setting,
i.e., many-to-many summarization (M2MS), where a single model could process
documents in any language and generate their summaries also in any language. As
the first step towards M2MS, we conduct preliminary studies to show that M2MS
can better transfer task knowledge across different languages than MLS and CLS.
Furthermore, we propose Pisces, a pre-trained M2MS model that learns language
modeling, cross-lingual ability and summarization ability via three-stage
pre-training. Experimental results indicate that our Pisces significantly
outperforms the state-of-the-art baselines, especially in the zero-shot
directions, where there is no training data from the source-language documents
to the target-language summaries.
Related papers
- Think Carefully and Check Again! Meta-Generation Unlocking LLMs for Low-Resource Cross-Lingual Summarization [108.6908427615402]
Cross-lingual summarization ( CLS) aims to generate a summary for the source text in a different target language.
Currently, instruction-tuned large language models (LLMs) excel at various English tasks.
Recent studies have shown that LLMs' performance on CLS tasks remains unsatisfactory even with few-shot settings.
arXiv Detail & Related papers (2024-10-26T00:39:44Z) - A Survey on Multilingual Large Language Models: Corpora, Alignment, and Bias [5.104497013562654]
We present an overview of MLLMs, covering their evolution, key techniques, and multilingual capacities.
We explore widely utilized multilingual corpora for MLLMs' training and multilingual datasets oriented for downstream tasks.
We discuss bias on MLLMs including its category and evaluation metrics, and summarize the existing debiasing techniques.
arXiv Detail & Related papers (2024-04-01T05:13:56Z) - D$^2$TV: Dual Knowledge Distillation and Target-oriented Vision Modeling
for Many-to-Many Multimodal Summarization [113.72253589338472]
Many-to-many multimodal summarization (M$3$S) task aims to generate summaries in any language with document inputs in any language and the corresponding image sequence.
We propose a dual knowledge distillation and target-oriented vision modeling framework for the M$3$S task.
arXiv Detail & Related papers (2023-05-22T06:47:35Z) - Understanding Translationese in Cross-Lingual Summarization [106.69566000567598]
Cross-lingual summarization (MS) aims at generating a concise summary in a different target language.
To collect large-scale CLS data, existing datasets typically involve translation in their creation.
In this paper, we first confirm that different approaches of constructing CLS datasets will lead to different degrees of translationese.
arXiv Detail & Related papers (2022-12-14T13:41:49Z) - A Variational Hierarchical Model for Neural Cross-Lingual Summarization [85.44969140204026]
Cross-lingual summarization () is to convert a document in one language to a summary in another one.
Existing studies on CLS mainly focus on utilizing pipeline methods or jointly training an end-to-end model.
We propose a hierarchical model for the CLS task, based on the conditional variational auto-encoder.
arXiv Detail & Related papers (2022-03-08T02:46:11Z) - Mixed-Lingual Pre-training for Cross-lingual Summarization [54.4823498438831]
Cross-lingual Summarization aims at producing a summary in the target language for an article in the source language.
We propose a solution based on mixed-lingual pre-training that leverages both cross-lingual tasks like translation and monolingual tasks like masked language models.
Our model achieves an improvement of 2.82 (English to Chinese) and 1.15 (Chinese to English) ROUGE-1 scores over state-of-the-art results.
arXiv Detail & Related papers (2020-10-18T00:21:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.