IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural
Language Generation
- URL: http://arxiv.org/abs/2104.08200v1
- Date: Fri, 16 Apr 2021 16:16:44 GMT
- Title: IndoNLG: Benchmark and Resources for Evaluating Indonesian Natural
Language Generation
- Authors: Samuel Cahyawijaya, Genta Indra Winata, Bryan Wilie, Karissa
Vincentio, Xiaohong Li, Adhiguna Kuncoro, Sebastian Ruder, Zhi Yuan Lim,
Syafri Bahar, Masayu Leylia Khodra, Ayu Purwarianti, Pascale Fung
- Abstract summary: We introduce IndoNLG, the first such benchmark for the Indonesian language for natural language generation (NLG)
We provide a vast and clean pre-training corpus of Indonesian, Sundanese, and Javanese datasets called Indo4B-Plus.
We evaluate the effectiveness and efficiency of IndoBART by conducting extensive evaluation on all IndoNLG tasks.
- Score: 45.90242600586664
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A benchmark provides an ecosystem to measure the advancement of models with
standard datasets and automatic and human evaluation metrics. We introduce
IndoNLG, the first such benchmark for the Indonesian language for natural
language generation (NLG). It covers six tasks: summarization, question
answering, open chitchat, as well as three different language-pairs of machine
translation tasks. We provide a vast and clean pre-training corpus of
Indonesian, Sundanese, and Javanese datasets called Indo4B-Plus, which is used
to train our pre-trained NLG model, IndoBART. We evaluate the effectiveness and
efficiency of IndoBART by conducting extensive evaluation on all IndoNLG tasks.
Our findings show that IndoBART achieves competitive performance on Indonesian
tasks with five times fewer parameters compared to the largest multilingual
model in our benchmark, mBART-LARGE (Liu et al., 2020), and an almost 4x and
2.5x faster inference time on the CPU and GPU respectively. We additionally
demonstrate the ability of IndoBART to learn Javanese and Sundanese, and it
achieves decent performance on machine translation tasks.
Related papers
- VLUE: A New Benchmark and Multi-task Knowledge Transfer Learning for Vietnamese Natural Language Understanding [1.813644606477824]
We introduce the first Vietnamese Language Understanding Evaluation (VLUE) benchmark.
The VLUE benchmark encompasses five datasets covering different NLU tasks, including text classification, span extraction, and natural language understanding.
We present CafeBERT, a new state-of-the-art pre-trained model that achieves superior results across all tasks in the VLUE benchmark.
arXiv Detail & Related papers (2024-03-23T16:26:49Z) - IndoToD: A Multi-Domain Indonesian Benchmark For End-to-End
Task-Oriented Dialogue Systems [26.094144160398447]
This paper introduces IndoToD, an end-to-end multi domain ToD benchmark in Indonesian.
We extend two English ToD datasets to Indonesian, comprising four different domains by delexicalization to efficiently reduce the size of annotations.
To ensure a high-quality data collection, we hire native speakers to manually translate the dialogues.
arXiv Detail & Related papers (2023-11-02T03:01:53Z) - NusaCrowd: Open Source Initiative for Indonesian NLP Resources [104.5381571820792]
NusaCrowd is a collaborative initiative to collect and unify existing resources for Indonesian languages.
Our work strives to advance natural language processing (NLP) research for languages that are under-represented despite being widely spoken.
arXiv Detail & Related papers (2022-12-19T17:28:22Z) - NusaX: Multilingual Parallel Sentiment Dataset for 10 Indonesian Local
Languages [100.59889279607432]
We focus on developing resources for languages in Indonesia.
Most languages in Indonesia are categorized as endangered and some are even extinct.
We develop the first-ever parallel resource for 10 low-resource languages in Indonesia.
arXiv Detail & Related papers (2022-05-31T17:03:50Z) - XDBERT: Distilling Visual Information to BERT from Cross-Modal Systems
to Improve Language Understanding [73.24847320536813]
This study explores distilling visual information from pretrained multimodal transformers to pretrained language encoders.
Our framework is inspired by cross-modal encoders' success in visual-language tasks while we alter the learning objective to cater to the language-heavy characteristics of NLU.
arXiv Detail & Related papers (2022-04-15T03:44:00Z) - GLGE: A New General Language Generation Evaluation Benchmark [139.25515221280767]
General Language Generation Evaluation (GLGE) is a new multi-task benchmark for evaluating the generalization capabilities of NLG models.
To encourage research on pretraining and transfer learning on NLG models, we make GLGE publicly available and build a leaderboard with strong baselines.
arXiv Detail & Related papers (2020-11-24T06:59:45Z) - IndoLEM and IndoBERT: A Benchmark Dataset and Pre-trained Language Model
for Indonesian NLP [41.57622648924415]
The Indonesian language is spoken by almost 200 million people and the 10th most spoken language in the world.
Previous work on Indonesian has been hampered by a lack of annotated datasets, a sparsity of language resources, and a lack of resource standardization.
We release the IndoLEM dataset comprising seven tasks for the Indonesian language, spanning morpho-syntax, semantics, and discourse.
We additionally release IndoBERT, a new pre-trained language model for Indonesian, and evaluate it over IndoLEM.
arXiv Detail & Related papers (2020-11-02T01:54:56Z) - IndoNLU: Benchmark and Resources for Evaluating Indonesian Natural
Language Understanding [41.691861010118394]
We introduce the first-ever vast resource for the training, evaluating, and benchmarking on Indonesian natural language understanding tasks.
IndoNLU includes twelve tasks, ranging from single sentence classification to pair-sentences sequence labeling with different levels of complexity.
The datasets for the tasks lie in different domains and styles to ensure task diversity.
We also provide a set of Indonesian pre-trained models (IndoBERT) trained from a large and clean Indonesian dataset Indo4B.
arXiv Detail & Related papers (2020-09-11T12:21:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.