Revisiting Pre-trained Language Models and their Evaluation for Arabic
Natural Language Understanding
- URL: http://arxiv.org/abs/2205.10687v1
- Date: Sat, 21 May 2022 22:38:19 GMT
- Title: Revisiting Pre-trained Language Models and their Evaluation for Arabic
Natural Language Understanding
- Authors: Abbas Ghaddar, Yimeng Wu, Sunyam Bagga, Ahmad Rashid, Khalil Bibi,
Mehdi Rezagholizadeh, Chao Xing, Yasheng Wang, Duan Xinyu, Zhefeng Wang,
Baoxing Huai, Xin Jiang, Qun Liu, Philippe Langlais
- Abstract summary: Existing Arabic PLMs are not well-explored and their pre-trainig can be improved significantly.
There is a lack of systematic and reproducible evaluation of these models in the literature.
We show that our models significantly outperform existing Arabic PLMs and achieve a new state-of-the-art performance on discriminative and generative Arabic NLU and NLG tasks.
- Score: 44.048072667378115
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: There is a growing body of work in recent years to develop pre-trained
language models (PLMs) for the Arabic language. This work concerns addressing
two major problems in existing Arabic PLMs which constraint progress of the
Arabic NLU and NLG fields.First, existing Arabic PLMs are not well-explored and
their pre-trainig can be improved significantly using a more methodical
approach. Second, there is a lack of systematic and reproducible evaluation of
these models in the literature. In this work, we revisit both the pre-training
and evaluation of Arabic PLMs. In terms of pre-training, we explore improving
Arabic LMs from three perspectives: quality of the pre-training data, size of
the model, and incorporating character-level information. As a result, we
release three new Arabic BERT-style models ( JABER, Char-JABER, and SABER), and
two T5-style models (AT5S and AT5B). In terms of evaluation, we conduct a
comprehensive empirical study to systematically evaluate the performance of
existing state-of-the-art models on ALUE that is a leaderboard-powered
benchmark for Arabic NLU tasks, and on a subset of the ARGEN benchmark for
Arabic NLG tasks. We show that our models significantly outperform existing
Arabic PLMs and achieve a new state-of-the-art performance on discriminative
and generative Arabic NLU and NLG tasks. Our models and source code to
reproduce of results will be made available shortly.
Related papers
- ALLaM: Large Language Models for Arabic and English [9.881560166505452]
We present ALLaM: Arabic Large Language Model, a series of large language models to support the ecosystem of Arabic Language Technologies (ALT)
Our autoregressive decoder-only architecture models demonstrate how second-language acquisition via vocabulary expansion and pretraining can steer a model towards a new language (Arabic) without any catastrophic forgetting in the original language (English)
We show that extensive alignment with human preferences can significantly enhance the performance of a language model compared to models of a larger scale with lower quality alignment.
arXiv Detail & Related papers (2024-07-22T05:35:17Z) - GemmAr: Enhancing LLMs Through Arabic Instruction-Tuning [0.0]
We introduce InstAr-500k, a new Arabic instruction dataset created by generating and collecting content.
We assess this dataset by fine-tuning an open-source Gemma-7B model on several downstream tasks to improve its functionality.
Based on multiple evaluations, our fine-tuned model achieves excellent performance on several Arabic NLP benchmarks.
arXiv Detail & Related papers (2024-07-02T10:43:49Z) - ArabicMMLU: Assessing Massive Multitask Language Understanding in Arabic [51.922112625469836]
We present datasetname, the first multi-task language understanding benchmark for the Arabic language.
Our data comprises 40 tasks and 14,575 multiple-choice questions in Modern Standard Arabic (MSA) and is carefully constructed by collaborating with native speakers in the region.
Our evaluations of 35 models reveal substantial room for improvement, particularly among the best open-source models.
arXiv Detail & Related papers (2024-02-20T09:07:41Z) - On the importance of Data Scale in Pretraining Arabic Language Models [46.431706010614334]
We conduct a comprehensive study on the role of data in Arabic Pretrained Language Models (PLMs)
We reassess the performance of a suite of state-of-the-art Arabic PLMs by retraining them on massive-scale, high-quality Arabic corpora.
Our analysis strongly suggests that pretraining data by far is the primary contributor to performance, surpassing other factors.
arXiv Detail & Related papers (2024-01-15T15:11:15Z) - AceGPT, Localizing Large Language Models in Arabic [73.39989503874634]
The paper proposes a comprehensive solution that includes pre-training with Arabic texts, Supervised Fine-Tuning (SFT) utilizing native Arabic instructions, and GPT-4 responses in Arabic.
The goal is to cultivate culturally cognizant and value-aligned Arabic LLMs capable of accommodating the diverse, application-specific needs of Arabic-speaking communities.
arXiv Detail & Related papers (2023-09-21T13:20:13Z) - Jais and Jais-chat: Arabic-Centric Foundation and Instruction-Tuned Open
Generative Large Language Models [57.76998376458017]
We introduce Jais and Jais-chat, new state-of-the-art Arabic-centric foundation and instruction-tuned open generative large language models (LLMs)
The models are based on the GPT-3 decoder-only architecture and are pretrained on a mixture of Arabic and English texts.
We provide a detailed description of the training, the tuning, the safety alignment, and the evaluation of the models.
arXiv Detail & Related papers (2023-08-30T17:07:17Z) - Cross-Lingual NER for Financial Transaction Data in Low-Resource
Languages [70.25418443146435]
We propose an efficient modeling framework for cross-lingual named entity recognition in semi-structured text data.
We employ two independent datasets of SMSs in English and Arabic, each carrying semi-structured banking transaction information.
With access to only 30 labeled samples, our model can generalize the recognition of merchants, amounts, and other fields from English to Arabic.
arXiv Detail & Related papers (2023-07-16T00:45:42Z) - JABER: Junior Arabic BERt [37.174723137868675]
We present JABER, Junior Arabic BERt, our pretrained language model prototype dedicated for Arabic.
We conduct an empirical study to systematically evaluate the performance of models across a diverse set of existing Arabic NLU tasks.
arXiv Detail & Related papers (2021-12-08T15:19:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.