Optical Character Recognition and Transcription of Berber Signs from
Images in a Low-Resource Language Amazigh
- URL: http://arxiv.org/abs/2303.13549v1
- Date: Tue, 21 Mar 2023 21:38:44 GMT
- Title: Optical Character Recognition and Transcription of Berber Signs from
Images in a Low-Resource Language Amazigh
- Authors: Levi Corallo and Aparna S. Varde
- Abstract summary: The Berber, or Amazigh language family is a low-resource North African vernacular language spoken by the indigenous Berber ethnic group.
It has its own unique alphabet called Tifinagh used across Berber communities in Morocco, Algeria, and others.
The Afroasiatic language Berber is spoken by 14 million people, yet lacks adequate representation in education, research, web applications etc.
- Score: 2.132096006921048
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The Berber, or Amazigh language family is a low-resource North African
vernacular language spoken by the indigenous Berber ethnic group. It has its
own unique alphabet called Tifinagh used across Berber communities in Morocco,
Algeria, and others. The Afroasiatic language Berber is spoken by 14 million
people, yet lacks adequate representation in education, research, web
applications etc. For instance, there is no option of translation to or from
Amazigh / Berber on Google Translate, which hosts over 100 languages today.
Consequently, we do not find specialized educational apps, L2 (2nd language
learner) acquisition, automated language translation, and remote-access
facilities enabled in Berber. Motivated by this background, we propose a
supervised approach called DaToBS for Detection and Transcription of Berber
Signs. The DaToBS approach entails the automatic recognition and transcription
of Tifinagh characters from signs in photographs of natural environments. This
is achieved by self-creating a corpus of 1862 pre-processed character images;
curating the corpus with human-guided annotation; and feeding it into an OCR
model via the deployment of CNN for deep learning based on computer vision
models. We deploy computer vision modeling (rather than language models)
because there are pictorial symbols in this alphabet, this deployment being a
novel aspect of our work. The DaToBS experimentation and analyses yield over 92
percent accuracy in our research. To the best of our knowledge, ours is among
the first few works in the automated transcription of Berber signs from
roadside images with deep learning, yielding high accuracy. This can pave the
way for developing pedagogical applications in the Berber language, thereby
addressing an important goal of outreach to underrepresented communities via AI
in education.
Related papers
- Harnessing Transfer Learning from Swahili: Advancing Solutions for Comorian Dialects [0.0]
We aim to pioneer NLP technologies for Comorian, a group of four languages or dialects belonging to the Bantu family.
Our approach is motivated by the hypothesis that if a human can understand a different language from their native language with little or no effort, it would be entirely possible to model this process on a machine.
arXiv Detail & Related papers (2024-12-09T22:47:41Z) - AceGPT, Localizing Large Language Models in Arabic [73.39989503874634]
The paper proposes a comprehensive solution that includes pre-training with Arabic texts, Supervised Fine-Tuning (SFT) utilizing native Arabic instructions, and GPT-4 responses in Arabic.
The goal is to cultivate culturally cognizant and value-aligned Arabic LLMs capable of accommodating the diverse, application-specific needs of Arabic-speaking communities.
arXiv Detail & Related papers (2023-09-21T13:20:13Z) - NusaWrites: Constructing High-Quality Corpora for Underrepresented and
Extremely Low-Resource Languages [54.808217147579036]
We conduct a case study on Indonesian local languages.
We compare the effectiveness of online scraping, human translation, and paragraph writing by native speakers in constructing datasets.
Our findings demonstrate that datasets generated through paragraph writing by native speakers exhibit superior quality in terms of lexical diversity and cultural content.
arXiv Detail & Related papers (2023-09-19T14:42:33Z) - Lip Reading for Low-resource Languages by Learning and Combining General
Speech Knowledge and Language-specific Knowledge [57.38948190611797]
This paper proposes a novel lip reading framework, especially for low-resource languages.
Since low-resource languages do not have enough video-text paired data to train the model, it is regarded as challenging to develop lip reading models for low-resource languages.
arXiv Detail & Related papers (2023-08-18T05:19:03Z) - Cross-Lingual NER for Financial Transaction Data in Low-Resource
Languages [70.25418443146435]
We propose an efficient modeling framework for cross-lingual named entity recognition in semi-structured text data.
We employ two independent datasets of SMSs in English and Arabic, each carrying semi-structured banking transaction information.
With access to only 30 labeled samples, our model can generalize the recognition of merchants, amounts, and other fields from English to Arabic.
arXiv Detail & Related papers (2023-07-16T00:45:42Z) - Design of Arabic Sign Language Recognition Model [0.0]
The model is tested on ArASL 2018, consisting of 54,000 images for 32 alphabet signs gathered from 40 signers.
The future work will be a model that converts Arabic sign language into Arabic text.
arXiv Detail & Related papers (2023-01-06T19:19:25Z) - Phonemic Representation and Transcription for Speech to Text
Applications for Under-resourced Indigenous African Languages: The Case of
Kiswahili [0.0]
It has emerged that several African indigenous languages, including Kiswahili, are technologically under-resourced.
This paper explores the transcription process and the development of a Kiswahili speech corpus.
It provides an updated Kiswahili phoneme dictionary for the ASR model that was created using the CMU Sphinx speech recognition toolbox.
arXiv Detail & Related papers (2022-10-29T09:04:09Z) - No Language Left Behind: Scaling Human-Centered Machine Translation [69.28110770760506]
We create datasets and models aimed at narrowing the performance gap between low and high-resource languages.
We propose multiple architectural and training improvements to counteract overfitting while training on thousands of tasks.
Our model achieves an improvement of 44% BLEU relative to the previous state-of-the-art.
arXiv Detail & Related papers (2022-07-11T07:33:36Z) - Can Character-based Language Models Improve Downstream Task Performance
in Low-Resource and Noisy Language Scenarios? [0.0]
We focus on North-African colloquial dialectal Arabic written using an extension of the Latin script, called NArabizi.
We show that a character-based model trained on only 99k sentences of NArabizi and fined-tuned on a small treebank leads to performance close to those obtained with the same architecture pre-trained on large multilingual and monolingual models.
arXiv Detail & Related papers (2021-10-26T14:59:16Z) - DziriBERT: a Pre-trained Language Model for the Algerian Dialect [2.064612766965483]
We study the Algerian dialect which has several specificities that make the use of Arabic or multilingual models inappropriate.
To address this issue, we collected more than one Million Algerian tweets, and pre-trained the first Algerian language model: DziriBERT.
arXiv Detail & Related papers (2021-09-25T11:51:35Z) - Read Like Humans: Autonomous, Bidirectional and Iterative Language
Modeling for Scene Text Recognition [80.446770909975]
Linguistic knowledge is of great benefit to scene text recognition.
How to effectively model linguistic rules in end-to-end deep networks remains a research challenge.
We propose an autonomous, bidirectional and iterative ABINet for scene text recognition.
arXiv Detail & Related papers (2021-03-11T06:47:45Z) - Can Multilingual Language Models Transfer to an Unseen Dialect? A Case
Study on North African Arabizi [2.76240219662896]
We study the ability of multilingual language models to process an unseen dialect.
We take user generated North-African Arabic as our case study.
We show in zero-shot and unsupervised adaptation scenarios that multilingual language models are able to transfer to such an unseen dialect.
arXiv Detail & Related papers (2020-05-01T11:29:23Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.