English to Bangla Machine Translation Using Recurrent Neural Network
- URL: http://arxiv.org/abs/2106.07225v1
- Date: Mon, 14 Jun 2021 08:26:50 GMT
- Title: English to Bangla Machine Translation Using Recurrent Neural Network
- Authors: Shaykh Siddique, Tahmid Ahmed, Md. Rifayet Azam Talukder, and Md.
Mohsin Uddin
- Abstract summary: This paper describes an architecture of English to Bangla machine translation system.
The model uses a knowledge-based context vector for the mapping of English and Bangla words.
The approach of the model outperforms the previous state-of-the-art systems in terms of cross-entropy loss metrics.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The applications of recurrent neural networks in machine translation are
increasing in natural language processing. Besides other languages, Bangla
language contains a large amount of vocabulary. Improvement of English to
Bangla machine translation would be a significant contribution to Bangla
Language processing. This paper describes an architecture of English to Bangla
machine translation system. The system has been implemented with the
encoder-decoder recurrent neural network. The model uses a knowledge-based
context vector for the mapping of English and Bangla words. Performances of the
model based on activation functions are measured here. The best performance is
achieved for the linear activation function in encoder layer and the tanh
activation function in decoder layer. From the execution of GRU and LSTM layer,
GRU performed better than LSTM. The attention layers are enacted with softmax
and sigmoid activation function. The approach of the model outperforms the
previous state-of-the-art systems in terms of cross-entropy loss metrics. The
reader can easily find out the structure of the machine translation of English
to Bangla and the efficient activation functions from the paper.
Related papers
- Quantum-RAG and PunGPT2: Advancing Low-Resource Language Generation and Retrieval for the Punjabi Language [0.0]
We present PunGPT2, the first fully open-source suite of Punjabi large language models.<n>We also present Pun-RAG, a retrieval-augmented generation framework combining PunGPT2 with a dense FAISS retriever.<n>We propose Quantum-RAG, a novel hybrid retrieval system that fuses sparse (BM25) and dense methods with quantum-inspired semantic matching.
arXiv Detail & Related papers (2025-08-03T21:03:22Z) - Low-Resource Neural Machine Translation Using Recurrent Neural Networks and Transfer Learning: A Case Study on English-to-Igbo [0.0]
We develop Neural Machine Translation (NMT) and Transformer-based transfer learning models for English-to-Igbo translation.
Our models are trained on a curated and benchmarked dataset compiled from Bible corpora, local news, Wikipedia articles, and Common Crawl.
With transfer learning, we observe a performance gain of +4.83 BLEU points, reaching an estimated translation accuracy of 70%.
arXiv Detail & Related papers (2025-04-24T05:02:26Z) - Training Neural Networks as Recognizers of Formal Languages [87.06906286950438]
Formal language theory pertains specifically to recognizers.
It is common to instead use proxy tasks that are similar in only an informal sense.
We correct this mismatch by training and evaluating neural networks directly as binary classifiers of strings.
arXiv Detail & Related papers (2024-11-11T16:33:25Z) - Interpretable Language Modeling via Induction-head Ngram Models [74.26720927767398]
We propose Induction-head ngram models (Induction-Gram) to bolster modern ngram models with a hand-engineered "induction head"
This induction head uses a custom neural similarity metric to efficiently search the model's input context for potential next-word completions.
Experiments show that this simple method significantly improves next-word prediction over baseline interpretable models.
arXiv Detail & Related papers (2024-10-31T12:33:26Z) - Hindi to English: Transformer-Based Neural Machine Translation [0.0]
We have developed a Machine Translation (NMT) system by training the Transformer model to translate texts from Indian Language Hindi to English.
We implemented back-translation to augment the training data and for creating the vocabulary.
This led us to achieve a state-of-the-art BLEU score of 24.53 on the test set of IIT Bombay English-Hindi Corpus.
arXiv Detail & Related papers (2023-09-23T00:00:09Z) - LAMASSU: Streaming Language-Agnostic Multilingual Speech Recognition and
Translation Using Neural Transducers [71.76680102779765]
Automatic speech recognition (ASR) and speech translation (ST) can both use neural transducers as the model structure.
We propose LAMASSU, a streaming language-agnostic multilingual speech recognition and translation model using neural transducers.
arXiv Detail & Related papers (2022-11-05T04:03:55Z) - Summarize and Generate to Back-translate: Unsupervised Translation of
Programming Languages [86.08359401867577]
Back-translation is widely known for its effectiveness for neural machine translation when little to no parallel data is available.
We propose performing back-translation via code summarization and generation.
We show that our proposed approach performs competitively with state-of-the-art methods.
arXiv Detail & Related papers (2022-05-23T08:20:41Z) - Improving English to Sinhala Neural Machine Translation using
Part-of-Speech Tag [1.1470070927586016]
Most people in Sri Lanka are unable to read and understand English properly.
There is a huge requirement of translating English content to local languages to share information among locals.
arXiv Detail & Related papers (2022-02-17T19:45:50Z) - Cross-lingual Transferring of Pre-trained Contextualized Language Models [73.97131976850424]
We propose a novel cross-lingual model transferring framework for PrLMs: TreLM.
To handle the symbol order and sequence length differences between languages, we propose an intermediate TRILayer" structure.
We show the proposed framework significantly outperforms language models trained from scratch with limited data in both performance and efficiency.
arXiv Detail & Related papers (2021-07-27T06:51:13Z) - End-to-End Natural Language Understanding Pipeline for Bangla
Conversational Agents [0.43012765978447565]
We propose a novel approach to build a business assistant which can communicate in Bangla and Bangla Transliteration in English with high confidence consistently.
We use Rasa Open Source Framework, fastText embeddings, Polyglot embeddings, Flask, and other systems as building blocks.
We present a pipeline for intent classification and entity extraction which achieves reasonable performance.
arXiv Detail & Related papers (2021-07-12T16:09:22Z) - Paraphrastic Representations at Scale [134.41025103489224]
We release trained models for English, Arabic, German, French, Spanish, Russian, Turkish, and Chinese languages.
We train these models on large amounts of data, achieving significantly improved performance from the original papers.
arXiv Detail & Related papers (2021-04-30T16:55:28Z) - Neural Machine Translation model for University Email Application [1.4731169524644787]
A state-of-the-art Sequence-to-Sequence Neural Network for ML -> EN and EN -> ML translations is compared with Google Translate.
The low BLEU score of Google Translation indicates that the application based regional models are better.
arXiv Detail & Related papers (2020-07-20T15:05:16Z) - HausaMT v1.0: Towards English-Hausa Neural Machine Translation [0.012691047660244334]
We build a baseline model for English-Hausa machine translation.
The Hausa language is the second largest Afro-Asiatic language in the world after Arabic.
arXiv Detail & Related papers (2020-06-09T02:08:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.