Enhanced Auto Language Prediction with Dictionary Capsule -- A Novel
Approach
- URL: http://arxiv.org/abs/2403.05982v1
- Date: Sat, 9 Mar 2024 18:43:48 GMT
- Title: Enhanced Auto Language Prediction with Dictionary Capsule -- A Novel
Approach
- Authors: Pinni Venkata Abhiram, Ananya Rathore, Abhir Mirikar, Hari Krishna S,
Sheena Christabel Pravin, Vishwanath Kamath Pethri, Manjunath Lokanath
Belgod, Reetika Gupta, K Muthukumaran
- Abstract summary: The paper presents a novel Auto Language Prediction Dictionary Capsule framework for language prediction and machine translation.
The model uses a combination of neural networks and symbolic representations to predict the language of a given input text and then translate it to a target language using pre-built dictionaries.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The paper presents a novel Auto Language Prediction Dictionary Capsule
(ALPDC) framework for language prediction and machine translation. The model
uses a combination of neural networks and symbolic representations to predict
the language of a given input text and then translate it to a target language
using pre-built dictionaries. This research work also aims to translate the
text of various languages to its literal meaning in English. The proposed model
achieves state-of-the-art results on several benchmark datasets and
significantly improves translation accuracy compared to existing methods. The
results show the potential of the proposed method for practical use in
multilingual communication and natural language processing tasks.
Related papers
- Decomposed Prompting for Machine Translation Between Related Languages
using Large Language Models [55.35106713257871]
We introduce DecoMT, a novel approach of few-shot prompting that decomposes the translation process into a sequence of word chunk translations.
We show that DecoMT outperforms the strong few-shot prompting BLOOM model with an average improvement of 8 chrF++ scores across the examined languages.
arXiv Detail & Related papers (2023-05-22T14:52:47Z) - Few-shot Subgoal Planning with Language Models [58.11102061150875]
We show that language priors encoded in pre-trained language models allow us to infer fine-grained subgoal sequences.
In contrast to recent methods which make strong assumptions about subgoal supervision, our experiments show that language models can infer detailed subgoal sequences without any fine-tuning.
arXiv Detail & Related papers (2022-05-28T01:03:30Z) - Constrained Language Models Yield Few-Shot Semantic Parsers [73.50960967598654]
We explore the use of large pretrained language models as few-shot semantics.
The goal in semantic parsing is to generate a structured meaning representation given a natural language input.
We use language models to paraphrase inputs into a controlled sublanguage resembling English that can be automatically mapped to a target meaning representation.
arXiv Detail & Related papers (2021-04-18T08:13:06Z) - Revisiting Language Encoding in Learning Multilingual Representations [70.01772581545103]
We propose a new approach called Cross-lingual Language Projection (XLP) to replace language embedding.
XLP projects the word embeddings into language-specific semantic space, and then the projected embeddings will be fed into the Transformer model.
Experiments show that XLP can freely and significantly boost the model performance on extensive multilingual benchmark datasets.
arXiv Detail & Related papers (2021-02-16T18:47:10Z) - Multi-Sense Language Modelling [19.396806939258806]
We propose a language model which not only predicts the next word, but also its sense in context.
This higher prediction granularity may be useful for end tasks such as assistive writing.
For sense prediction, we utilise a Graph Attention Network, which encodes definitions and example uses of word senses.
arXiv Detail & Related papers (2020-12-10T16:06:05Z) - Are Some Words Worth More than Others? [3.5598388686985354]
We propose two new intrinsic evaluation measures within the framework of a simple word prediction task.
We evaluate several commonly-used large English language models using our proposed metrics.
arXiv Detail & Related papers (2020-10-12T23:12:11Z) - Grounded Compositional Outputs for Adaptive Language Modeling [59.02706635250856]
A language model's vocabulary$-$typically selected before training and permanently fixed later$-$affects its size.
We propose a fully compositional output embedding layer for language models.
To our knowledge, the result is the first word-level language model with a size that does not depend on the training vocabulary.
arXiv Detail & Related papers (2020-09-24T07:21:14Z) - Learning Spoken Language Representations with Neural Lattice Language
Modeling [39.50831917042577]
We propose a framework that trains neural lattice language models to provide contextualized representations for spoken language understanding tasks.
The proposed two-stage pre-training approach reduces the demands of speech data and has better efficiency.
arXiv Detail & Related papers (2020-07-06T10:38:03Z) - On the Importance of Word Order Information in Cross-lingual Sequence
Labeling [80.65425412067464]
Cross-lingual models that fit into the word order of the source language might fail to handle target languages.
We investigate whether making models insensitive to the word order of the source language can improve the adaptation performance in target languages.
arXiv Detail & Related papers (2020-01-30T03:35:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.