Example-Based Machine Translation from Text to a Hierarchical
Representation of Sign Language
- URL: http://arxiv.org/abs/2205.03314v1
- Date: Fri, 6 May 2022 15:48:43 GMT
- Title: Example-Based Machine Translation from Text to a Hierarchical
Representation of Sign Language
- Authors: \'Elise Bertin-Lem\'ee, Annelies Braffort, Camille Challant, Claire
Danet, Michael Filhol
- Abstract summary: This article presents an original method for Text-to-Sign Translation.
It compensates data scarcity using a domain-specific parallel corpus of alignments between text and hierarchical formal descriptions of Sign Language videos in AZee.
Based on the detection of similarities present in the source text, the proposed algorithm exploits matches and substitutions of aligned segments to build multiple candidate translations.
The resulting translations are in the form of AZee expressions, designed to be used as input to avatar systems.
- Score: 1.3999481573773074
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This article presents an original method for Text-to-Sign Translation. It
compensates data scarcity using a domain-specific parallel corpus of alignments
between text and hierarchical formal descriptions of Sign Language videos in
AZee. Based on the detection of similarities present in the source text, the
proposed algorithm recursively exploits matches and substitutions of aligned
segments to build multiple candidate translations for a novel statement. This
helps preserving Sign Language structures as much as possible before falling
back on literal translations too quickly, in a generative way. The resulting
translations are in the form of AZee expressions, designed to be used as input
to avatar synthesis systems. We present a test set tailored to showcase its
potential for expressiveness and generation of idiomatic target language, and
observed limitations. This work finally opens prospects on how to evaluate
translation and linguistic aspects, such as accuracy and grammatical fluency.
Related papers
- From Rule-Based Models to Deep Learning Transformers Architectures for Natural Language Processing and Sign Language Translation Systems: Survey, Taxonomy and Performance Evaluation [0.0]
There is a pressing need for an efficient, signs-driven, integrated end-to-end translation system.
We present the requirements of a real-time Quality-of-Service sign language ma-chine translation system underpinned by accurate deep learning algorithms.
arXiv Detail & Related papers (2024-08-27T07:11:45Z) - Explicit Syntactic Guidance for Neural Text Generation [45.60838824233036]
Generative Grammar suggests that humans generate natural language texts by learning language grammar.
We propose a syntax-guided generation schema, which generates the sequence guided by a constituency parse tree in a top-down direction.
Experiments on paraphrase generation and machine translation show that the proposed method outperforms autoregressive baselines.
arXiv Detail & Related papers (2023-06-20T12:16:31Z) - Decomposed Prompting for Machine Translation Between Related Languages
using Large Language Models [55.35106713257871]
We introduce DecoMT, a novel approach of few-shot prompting that decomposes the translation process into a sequence of word chunk translations.
We show that DecoMT outperforms the strong few-shot prompting BLOOM model with an average improvement of 8 chrF++ scores across the examined languages.
arXiv Detail & Related papers (2023-05-22T14:52:47Z) - Dual-Alignment Pre-training for Cross-lingual Sentence Embedding [79.98111074307657]
We propose a dual-alignment pre-training (DAP) framework for cross-lingual sentence embedding.
We introduce a novel representation translation learning (RTL) task, where the model learns to use one-side contextualized token representation to reconstruct its translation counterpart.
Our approach can significantly improve sentence embedding.
arXiv Detail & Related papers (2023-05-16T03:53:30Z) - Multilingual Extraction and Categorization of Lexical Collocations with
Graph-aware Transformers [86.64972552583941]
We put forward a sequence tagging BERT-based model enhanced with a graph-aware transformer architecture, which we evaluate on the task of collocation recognition in context.
Our results suggest that explicitly encoding syntactic dependencies in the model architecture is helpful, and provide insights on differences in collocation typification in English, Spanish and French.
arXiv Detail & Related papers (2022-05-23T16:47:37Z) - Unsupervised Word Translation Pairing using Refinement based Point Set
Registration [8.568050813210823]
Cross-lingual alignment of word embeddings play an important role in knowledge transfer across languages.
Current unsupervised approaches rely on similarities in geometric structure of word embedding spaces across languages.
This paper proposes BioSpere, a novel framework for unsupervised mapping of bi-lingual word embeddings onto a shared vector space.
arXiv Detail & Related papers (2020-11-26T09:51:29Z) - Neural Syntactic Preordering for Controlled Paraphrase Generation [57.5316011554622]
Our work uses syntactic transformations to softly "reorder'' the source sentence and guide our neural paraphrasing model.
First, given an input sentence, we derive a set of feasible syntactic rearrangements using an encoder-decoder model.
Next, we use each proposed rearrangement to produce a sequence of position embeddings, which encourages our final encoder-decoder paraphrase model to attend to the source words in a particular order.
arXiv Detail & Related papers (2020-05-05T09:02:25Z) - Sign Language Transformers: Joint End-to-end Sign Language Recognition
and Translation [59.38247587308604]
We introduce a novel transformer based architecture that jointly learns Continuous Sign Language Recognition and Translation.
We evaluate the recognition and translation performances of our approaches on the challenging RWTH-PHOENIX-Weather-2014T dataset.
Our translation networks outperform both sign video to spoken language and gloss to spoken language translation models.
arXiv Detail & Related papers (2020-03-30T21:35:09Z) - Multilingual Alignment of Contextual Word Representations [49.42244463346612]
BERT exhibits significantly improved zero-shot performance on XNLI compared to the base model.
We introduce a contextual version of word retrieval and show that it correlates well with downstream zero-shot transfer.
These results support contextual alignment as a useful concept for understanding large multilingual pre-trained models.
arXiv Detail & Related papers (2020-02-10T03:27:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.