ASL-Skeleton3D and ASL-Phono: Two Novel Datasets for the American Sign
Language
- URL: http://arxiv.org/abs/2201.02065v1
- Date: Thu, 6 Jan 2022 14:10:03 GMT
- Title: ASL-Skeleton3D and ASL-Phono: Two Novel Datasets for the American Sign
Language
- Authors: Cleison Correia de Amorim and Cleber Zanchettin
- Abstract summary: The Sign Language Recognition field aims to bridge the gap between users and non-users of sign languages.
This paper contributes by introducing two new datasets for the American Sign Language.
- Score: 3.974836127188525
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sign language is an essential resource enabling access to communication and
proper socioemotional development for individuals suffering from disabling
hearing loss. As this population is expected to reach 700 million by 2050, the
importance of the language becomes even more essential as it plays a critical
role to ensure the inclusion of such individuals in society. The Sign Language
Recognition field aims to bridge the gap between users and non-users of sign
languages. However, the scarcity in quantity and quality of datasets is one of
the main challenges limiting the exploration of novel approaches that could
lead to significant advancements in this research area. Thus, this paper
contributes by introducing two new datasets for the American Sign Language: the
first is composed of the three-dimensional representation of the signers and,
the second, by an unprecedented linguistics-based representation containing a
set of phonological attributes of the signs.
Related papers
- Reconsidering Sentence-Level Sign Language Translation [2.099922236065961]
We show that for 33% of sentences in our sample, our fluent Deaf signer annotators were only able to understand key parts of the clip in light of discourse-level context.
These results underscore the importance of understanding and sanity checking examples when adapting machine learning to new domains.
arXiv Detail & Related papers (2024-06-16T19:19:54Z) - SignDiff: Learning Diffusion Models for American Sign Language
Production [27.899654531461238]
The field of Sign Language Production lacked a large-scale, pre-trained model based on deep learning for continuous American Sign Language (ASL) production in the past decade.
We propose SignDiff, a dual-condition diffusion pre-training model that can generate human sign language speakers from a skeleton pose.
Our ASLP method proposes two new improved modules and a new loss function to improve the accuracy and quality of sign language skeletal posture.
arXiv Detail & Related papers (2023-08-30T15:14:56Z) - Slovo: Russian Sign Language Dataset [83.93252084624997]
This paper presents the Russian Sign Language (RSL) video dataset Slovo, produced using crowdsourcing platforms.
The dataset contains 20,000 FullHD recordings, divided into 1,000 classes of isolated RSL gestures received by 194 signers.
arXiv Detail & Related papers (2023-05-23T21:00:42Z) - LSA-T: The first continuous Argentinian Sign Language dataset for Sign
Language Translation [52.87578398308052]
Sign language translation (SLT) is an active field of study that encompasses human-computer interaction, computer vision, natural language processing and machine learning.
This paper presents the first continuous Argentinian Sign Language (LSA) dataset.
It contains 14,880 sentence level videos of LSA extracted from the CN Sordos YouTube channel with labels and keypoints annotations for each signer.
arXiv Detail & Related papers (2022-11-14T14:46:44Z) - WLASL-LEX: a Dataset for Recognising Phonological Properties in American
Sign Language [2.814213966364155]
We build a large-scale dataset of American Sign Language signs annotated with six different phonological properties.
We investigate whether data-driven end-to-end and feature-based approaches can be optimised to automatically recognise these properties.
arXiv Detail & Related papers (2022-03-11T17:21:24Z) - All You Need In Sign Language Production [50.3955314892191]
Sign language recognition and production need to cope with some critical challenges.
We present an introduction to the Deaf culture, Deaf centers, psychological perspective of sign language.
Also, the backbone architectures and methods in SLP are briefly introduced and the proposed taxonomy on SLP is presented.
arXiv Detail & Related papers (2022-01-05T13:45:09Z) - Including Signed Languages in Natural Language Processing [48.62744923724317]
Signed languages are the primary means of communication for many deaf and hard of hearing individuals.
This position paper calls on the NLP community to include signed languages as a research area with high social and scientific impact.
arXiv Detail & Related papers (2021-05-11T17:37:55Z) - AM2iCo: Evaluating Word Meaning in Context across Low-ResourceLanguages
with Adversarial Examples [51.048234591165155]
We present AM2iCo, Adversarial and Multilingual Meaning in Context.
It aims to faithfully assess the ability of state-of-the-art (SotA) representation models to understand the identity of word meaning in cross-lingual contexts.
Results reveal that current SotA pretrained encoders substantially lag behind human performance.
arXiv Detail & Related papers (2021-04-17T20:23:45Z) - Sign Language Production: A Review [51.07720650677784]
Sign Language is the dominant yet non-primary form of communication language used in the deaf and hearing-impaired community.
To make an easy and mutual communication between the hearing-impaired and the hearing communities, building a robust system capable of translating the spoken language into sign language is fundamental.
To this end, sign language recognition and production are two necessary parts for making such a two-way system.
arXiv Detail & Related papers (2021-03-29T19:38:22Z) - How2Sign: A Large-scale Multimodal Dataset for Continuous American Sign
Language [37.578776156503906]
How2Sign is a multimodal and multiview continuous American Sign Language (ASL) dataset.
It consists of a parallel corpus of more than 80 hours of sign language videos and a set of corresponding modalities including speech, English transcripts, and depth.
A three-hour subset was recorded in the Panoptic studio enabling detailed 3D pose estimation.
arXiv Detail & Related papers (2020-08-18T20:22:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.