On the Importance of Sign Labeling: The Hamburg Sign Language Notation
System Case Study
- URL: http://arxiv.org/abs/2302.10768v1
- Date: Thu, 19 Jan 2023 11:11:57 GMT
- Title: On the Importance of Sign Labeling: The Hamburg Sign Language Notation
System Case Study
- Authors: Maria Ferlin and Sylwia Majchrowska and Marta Plantykow and Alicja
Kwa\'sniwska and Agnieszka Miko{\l}ajczyk-Bare{\l}a and Milena Olech and
Jakub Nalepa
- Abstract summary: We analyze the HamNoSys labels provided by various maintainers of open sign language corpora in five sign languages.
Our findings provide valuable insights into the limitations of the current labeling methods.
- Score: 5.849513679510834
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Labeling is the cornerstone of supervised machine learning, which has been
exploited in a plethora of various applications, with sign language recognition
being one of them. However, such algorithms must be fed with a huge amount of
consistently labeled data during the training process to elaborate a
well-generalizing model. In addition, there is a great need for an automated
solution that works with any nationally diversified sign language. Although
there are language-agnostic transcription systems, such as the Hamburg Sign
Language Notation System (HamNoSys) that describe the signer's initial position
and body movement instead of the glosses' meanings, there are still issues with
providing accurate and reliable labels for every real-world use case. In this
context, the industry relies heavily on manual attribution and labeling of the
available video data. In this work, we tackle this issue and thoroughly analyze
the HamNoSys labels provided by various maintainers of open sign language
corpora in five sign languages, in order to examine the challenges encountered
in labeling video data. We also investigate the consistency and objectivity of
HamNoSys-based labels for the purpose of training machine learning models. Our
findings provide valuable insights into the limitations of the current labeling
methods and pave the way for future research on developing more accurate and
efficient solutions for sign language recognition.
Related papers
- Improving Continuous Sign Language Recognition with Cross-Lingual Signs [29.077175863743484]
We study the feasibility of utilizing multilingual sign language corpora to facilitate continuous sign language recognition.
We first build two sign language dictionaries containing isolated signs that appear in two datasets.
Then we identify the sign-to-sign mappings between two sign languages via a well-optimized isolated sign language recognition model.
arXiv Detail & Related papers (2023-08-21T15:58:47Z) - Learnt Contrastive Concept Embeddings for Sign Recognition [33.72708697077754]
We focus on explicitly creating sign embeddings that bridge the gap between sign language and spoken language.
We train a vocabulary of embeddings that are based on the linguistic labels for sign video.
We develop a conceptual similarity loss which is able to utilise word embeddings from NLP methods to create sign embeddings that have better sign language to spoken language correspondence.
arXiv Detail & Related papers (2023-08-18T12:47:18Z) - Label Aware Speech Representation Learning For Language Identification [49.197215416945596]
We propose a novel framework of combining self-supervised representation learning with the language label information for the pre-training task.
This framework, termed as Label Aware Speech Representation (LASR) learning, uses a triplet based objective function to incorporate language labels along with the self-supervised loss function.
arXiv Detail & Related papers (2023-06-07T12:14:16Z) - Imprecise Label Learning: A Unified Framework for Learning with Various Imprecise Label Configurations [91.67511167969934]
imprecise label learning (ILL) is a framework for the unification of learning with various imprecise label configurations.
We demonstrate that ILL can seamlessly adapt to partial label learning, semi-supervised learning, noisy label learning, and, more importantly, a mixture of these settings.
arXiv Detail & Related papers (2023-05-22T04:50:28Z) - Natural Language-Assisted Sign Language Recognition [28.64871971445024]
We propose the Natural Language-Assisted Sign Language Recognition framework.
It exploits semantic information contained in glosses (sign labels) to mitigate the problem of visually indistinguishable signs (VISigns) in sign languages.
Our method achieves state-of-the-art performance on three widely-adopted benchmarks: MSASL, WLASL, and NMFs-CSL.
arXiv Detail & Related papers (2023-03-21T17:59:57Z) - On the Importance of Signer Overlap for Sign Language Detection [65.26091369630547]
We argue that the current benchmark data sets for sign language detection estimate overly positive results that do not generalize well.
We quantify this with a detailed analysis of the effect of signer overlap on current sign detection benchmark data sets.
We propose new data set partitions that are free of overlap and allow for more realistic performance assessment.
arXiv Detail & Related papers (2023-03-19T22:15:05Z) - Scaling up sign spotting through sign language dictionaries [99.50956498009094]
The focus of this work is $textitsign spotting$ - given a video of an isolated sign, our task is to identify $textitwhether$ and $textitwhere$ it has been signed in a continuous, co-articulated sign language video.
We train a model using multiple types of available supervision by: (1) $textitwatching$ existing footage which is sparsely labelled using mouthing cues; (2) $textitreading$ associated subtitles which provide additional translations of the signed content.
We validate the effectiveness of our approach on low
arXiv Detail & Related papers (2022-05-09T10:00:03Z) - Open Source HamNoSys Parser for Multilingual Sign Language Encoding [3.867363075280544]
This paper presents an automated tool to convert HamNoSys annotations into numerical labels.
Our proposed numerical multilabels greatly simplify the structure of HamNoSys annotation without significant loss of gloss meaning.
These numerical multilabels can potentially be used to feed the machine learning models, which would accelerate the development of vision-based sign language recognition.
arXiv Detail & Related papers (2022-04-14T12:33:33Z) - Sign Language Video Retrieval with Free-Form Textual Queries [19.29003565494735]
We introduce the task of sign language retrieval with free-form textual queries.
The objective is to find the signing video in the collection that best matches the written query.
We propose SPOT-ALIGN, a framework for interleaving iterative rounds of sign spotting and feature alignment to expand the scope and scale of available training data.
arXiv Detail & Related papers (2022-01-07T15:22:18Z) - Watch, read and lookup: learning to spot signs from multiple supervisors [99.50956498009094]
Given a video of an isolated sign, our task is to identify whether and where it has been signed in a continuous, co-articulated sign language video.
We train a model using multiple types of available supervision by: (1) watching existing sparsely labelled footage; (2) reading associated subtitles which provide additional weak-supervision; and (3) looking up words in visual sign language dictionaries.
These three tasks are integrated into a unified learning framework using the principles of Noise Contrastive Estimation and Multiple Instance Learning.
arXiv Detail & Related papers (2020-10-08T14:12:56Z) - BSL-1K: Scaling up co-articulated sign language recognition using
mouthing cues [106.21067543021887]
We show how to use mouthing cues from signers to obtain high-quality annotations from video data.
The BSL-1K dataset is a collection of British Sign Language (BSL) signs of unprecedented scale.
arXiv Detail & Related papers (2020-07-23T16:59:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.