Interpretation of Swedish Sign Language using Convolutional Neural
Networks and Transfer Learning
- URL: http://arxiv.org/abs/2010.07827v1
- Date: Thu, 15 Oct 2020 15:34:09 GMT
- Title: Interpretation of Swedish Sign Language using Convolutional Neural
Networks and Transfer Learning
- Authors: Gustaf Halvardsson, Johanna Peterson, C\'esar Soto-Valero, Benoit
Baudry
- Abstract summary: We use Convolutional Neural Networks (CNNs) and transfer learning in order to make computers able to interpret signs of the Swedish Sign Language (SSL) hand alphabet.
Our model consist of the implementation of a pre-trained InceptionV3 network, and the usage of the mini-batch gradient descent optimization algorithm.
The final accuracy of the model, based on 8 study subjects and 9,400 images, is 85%.
- Score: 2.7629216089139934
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The automatic interpretation of sign languages is a challenging task, as it
requires the usage of high-level vision and high-level motion processing
systems for providing accurate image perception. In this paper, we use
Convolutional Neural Networks (CNNs) and transfer learning in order to make
computers able to interpret signs of the Swedish Sign Language (SSL) hand
alphabet. Our model consist of the implementation of a pre-trained InceptionV3
network, and the usage of the mini-batch gradient descent optimization
algorithm. We rely on transfer learning during the pre-training of the model
and its data. The final accuracy of the model, based on 8 study subjects and
9,400 images, is 85%. Our results indicate that the usage of CNNs is a
promising approach to interpret sign languages, and transfer learning can be
used to achieve high testing accuracy despite using a small training dataset.
Furthermore, we describe the implementation details of our model to interpret
signs as a user-friendly web application.
Related papers
- Deep Neural Network-Based Sign Language Recognition: A Comprehensive Approach Using Transfer Learning with Explainability [0.0]
We suggest a novel solution that uses a deep neural network to fully automate sign language recognition.
This methodology integrates sophisticated preprocessing methodologies to optimise the overall performance.
Our model's ability to provide informational clarity was assessed using the SHAP (SHapley Additive exPlanations) method.
arXiv Detail & Related papers (2024-09-11T17:17:44Z) - Enhancing Sign Language Detection through Mediapipe and Convolutional Neural Networks (CNN) [3.192629447369627]
This research combines MediaPipe and CNNs for the efficient and accurate interpretation of ASL dataset.
The accuracy achieved by the model on ASL datasets is 99.12%.
The system will have applications in the communication, education, and accessibility domains.
arXiv Detail & Related papers (2024-06-06T04:05:12Z) - Seeing in Words: Learning to Classify through Language Bottlenecks [59.97827889540685]
Humans can explain their predictions using succinct and intuitive descriptions.
We show that a vision model whose feature representations are text can effectively classify ImageNet images.
arXiv Detail & Related papers (2023-06-29T00:24:42Z) - SignBERT+: Hand-model-aware Self-supervised Pre-training for Sign
Language Understanding [132.78015553111234]
Hand gesture serves as a crucial role during the expression of sign language.
Current deep learning based methods for sign language understanding (SLU) are prone to over-fitting due to insufficient sign data resource.
We propose the first self-supervised pre-trainable SignBERT+ framework with model-aware hand prior incorporated.
arXiv Detail & Related papers (2023-05-08T17:16:38Z) - Sign Language to Text Conversion in Real Time using Transfer Learning [0.0]
We propose a deep learning model trained on the American Sign Language.
There has been an improvement in accuracy from 94% of CNN to 98.7% by Transfer Learning.
arXiv Detail & Related papers (2022-11-13T17:20:19Z) - Reasoning-Modulated Representations [85.08205744191078]
We study a common setting where our task is not purely opaque.
Our approach paves the way for a new class of data-efficient representation learning.
arXiv Detail & Related papers (2021-07-19T13:57:13Z) - Application of Transfer Learning to Sign Language Recognition using an
Inflated 3D Deep Convolutional Neural Network [0.0]
Transfer learning is a technique to utilize a related task with an abundance of data available to help solve a target task lacking sufficient data.
This paper investigates how effectively transfer learning can be applied to isolated sign language recognition.
arXiv Detail & Related papers (2021-02-25T13:37:39Z) - Learning to Learn Parameterized Classification Networks for Scalable
Input Images [76.44375136492827]
Convolutional Neural Networks (CNNs) do not have a predictable recognition behavior with respect to the input resolution change.
We employ meta learners to generate convolutional weights of main networks for various input scales.
We further utilize knowledge distillation on the fly over model predictions based on different input resolutions.
arXiv Detail & Related papers (2020-07-13T04:27:25Z) - Auto-Rectify Network for Unsupervised Indoor Depth Estimation [119.82412041164372]
We establish that the complex ego-motions exhibited in handheld settings are a critical obstacle for learning depth.
We propose a data pre-processing method that rectifies training images by removing their relative rotations for effective learning.
Our results outperform the previous unsupervised SOTA method by a large margin on the challenging NYUv2 dataset.
arXiv Detail & Related papers (2020-06-04T08:59:17Z) - ASL Recognition with Metric-Learning based Lightweight Network [0.0]
We propose a lightweight network for ASL gesture recognition with a performance sufficient for practical applications.
The training code is available as part of Intel OpenVINO Training Extensions.
arXiv Detail & Related papers (2020-04-10T14:41:30Z) - Curriculum By Smoothing [52.08553521577014]
Convolutional Neural Networks (CNNs) have shown impressive performance in computer vision tasks such as image classification, detection, and segmentation.
We propose an elegant curriculum based scheme that smoothes the feature embedding of a CNN using anti-aliasing or low-pass filters.
As the amount of information in the feature maps increases during training, the network is able to progressively learn better representations of the data.
arXiv Detail & Related papers (2020-03-03T07:27:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.