Sign Language Recognition Using Deep Learning
Sign Language Recognition Using Deep Learning
https://doi.org/10.22214/ijraset.2022.40627
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue III Mar 2022- Available at www.ijraset.com
Abstract: Millions of people with speech and hearing impairments communicate with sign languages every day. For hearing-
impaired people, gesture recognition is a natural way of communicating, much like voice recognition is for most people. In this
study, we look at the issue of translating/converting sign language to text and propose a better solution based on machine
learning techniques. We want to establish a system that hearing-impaired people may utilise in their everyday lives to promote
communication and collaboration between hearing-impaired people and people who aren't trained in American Sign Language
(ASL). To develop a deep learning model for the ASL dataset, we'll use a technique called Transfer Learning in combination
with Data Augmentation.
Keywords: Sign language, machine leaning, Transfer learning, ASL, Inception v3
I. INTRODUCTION
Sign language is used to communicate by those who are deaf or hard of hearing. People employ nonverbal communication such as
sign language movements to communicate their thoughts and emotions. Non-signers, on the other hand, have a hard time
understanding it, which is why skilled sign language interpreters are required for medical and legal consultations, as well as
educational and training sessions. Over the last few years, the demand for translation services has significantly increased. Other
methods have been devised, such as video remote human interpreting using high-speed Internet connections. As a result, they will
provide a simple sign language translating service that may be used but has significant restrictions.
In the literature, certain studies [1]–[3] for automated ASL recognition have previously been published. Some of these algorithms
have only been tested on a small sample dataset, while others rely on the typical shallow neural network approach to classification.
Shallow neural networks need feature identification and appropriate feature selection by hand. Deep learning (DL) approaches have
considerably enhanced the performance of classic shallow neural networks for machine learning applications, particularly for image
recognition and computer vision issues.
The rest of this paper is organised as follows: A brief summary of comparable investigations described in the literature is provided
in section II. A brief overview of the dataset utilised in this study is provided in section III. In section IV, the proposed approach is
described. The study's findings are provided in part V, and the study's primary findings are summarised in section VI.
II. BACKGROUND
To develop a deep learning model for the ASL dataset, we'll use a technique called Transfer Learning in combination with Data
Augmentation. Transfer learning is a machine learning approach in which a model developed for one task is utilised as the
foundation for another task's model. Given the massive computing and time resources necessary to construct neural network models
for these challenges, as well as the significant leaps in skill that they give on related problems, pre-trained models are a common
strategy in deep learning for computer vision and natural language processing applications.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 246
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue III Mar 2022- Available at www.ijraset.com
III. DATASET
This Kaggle dataset of ASL Alphabet was used to train the network. The dataset consists of 87,000 200x200 pixel photos organised
into 29 types (26 English Alphabets and 3 additional signs of SPACE, DELETE, and NOTHING).
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 247
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue III Mar 2022- Available at www.ijraset.com
V. EXPERIMENTAL RESULTS
To train our model, we utilized Categorical Cross entropy to assess the loss and a Stochastic Gradient Descent optimizer (with a
learning rate of 0.0001 and a momentum of 0.9). The model is trained over a period of 24 epochs. The following are the outcomes:
TABLE I
Metric Value
Training accuracy 0.9888 (98.88%)
Training loss 0.1100
Validation accuracy 0.9576 (95.76%)
Validation loss 0.1926
Test accuracy 96.43%
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 248
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue III Mar 2022- Available at www.ijraset.com
REFERENCES
[1] V. Bheda and D. Radpour, “Using deep convolutional networks for gesture recognition in american sign language,” arXiv:1710.06836, 2017.
[2] B. Garcia and S. A. Viesca, “Real-time american sign language recognition with convolutional neural networks,” Convolutional Neural Networks for Visual
Recognition, vol. 2, 2016.
[3] A. Barczak, N. Reyes, M. Abastillas, A. Piccio, and T. Susnjak, “A new 2d static hand gesture colour image dataset for asl gestures,” 2011.
[4] https://machinelearningmastery.com/transfer-learning-for-deep-learning/
[5] K. Bantupalli and Y. Xie, "American Sign Language Recognition using Deep Learning and Computer Vision," 2018 IEEE International Conference on Big
Data (Big Data), 2018, pp. 4896-4899, doi: 10.1109/BigData.2018.8622141.
[6] R. Fatmi, S. Rashad and R. Integlia, "Comparing ANN, SVM, and HMM based Machine Learning Methods for American Sign Language Recognition using
Wearable Motion Sensors," 2019 IEEE 9th Annual Computing and Communication Workshop and Conference (CCWC), 2019,pp.0290-0297, doi:
10.1109/CCWC.2019.8666491.
[7] M. M. Rahman, M. S. Islam, M. H. Rahman, R. Sassi, M. W. Rivolta and M. Aktaruzzaman, "A New Benchmark on American Sign Language Recognition
using Convolutional Neural Network," 2019 International Conference on Sustainable Technologies for Industry 4.0 (STI),2019,pp.1-6,doi:
0.1109/STI47673.2019.9067974.
[8] http://cs231n.stanford.edu/reports/2016/pdfs/214_Report.pdf
[9] Sharma, S., Kumar, K. ASL-3DCNN: American sign language recognition technique using 3-D convolutional neural networks. Multimed Tools Appl 80,
26319–26331 (2021). https://doi.org/10.1007/s11042-021-10768-5
[10] Y. Ye, Y. Tian, M. Huenerfauth and J. Liu, "Recognizing American Sign Language Gestures from Within Continuous Videos," 2018 IEEE/CVF Conference
on Computer Vision and Pattern Recognition Workshops (CVPRW), 2018, pp. 2145-214509, doi: 10.1109/CVPRW.2018.00280.
[11] C.K.M. Lee, Kam K.H. Ng, Chun-Hsien Chen, H.C.W. Lau, S.Y. Chung, Tiffany Tsoi, American sign language recognition and training method with
recurrent neural network, Expert Systems with Applications, Volume 167, 2021, 114403
[12] M. Taskiran, M. Killioglu and N. Kahraman, "A Real-Time System for Recognition of American Sign Language by using Deep Learning," 2018 41st
International Conference on Telecommunications and Signal Processing (TSP), 2018, pp. 1-5, doi: 10.1109/TSP.2018.8441304.
[13] https://www.irjet.net/archives/V7/i3/IRJET-V7I3418.pdf
[14] https://towardsdatascience.com/sign-language-recognition-using-deep-learning-6549268c60bd.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 249