Sign Language Recognition
Sign Language Recognition
https://doi.org/10.22214/ijraset.2023.48586
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com
Abstract: People communicate using sign language by visually conveying sign patterns to portray purpose. One method of
communicating with deaf-mute people is to use sign language mechanisms. One of the nonverbal communication strategies
used in sign language is the hand gesture. Many manufacturers all over the world have created various sign language systems,
but they are neither adaptable nor cost-effective for end users. We present a design that can recognize various American sign
language static hand motions in real-time using transfer learning, Python, and OpenCV in this paper. “Hello, Yes, No, Thank
You, and I Love You" are all prevalent sign language terms that our system correctly acknowledges. The following are the key
steps in system design; we created our own dataset taking prominent gestures of the American Sign Language, captured
images with OpenCV and webcam, the images were then labelled for object detection, training and testing of dataset was done
with transfer learning using SSD MobileNet, and eventually the gestures were successfully determined in real-time.
Keywords: Sign Language, Python, OpenCV, Transfer Learning
I. INTRODUCTION
Sign languages are regarded as a systematic compendium of hand gestures with strong connotations used by hearing impaired
individuals to socialize in everyday experiences. The impairments effecting hearing and speeches affect over 360 million people
worldwide. Non - verbal communication is used by deaf and dumb people to communicate with one another and with others.
Because once you can't talk or listen, it's hard to communicate with one another.
They facilitate communique without the usage of words. Even so, there may be one ultimate problem: only a few humans appear
to recognize signal language. Despite the truth that deaf and dumb humans can have interaction the use of signal languages, it's far
difficult for all of them to communicate with people who have regular listening to and vice versa main to a scarcity of signal
linguistic knowledge. This problem may be constant through imposing a technological solution. Using that solution, you'll fast
rework signal language gestures into the extensively spoken language, English.
They use a concurrently and particular mixture of gestures, hand forms, and alignment to express detailed info. The American
Sign Language (ASL) system is however one widely used set of languages. Speech-impaired individuals learn sign language so
they can converse with others and meet their daily needs.
This is an open cv execution of strategic that makes use of an internet digital digicam to take pics of hand gestures. Image
labelling is needed after photo capture, and a pre-educated version SSD Mobile internet v2 is used for signal recognition. Like a
consequence, a robust communique path may be fashioned amongst deaf and listening to publics. This machine lets in someone to
explicit themselves nonverbally through transferring their palms speedy instead of losing time seeking to shape every phonics for
each word.
This reduces the probability of misinterpretation and mistranslation. Similarly, hundreds of phrases that we use on a normal
foundation can certainly be delegated through hand gestures and recognized.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 386
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com
Sign language recognition architecture has been divided into four segments, and they are as follows;
1) “Creating our own dataset”.
2) “Labelling captured images”.
3) “Training the models using tensor flow object detection”.
4) “Detecting the hand gesture”.[1]
B. Architecture
C. Data Acquisition
It is necessary to create a proper database of sign language gestures so that the images captured while communicating with this
system can be compared. We collected 400 images to create the dataset for this project. This dataset contains five gestures,
namely "Hello, Yes, No, I Love You, and Thank You," which are very useful when dealing with real-time applications.
To enhance accuracy, a couple of pix of various signal language gestures have been taken from extraordinary angles and below
extraordinary lights conditions. To create our dataset, we used the Open pc vision (OpenCV) library.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 387
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com
D. Data Annotation
Data annotation is the categorization and labelling of data for AI applications. Training data must be properly categorized and
annotated for a specific use case. With high-quality, human-powered data annotation, companies can build and improve AI
implementations.
LabelImg is a data annotation tool that is used for data annotation. LabelImg is indeed a free and open-source image labelling
programme. It's implemented in Python and has a graphical interface built with QT. It's a quick and easy way to label some few
hundred images. They assist in identifying elements in your data that you want to train your model to determine in unlabeled
datasets. Elevated datasets are required for computer vision and the development of a developing competence model. The garbage
in, garbage out philosophy is followed when creating computer vision models, which means labelling images carefully and
accurately is critical.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 388
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 389
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com
After a model detects the accuracy of a class in an image, it is stored in the form of a NumPy array in the ‘det’ tensor. That key of
the array is named ‘detection_scores’ here. The first element of the named array contains the highest accuracy element. This
accuracy is plotted against the class in the same tensor. This information is stored as another NumPy array inside another key-
value pair, with a key named ‘detection_classes.’ We take this information to determine the highest accuracy and the consequent
mapped class.
When we had a closer look at the data and the prediction scores of our model related to every class, we made an insightful
observation. The system captures images with the right label when it has a prediction score of more than 85% against the label for
the image. Below which, it is either confused about the label or the image is faulty, either way the detection is not accurate.
Hence, the assumed accuracy of the model determines all the times the model got a prediction score of more than 85% for a
particular image. The percentage of times the model has satisfied this condition, is the assumed accuracy of the system. This is
mapped by using a truth table and assigning positive-negative values against each prediction score. It helps in determining
accuracy that can be safely assumed to be the detection accuracy of the model.
I. Real-Time Detection
This model purposes a pipeline that takes through a web camera from a client who is marking a motion and afterward by
extricating various casings of video, it creates communication through signing opportunities for each signal.
The ongoing recognition is finished utilizing OpenCV and webcam once more. For ongoing identification, cv2, and NumPy
conditions are utilized. The framework recognizes signs continuously and deciphers what each motion implies into English.
III. OUTPUTS
A. Class-wise Accuracy Detection
Matplotlib is used to plot the above result. For better visualization, colors red and green have been chosen. The x-axis represents
the 5 classes and the y-axis represents. Highest accuracy achieved is of the class ‘Thank You’ i.e., 99.13 percent.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 390
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com
C. Real-time Detection
The constant location is finished utilizing OpenCV and webcam once more. For, continuous discovery, cv2, and NumPy
conditions are utilized. The framework recognizes signs progressively and deciphers what each motion implies into English as
displayed:
.
Figure 11. Detection ‘Yes’ Figure 12. Detection ‘I Love You’
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 391
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com
For each gesture, the recognition rate is recorded and tabulated in the result as shown:
The 'Yes' sign has the highest recognition of 100 percent. Because of its resemblance to the 'I Love You' hand gesture, the
estimation of the 'hello' sign has the least recognition performance of 88 percent.
The project has yielded acceptable results, and the word signs in static position are recognized. The overall output rate was
94.7368 percent.
REFERENCES
[1] Shivani Y, Jarul R, Vandana N, “Sign Language Recognition System using Transfer Learning Technique”, Artificial & Computational Intelligence,
December-2021.
[2] Sharvani Srivastava, Amisha Gangwar, Richa Mishra & Sudhakar Singh, “Sign Language Recognition System Using TensorFlow Object Detection API”,
CCIS Vol.1534, Feburary-2022.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 392