Sign Language To Text-Speech Translator Using Machine Learning
Sign Language To Text-Speech Translator Using Machine Learning
912
Akshatha Rani K et al., International Journal of Emerging Trends in Engineering Research, 9(7), July 2021, 912 – 916
detection by segmentation. They used SVM end KNN [10] proposed sign language recognition system using
supervised machine learning techniques for image CNN and computer vision. The system used HSV color
classification to evaluate the accuracy. algorithm for hand gesture detection and they set the
background black. Image pre-processing consists of grayscale
[4] proposed hand gesture recognition using PCA. This conversion, dilation, mask operation and hand gesture was
system color model approach and thresholding method with segmented. The CNN architecture was used for feature
effective template matching for hand detection. Hand extraction in the first layer and then for image classification.
recognition is segmented with skin color modelling in YCbCr This system was able to recognize 10 alphabets and it
color space. Otsu thresholding is used for foreground and achieved 90% of accuracy.
background separation. PCA is used for template matching
for gesture recognition the system achieved accuracy of 3. SYSTEM METHOD
91.43% for low brightness images.
The below figure 1 shows the block diagram of the proposed
[5] proposed ASL gesture recognition by using deep CNN system.
for letters and digits. In this paper, images were pre-processed
in which image background was removed by using
background subtraction technique. The dataset was split into
two, one for training and other for testing, and they have used
CNN to classify images. The system achieved 82.5% accuracy
on the alphabet gestures.
913
Akshatha Rani K et al., International Journal of Emerging Trends in Engineering Research, 9(7), July 2021, 912 – 916
914
Akshatha Rani K et al., International Journal of Emerging Trends in Engineering Research, 9(7), July 2021, 912 – 916
4. RESULT AND DISCUSSIONS The model has achieved 74% of validation accuracy with
efficient hand tracking technique. Its graph is plotted with
In the training phase, the system is trained by using 2000 number of epochs against validation accuracy of the model, as
images with the ANN architecture and the model is saved. shown in the below fig. 7.
now the prediction of letters takes place by using the model.
the system first detects the hand in the live video frame, when
the hand tracking is done then it recognises the sign and
display it on the screen in the text format.
The fig. 5 shows the sign of letter ‘A’. The system which
tracks the hand, and it compares the hand pattern with trained 5. CONCLUSION
images and then it predicts the sign is letter ‘A’ with the
probability percentage of prediction. There are many researches have been carried out in the
field of machine learning and computer vision. They have
In this system, we can also create the words by joining the contributed effective works which are very necessary and
letters one after other. The fig. 6 shows that first the system helpful for everyday life. Likewise various research has been
tracks the hand pattern, and it predicts the sign is ‘A’ then it done on sign language recognition using different methods
predicts the next sign ‘I’ which is shown to the camera and like neural networks, KNN, SVM and LSTM. In this paper,
after that, predicts the sign ‘M’, so it formed the word ‘AIM’. the proposed system concentrated on hand tracking technique
Likewise, we can form any word. which is very effective technique. It also detects the hand for
different skin colours and lighting condition, and it also
By using the ‘space’ which is also trained under the model, detects the hand in low-light condition. We used ANN to
we can form a sentence. After the sign to text conversion, the classify images of asl alphabet, the system recognises almost
text can be converted to speech which is helpful for blind all the letters and achieved 74% of accuracy. The system
people. The system pronounces the word or text. which also incorporates the speech which converts the
recognized sign text to speech, so that it will also be helpful
for blind people.
6. FUTURE WORK
REFERENCES
1. https://google.github.io/mediapipe/solutions/hands.html
2. Singha, J. and Das, K. Hand Gesture Recognition
Based on Karhunen-Loeve Transform, Mobile and
Embedded 232 Technology International Conference,
January 17-18, 2013.
Figure 6: Word formation
915
Akshatha Rani K et al., International Journal of Emerging Trends in Engineering Research, 9(7), July 2021, 912 – 916
916