Sign Language Recognition Using Deep Learning
Sign Language Recognition Using Deep Learning
Sign Language Recognition Using Deep Learning
https://doi.org/10.22214/ijraset.2022.45913
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue VII July 2022- Available at www.ijraset.com
Abstract: Sign language is a way of communicating using hand gestures and movements, body language and facial expressions,
instead of spoken words. It can also be defined as any of various formal languages employing a system of hand gestures and
their placement relative to the upper body, facial expressions, body postures, and finger spelling especially for communication by
and with deaf people.
The project that is being built is to recognize the action performed by the person/user in sign language using Deep learning.
Ordinary people are not well versed in sign language.
The project tries to solve this problem using deep learning that is precisely using TensorFlow.
In the project a LSTM (long-short term memory) model in deep Learning is built using TensorFlow to categories the action the
user is doing. This will help the user with special needs to communicate with other people using the application we built.
By this we can bridge the gap between the especially abled people and ordinary people.
Keywords: Sign language, Deep learning, LSTM, TensorFlow
I. INTRODUCTION
Deafness has varying descriptions in artistic and medical terms. In medical terms, the meaning of deafness is hearing loss that
precludes a person from understanding spoken language, an audiological condition. In this term it's written with a lower cased. In a
medical terms, deafness is defined as a degree of sound loss similar that a person is unfit to understand speech, indeed in the
presence of modification.
In profound deafness, indeed the loftiest intensity sounds produced by an audiometer (an instrument used to measure sound by
producing pure tone sounds through a range of frequency) may not be detected. In total deafness, no sounds at each, anyhow of
modification or system of product, can be heard. A mute is a person who doesn't speak, either from an incapability to speak or
reluctance to speak. The term" mute" is specifically applied to a person who, due to profound natural (or beforehand) deafness, is
unfit to use eloquent language and so is deaf-mute.
The problem is that there exists a communication hedge between normal people and especially- abled people as the normal person
aren't clued in sign language and isn't suitable to communicate with especially- abled person. The ideal of this design is to give a
communication result in- form of an operation that can fete the sign- language and give the affair in form of textbook that can be
fluently understood by the normal person. can be fluently understood by the normal person. We prognosticate the sign language
deep literacy that's using long term short memory algorithm this algorithm is a neural network that helps us to prognosticate the
action performed by this especially- abled person dictation. In this way it decreases the communication hedge between a normal
person and an especially- abled person (Deaf and Mute person). A homemade translator cannot always be present to restate the
conduct of an especially- abled person and help him to overcome the difficulties faced by him in the communication with others
who don't know sign- language used by the person. Our proposed system will help the deaf and hard- of- hail communicate better
with members of the community. For illustration, there have been incidents where those who are deaf have had trouble
communicating with first askers when in need., it's unrealistic to anticipate everyone to come completely fluent in sign language.
Down the line, advancements like these in computer recognition could prop a first polled in understanding and helping those that are
unfit to communicate through speech.
Another operation is to enable the deaf and hardy- of- hard equal access to videotape consultations, whether in a professional
environment or while trying to communicate with their healthcare providers via telehealth. rather of using introductory converse,
these advancements would allow the hearing- bloodied access to effective videotape communication.
The design being erected is an operation that can fete the stoner's conduct and restate that action to textbook and speech. The
operation is doing this using deep literacy, that's we're erecting a model that will fete the conduct and orders that action and
translates it to textbook and speech.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 4173
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue VII July 2022- Available at www.ijraset.com
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 4174
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue VII July 2022- Available at www.ijraset.com
In this way we are able to detect the sign performed in front of the camera in real time using open CV and LSTM
Long short-term memory (LSTM) is an artificial recurrent neural network (RNN) architecture used in the field of deep learning.
Unlike standard feedforward neural networks, LSTM has feedback connections. It can process not only single data points (such as
images), but also entire sequences of data (such as speech or video). For example, LSTM is applicable to tasks such as
unsegmented, connected handwriting recognition, speech recognition and anomaly detection in network traffic or IDSs (intrusion
detection systems). A common LSTM unit is composed of a cell, an input gate, an output gate and a forget gate. The cell remembers
values over arbitrary time intervals and the three gates regulate the flow of information into and out of the cell.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 4175
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue VII July 2022- Available at www.ijraset.com
IV. IMPLEMENTATION
A. Collecting the Data for Creating the data Samples
The implementation for the proposed system is done in jupyter notebook. The language used is python 3.9 for the proposed system.
The keras library from the TensorFlow library is used to import the LSTM model required for training. The OpenCV is used to
capture the actions for the training and testing. Mediapipe Holistic is a pipeline used to create the landmarks that serves as the
keypoints. The landmarks from the user's hands are captured and saved in the file. This process repeated for 30 times for each action
that is to be included.
V. CONCLUSION
The work successfully covers the commonly used gestures and interprets them into a sentence at high speed and accuracy.
Recognition of the gestures does not get affected by the lighting of the environment, color or size of the person. This Application
requires less data when compared to the applications that were built on the CNN algorithm. It is also faster to train as it takes less
data as the input. It also performs faster detections when compared to a CNN model. And also got good accuracy score for the
validation data. We are including as many words as possible in the near future. Model training gets complex as the number of
different words increases. As this work can bridge a gap between normal people and disabled people. Hence, our future
enhancements or work would primarily focus on two things.
REFERENCES
[1] S. Nikam and A. G. Ambekar, "Sign language recognition using image-based hand gesture recognition techniques," 2016 Online International Conference on
Green Engineering and Technologies (IC-GET), 2016, pp. 1-5, doi: 10.1109/GET.2016.7916786.
[2] S. Suresh, H. T. P. Mithun and M. H. Supriya, "Sign Language Recognition System Using Deep Neural Network," 2019 5th International Conference on
Advanced Computing & Communication Systems (ICACCS), 2019, pp. 614-618, doi: 10.1109/ICACCS.2019.8728411.
[3] S. Gupta, R. Thakur, V. Maheshwari and N. Pulgam, "Sign Language Converter Using Hand Gestures," 2020 3rd International Conference on Intelligent
Sustainable Systems (ICISS), 2020, pp. 251-256, doi: 10.1109/ICISS49785.2020.9315964.
[4] Jayaprakash, Rekha & Majumder, Somajyoti. (2011). Hand Gesture Recognition for Sign Language: A New Hybrid Approach. 1.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 4176
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 10 Issue VII July 2022- Available at www.ijraset.com
[5] Starner T, Weaver J, Pentland A (1998) Real-time American sign language recognition using desk and wearable computer-based video. IEEE Trans Pattern
Anal Mach Intell 20:1371–1375
[6] Moryossef, Amit & Tsochantaridis, Ioannis & Aharoni, Roee & Ebling, Sarah & Narayanan, Srini. (2020). Real-Time Sign Language Detection using Human
Pose Estimation.
[7] Pose-based Sign Language Recognition using GCN and BERT Anirudh Tunga* Purdue University [email protected] Sai Vidyaranya Nuthalapati*
[email protected] Juan Wachs Purdue University
[8] Z. Yao and X. Song, "Vehicle Pose Detection and Application Based on Grille Net," 2019 3rd International Conference on Electronic Information Technology
and Computer Engineering (EITCE), 2019, pp. 789-793, doi: 10.1109/EITCE47263.2019.9094787.
[9] J. Su, X. Huang and M. Wang, "Pose detection of partly covered target in micro-vision system," Proceedings of the 10th World Congress on Intelligent Control
and Automation, 2012, pp. 4721-4725, doi: 10.1109/WCICA.2012.6359373.
[10] Siddhartha Pratim Das, Anjan Kumar Talukdar, Kandarpa Kumar Sarma, Sign Language Recognition Using Facial Expression, Procedia Computer Science,
Volume 58, 2015, Pages 210-216,
[11] Das, Siddhartha & Talukdar, Anjan & Sarma, Kandarpa. (2015). Sign Language Recognition Using Facial Expression. Procedia Computer Science.58.
10.1016/j.procs.2015.08.056.
[12] Vahdani∗ · Matt Huenerfauth · Yingli Tian
[13] Sign Language Recognition Helen Cooper, Brian Holt and Richard Bowden
[14] Aditya Kanodia, Prince Singh, Durgi Rajesh, G Malathi: INDIAN SIGN LANGUAGE USING HOLISTIC POSE DETECTION.
[15] Agarwal A, Thakur MK (2013) Sign language recognition using Microsoft Kinect. In: IEEE Sixth International Conference on Contemporary Computing (IC3),
pp 181–185.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 4177