Sat - 28.Pdf - Sign Language Recognition Using Machine Learning
Sat - 28.Pdf - Sign Language Recognition Using Machine Learning
Sat - 28.Pdf - Sign Language Recognition Using Machine Learning
5
Table of Contents
1. INTRODUCTION .......................................................................................................................... 7
1.1 OUTLINE OF THE PROJECT ........................................................................................... 7
1.2 LITERATURE VIEW ............................................................................................................... 7
1.3 SYSTEM IMPLEMENTATION ............................................................................................ 10
1.4 OBJECTIVES ..................................................................................................................... 10
2. AIM AND SCOPE OF THE PRESENT INVESTIGATION ................................................... 11
2.1 SIGN LANGUAGE RECOGINITION .............................................................................. 11
2.2 SERVERPORTAL .............................................................................................................. 11
3. SYSTEM REQUIRMENT .......................................................................................................... 12
3.1 REQUIRMENTS SPECIFICATION ................................................................................. 12
3.1.1 HARDWARE REQUIRMENTS ................................................................................ 12
3.1.2 SOFTWARE REQUIRMENTS ................................................................................. 12
3.2 ABOUT THE SOFTWARE ............................................................................................... 12
3.2.1 SPYDER ...................................................................................................................... 12
4. EXPERIMENTAL OR MATERIALS AND METHODS ALGORITHM USED ..................... 13
4.1 METHODOLOGY ............................................................................................................... 13
4.2 ALGORITHM ...................................................................................................................... 16
5. RESULTS AND DISCUSSION, PERFORMANCE ANALYSIS ........................................... 22
5.1 OVERVIEW OF THE PLATFORM .................................................................................. 22
5.2 TESTING ............................................................................................................................. 23
6. SUMMARY AND CONCLUSIONS .......................................................................................... 23
6.1 CONCLUSION .................................................................................................................... 23
6.2 FURTHER ENHANCEMENT ............................................................................................... 23
7. REFRENCES .............................................................................................................................. 23
8. APPENDIX .................................................................................................................................. 24
A) SAMPLE CODE: - ..................................................................................................................... 24
B) SCREEN SHOTS: - .................................................................................................................. 27
6
1. INTRODUCTION
1.1 OUTLINE OF THE PROJECT
Indian sign language is a predominant sign language Since the only disability
D&M people have is communication related and they cannot use spoken
languages hence the only way for them to communicate is through sign
language. Communication is the process of exchange of thoughts and messages
in various ways such as speech, signals, behaviour and visuals. Deaf and
dumb(D&M) people make use of their hands to express different gestures to
express their ideas with other people. Gestures are the nonverbally exchanged
messages and these gestures are understood with vision. This nonverbal
communication of deaf and dumb people is called sign language.
American sign language is a predominant sign language Since the only disability
D&M people have is communication related and they cannot use spoken
languages hence the only way for them to communicate is through sign
language. Communication is the process of exchange of thoughts and messages
in various ways such as speech, signals, behaviour and visuals. Deaf and
dumb(D&M) people make use of their hands to express different gestures to
express their ideas with other people. Gestures are the nonverbally exchanged
messages and these gestures are understood with vision. This nonverbal
communication of deaf and dumb people is called sign language.
For interaction between normal people and D&M people a language barrier is
created as sign language structure which is different from normal text. So, they
depend on vision-based communication for interaction. If there is a common
interface that converts the sign language to text, the gestures can be easily
understood by the other people. So, research has been made for a vision-based
interface system where D&M people can enjoy communication without really
knowing each other's language. The aim is to develop a user-friendly human
computer interface (HCI) where the computer understands the human sign
language. There are various sign languages all over the world, namely American
Sign Language (ASL), French Sign Language, British Sign Language (BSL),
Indian Sign language, Japanese Sign Language and work has been done on
other languages all around the world
The importance of the very use of this method is increasing day by day as this
gives the people with disabilities and opportunity to try their hands in
different fields that require communication. With the help of this project, they can
communicate with the majority in any industry, thus giving them an even playing
field to cope up with. Here, we use the method of object identification followed by
recognition to help us differentiate between the many symbols used in the sign.
7
Fingerspelling Word level sign Non-manual features
vocabulary
gestures in order to form a complete word by combining each gesture. The gestures
we aim to train are as given in the image below.
In the recent years there has been tremendous research done on the hand gesture
recognition.
With the help of literature survey done we realized the basic steps in hand gesture
recognition are:-
i. Data-acquisition
ii. Data-pre-processing
iii. Gesture-classification
Data acquisition:
The different approaches to acquire data about the hand gesture can be done in the
following ways:
8
It uses electromechanical devices to provide exact hand configuration, and
position. Different glove-based approaches can be used to extract information.
But it is expensive and not user friendly.
In vision-based methods computer camera is the input device for observing the
information of hands or fingers. The Vision Based methods require only a
camera, thus realizing a natural interaction between humans and computers
without the use of any extra devices. These systems tend to complement
biological vision by describe in artificial vision systems that are implemented in
software and/or hardware.
The main challenge of vision-based hand detection is to cope with the large
variability of human hand‘s appearance due to a huge number of hand
movements, to different skin colour possibilities as well as to the variations in
viewpoints, scales, and speed of the camera capturing the scene.
In [1] the approach for hand detection combines threshold-based colour detection
with background subtraction. We can use Ad boost face detector to differentiate
between faces and hands as both involve similar skin-color.
Gesture Classification:
In [1] Hidden Markov Models (HMM) is used for the classification of the gestures.
This model deals with dynamic aspects of gestures. Gestures are extracted from
a sequence of video images by tracking the skin-color blobs corresponding to the
hand into a body– face space cantered on the face of the user. The goal is to
9
recognize two classes of gestures: deictic and symbolic. An image is filtered
using a fast look–up indexing table. After filtering, skin colour pixels are gathered
into blobs. Blobs are statistical objects based on the location (x, y) and the
colorimetry (Y,U,V) of the skin colour pixels in order to determine homogeneous
areas.
In Naïve Bayes Classifier is used which is an effective and fast method for static
hand gesture recognition. It is based on classifying the different gestures
according to geometric based invariants which are obtained from image data
after segmentation. Thus, unlike many other recognition methods, this method is
not dependent on skin colour. The gestures are extracted from each frame of the
video, with a static background. The first step is to segment and label the objects
of interest and to extract geometric invariants from them. Next step is the
classification of gestures by using a K nearest neighbour algorithm aided with
distance weighting algorithm (KNNDW) to provide suitable data for a locally
weighted Naïve Bayes ―classifier.
For interaction between normal people and D&M people a language barrier is
created as sign language structure which is different from normal text. So, they
depend on vision-based communication for interaction.
If there is a common interface that converts the sign language to text the
gestures can be easily understood by the other people. So, research has been
made for a vision-based interface system where D&M people can enjoy
communication without really knowing each other's language.
If there is a common interface that converts the sign language to text the
gestures can be easily understood by the other people. So, research has been
made for a vision-based interface system where D&M people can enjoy
communication without really knowing each other's language.
1.4 OBJECTIVES
10
4) To increase efficiency proposed system is dependent on classification method.
5) Proposed system is used to reduce confusion at the time of processing data average.
Sign language recognition and translation is a research area with high potential
impact. There are over 300 sign languages used around the world, and 70 million
deaf people are using them. Sign language processing would break down all the
barriers for sign language users. This can be very helpful for the deaf and dumb
people in communicating with others as knowing sign language is not something that
is common to all, moreover, this can be extended to creating automatic editors,
where the person can easily write by just their hand gestures. The types of data
available and the relative merits are explored allowing examination of the features
which can be extracted. Classifying the manual aspects of sign (similar to gestures)
is then discussed from a tracking and non-tracking viewpoint before summarizing
some of the approaches to the non-manual aspects of sign languages. Methods for
combining the sign classification results into full SLR are given showing the
progression towards speech recognition techniques and the further adaptations
required for the sign specific case. Finally, the current frontiers are discussed, and
the recent research presented. This covers the task of continuous sign recognition,
the work towards true signer independence, how to effectively combine the different
modalities of sign, making use of the current linguistic research and adapting to
larger more noisy data sets.
2.2 SERVERPORTAL
3. SYSTEM REQUIRMENT
3.1 REQUIRMENTS SPECIFICATION
Processor – Intel i5
Ram – 4GB
Hard Disk Dive – 40GB
Monitor – LCD
3.1.2 SOFTWARE REQUIRMENTS
Operating System – Windows 10
Software – SPYDER
3.2.1 SPYDER
Spyder, the Scientific Python Development Environment, is a free integrated
development environment (IDE) that is included with Anaconda. It includes
editing, interactive testing, debugging, and introspection features.
Spyder is an open-source cross-platform IDE. The PYTHON Spyder IDE is
written completely in Python. It is designed by scientists and is exclusively for
scientists, data-analyst, and engineers. It is also known as the Scientific
Python Development IDE and has a huge set of remarkable features,
WHY SPYDER?
12
iii) Spyder contains features like a text editor with syntax highlighting, code
completion and variable exploring, which you can edit its values using a
Graphical User Interface (GUI).
4.1 METHODOLOGY
A. DATA ACQUISITION
The different approaches to acquire data about the hand gesture can be done in
the following ways:
In vision-based methods computer camera is the input device for observing the
information of hands or fingers. The Vision Based methods require only a camera, thus
realizing a natural interaction between humans and computers without the use of any
extra devices. These systems tend to complement biological vision by describing 7
artificial vision systems that are implemented in software and/or hardware. The main
challenge of vision-based hand detection is to cope with the large variability of human
hand‘s appearance due to a huge number of hand movements, to different skin-color
possibilities as well as to the variations in viewpoints, scales, and speed of the camera
capturing the scene.
However, due to the rarity of the chosen project, there were a little to no datasets
available online. As a result, we had to make our own datasets to help train the code to
give us the best results possible. It was made possible with the use of Open computer
vision also known as the Open CV library to obtain the customized datasets we use in
this program.
13
Fig 4.1. Image captured using figure.
14
C. OBJECT DETECTION
A bounding box around the object is used to indicate its location. The drawn
bounding box may or may not pinpoint the exact location of the given piece. The
potential to spot the provided object inside a photo explain the performance of the
algorithm that has been used for detection. Sign language detection is one of the
examples of object detection.
Generation of small portion in the input as shown in the image below. As shown
below in the figure large set of boxes are spanning the full image.
Feature extraction is the next step that will be done for each rectangular area to
check whether or whether not the rectangle area contains a valid object.
15