Sign Language
Sign Language
Sign Language
Smartphones
angle filters [27]. Detecting the position and orientation of plied. Support Vector Machines (SVM) define decision bound-
the hand at each frame accurately also allows us to detect aries between classes, which are linear in some transformed
the motion of the hand for dynamic signs. feature space, but can be highly nonlinear in the original
Skin segmentation algorithms, which often depend on spec- feature space [31]. Several papers use SVMs, e.g., [4, 9, 10,
ifying thresholds [18], are widely used in Computer Vision 12, 16]. Hakkun 2015 [7] use K-Nearest Neighbor (KNN)
applications. The researchers either specify skin thresholds for classification. Another simple technique for classifica-
manually or automatically by taking a skin color sample be- tion is template matching, used by [5, 13, 30, 21, 12]. The
fore the experiment. Several available models use RGB color Backpropagation algorithm [23] can lead to very efficient
space, e.g., [10, 13, 16, 25, 30]. To solve brightness and light- classification timewise, but it needs more training data to
ing problems, [9] use YCrCb color space, [4] employ HSV minimize error rate. Backpropagation is used by [24] as the
color space, and [19] benefit from HUV color space. recognition method. In Rao 2016 [22], because the speed
The Viola-Jones detection method [27], which uses cas- of processing in portable devices is a major factor, a mini-
cades of boosted rectangle filters, is a well-known method, mum distance classifier (MDC) was chosen as a classification
that is commonly used for detecting hands. Some researchers method. The experiments use sentences of signs as training
[7, 4] implement the Viola-Jones method on portable plat- and test data.
forms, as Viola-Jones is relatively easy to implement and Some systems assume that the only visible object in the
has low hardware requirements. Another alternative, used captured image is the hand [7, 4], while the more advanced
by [12, 21, 22, 25], is Principal Component Analysis (PCA). models manage to capture both hands and face. One way
Additional hand details are also extracted by various meth- to remove the confusion between a face and hand area is to
ods. Examples of such details are the number of open fingers subtract or isolate the face, so that the detection of hand
(measured by finding contours), finding the palm area (by details will be more precise [4]. Another issue that can be
finding the largest circle that fits in the hand region), de- considered is hand angle and hand distance from the mo-
tecting the convex hull, and getting convexity defects [4, 9, bile device. In tests conducted in [7], optimal results were
16]. Canny edge detection [1] can also be used to identify achieved with no more than 50 cm distance between the
the hand area [10]. Likewise, a Sobel Edge filter, which mea- hand and the camera, and the hand being in the upright
sures the changes in value in the highest moving direction, state.
has been used [21, 22, 24]. Prasuhn 2014 [19] apply a Due to slow processing time in some models, a client-
Histogram of Orientation Gradients (HOG) method, which server framework is used. In such a framework, the phone is
is sensitive to the angle of the object, to extract the fea- connected to a regular computer via wireless network. Such
tures from the input image. Another method, used by [5], is an approach was implemented in [19, 21, 24, 25, 30]. A cloud
background subtraction using a motion detection method. service can be used to execute part of recognition operations,
In Jin 2016 [10], Speeded-Up Robust Features (SURF) is as done in [9]. Moreover, Elleuch 2015 [4] implement a
used as an extra feature to improve accuracy. multithreading technique by running face subtraction and
Once the features describing a sign have been extracted, hand pre-processing at the same time, thus decreasing the
there are numerous recognition procedures that can be ap- processing time by half.
4. CONCLUSIONS [7] R. Y. Hakkun, A. Baharuddin, et al. Sign language
In this paper, we have provided a survey of existing tech- learning based on android for deaf and speech
niques for sign language recognition in smartphones. We impaired people. In Electronics Symposium (IES),
discussed sensor-based approaches, which track hand motion 2015 International, pages 114–117. IEEE, 2015.
and/or posture using hardware-based trackers installed in a [8] S. Hamrick, L. Jacobi, P. Oberholtzer, E. Henry, and
glove or inside a smartphone. We also discussed vision-based J. Smith. Libguides. deaf statistics. deaf population of
approaches, which use the phone camera for observing the the us. Montana, 16(616,796):2–7, 2010.
hand. In discussing both types of approaches, we focused on [9] P. Hays, R. Ptucha, and R. Melton. Mobile device to
the detection and feature extraction module as well as the cloud co-processing of asl finger spelling to text
classification module of each approach. conversion. In Image Processing Workshop
Regarding vision-based methods, significant challenges re- (WNYIPW), 2013 IEEE Western New York, pages
main to be overcome by future research, regarding accuracy 39–43. IEEE, 2013.
of hand detection and articulated hand pose estimation, as [10] C. M. Jin, Z. Omar, and M. H. Jaward. A mobile
well as classification accuracy. Most existing vision-based application of american sign language translation via
methods only recognize static gestures, and we expect new image processing algorithms. In Region 10 Symposium
methods to be proposed for handling dynamic gestures. Sim- (TENSYMP), 2016 IEEE, pages 104–109. IEEE, 2016.
ilarly, existing methods typically cover no more than a few [11] M. Joselli and E. Clua. grmobile: A framework for
tens of signs, and there is significant room for improvement touch and accelerometer gesture recognition for mobile
until methods can cover the several thousands of signs that games. In 2009 VIII Brazilian Symposium on Games
users of a sign language employ in their daily usage. Ex- and Digital Entertainment, pages 141–150. IEEE,
tending vision-based recognition systems to cover dynamic 2009.
gestures and thousands of signs may strain the hardware [12] T. J. Joshi, S. Kumar, N. Tarapore, and V. Mohile.
capabilities of smartphones. While smartphone hardware Static hand gesture recognition using an android
specs are expected to continue to improve rapidly, cloud device. International Journal of Computer
processing could push the boundaries further ahead by al- Applications, 120(21), 2015.
leviating the hardware requirements on the mobile device. [13] R. Kamat, A. Danoji, A. Dhage, P. Puranik, and
However, maintaining interactivity and low latency while us- S. Sengupta. Monvoix-an android application for
ing cloud processing can also be challenging, and these are hearing impaired people. Journal of Communications
also issues that we expect future research to focus on. Technology, Electronics and Computer Science,
8:24–28, 2016.
5. ACKNOWLEDGMENTS [14] L.-J. Kau, W.-L. Su, P.-J. Yu, and S.-J. Wei. A
This work was partially supported by National Science real-time portable sign language translation system. In
Foundation grants IIS-1055062 and IIS-1565328. Any opin- 2015 IEEE 58th International Midwest Symposium on
ions, findings, and conclusions or recommendations expressed Circuits and Systems (MWSCAS), pages 1–4. IEEE,
in this publication are those of the authors, and do not nec- 2015.
essarily reflect the views of the National Science Foundation. [15] J. Kruskall and M. Liberman. The symmetric time
warping algorithm: From continuous to discrete. time
6. REFERENCES warps, string edits and macromolecules, 1983.
[1] J. Canny. A computational approach to edge [16] H. Lahiani, M. Elleuch, and M. Kherallah. Real time
detection. IEEE Transactions on pattern analysis and hand gesture recognition system for android devices.
machine intelligence, (6):679–698, 1986. In Intelligent Systems Design and Applications
[2] B. Choe, J.-K. Min, and S.-B. Cho. Online gesture (ISDA), 2015 15th International Conference on, pages
recognition for user interface on accelerometer built-in 591–596. IEEE, 2015.
mobile phones. In International Conference on Neural [17] G. Niezen and G. P. Hancke. Gesture recognition as
Information Processing, pages 650–657. Springer, 2010. ubiquitous input for mobile phones. In International
[3] E. Costello. American sign language dictionary. Workshop on Devices that Alter Perception (DAP
Random House Reference &, 2008. 2008), in conjunction with Ubicomp, pages 17–21.
[4] H. Elleuch, A. Wali, A. Samet, and A. M. Alimi. A Citeseer, 2008.
static hand gesture recognition system for real time [18] S. L. Phung, A. Bouzerdoum, and D. Chai. Skin
mobile device monitoring. In Intelligent Systems segmentation using color pixel classification: analysis
Design and Applications (ISDA), 2015 15th and comparison. IEEE transactions on pattern
International Conference on, pages 195–200. IEEE, analysis and machine intelligence, 27(1):148–154,
2015. 2005.
[5] P. Gandhi, D. Dalvi, P. Gaikwad, and S. Khode. [19] L. Prasuhn, Y. Oyamada, Y. Mochizuki, and
Image based sign language recognition on android. H. Ishikawa. A hog-based hand gesture recognition
International Journal of Engineering and Techniques, system on a mobile device. In 2014 IEEE
1(5):55–60, 2015. International Conference on Image Processing (ICIP),
[6] H. P. Gupta, H. S. Chudgar, S. Mukherjee, T. Dutta, pages 3973–3977. IEEE, 2014.
and K. Sharma. A continuous hand gestures [20] C. Preetham, G. Ramakrishnan, S. Kumar, A. Tamse,
recognition technique for human-machine interaction and N. Krishnapura. Hand talk-implementation of a
using accelerometer and gyroscope sensors. IEEE gesture recognizing glove. In India Educators’
Sensors Journal, 16(16):6425–6432, 2016. Conference (TIIEC), 2013 Texas Instruments, pages
328–331. IEEE, 2013.
[21] J. L. Raheja, A. Singhal, and A. Chaudhary. Android
based portable hand sign recognition system. arXiv
preprint arXiv:1503.03614, 2015.
[22] G. A. Rao and P. Kishore. Sign language recognition
system simulated for video captured with smart phone
front camera. International Journal of Electrical and
Computer Engineering (IJECE), 6(5):2176–2187, 2016.
[23] D. E. Rumelhart, G. E. Hinton, and R. J. Williams.
Learning representations by back-propagating errors.
Cognitive modeling, 5(3):1, 1988.
[24] A. Saxena, D. K. Jain, and A. Singhal. Hand gesture
recognition using an android device. In
Communication Systems and Network Technologies
(CSNT), 2014 Fourth International Conference on,
pages 819–822. IEEE, 2014.
[25] A. Saxena, D. K. Jain, and A. Singhal. Sign language
recognition using principal component analysis. In
Communication Systems and Network Technologies
(CSNT), 2014 Fourth International Conference on,
pages 810–813. IEEE, 2014.
[26] M. Seymour and M. Tšoeu. A mobile application for
south african sign language (sasl) recognition. In
AFRICON, 2015, pages 1–5. IEEE, 2015.
[27] P. Viola and M. J. Jones. Robust real-time face
detection. International journal of computer vision,
57(2):137–154, 2004.
[28] H. Wang, A. Stefan, S. Moradi, V. Athitsos, C. Neidle,
and F. Kamangar. A system for large vocabulary sign
search. In European Conference on Computer Vision,
pages 342–353. Springer, 2010.
[29] X. Wang, P. Tarrı́o, E. Metola, A. M. Bernardos, and
J. R. Casar. Gesture recognition using mobile phone’s
inertial sensors. In Distributed Computing and
Artificial Intelligence, pages 173–184. Springer, 2012.
[30] K. S. Warrier, J. K. Sahu, H. Halder, R. Koradiya,
and V. K. Raj. Software based sign language
converter. In Communication and Signal Processing
(ICCSP), 2016 International Conference on, pages
1777–1780. IEEE, 2016.
[31] J. Weston and C. Watkins. Multi-class support vector
machines. Technical report, Citeseer, 1998.
[32] Wikipedia. American manual alphabet,
https://en.wikipedia.org/wiki/american manual alphabet,
2016.