A Sign Language Convention Using YOLOv5 Updatd

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 5

A Sign Language Convention Using YOLOv5

Siddharth Kharche , Umesh Kardile , Sahil Wadaskar , Dashrath


1 2 3

Bogati , Neha Sharma


4 5

1,2,3,4: Second Year Engineering students, Department of Computer Engineering,


Ajeenkya D. Y. Patil School of Engineering, Pune, Maharashtra, India
5: Assistant Professor, Computer Engineering Department,Ajeenkya D. Y. Patil
School of Engineering, Pune, Maharashtra, India

ABSTRACT
To detect sign language, YOLOv5 technology is used to train a neural network to identify hand
gestures and motions connected to sign language. A cutting-edge object identification system
called YOLOv5 is capable of effectively identifying things in pictures and video frames. The
project's goal might be to create a real-time system for recognizing sign language that can
transform movements into spoken or written language, facilitating communication for those who
are hard of hearing or deaf. Additionally, this technology could be applied to the development of
assistive equipment, accessibility in public areas, and education.As part of the project, a dataset
of sign language gestures would be gathered, labeled, and used to train the YOLOv5 model. The
trained model would then be able to translate or provide immediate feedback by detecting and
recognizing these motions in real-time. To guarantee the model's efficacy in practical
applications, assessment of its performance and correctness is essential.

KEYWORDS: Gestures, Sign language, YOLOv5.

I.INTRODUCTION
Communication plays a vital role in our lives, impacting social and emotional well-being. For
individuals with hearing impairment, communication can be challenging due to their inability to
hear sounds, including their own voices. Sign language, a structural system of hand gestures and
visual motions, serves as a powerful means of communication for the deaf community. However,
it remains underutilized by those without hearing Important. To address this gap, sign language
recognition systems have emerged. These technologies automatically convert sign language
gestures into text or voice formats, facilitating human-computer interaction. .. While the dynamic
gesture is utilized for particular concepts, the static gesture is used to indicate the alphabet and
numbers. Additionally, words, sentences, etc. are dynamic. Hand movements make up the static
gesture, while head, hands, or both can move in the latter. Three main components make up sign
language, which is a visual language: finger printing, word-level sign vocabulary, and non-
manual aspects. In contrast to the latter, which is keyword-based, finger spelling is used to spell
words letter by letter and convey the meaning. Though there have been several research efforts
over the past few decades, designing a sign language translator remains extremely difficult.
II.SYSTEM MODEL AND ASSUMPTIONS
The system model consists of a camera-equipped device that can record video input in real time.
The YOLOv5 object detection model, a deep learning framework well-known for its
effectiveness in identifying objects in photos and videos, is used to process this data. Within the
video frames, the trained YOLOv5 model recognizes human hands making distinct sign signals.
The availability of a trustworthy dataset for training the YOLOv5 model particularly for sign
language hand motions is one of the fundamental presumptions behind this system architecture.
The limitations of real-time processing, such as hardware capabilities and computational
resources, are also taken into consideration by assumptions. Furthermore, the system's
performance assumptions are based on the assumption that hand motions in the video feed are
unambiguous and unobstructed. After a sign gesture is successfully detected, the system
proceeds through a series of steps for recognition and interpretation. Preprocessing the identified
hand regions, obtaining pertinent features, and mapping them to matching sign language words
or symbols are some of these processes. Using a pre-trained machine learning model, the
interpretation phase may involve adding methods such as sequence prediction or classification. A
sign language recognition system that uses YOLOv5 technology can be designed and
implemented in an efficient manner by defining important assumptions and creating a clear
system model. These elements and presumptions offer a foundation for creating a reliable and
useful application that can read sign language in real time. Depending on certain use cases and
deployment conditions, improvements and adjustments can be performed.
III.EFFICIENT COMMUNICATION

A strong and simple method is required to develop effective communication for a sign language
convention utilizing YOLOv5.
Gathering and Preparing Data
Gathering Datasets: assemble a varied collection of sign language movements that includes a
large number of different signs and variants.
Data Preprocessing: Make that the lighting, background, and hand locations are consistent by
cleaning and preprocessing the dataset.
Instruction of Models
Model Choice: Because YOLOv5 is an accurate and efficient object identification model, it
should be used.
Optimizing YOLOv5: Utilize the sign language dataset to train the YOLOv5 model for hand
motion detection and localization.
Camera Integration for Real-time Object Detection: Put in place real-time webcam or camera
video capture.

IV.SECURITY
Make sure users are aware of the system's privacy policy, which includes information on data
collecting, usage, and storage procedures. Before collecting or using any personal data, including
gesture or video data, get users' express approval. Keep all software components ,including all
the technologies upto date. YOLOv5 sign language convention systems require a complete
approach to ensure their security, which includes secure communication, data encryption, access
control, and secure deployment techniques. You may secure user data and privacy while
improving the system's confidentiality, integrity, and availability by putting these security
measures into place. Maintaining a safe and dependable system also requires being up to date on
the most recent security threats and best practices.
V. RESULT AND DISCUSSION

Performance of Resulting Object Detection:YOLOv5 showed effective and precise object


recognition, especially when it came to hand gesture localization within sign language frames.
To evaluate the performance of the model, measures for precision and recall for hand detection
were taken.
Accurate Gesture Recognition: Based on the identified hand regions, the gesture recognition
model classified sign language gestures with a given degree of accuracy.
The effectiveness of the recognition model was assessed using accuracy metrics, such as
classification accuracy and confusion matrix. Processor Speed in Real Time:
With little delay between gesture detection and translation, the system demonstrated real-time
processing capabilities appropriate for interactive applications. Obstacles and Upcoming Tasks:
Determine the obstacles that the project faced, such as limited computing resources, poor data
quality, or security flaws.Make suggestions for possible directions for future development,
including adding multi-hand motions, enhancing the capacity for gesture detection, or adding
more sophisticated security features. Effects and Uses: Talk about the possible benefits and real-
world effects of the sign language convention system for improving accessibility and
communication for those who have hearing loss.
Examine the technology's wider uses in the domains of assistive technology, healthcare, and
education.

VI.CONCLUSION

In summary, the construction of a sign language convention system with security considerations
utilizing YOLOv5 technology shows encouraging outcomes in terms of data security, real-time
processing speed, gesture recognition accuracy, and object identification. We can learn more
about the efficacy and potential implications of this cutting-edge technology in promoting
inclusive communication and accessibility for a variety of demographics by assessing the
system's results and having meaningful conversations about model performance, user experience,
security procedures,futuredirections.
References

[1]Survey on Different Gesture Recognition Technologies and Techniques, Noor Adnan


Ibraheem and Rafiqul Zaman Khan, International Journal of Computer Applications, Vol. 50,
No. 7, 2012.
[2]A Fast Algorithm for Multilevel Thresholding, Ping-Sung Liao, Tse-Sheng Chen, and Pau-
Choo Chung, 2001, Journal of Information Science and Engineering 17, pp. 713-727
[3]Background subtraction techniques by Dr. Alan M. McIvor, Image and Vision Computing
New Zealand 2000 (IVCNZ00)
[4]Skin Segmentation Using Color and Edge Information, Proceedings of the International
Symposium on Signal Processing and its Applications, July 1-4, 2003, Paris, France, Son Lam
Phung, Abdesselam Bouzerdoum, and Douglas Chai
[5]Motion-based Segmentation and Region Tracking in Image Sequences, Jorge Badenas, Josee
Miguel Sanchiz, and Filiberto Pla, 2001, Pattern recognition 34, pp. 661-670Visual Hand
Gestures categorization using Wavelet Transform and Moment Based Features was published in
Sanjay Kumar and Dinesh K. Kumar's 2005 International Journal of Wavelets, Multiresolution
and Information Processing, Volume 3, Issue 1.

[6]Qing Chen, Nicolas D. Georganas, Emil M. Petriu, Hand Gesture Reconition Using Real-
Time Vision in Real Time Instrumentation and Measurement Technology Conference-IMTC
2007, Parsaw, Poland.
[7] F. N. H. Al-Nuaimy, "Design and implementation of deaf and mute people interaction
system," in Proceedings of the International Conference on Engineering and Technology (ICET),
August 2017, pp. 1–6.
[8] Human—Hand posture classification for robotic teleoperation using wearable sensor, P. P.
Devnath, A. S. Kundu, O. Mazumder, and S. Bhaumik, Proc. IEEE Region 10 Symp.
(TENSYMP), Jun. 2019, pp. 114–119.
[9] In Proc. 3rd Int. Conf. Comput. Inform. Commun. Technol. (CICT), New York, NY, USA,
Feb. 2017, pp. 1–5, S. Devi and S. Deb, "Low cost tangible glove for translating sign gestures to
speech and text in Hindi language."
[10] "High-speed gesture modelling through boundary analysis of active signals from wearable
data glove," by A. Samraj, R. Kumarasamy, K. Rajendran, and K. Selvaraj, is published in the
International Journal of Grid Utility Computing, volume 10, issue 1, pages 29–35, 2019.
[11] M. S. Verdadero and J. C. D. Cruz, "An assistive hand glove for persons with hearing and
speech impairments," in Proceedings of the 11th International Conference on Humanities,
Nanotechnology, Information Technology, Communication Control, Environment, and
Management (HNICEM), November 2019, pp. 1-6.

You might also like