Gesture Controlled Virtual Mouse

Download as pdf or txt
Download as pdf or txt
You are on page 1of 8

11 I January 2023

https://doi.org/10.22214/ijraset.2023.48879
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com

Gesture Controlled Virtual Mouse


Rajat Dudhapachare5, Mayur Awatade2, Mayur Kapgate3, Nihal Vaidya4, Pushpak Kakde1, Prof. Rajesh Nakhate6
1, 2, 3, 4, 5
Student, 6Mentor, Department of Information Technology, G H Raisoni Institute of Engineering, Nagpur

Abstract: As a result of the pandemic, a technological paradigm shift has happened. One such field of research is the
development and usage of a virtual mouse in place of a real one. Despite the fact that a remote mouse or Bluetooth mouse is
popular nowadays, it requires various factors such as cost and power. There would be no such cutoff points in the proposed
system, assuming signal acknowledgement. The primary breakthroughs used in this review are object recognizable proof and
image management. Thus, this study proposes a cutting-edge method for dealing with mouse developments employing an
ongoing camera with the help of the hand. It does not necessitate the use of any electrical device. It will use various hand
gestures to direct activities such as left click and hauling. For the product's execution, Python and OpenCV will be used. As a
result, the proposed mouse framework eliminates the need for a device.
Keywords: Human-Computer Interaction (HCI), Object detection, Image processing, OpenCV.

I. INTRODUCTION
Because of virtualization and the gradual move toward dramatic breakthroughs such as the metaverse, ordinary human-machine
interaction has evolved. As technology advances, devices get smaller and more affordable. A few gadgets have gone remote, while
others have slipped through the cracks. [1]. Motions can convey something while drawing in with others. From easy to
unimaginably confounded hand developments. For instance, we can highlight something (an item or individuals) or use various
straightforward signals or movements that are conveyed in gesture-based communication that are coordinated with their punctuation
and word reference, all the more frequently known as gesture-based communications. Therefore, people can convey all the more
actually by utilizing hand movements as a gadget with the assistance of PCs. [3] In this work, we are illustrating, genuine signal.
The fundamental arrangement includes a low-cost web camera that may be used to provide data to the framework. The overall
interaction is divided into four stages: outline capturing, image handling, region extraction, and include coordination. It can also be
referred to as equipment because it has a camera for tracking hands. [7] Notwithstanding the limits, the PC innovation actually
keeps on developing, so does the significance of the human PC associations. Starting from the presentation of a cell phone that can
be connect with contact screen innovation, the world is beginning to request a similar innovation to be applied on each mechanical
gadget, this incorporates the work area framework. In any event, despite the fact that touch screen innovation for the work area
framework exists, the expense might be extremely high. As a result, a virtual human PC collaboration gadget that replaces the real
mouse or console by using a camera or other image capturing gadgets might be an alternative means for the touch screen. This
device, the camera, will be constantly employed by a product that screens the signals sent by the client in order to deal with it and
signify movement of a pointes, similar to a real mouse. We investigate which motions are appropriate, how to recall them, and
which instructions they should control. There are several mechanical advancements occurring in today's culture, such as regular
language handling, biometric confirmation, and facial recognition, all of which can be found on our tablets, iPads, PCs, and mobile
phones. Since human hand stances and motions are as yet a strong entomb human correspondence mode, the exploration local area
is keen on utilizing them to control or connect with artificial frameworks. The utilization of hand motions as control or
correspondence methodology proceeds to invigorate mainstream researchers' interest. With the objective that we can similarly make
workspaces as simple to use as mobile phones.

A. Scope and Proposed Model


There are different existing frameworks. One with customary mouse (equipment apparatus) to explore around in the screen. It is
beyond the realm of possibilities to use hand movements to get to the screen.
Other is the signal framework which utilizations variety tapes to distinguish the motions. What's more, the functionalities performed
are static in nature which are fundamental. Utilizing the ongoing framework, we can use a PC/PC with a web camera and
mouthpiece to control the mouse and execute basic tasks without extra PC equipment. Moreover, a voice partner is utilized to play
out extra errands.[3]

©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1593
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com

B. Related Works
Related work in the area is talked about which gives progress into the hypothetical foundation. Signal acknowledgment may be a
modern way to the PCs to figure out human motion (Body Language). It will fabricate a more prominent collaboration between
human and PC machines as opposed to crude text-based association. The vast majority of the marker-based motion acknowledgment
mouse utilizes something like two tone markers to follow. Due to identifying different varieties, the framework gets slow and a few
laggings show up on the framework during the hour of execution[2].
Neethu. P.S et al. provided a Constant Static and Dynamic Hand Motion Acknowledgement in which they planned, created, and
focused on an acceptable motion acknowledgement that can be used in a variety of human-PC cooperation applications structure for
continuous. Under any event, it couldn't work at a mind-boggling foundation and could only be processed in bright light.
Patil et al. proposed a machine-UI that confirms hand movement utilizing core PC vision and media methods. However, a
significant break point must be reached before working with movement connection estimates, skin pixel recognition, and hand
division from set apart methods can be completed.
The observed hand signals control various virtual things created with Solidarity using a virtual-mouse-based object controlling
framework. The framework was tested using six hand signals, and it was discovered that the framework may be used to operate
several virtual items. Hand motion recognition, which we see frequently on our PDAs, is a cutting-edge way of human-PC
communication, i.e., we have some control over important for a wide variety of folks. This document is presented on Page 5/14 in
light of this concept. This paper explains the computations and procedures for variety identification and virtual mouse in detail. Our
system may be validated by displaying our hands in front of the webcam and recognizing hand signals.
Varun et al. demonstrated that Hand Signal Acknowledgment plays an important role in human-PC communication. Creators
demonstrated that there are several new mechanical advancements occurring, such as biometric verification.
Boruah et al. demonstrated that a hand motion recognition framework is a typical and simple technique of communicating in this
day and age. The expansion of teaching methodologies via the use of technology subordinate useful items to enhance
correspondence and collaboration between the instructor and the understudy is an important component of current e-learning. In this
study, we present an intelligent learning-assistance gadget based on a vision-based hand motion acknowledgement architecture that
uses Media Pipe. [4]

II. METHODOLOGY
The Following Object In this paper, we will create A Virtual Mouse that makes use of Python, OpenCV, Autopy, and
MediaPipesystem. We'll go over how this framework was created piece by piece:

A. OpenCV
OpenCV stands for Open-Source PC Vision, and it is a C++ library used for PC Vision. It is cross-stage and usable. It offers
continuous GPU speed increase characteristics that are used in a variety of areas such as 3D and 2D element toolboxes, Faces, and
motion acknowledgment frameworks.

B. Autopy
Autopy is a Python cross-stage GUI motorization package that displays our fingers. It indicates whether a finger is pointing up or
down in a 0 or 1 result plan. OpenCV anticipates the outcome and creates a true.

C. MediaPipe
A cross-stage structure called MediaPipe is used to create multimodal ML pipelines, and it is an open source technology that Google
has made available. The MediaPipe Hands Model learns the precise identification of hands from a large number of historical data
sources, making hand discovery possible. By using hand motion control and signing to receive information, MediaPipe Hands may
establish the purpose of communication. Additionally, it can enable the overlay of computerized real-world material in augmented
reality (artificial intelligence) image frames. [5]
It employs AI (ML) to imprint 2D 3D milestones of a hand from a single edge. MediaPipe Hands makes use of several models that
are available in an ML pipeline. A Hand Milestone Model operates on the cropped picture area, but a Palm Location Model works
on the whole image and returns a placed hand leaping box that is used for the hand location procedure independently. The pipeline
created may operate in phases, allowing for flexibility in flexible and work areas.

©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1594
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com

The following advances are included to help with the calculation:


1) The first stage is to take the photograph using the camera.
2) The camera then detects and eliminates the human hand from the information image.
3) At that point, the position of the human hand is saved in the framework utilizing the standard" coordinate- framework".
4) Then, when the following casing is caught. The position of the hand from the succeeding casing is captured and saved in the
framework.
5) The position of two hands is then examined, and the cursor moves accordingly.
6) For the tapping operation, the point between the two hands of the finger is evaluated, and if the point is less than 15 degrees, the
framework responds with a left-click. Along these lines, complete mouse operation should be available with uncovered hands.

D. PyCharm
An Integrated Development Environment (IDE) used for Python programming is called PyCharm. It supports web development
with Django and offers code analysis, a graphical debugger, an integrated unit analyzer, and compatibility with form control
frameworks (VCSes). JetBrains, a company based in the Czech Republic, developed PyCharm. [5]
It is cross-stage dealing with Windows, Macintosh operating system X and Linux. PyCharm has an Expert Version, delivered under
an exclusive permit and a Local area Release delivered under the Apache License.[6] PyCharm People group Release is less broad
than the Expert Release.

E. PyAutogui
To replicate mouse pointer movements and snaps as well as console button clicks, PyAutoGUI is a Python package that runs on
Windows, MacOS X, and Linux.
Several features of PyAutoGUI include:
1) Clicking and dragging the mouse over windows for another program.
2) keystrokes being sent to applicants (for example, to fill out forms).
3) Capture screenshots of the screen whenever you come across an image, such as a button or checkbox.
4) You may move, resize, maximize, minimize, or close an application's window by finding it (Windows-only, currently).
5) Display message and alarm boxes.

III. WORKING

Hand image

Identifying the
palm on hand

Mark 3D hand
key points on
palm(Hand
Landmark)

Identify hand gesture based on mapping


hand key points with menus of application

Display menu based


on hand gesture

©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1595
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com

1) Hold the mouse pointer still

2) For performing Left Click

3) For performing Right Click

©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1596
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com

4) For Scrolling and moving pointer.

5) To select and drag the files

IV. RESULT
The Hand signal acknowledgment and partner framework plays become a significant part in building proficient human-machine
cooperation. Execution utilizing hand signal acknowledgment guarantees far reaching in innovation industry. The MediaPipe as one
system in view of AI assumes a powerful part in fostering this application utilizing hand motion acknowledgment.

V. CONCLUSION
The Virtual Mouse Utilizing Hand Motions project was created to enable the usage of the ideal PC mouse without the need for a
physical object or surface. using a camera to follow hand gestures and then translate them into cursor actions that respond. The
project is still in its early stages, but the team is optimistic that the Virtual Mouse will eventually have the option to replace the
conventional PC mouse. The project may be able to help those who have limited mobility or are unable to operate a standard mouse.

REFERENCES
[1] Hritik Joshi Medi-Caps University Nitin waybhase Medi-Caps University Ratnesh Litoriya Medi-Caps University Dharmendra Mangal Medi-Caps University.
[2] CN Sujatha, S.P.V Subbarao, P.Preetham, P.Surajvarma, Y.Upendra 1.Professor, ECE, Sreenidhi Institute of Science and Technology, Ghatkesar, Hyderabad
2.HOD, ECE, Sreenidhi Institute of Science and Technology, Ghatkesar, Hyderabad.
[3] Cherukupally Karunakar Reddy , Suraj Janjirala , Kevulothu Bhanu Prakash Department of Computer Science and Technology, Sreenidhi Institute of Science
and Technology

©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1597
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue I Jan 2023- Available at www.ijraset.com

[4] VIRTUAL MOUSE USING HAND GESTURES Prof. D.L. Falak*


[5] Virtual Mouse Using Hand Gesture and Voice Assistant Khushi Patel1 , Snehal Solaunde2 , ShivaniBhong3 , Prof. Sairabanu Pansare Nutan College of
Engineering and Research
[6] Virtual Mouse Control Using Hand Class Gesture Vijay Kumar Sharma, Vimal Kumar, Md. Iqbal, Sachin Tawara, Vishal Jayaswal Department of Computer
Science and Engineering MIET, Meerut
[7] Mr. Gajendra Moroliya1 , Mr. Sahil Patwekar2 , Prof. S.P. Gopnarayan3 1,2,3Electronics and telecommunication, AISSMS IOIT, Pune, India “Virtual Mouse
Using Hand Gesture”
[8] Banerjee, A., Ghosh, A., Bharadwaj, K., & Saikia, H. (2014). Mouse control using a web camera based on color detection preprint arXiv:1403.4722.
[9] Balamurugan. C1, Arumuga Kumar. M2, Arun N3 and Deepak proposed a study on “HCI SYSTEM WITH HAND GESTURE” in International Research
Journal of Engineering and Technology (IRJET).

©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1598

You might also like