Gesture Based Mouse
Gesture Based Mouse
Gesture Based Mouse
Abstract—Researchers all throughout the world are Distinguishing between gesture and posture, gesture focuses
working to improve the responsiveness of our gadgets on hand movement, whilst posture highlights the shape of
and to make it possible for them to operate with the the hand. The mouse is a key instrument for Human-
least amount of physical interaction. There are three Computer Interaction (HCI), however in some real-world
primary varieties of mice—wired, wireless, and situations, such Human-Robot Interaction (HRI), it might not
Bluetooth—that need electricity to function. Mice are be practical. Numerous studies have looked at alternative
an essential tool in human-computer interaction. This HCI methods, and hand gestures have come to light as a
study presents a novel computer system that can be natural and intuitive replacement for the conventional
used interactively without a real mouse. The device mouse, especially useful in situations such as human-
recognizes hand motions without any additional machine interfaces. While many people take technology for
hardware thanks to advanced machine learning and granted, utilizing standard gadgets can be challenging for
computer vision algorithms. This approach, which is those who have physical disabilities. Extreme impairments in
compatible with CNN models provided by mediapipe, mobility can severely impair fine motor function, making it
has tremendous potential for a variety of users, difficult to interact with a regular mouse. Sufficient assistive
especially those who struggle with conventional mouse technology are essential for ensuring accessibility for those
operation due to partial paralysis. The technology uses who require it. The idea is to design a virtual mouse system
computer vision to let people use the mouse by showing that replaces touch displays with a camera enabling easier-
various finger counts. Various hand motions are to-use connection with the device. Using vision-based
required for actions like clicking and dragging objects. Computer Vision (CC) can help a webcam reach its full
Additionally, this system uses a camera as an input potential. Using specified labels, this method effectively
device and runs on a single computer. The proposed records and predicts hand motions. The ultimate objective is
solution, which shows the camera's output on the to create a user-friendly virtual mouse system that
screen for user adjustment, utilizes Python and communicates with devices via webcam-based vision,
OpenCV. providing a good replacement for touch displays and
improving accessibility, particularly for those with physical
Keywords- Mediapipe, Convolutional Neural disabilities.
Network, Computer Vision, Hand Recognition, Partial
Paralysis, Human-Computer Interaction, Human- II. LITERATURE REVIEW
Robot Interaction. Researchers and automobile organizations have spent the
previous five years carefully examining the proposed
I. INTRODUCTION problem in an effort to come up with a solution. The
suggested remedies take multiple tacks, focusing on the
Within the field of computer vision, gesture recognition, examination of various patterns.
which focuses on recognizing hand movements as a kind of In order to enhance user-computer interaction, Aashni Hariaa
visual communication has long been an intriguing et al. [1] presented "Hand Gesture Recognition for Human
challenge. These hand gestures involve motions across the Computer Interaction," which makes use of a backdrop
middle of the palm, finger placement, and hand form. Hand extraction and contours recognition algorithm. A variety of
gestures can be categorized as either static or dynamic. sensor modalities, such as gestures, voice, facial expressions,
Static gestures consist of a single, steady hand form, and body language, can be used to facilitate human contact.
whereas dynamic gestures include a series of hand motions, Horatiu-Stefan et al. presented a paper titled "Human Hand
such as waving. Gesture Based System for Mouse Cursor Control." This
model makes use of a computer, a webcam, and a blue hand traditional techniques such as manually entering buttons or
pad. Hand pad postures are used to guide operations. The adjusting the placements of a physical computer mouse.
OpenCV library and Visual C++ 2008 were utilized in the Instead of using a real computer mouse to do similar tasks, it
development of the model. uses computer vision and a camera to monitor and control
A "Mouse Control Using a Web Camera Based on Color different mouse movements.
Detection" was proposed by Abhik Banerjee et al. [2] to Firstly, the hand and its key characteristics are identified by
control click events and cursor movements by sensing the means of Mediapipe. Mediapipe detects a total of 21
color of the image. distinct colors correspond to distinct important points. To recognize the hand, the process uses
cursor controls, and clicking operations are carried out by Mediapipe's palm detection model and hand landmark
simultaneously recognizing these colors. This methodology model. When the hand landmarks are taken into
was developed with the help of MATLAB software and the consideration, palm detection is a simple process.
MATLAB Image Processing Toolbox.
Abhilash et al. [3] introduced the "Virtual Mouse Using
Hand Gesture" approach for manipulating the mouse
pointer and clicking actions. In this example, a red object
was used to build a mask. This approach was implemented
using Python software, which included a number of
modules and functions.
To deal with the mouse pointer and click actions, Alisha
Pradhana et al. [4] proposed the "Design of Intangible
Interface for Mouseless Computer Handling Using Hand
Gestures". The Microsoft Visual Studio integrated
development environment was used to apply this strategy,
which uses a convex hull algorithm. To minimize noise and
identify the user's hand, a red glove is utilized.
"A Real Time Hand Gesture Recognition System Using
Motion History Image" was proposed by Chiung Hsieh et
al. [5] to manipulate the mouse cursor. To minimize the
number of incorrect classifications, the recommended
method makes use of an adaptive skin color detection
model. A C++ software platform with the OpenCV image
processing package loaded was used to construct these
techniques.
"Cursor Control Using Hand Gestures" was the idea put out
by Pooja Kumari et al. [6] to operate a mouse using hand
motions that are recorded by a camera. Using this method,
the camera functions as a sensor to record and detect color
tips applied on the hand. Known as the marker-based
technique, it requires users to hold color-tip mouse cursors Figure 1 Co-ordinates or landmark in hand
in their hands. The MATLAB environment, the MATLAB
1. CAMERA
Image Processing Toolbox, and the OpenCV library were
all utilized in the implementation of this technique. The webcam that is connected to the computer or the
laptop's built-in camera provides the frames that the
system uses to operate. In order to accomplish this, a
III. MATERIALS AND METHODS
video capture object is set up, allowing for real-time
The program relies on specific software requirements
webcam video capturing. Frames are gradually
crucial for its proper operation. These essential
captured by the camera and then sent to the system
prerequisites ensure the functionality, performance, and
for additional processing.
usability of the project:
2. CAPTURING
Python (version 3.6 – 3.8.5)
Through the use of an infinite loop, the camera
pycaw captures every frame until the application ends.
PyAutoGUI Following their extraction from the real-time video,
AutoPy the frames undergo processing that changes their
MediaPipe color space from BGR to HSV.
NumPy
3. COLOR DETECTION AND MASKING
The project development process is based on these Color detection works by identifying color pixels in
predefined software requirements, which ensures the final the webcam frames that correspond to fingertips that
product satisfies anticipated functional, performance, and have colored caps on them. This is the first and
usability criteria. It is essential to remember that, depending crucial stage in the suggested system. This process
on the project's particular objectives and use cases, more yields a grayscale image in which the color cap is
extensive and specialized requirements can be necessitated. highlighted by pixel intensity, setting it out from the
Using a real-time camera to control cursor movement, this surrounding frame. The color cap itself and the
project proposes an innovative approach to Human- rectangle enclosing it are then monitored. Then, by
Computer Interaction (HCI) that offers an alternative to keeping an eye on these color caps, gesture detection
becomes simpler.
4. GESTURE RECOGNITION
4.1 Mouse Movements
First, by extracting the coordinates of the centers of
the identified rectangles, the computation may be
performed to find the center of the two color objects
that were successfully detected. The built-in
OpenCV function is used to draw a line between
these two locations. Next, the following equation is
used to find the line's midpoint:
M=(Xa+Xb/2, Ya+Yb/2)
REFERENCES
[1] Aashni Hariaa , Archanasri Subramaniana ,
Nivedhitha Asokkumara , Shristi Poddara and Jyothi S
Nayak “Hand Gesture Recognition for Human Computer
Interaction” Proc.ICACC(International Conference on
Advances in Computing & Communications), 2017
August,pp367–374.
[2] Abhik Banerjee, Abhirup Ghosh, Koustuvmoni
Bharadwaj,” Mouse Control using a Web Camera based
on Color Detection”,IJCTT,vol.9, Mar 2019.
[3] Abhilash S , Lisho Thomas, Naveen Wilson, and
Chaithanya “VIRTUAL MOUSE USING HAND
GESTURE” Proc. International Research Journal of
Engineering and Technology (IRJET), e-ISSN: 2395-0056
p-ISSN: 2395-0072,Apr-2018, Volume: 05 Issue:04.
[4] Alisha Pradhana , B.B.V.L. Deepak “Design of
Intangible Interface for Mouseless Computer Handling
using Hand Gestures” ICCCV(International Conference
on Communication, Computing and Virtualization), 2016,
oi: 10.1016/j.procs.2016.03.037.
[5] Chiung Hsieh, D.-Hua Liou, & D. Lee, , “A real time
hand gesture recognition system using motion history
image,” Proc. IEEE Int'l Conf. Signal Processing Systems
(ICSPS), 2.10.1109/ICSPS.2010.5555462, 2020.