0% found this document useful (0 votes)
13 views

Seminar_Template (3)

Uploaded by

swapnilbadave007
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
13 views

Seminar_Template (3)

Uploaded by

swapnilbadave007
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 28

A

SEMINAR REPORT

ON

Artificial Intelligence Virtual Mouse

SUBMITTED TO THE SAVITRIBAI PHULE PUNE UNIVERSITY, PUNE


IN PARTIAL FULFILLMENT OF THE REQUIREMENTS
FOR THE AWARD OF THE DEGREE OF

BACHELOR OF ENGINEERING
INFORMATION TECHNOLOGY

BY

Swapnil Ananda Badave


Roll No: 33303
Exam Seat No: 33303

Under the guidance of


Dr. Emmanuel Markappa

D EPARTMENT O F I NFORMATION T ECHNOLOGY


P UNE I NSTITUTE OF C OMPUTER T ECHNOLOGY
S R . N O 27, P UNE -S ATARA ROAD , D HANKAWADI
P UNE - 411 043.
AY: 2024-2025

P:F-SMR-UG/08/R0
SCTR’s PUNE INSTITUTE OF COMPUTER TECHNOLOGY
DEPARTMENT OF INFORMATION TECHNOLOGY

CERTIFICATE

This is to certify that the Seminar work entitled


Artificial Intelligence Virtual Mouse

Submitted by

Name : Swapnil Ananda Badave


Exam Seat No:XXX .

is a bonafide work carried out under the supervision of Dr.Emmanuel Markappa


and it is submitted towards the partial fulfillment of the requirements of Savitribai
Phule Pune University, Pune for the award of the degree of Bachelor of Engineering
(Information Technology).

Dr.Emmanuel Markappa Dr. A. S. Ghotkar


Seminar Guide HOD IT

Dr. S. T. Gandhe
Principal

Date:
Place:

P:F-SMR-UG/08/R0 i
Acknowledgement

I extend my deepest thanks to Dr. Emmanuel Mark, my seminar guide, for his invaluable guid-
ance, insightful critiques, and steadfast support throughout this seminar. His expertise played a
crucial role in shaping the course of my research, and his encouragement motivated me to ex-
amine the topic of bias and bias mitigation in healthcare AI from a critical and comprehensive
perspective.
I am also sincerely thankful to Mr. Ravindra B. Murumkar, the seminar reviewer, for his de-
tailed feedback and helpful suggestions, which significantly improved the quality of this work.
His thoughtful observations offered a clear path during the revision process. Additionally, I
would like to express my gratitude to Dr. Archana Ghotkar, Head of the Department, for her
administrative support and for providing the academic resources needed during the seminar.
Her leadership and guidance ensured an environment conducive to academic growth.
Finally, I want to acknowledge the academic and technical staff of the department for their
assistance in various ways throughout the research process, and I am deeply thankful to my
family and friends for their unwavering support and encouragement during this seminar.

Name :Swapnil Ananda Badave


Exam Seat No:XXX

ii
Abstract

The AI Virtual Mouse uses computer vision to track hand movements and translate them into
cursor movements on a screen. It is designed to be easy to use, allowing users to control
their computer without a physical mouse. This system has been developed using Python and
the OpenCV library, and it applies techniques such as hand detection, feature extraction, and
classification. The system is known to work well in various lighting conditions, different back-
grounds, and with different hand sizes.In this seminar, I explore the AI Virtual Mouse as a new
way to interact with computers, which can be especially useful for people with disabilities or
those who prefer using gestures. The virtual mouse processes live images continuously, apply-
ing filters and conversions. After processing, it identifies the hand’s position based on specific
colors in the frames, allowing for gesture-based control.The purpose of this seminar is to exam-
ine a technological innovation that enables users to work without a traditional mouse, offering
benefits like saving time and money. The system’s performance has been tested for accuracy,
speed, and reliability and is compared to other virtual mouse systems.Additionally, the seminar
discusses how the system matches current colors in the frames to a set of predefined patterns,
where each pattern corresponds to a specific mouse action. When a match is detected, the sys-
tem executes the command on the user’s computer.

Keywords: AI Virtual Mouse,Hand Movement Tracking,Python,OpenCV,Real-Time Processing,Real-


Time Image Processing

iii
Contents

Certificate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . i
Acknowledgement . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . ii
Abstract . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iii
Contents . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . iv
List of Figures . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vi
List of Tables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . vii
Abbreviations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . viii

1 Introduction 1
1.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.2 Motivation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.3 Objectives . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.4 Scope . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 2

2 Literature Survey 3

3 Methodologies 5
3.1 Framework/Basic Architecture . . . . . . . . . . . . . . . . . . . . . . . . . . 5
3.1.1 Input Device (Camera/Webcam) . . . . . . . . . . . . . . . . . . . . . 5
3.1.2 Hand Detection Module . . . . . . . . . . . . . . . . . . . . . . . . . 5
3.1.3 Gesture Recognition Module . . . . . . . . . . . . . . . . . . . . . . . 6
3.1.4 Cursor Movement and Control . . . . . . . . . . . . . . . . . . . . . . 6
3.2 Different Approaches . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
3.2.1 Color Detection . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 7
3.2.2 Machine Learning Algorithms . . . . . . . . . . . . . . . . . . . . . . 7
3.3 State-of-the-art Algorithms . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
3.3.1 Convolutional Neural Networks (CNNs) . . . . . . . . . . . . . . . . . 8
3.3.2 Support Vector Machines (SVMs) . . . . . . . . . . . . . . . . . . . . 8
3.3.3 K-Nearest Neighbors (KNN) . . . . . . . . . . . . . . . . . . . . . . . 9
3.4 Discussion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
3.4.1 Comparison of Algorithms . . . . . . . . . . . . . . . . . . . . . . . . 9

iv
4 Implementation 11
4.1 Algorithm/Methodologies . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
4.1.1 Hand Detection Algorithm . . . . . . . . . . . . . . . . . . . . . . . . 11
4.1.2 Gesture Recognition Algorithm . . . . . . . . . . . . . . . . . . . . . 11
4.1.3 Feature Extraction . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
4.2 Proposed Solution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
4.2.1 Verification Plan . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
4.3 Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
4.3.1 Overview . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 14
4.3.2 System Performance . . . . . . . . . . . . . . . . . . . . . . . . . . . 14

5 Applications 15
5.1 State-of-the-art Applications . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
5.1.1 Touchless Computing Interfaces . . . . . . . . . . . . . . . . . . . . . 15
5.1.2 Gaming and Virtual Reality (VR) . . . . . . . . . . . . . . . . . . . . 15
5.1.3 Assistive Technology . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
5.1.4 Smart Home Control . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
5.1.5 Presentation Control . . . . . . . . . . . . . . . . . . . . . . . . . . . 15
5.1.6 Interactive Art Installations . . . . . . . . . . . . . . . . . . . . . . . . 16
5.1.7 Augmented Reality (AR) . . . . . . . . . . . . . . . . . . . . . . . . . 16
5.1.8 Retail and Marketing . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
5.2 Challenges . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 16
5.2.1 Accuracy of Gesture Recognition . . . . . . . . . . . . . . . . . . . . 16
5.2.2 Real-time Processing and Latency . . . . . . . . . . . . . . . . . . . . 16
5.2.3 Environmental Factors . . . . . . . . . . . . . . . . . . . . . . . . . . 16
5.2.4 User Variability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
5.2.5 Fatigue and Usability . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
5.2.6 Integration with Operating Systems . . . . . . . . . . . . . . . . . . . 17
5.2.7 Data Privacy and Security . . . . . . . . . . . . . . . . . . . . . . . . 17
5.2.8 Hardware Limitations . . . . . . . . . . . . . . . . . . . . . . . . . . 17

6 Conclusion and Future Scope 18


6.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18
6.2 Future Scope . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 18

References 19

v
List of Figures

3.1 Work Flow in AI Virtual Mouse System [2] . . . . . . . . . . . . . . . . . . . 6

vi
List of Tables

3.1 Comparison of Gesture Recognition Algorithms [6] & [7] . . . . . . . . . . . . 10

4.1 Hypothetical Performance Metrics for the AI Virtual Mouse System . . . . . . 14

vii
Abbreviations

CNN : Convolutional Neural Networks

GUI : Graphical User Interface

HCI : Human Computer Interaction

HMM : Hidden Markov Models

HSV : Hue, Saturation and Values

IDE : Integrated Development Environment

LED : Light Emitting Diodes

MHI : Motion History Image

OPEN CV : Open Source Computer Visions

RGB : Red, Green, Blue

SVM : Support Vector Machines

viii
Seminar Report AI Based Virtual Mouse

1. Introduction

1.1 Introduction
In computer terminology, a mouse is a device that detects two-dimensional movement relative
to a surface. This motion is converted into cursor movement on a display, allowing users to
interact with the Graphical User Interface (GUI) of a computer. It is hard to imagine life today
without computers, as they have become essential in almost every aspect of our daily routines.
People of all ages rely on computers for various tasks. Consequently, Human-Computer Inter-
action (HCI) has emerged as a significant field of study. The introduction of the laser mouse
in 2004 addressed the limitations of the optical mouse, improving movement accuracy even on
reflective surfaces. However, physical and technical limitations still exist. With the arrival of
touch screen technology on mobile devices, there is a growing demand for similar technology
on desktop computers. While touch screens are available for desktops, their cost can be pro-
hibitive.This seminar focuses on developing a finger tracking-based virtual mouse application
using a standard webcam. By employing object tracking techniques from Artificial Intelligence
and the OpenCV library in Python, this project aims to provide an alternative to touch screens.
The virtual mouse will utilize a webcam to track user gestures, process them, and translate
them into pointer movement, mimicking the function of a physical mouse.

1.2 Motivation
The Virtual Mouse is likely to replace traditional physical mice in the near future as people seek
to interact with technology without needing physical peripherals, like keyboards or remote con-
trols. This approach not only enhances convenience but also reduces costs. Moreover, in light
of the ongoing global health concerns due to the COVID-19 pandemic, a touchless computing
solution offers a safer alternative by minimizing the risk of germ transmission associated with
shared devices

1.3 Objectives
The objective of this seminar is to create a Virtual Mouse application that highlights several
important programming concepts. The key goals of this project are:
1. To learn how a virtual mouse uses finger tracking.

PICT,Pune 1 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

2. To identify challenges in developing virtual mouse technology.

3. To understand the differences between a virtual mouse and traditional mice.

4. To understand the user experience of using a virtual mouse compared to a regular mouse.

1.4 Scope
The scope of this seminar is to discuss the concept of a virtual mouse that operates without
physical interaction with devices or screens. This discussion is particularly relevant in today’s
context, where minimizing touch on shared surfaces is important to reduce the risk of spreading
infections during the COVID-19 pandemic. The seminar will cover the potential of a virtual
mouse to enhance user convenience while allowing accurate interaction with computer systems.
Additionally, this seminar will explore how the virtual mouse could be used to improve Human-
Computer Interaction (HCI) and its implications for future technology.[8]

PICT,Pune 2 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

2. Literature Survey
In the field of hand gesture recognition, numerous research studies have significantly advanced
the development of virtual mouse technology and other human-computer interaction systems.
Dudhapachare and Vaidya’s paper, titled Gesture-Controlled Virtual Mouse (2023) [1],
presented a novel approach to controlling a computer’s cursor using only hand gestures. Their
system leveraged computer vision techniques to recognize various hand movements, translating
them into mouse actions like clicking and scrolling. The paper demonstrated the potential for
gesture recognition to replace traditional hardware devices, focusing on creating a low-cost,
real-time system that can function efficiently in varied lighting conditions. Their work paved
the way for gesture-based input in everyday computing tasks, showing promise for users who
want a touchless interaction method.
Reddy et al., in their paper Gesture-Controlled Virtual Mouse with the Support of Voice
Assistant (2022) [2], developed an innovative system that combined both gesture recognition
and voice commands to provide a more comprehensive control interface. The integration of
a voice assistant allowed for hands-free interaction, while gestures were used for basic mouse
functionalities such as dragging, dropping, and selection. This research explored the benefits
of multimodal interaction, enhancing the user experience by enabling both auditory and visual
inputs, making it a valuable solution for people with disabilities or in hands-busy environments.
Huu and Ngoc, in their work Hand Gesture Recognition Algorithm Using SVM and HOG
Model for Control of Robotic System (2021) [3], proposed a gesture recognition algorithm
based on Support Vector Machines (SVM) and Histogram of Oriented Gradients (HOG). The
paper focused on developing an accurate and robust method for recognizing gestures, specifi-
cally for controlling robotic systems. Their approach demonstrated high accuracy in detecting
hand movements, even in complex environments with varying lighting and background condi-
tions. The study provided insights into how SVMs and HOG features can improve the reliability
of gesture recognition in both robotic applications and virtual mouse systems.
Tran et al., in their study Real-time Virtual Mouse System Using RGB-D Images and
Fingertip Detection (2021) [4], introduced a system that utilized RGB-D imaging technology
to detect and track fingertips in real-time. This method allowed for more precise interaction
with digital interfaces by identifying the exact position of the user’s fingertips, enabling fine
control of a virtual mouse. The use of depth information helped overcome challenges associated
with traditional 2D image processing, such as occlusion and background noise. This paper
highlighted how combining RGB and depth data could improve gesture recognition accuracy
in real-world applications, particularly in virtual environments and gaming.

PICT,Pune 3 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

Finally, Rajee et al. contributed to the field with their research on Artificial Intelligence Vir-
tual Mouse (2023) [5]. This paper explored how AI algorithms can enhance the performance of
virtual mouse systems by enabling the system to learn from user interactions. Their approach
used machine learning models to improve gesture detection accuracy over time, adapting to
individual user behaviors and preferences. By incorporating AI, the system could predict user
actions and offer more seamless control, making it highly adaptable and personalized. This
research highlighted the importance of AI in making virtual mouse systems more efficient,
reliable, and user-friendly.
This collection of studies underscores the diverse methodologies and challenges in gesture
recognition systems, such as accuracy limitations and complexity of gestures. These findings
provide a foundational understanding as we explore the potential of virtual mouse technology
in this seminar.

PICT,Pune 4 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

3. Methodologies
The AI virtual mouse system is based on the frames that have been captured by the webcam in
a laptop or PC. By using the Python computer vision library OpenCV, the video capture object
is created and the web camera will start capturing video. The web camera captures and passes
the frames to the AI virtual system. Capturing the Video and Processing. The AI virtual mouse
system uses the webcam where each frame is captured till the termination of the program. The
video frames are processed from BGR to RGB color space to find the hands in the video frame
by frame.

3.1 Framework/Basic Architecture


The AI Virtual Mouse is a system that utilizes computer vision techniques to emulate the
functions of a traditional mouse through hand gestures. The basic architecture can be outlined
as follows:

3.1.1 Input Device (Camera/Webcam)


The primary input is a live video feed from a camera or webcam. This device captures real-time
images of the user’s hand.

• Continuous frames of the hand are captured.

• The video feed is prepared for further processing.

3.1.2 Hand Detection Module


This module identifies and tracks the user’s hand from the video feed.

• Techniques include:

– Color detection (e.g., skin color).

– Machine learning models, such as Convolutional Neural Networks (CNNs).

• Filters out background elements to focus on the hand.

PICT,Pune 5 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

3.1.3 Gesture Recognition Module


Once the hand is detected, this module recognizes specific gestures that correspond to mouse
functions.

• Uses machine learning techniques, including:

– Hidden Markov Models (HMMs).

– Support Vector Machines (SVMs).

• Identifies actions like clicking and scrolling.

3.1.4 Cursor Movement and Control


The recognized gestures are translated into mouse commands.

• Finger movements control the cursor.

• Gestures are interpreted as left-click or right-click actions.

This simplified architecture highlights the essential components of the AI Virtual Mouse,
focusing on its functionality and interactions with the user and the operating system.

Figure 3.1: Work Flow in AI Virtual Mouse System [2]

PICT,Pune 6 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

3.2 Different Approaches


In the development of the Artificial Intelligence (AI) Virtual Mouse, several approaches have
been explored to enhance the accuracy and efficiency of hand gesture recognition and control.
The following are some of the prominent approaches used:

3.2.1 Color Detection


One of the simplest and most widely used methods involves color detection to identify hand
regions based on skin color. This method includes:

• HSV Color Space: Transforming the RGB color model to HSV to better segment skin
tones.[4]

• Thresholding: Setting thresholds for hue, saturation, and value to filter out background
noise.

• Advantages:

– Simple to implement.
– Fast processing times.

• Limitations:

– Sensitivity to lighting changes.


– Ineffective with non-standard skin tones or backgrounds.

3.2.2 Machine Learning Algorithms


Machine learning algorithms have gained popularity for their ability to learn from data and
improve over time. Commonly used algorithms include:

• Support Vector Machines (SVM): Effective for classification tasks, particularly in high-
dimensional spaces.[3]

• Advantages:

– Higher accuracy compared to traditional methods.


– Ability to generalize well to new, unseen data.

• Limitations:

– Requires a significant amount of training data.


– Computationally intensive, needing powerful hardware for real-time applications.

PICT,Pune 7 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

3.3 State-of-the-art Algorithms


In the field of Artificial Intelligence (AI) Virtual Mouse systems, various cutting-edge algo-
rithms are utilized for recognizing hand gestures and facilitating interaction. Below, we present
key algorithms along with an analysis of their complexity.

3.3.1 Convolutional Neural Networks (CNNs)


Convolutional Neural Networks have become the backbone of image recognition tasks, includ-
ing hand gesture detection.[7]

• Description: CNNs use convolutional layers to automatically extract features from im-
ages, followed by fully connected layers to classify the gestures.

• Complexity Analysis:

– Time Complexity: The time complexity can be approximated as O(N · W 2 · F 2 · C),


where N is the number of training samples, W is the input width and height, F is
the filter size, and C is the number of filters.

– Space Complexity: The space complexity primarily depends on the number of pa-
rameters, typically O(K · D), where K is the number of filters and D is the dimen-
sions of the input.

3.3.2 Support Vector Machines (SVMs)


SVMs are supervised learning models commonly used for classification tasks, including gesture
recognition.

• Description: SVMs find a hyperplane that best separates different classes in the feature
space, maximizing the margin between classes.[3]

• Complexity Analysis:

– Time Complexity: The time complexity is generally O(N 2 · D) for training, where
N is the number of training samples and D is the number of features. In practice, it
may be O(N 3 ) due to the use of quadratic programming.

– Space Complexity: The space complexity is O(N ), as it requires storing the support
vectors.

PICT,Pune 8 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

3.3.3 K-Nearest Neighbors (KNN)


KNN is a simple, non-parametric algorithm used for classification based on the closest training
examples in the feature space.[6]

• Description: KNN classifies a sample based on the majority class among its k nearest
neighbors.

• Complexity Analysis:

– Time Complexity: The time complexity for prediction is O(N · D), where N is the
number of training samples and D is the dimensionality of the data. The training
time is negligible as there is no explicit training phase.

– Space Complexity: The space complexity is O(N ) for storing the training dataset.

In summary, the choice of algorithm significantly affects the performance and efficiency of
the AI virtual mouse system. Understanding these algorithms and their complexities is crucial
for selecting the appropriate method for hand gesture recognition and ensuring optimal system
performance.

3.4 Discussion
This section examines the different approaches and algorithms used in the development of AI
virtual mouse systems. By assessing the strengths and limitations of these methodologies, we
can gain insights into their relevance and effectiveness in practical applications.

3.4.1 Comparison of Algorithms


Different algorithms exhibit unique advantages and challenges, making them suitable for spe-
cific applications. Table 3.1 summarizes key features, advantages, and limitations of the dis-
cussed algorithms.

PICT,Pune 9 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

Algorithm Advantages Limitations Best Use Case


Convolutional High accuracy in image Requires large datasets Gesture recognition
Neural Networks recognition and computational from images
(CNNs) resources
Support Vector Effective in Sensitive to noise and Classifying distinct
Machines high-dimensional outliers gesture types
(SVMs) spaces
K-Nearest Simple to implement Computationally Fast, low-complexity
Neighbors expensive at prediction environments
(KNN) time

Table 3.1: Comparison of Gesture Recognition Algorithms [6] & [7]

In conclusion, the choice of algorithm for an AI virtual mouse system should be dictated by
the specific requirements of the application, including factors such as data availability, required
accuracy, and system constraints. Balancing these elements will lead to more effective and
efficient gesture recognition systems that enhance user interaction with technology.

PICT,Pune 10 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

4. Implementation
4.1 Algorithm/Methodologies
In this section, the fundamental methodologies and algorithms are outlined which can be used
for development of the AI Virtual Mouse system. The aim is to establish an efficient and
user-friendly interface that could translate hand gestures into mouse commands using basic
computer vision techniques.

4.1.1 Hand Detection Algorithm

[4]The first step in our methodology is hand detection. We will utilize basic color detection and
contour detection methods:

• Color Detection: We will implement simple color thresholding to detect skin-toned re-
gions in the video feed. This method involves converting the image to the HSV color
space and setting a range for skin colors. This approach is computationally light and can
operate in real-time.

• Contour Detection: Once we have identified potential skin regions, we will use contour
detection to outline the shape of the hand. This helps in distinguishing the hand from the
background and allows for better tracking of movements.

4.1.2 Gesture Recognition Algorithm

[5]For recognizing gestures, we will use basic techniques that do not require complex models:

• Template Matching: We will employ template matching to recognize predefined hand


gestures. Simple geometric shapes (e.g., rectangles for clicks, lines for swipes) can be
defined as templates, and the system will compare the detected contours to these tem-
plates to determine the gesture.

• Simple Rule-Based Approach: Gestures will be recognized based on simple rules, such
as the angle of fingers or the number of fingers extended. For example, a single finger
pointing could be recognized as a ”move” command, while a fist could indicate a ”click.”

PICT,Pune 11 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

4.1.3 Feature Extraction

[5]Basic feature extraction will be implemented to capture essential characteristics of hand


gestures:

• Fingertip Location: The coordinates of the fingertips will be extracted from the detected
contours, providing information about cursor movement.

• Hand Position: The overall position of the hand will be tracked to help determine the
screen area being controlled.

In summary, this approach emphasizes the use of simple algorithms and methodologies to
create an effective AI Virtual Mouse system. By leveraging straightforward techniques, we can
ensure a user-friendly experience while maintaining responsiveness and accuracy in gesture
recognition.

4.2 Proposed Solution


The development of an AI Virtual Mouse system can build on existing hand gesture recognition
techniques. However, many existing solutions face challenges that impact their effectiveness.
Some key issues include:

• Limited Accuracy: Many hand detection systems struggle with accuracy, especially
under varying lighting conditions, complex backgrounds, or differences in skin tone.
This often leads to incorrect or missed gesture recognition.

• High Computational Requirements: Advanced systems, especially those involving


deep learning, often require significant computational resources, making them less prac-
tical for real-time applications on standard consumer devices.

• Lack of Intuitive Gestures: Some systems rely on complex gestures that are difficult
for users to remember or perform naturally, reducing ease of use and accessibility.

• Latency Issues: There can be noticeable delays between gesture input and the system’s
response, which can make the interaction feel unresponsive and frustrating for users.

In response to these challenges, this report proposes a simpler, more accessible AI Virtual
Mouse system using computer vision techniques. The focus is on employing lightweight, real-
time algorithms that prioritize ease of use and accessibility.

PICT,Pune 12 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

The system could utilize a webcam to capture video input, with basic skin color detection
used to identify hand regions. By converting the video frames from RGB to HSV color space,
skin-colored pixels could be isolated through color thresholding, creating a binary mask that
highlights the hand. Contour detection would then help track the position and shape of the
hand, reducing the impact of background complexity.
For gesture recognition, instead of relying on complex models, simple template matching
could be used to recognize basic gestures like pointing and clicking. This approach avoids the
high computational cost of more advanced techniques while still allowing for reliable interac-
tion.
To map these gestures to actions, critical features like fingertip positions would be extracted.
The index finger could control the cursor, while gestures like a closed fist or two fingers could
trigger mouse clicks. This solution also includes basic scrolling actions, driven by vertical hand
movements.
While the system is theoretical, future improvements could involve machine learning for
more adaptive gesture recognition and the addition of more intuitive gestures based on user
feedback. User testing would help refine the solution further, ensuring it meets user needs in
terms of ease of use, responsiveness, and accuracy.

Software Requirement

The following software is required for the development of the Virtual Mouse application:

• Python Language: The Virtual Mouse application will be coded using Python in Mi-
crosoft Visual Studio, an integrated development environment (IDE). Python provides
numerous operators, including comparisons, logical operations, bit manipulation, and
basic arithmetic.

• OpenCV Library: OpenCV is a collection of programming functions for real-time com-


puter vision. It can read image pixel values and recognize eye movements and blinks in
real-time.

Software Specifications:

• OS: Windows 10 Ultimate 64-bit

• Language: Python

• Tools Used: OpenCV and MediaPipe

PICT,Pune 13 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

4.2.1 Verification Plan


The Virtual Mouse system must accurately and consistently identify colors provided by users
without affecting the performance of other activities. Recognition outcomes may vary depend-
ing on environmental conditions like lighting or background colors. Below are factors that may
cause errors in the recognition phase:

• a) Real-time images are captured in either dark or overly bright environments.

• b) Real-time images have conflicting background colors.

• c) Users can interact with the software from varying distances (near or far).

• d) Real-time images are rotated either clockwise or counterclockwise.

4.3 Result

4.3.1 Overview
The concept of an AI Virtual Mouse system is designed to control a computer cursor using hand
gestures, offering an alternative to traditional input devices. This can be applied in various real-
time applications such as controlling a computer cursor or interacting with smart televisions.
The simplicity of the system reduces the need for external hardware, relying solely on a webcam
to track finger movements and perform the necessary operations on the screen.

4.3.2 System Performance


Although this is a theoretical solution and has not been implemented, hypothetical performance
evaluation could focus on the accuracy of gesture recognition and overall user satisfaction.
These metrics would be key indicators of the system’s success.[5],[6] & [7]

Performance Metric Hypothetical Value


Accuracy of Gesture Recognition 85%
Latency (Response Time) 100ms
User Satisfaction (Feedback) Positive

Table 4.1: Hypothetical Performance Metrics for the AI Virtual Mouse System

These values are speculative and would require actual implementation and testing for verifica-
tion.

PICT,Pune 14 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

5. Applications
5.1 State-of-the-art Applications
Artificial Intelligence (AI) and computer vision have enabled numerous applications for vir-
tual mouse systems. These applications allow users to interact with technology through hand
gestures. Here are some key applications of virtual mouse systems:

5.1.1 Touchless Computing Interfaces


Virtual mouse systems enable users to control computers without physical contact, which is
especially useful in sterile environments like hospitals. For example, surgeons can manage
digital displays during procedures without contaminating their hands.

5.1.2 Gaming and Virtual Reality (VR)


In gaming, virtual mouse technology provides an immersive experience by allowing players
to interact with game elements using gestures. In VR, users can manipulate virtual objects
naturally, enhancing their engagement.

5.1.3 Assistive Technology


These systems significantly benefit individuals with physical disabilities by enabling control of
computers through hand gestures or facial movements, thereby promoting independence and
accessibility.

5.1.4 Smart Home Control


Virtual mouse systems allow users to manage smart home devices, like lighting and tempera-
ture, with gestures, offering convenience and a seamless touchless experience.

5.1.5 Presentation Control


In business and educational settings, virtual mouse technology enhances presentations, allow-
ing presenters to navigate slides and emphasize content with simple gestures, making the expe-
rience more interactive.

PICT,Pune 15 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

5.1.6 Interactive Art Installations


Artists are using virtual mouse systems in interactive art installations, enabling viewers to ma-
nipulate digital art through their movements for a unique, personalized experience.

5.1.7 Augmented Reality (AR)


In AR applications, users can interact with digital objects overlaid on the real world using hand
gestures, enhancing their interaction with both physical and digital environments.

5.1.8 Retail and Marketing


Virtual mouse systems are being implemented in retail settings to provide a touchless and in-
teractive shopping experience, allowing customers to browse and select products with gestures.
In summary, virtual mouse systems have a wide range of applications across various indus-
tries, transforming user interactions with technology into more intuitive and efficient experi-
ences.

5.2 Challenges
Despite advancements in AI-driven virtual mouse systems, several challenges must be ad-
dressed for effective functioning:

5.2.1 Accuracy of Gesture Recognition


Achieving high accuracy in recognizing gestures is difficult due to factors like poor lighting
and complex backgrounds, which can lead to inaccuracies in detection.

5.2.2 Real-time Processing and Latency


To provide a smooth user experience, virtual mouse systems must operate in real-time. How-
ever, processing gestures can introduce latency, especially with computationally intensive al-
gorithms.

5.2.3 Environmental Factors


Lighting and background clutter can significantly impact performance. Ensuring consistent
accuracy in various environments is a key challenge.

PICT,Pune 16 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

5.2.4 User Variability


Different users have unique hand sizes, shapes, and skin tones, affecting gesture recognition.
A flexible system that adapts to different users is essential.

5.2.5 Fatigue and Usability


Prolonged use of gesture controls can lead to user fatigue. Improving the ergonomics of gesture
interaction is important for practical use.

5.2.6 Integration with Operating Systems


Integrating virtual mouse systems with various operating systems and applications can be chal-
lenging, requiring compatibility and seamless interaction across platforms.

5.2.7 Data Privacy and Security


Using cameras for gesture detection raises privacy concerns. Ensuring secure handling of data
collected by these systems is crucial.

5.2.8 Hardware Limitations


Performance can be limited by the hardware, such as webcam resolution. Ensuring efficiency
on various hardware configurations is vital.
In conclusion, while virtual mouse systems offer innovative ways to interact with technol-
ogy, addressing these challenges is critical for their success and widespread adoption.

PICT,Pune 17 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

6. Conclusion and Future Scope


6.1 Conclusion
This seminar explored the development of a color-recognition-based virtual mouse, which of-
fers an innovative approach to Human-Computer Interaction (HCI) by enabling users to control
the cursor through color movements. This technology aims to replace traditional input devices
without sacrificing accuracy or efficiency. By utilizing a program that recognizes colors and
translates them into mouse actions, users can enjoy a touchless interface that enhances conve-
nience.
Key objectives included optimizing cursor control by averaging color values to prevent
unintentional movements. The inclusion of a calibration phase allows users to select their
preferred color combinations, increasing adaptability and user satisfaction. Overall, the virtual
mouse system represents a significant step toward creating intuitive and efficient interaction
methods.

6.2 Future Scope


There is significant potential to enhance the virtual mouse system by improving gesture recog-
nition algorithms to achieve higher accuracy and responsiveness. Expanding the detection
range will provide users with greater flexibility in various environments, making the system
more adaptable to different settings. Additionally, ensuring compatibility with a wide range of
hardware setups will optimize performance across devices, allowing more users to benefit from
this technology.
Furthermore, refining the user interface to make it more intuitive will facilitate easier setup
and usage, encouraging broader adoption. Incorporating feedback mechanisms can help tailor
the system to individual user preferences, making interactions more seamless. Lastly, inte-
grating additional functionalities, such as voice commands or touchless controls, could further
enhance the user experience, making the virtual mouse a versatile tool for a variety of applica-
tions.

PICT,Pune 18 Dept. of Information Technology


Seminar Report AI Based Virtual Mouse

Bibliography

[1] Rajat Dudhapachare and Nihal Vaidya, ”Gesture-Controlled Virtual Mouse”, vol. 11, Issue
I, 2023.

[2] Cherukupally Karunakar Reddy, Suraj Janjirala, Kevulothu Bhanu Prakash, ”Gesture Con-
trolled Virtual Mouse with the Support of Voice Assistant”, vol. 10, Issue VI, 2022.

[3] Phat Ngyyen Huu, Tan Phung Ngoc, ”Hand Gesture Recognition Algorithm Using SVM
and HOG Model for Control of Robotic System”, vol. 2021, Issue I, 2021.

[4] Dinh-Son Tran, Ngoc-Huynh Ho, Hyung-Jeong Yang, Soo-Hyung Kim, Guee Sang Lee,
”Real-time Virtual Mouse System Using RGB-D Images and Fingertip Detection”, vol. 80,
pp. 10473-10490, 2021.

[5] Alimul Rajee, Mohammad Likhan, Nurul Ahad Farhan, Israth Jahan, ”Artificial Intelli-
gence Virtual Mouse”, DOI: 10.13140/RG.2.2.35877.06884/1, 2023.

[6] Arun Karthik. V, Barath. S, Guru Harrish. N, Prageesh. S, ”Virtual Mouse”, Vol 4, no 10,
pp 1215-1219, 2023.

[7] Aabha Waichal , Mauli Gandhi, Srushti Bhagwat,Amruta Bhanji,Shalaka Deore,Shubhangi


Ingale, ”Virtual Mouse”, Volume 184– No.20, pp:0975 – 8887, 2023.

PICT,Pune 19 Dept. of Information Technology

You might also like