SlideShare a Scribd company logo
Computer Science and Information Technologies
Vol. 1, No. 3, November 2020, pp. 116~120
ISSN: 2722-3221, DOI: 10.11591/csit.v1i3.p116-120  116
Journal homepage: http://iaesprime.com/index.php/csit
Hand gesture recognition using machine learning algorithms
Abhishek B1
, Kanya Krishi2
, Meghana M3
, Mohammed Daaniyaal4
, Anupama H S5
1,2,3,4
B. E, Computer Science and Engineering, BMS Institute of Technology, Bangalore, India
5
BMS Institute of Technology, Bangalore, India
Article Info ABSTRACT
Article history:
Received Apr 24, 2020
Revised Jun 14, 2020
Accepted Jun 29, 2020
Gesture recognition is an emerging topic in today’s technologies. The main
focus of this is to recognize the human gestures using mathematical algorithms
for human computer interaction. Only a few modes of human-computer
interaction exist, they are: through keyboard, mouse, touch screens etc. Each
of these devices has their own limitations when it comes to adapting more
versatile hardware in computers. Gesture recognition is one of the essential
techniques to build user-friendly interfaces. Usually, gestures can be
originated from any bodily motion or state, but commonly originate from the
face or hand. Gesture recognition enables users to interact with the devices
without physically touching them. This paper describes how hand gestures are
trained to perform certain actions like switching pages, scrolling up or down
in a page.
Keywords:
Gesture recognition
Human–computer interaction
User-friendly interface
This is an open access article under the CC BY-SA license.
Corresponding Author:
Abhishek B,
B.E, Computer Science and Engineering,
BMS Institute of Technology, Bangalore, India.
Email: anupamahs@bmsit.in
1. INTRODUCTION
Gesture recognition is a technique which is used to understand and analyze the human body language
and interact with the user accordingly. This in turn helps in building a bridge between the machine and the user
to communicate with each other. Gesture recognition is useful in processing the information which cannot be
conveyed through speech or text. Gestures are the simplest means of communicating something that is
meaningful. This paper involves implementation of the system that aims to design a vision-based hand gesture
recognition system with a high correct detection rate along with a high-performance criterion, which can work
in a real time human–computer interaction (HCI) system without having any of the limitations (gloves, uniform
background etc.) on the user environment. The system can be defined using a flowchart that contains three
main steps, they are: learning, detection, recognition as shown in Figure 1.
Learning involves two aspects such as:
− Training dataset: This is the dataset that consists of different types of hand gestures that are used to train
the system based on which the system performs the actions.
− Feature extraction: It involves determining the centroid that divides the image into two halves at its
geometric centre.
Detection involves three aspects:
− Capture scene: Captures the images through a web camera, which is used as an input to the system.
− Preprocessing: Images that are captured through the webcam are compared with the dataset to recognize
the valid hand movements that are needed to perform the required actions.
− Hand detection: The requirements for hand detection involve the input image from the webcam.
The image should be fetched with a speed of 20 frames per second. Distance should also be maintained
Comput. Sci. Inf. Technol. 
Hand gesture recognition using machine learning algorithms… (Abhishek B)
117
between the hand and the camera. Approximate distance that should be between hand the camera is around
30 to 100 cm. The video input is stored frame by frame into a matrix after preprocessing.
Recognition consists of:
− Gesture recognition: The number of fingers present in the hand gesture is determined by making use of
defect points present in the gesture. The resultant gesture obtained is fed through a 3-dimensional
convolutional neural network consecutively to recognize the current gesture.
− Performing action: The recognized gesture is used as an input to perform the actions required by the user.
Figure 1. Flowchart of HCI
2. LITERATURE SURVEY
The implementation is divided into four main steps: 1) image enhancement and segmentation, 2)
orientation detection, 3) feature extraction, and 4) classification [1]. This work was focused on above four
categories but main limitation was change of color was happening very rapidly by the change in the different
lighting condition, which may cause error or even failures. For example, due to insufficient light condition, the
existence of hand area is not detected but the non-skin regions are mistaken for the hand area because of same
color [2]. Involves three main steps for hand gesture recognition system: 1) segmentation, 2) feature
representation 3) recognition techniques. The system is based on hand gesture recognition by modeling of the
hand in spatial domain. The system uses various 2D and 3D geometric and non-geometric models for modeling.
It has used fuzzy c-means clustering algorithm which resulted in an accuracy of 85.83%. The main drawback
of the system is it does not consider gesture recognition of temporal space, i.e., motion of gestures and it is
unable to classify images with complex background i.e., where there are other objects in the scene with the
hand objects [3]. This survey focuses on the hand gesture recognition using different steps like data acquisition,
pre-processing, segmentation and so on. Suitable input device should be selected for the data acquisition. There
are a number of input devices for data acquisition. Some of them are data gloves, marker, and hand images
(from webcam/Kinect 3D Sensor). But the limitation with this work was change in the illumination, rotation
and orientation, scaling problem and special hardware which is pretty costlier [4]. The system implementation
is divided into three phases: 1) Hand gesture recognition using kinetic camera, 2) algorithms for hand detection
recognition, 3) hand gesture recognition. The limitation here is that the edge detection and segmentation
algorithms used here are not very efficient when compared to neural networks. The dataset being considered
here is very small and can be used to detect very few sign gestures.
The system architecture consists of: 1) image acquisition, 2) segmentation of hand region, 3) distance
transforms method for gesture recognition [5]. The limitations of this system involve 1) the numbers of gestures
 ISSN: 2722-3221
Comput. Sci. Inf. Technol., Vol. 1, No. 3, November 2020: 117 – 120
118
that are recognized are less and 2) the gestures recognized were not used to control any applications [6]. In this
implementation there are three main algorithms that are used: 1) Viola–Jones algorithm. 2) convex hull
algorithm, 3) the AdaBoost based learning algorithm. The work was accomplished by training a set of feature
set which is local contour sequence. The limitations of this system are that it requires two sets of images for
classification. One is the positive set that contains the required images, the other is the negative set that contains
contradicting images [7]. The system implementation consists of three components: 1) hand detection 2)
gesture recognition, and 3) HCI. It has implemented the following methodology: 1) the input image is
preprocessed and the hand detector tries to filter out the hand from the input image, 2) a CNN classifier is
employed to recognize gestures from the processed image, while a Kalman Filter is used to estimate the position
of the mouse cursor, and 3) the recognition and estimation results are submitted to a control centre which
decides the action to be taken. One of the limitations of this system is that it recognizes only the static images
[8]. This implementation focuses on detection of hand gestures using java and neural networks. It is divided
into two phases: 1) Detection module using java where in the hand is detected using background subtraction
and conversion of video feed into HSB video feed thus detecting skin pixels; 2) The second module is the
prediction module; a convolutional neural network is used. The input feed image is gained from Java. The input
image is fed into the neural network and is analyzed with respect to the dataset images. One of the limitations
of this system is that it requires socket programming in order to connect java and python modules.
3. IMPLEMENTATION
A hand gesture recognition system was developed to capture the hand gestures being performed by
the user and to control a computer system based on the incoming information. Many of the existing systems in
literature have implemented gesture recognition using only spatial modelling, i.e., recognition of a single
gesture and not temporal modelling i.e., recognition of motion of gestures. Also, the existing systems have not
been implemented in real time, they use a pre captured image as an input for gesture recognition. To overcome
these existing problems a new architecture has been developed which aims to design a vision-based hand
gesture recognition system with a high correct detection rate along with a high-performance criterion, which
can work in a real time HCI system without having any of the mentioned strict limitations (gloves, uniform
background, etc.) on the user environment. The design is composed of a HCI system which uses hand gestures
as input for communication as show in Figure 2.
Figure 2. Design of the proposed HCI system
Comput. Sci. Inf. Technol. 
Hand gesture recognition using machine learning algorithms… (Abhishek B)
119
Input to the system is from the web camera or a prerecorded video sequence. Later it detects the skin
color by using an adaptive algorithm in the beginning of the frames. For the current user skin color has to be
fixed based on the lighting and camera parameter and condition. Once it has been fixed, hand is localized with
a histogram clustering method. Then a machine learning algorithm has been used to detect the hand gestures
in consecutive frames to distinguish the current gesture. These gestures are used as an input for a computer
application as shown in Figure 3. The system is divided into 3 subsystems:
3.1. Hand and motion detection
The Web-camera captures the hand movement and provides it as input to OpenCV and TensorFlow
Object detector. Edge detection and skin detection are performed to obtain the boundary of the hand. This is
then sent to the 3D CNN.
3.2. Dataset
Dataset is used for training the 3D CNN. Two types of datasets are being used–one for the hand
detection and the other for the motion or gesture detection. Hand detection uses EGO dataset, Motion or
Gesture Recognition uses Jester dataset.
3.3. 3D CNN
CNN’s are a class of deep learning neural networks used for analyzing videos and images. It consists
of several layers–input layer, hidden layers, and output layer. It performs back propagation for better accuracy
and efficiency. It performs training and verification of the recognized gestures and HCIs take place–turning of
the pages, zooming in, and zooming out. The interactions with the computer take place with the help of
PyAutoGUI or System Calls.
Figure 3. System recognized hand gestures
4. CONCLUSION
The importance of gesture recognition lies in building efficient human-machine interaction. This
paper describes how the implementation of the system is done based upon the images captured. Hand detection
is done using OpenCV and TensorFlow object detector. And further it is enhanced for interpretation of gestures
by the computer to perform actions like switching the pages, scrolling up or down the page.
 ISSN: 2722-3221
Comput. Sci. Inf. Technol., Vol. 1, No. 3, November 2020: 117 – 120
120
ACKNOWLEDGEMENT
This work is done, supervised, and supported by the students and faculty members of the Department
of Computer Science and Engineering, BMS Institute of Technology, Bangalore.
REFERENCES
[1] M. Panwar and P. S. Mehra, “Hand gesture recognition for human computer interaction,” 2011 International
Conference on Image Information Processing, Shimla, pp. 1-7, 2011.
[2] R. Zaman Khan and N. A. Ibraheem, “Comparative Study of Hand Gesture Recognition System,” International
Conference of Advanced Computer Science & Information Technology, 2012.
[3] A. R. Sarkar, G. Sanyal, and S. Majumder, “Hand Gesture Recognition Systems: A Survey,” International Journal
of Computer Applications, vol. 71, no.15, pp. 25-37, May 2013.
[4] A. E. Manjunath, B. P. V. Kumar, H. Rajesh, “Comparative Study of Hand Gesture Recognition Algorithms,”
International Journal of Research in Computer and Communication Technology, vol 3, no. 4, April 2014.
[5] D. R. Jadhav and L. M. R. J. Lobo, “Navigation of PowerPoint Using Hand Gestures,” International Journal of
Science and Research (IJSR), vol. 4, no. 1, pp. 833-837, 2015.
[6] R. M. Gurav and P. K. Kadbe, "Real time finger tracking and contour detection for gesture recognition using
OpenCV," 2015 International Conference on Industrial Instrumentation and Control (ICIC), pp. 974-977, 2015.
[7] P. Xu, “A Real-time Hand Gesture Recognition and Human-Computer Interaction System,” 2017, [Online].
Available: https://arxiv.org/abs/1704.07296
[8] P. Suganya, R. Sathya, and K. Vijayalakshmi, “Detection and Recognition of Gestures to Control the System
Applications by Neural Networks,” International Journal of Pure and Applied Mathematics, vol. 118, no. 10, pp.
399-405, January 2018.

More Related Content

Similar to Hand gesture recognition using machine learning algorithms (20)

PDF
A Survey on Virtual Whiteboard-A Gesture Controlled Pen-free Tool
vivatechijri
 
PDF
Controlling Computer using Hand Gestures
IRJET Journal
 
PDF
Real Time Vision Hand Gesture Recognition Based Media Control via LAN & Wirel...
IJMER
 
PDF
COMPARATIVE STUDY OF HAND GESTURE RECOGNITION SYSTEM
cscpconf
 
DOC
Test
theextraaedge
 
PDF
40120140503005 2
IAEME Publication
 
PDF
VIRTUAL PAINT APPLICATION USING HAND GESTURES
IRJET Journal
 
PDF
Hand Gesture Recognition System for Human-Computer Interaction with Web-Cam
ijsrd.com
 
PDF
Hand Segmentation Techniques to Hand Gesture Recognition for Natural Human Co...
Waqas Tariq
 
PDF
A Survey Paper on Controlling Computer using Hand Gestures
IRJET Journal
 
PDF
IRJET - Paint using Hand Gesture
IRJET Journal
 
PDF
Smart hand gestures recognition using K-NN based algorithm for video annotati...
nooriasukmaningtyas
 
PPTX
qwerasdfzxcv
Ram Sharma
 
PDF
Human-machine interactions based on hand gesture recognition using deep learn...
IJECEIAES
 
PDF
Real-Time System of Hand Detection And Gesture Recognition In Cyber Presence ...
IJERA Editor
 
PDF
Ay4103315317
IJERA Editor
 
PDF
HAND GESTURE RECOGNITION FOR HCI (HUMANCOMPUTER INTERACTION) USING ARTIFICIAL...
International Journal of Technical Research & Application
 
PDF
G0342039042
ijceronline
 
PPTX
hand gestuer12233322233333333666666.pptx
ArasuP1
 
PDF
Smart Presentation Control by Hand Gestures Using Computer Vision and Google’...
IRJET Journal
 
A Survey on Virtual Whiteboard-A Gesture Controlled Pen-free Tool
vivatechijri
 
Controlling Computer using Hand Gestures
IRJET Journal
 
Real Time Vision Hand Gesture Recognition Based Media Control via LAN & Wirel...
IJMER
 
COMPARATIVE STUDY OF HAND GESTURE RECOGNITION SYSTEM
cscpconf
 
40120140503005 2
IAEME Publication
 
VIRTUAL PAINT APPLICATION USING HAND GESTURES
IRJET Journal
 
Hand Gesture Recognition System for Human-Computer Interaction with Web-Cam
ijsrd.com
 
Hand Segmentation Techniques to Hand Gesture Recognition for Natural Human Co...
Waqas Tariq
 
A Survey Paper on Controlling Computer using Hand Gestures
IRJET Journal
 
IRJET - Paint using Hand Gesture
IRJET Journal
 
Smart hand gestures recognition using K-NN based algorithm for video annotati...
nooriasukmaningtyas
 
qwerasdfzxcv
Ram Sharma
 
Human-machine interactions based on hand gesture recognition using deep learn...
IJECEIAES
 
Real-Time System of Hand Detection And Gesture Recognition In Cyber Presence ...
IJERA Editor
 
Ay4103315317
IJERA Editor
 
HAND GESTURE RECOGNITION FOR HCI (HUMANCOMPUTER INTERACTION) USING ARTIFICIAL...
International Journal of Technical Research & Application
 
G0342039042
ijceronline
 
hand gestuer12233322233333333666666.pptx
ArasuP1
 
Smart Presentation Control by Hand Gestures Using Computer Vision and Google’...
IRJET Journal
 

More from CSITiaesprime (20)

PDF
Vector space model, term frequency-inverse document frequency with linear sea...
CSITiaesprime
 
PDF
Electro-capacitive cancer therapy using wearable electric field detector: a r...
CSITiaesprime
 
PDF
Technology adoption model for smart urban farming-a proposed conceptual model
CSITiaesprime
 
PDF
Optimizing development and operations from the project success perspective us...
CSITiaesprime
 
PDF
Unraveling Indonesian heritage through pattern recognition using YOLOv5
CSITiaesprime
 
PDF
Capabilities of cellebrite universal forensics extraction device in mobile de...
CSITiaesprime
 
PDF
Company clustering based on financial report data using k-means
CSITiaesprime
 
PDF
Securing DNS over HTTPS traffic: a real-time analysis tool
CSITiaesprime
 
PDF
Adversarial attacks in signature verification: a deep learning approach
CSITiaesprime
 
PDF
Optimizing classification models for medical image diagnosis: a comparative a...
CSITiaesprime
 
PDF
Acoustic echo cancellation system based on Laguerre method and neural network
CSITiaesprime
 
PDF
Clustering man in the middle attack on chain and graph-based blockchain in in...
CSITiaesprime
 
PDF
Smart irrigation system using node microcontroller unit ESP8266 and Ubidots c...
CSITiaesprime
 
PDF
Development of learning videos for natural science subjects in junior high sc...
CSITiaesprime
 
PDF
Clustering of uninhabitable houses using the optimized apriori algorithm
CSITiaesprime
 
PDF
Improving support vector machine and backpropagation performance for diabetes...
CSITiaesprime
 
PDF
Video shot boundary detection based on frames objects comparison and scale-in...
CSITiaesprime
 
PDF
Machine learning-based anomaly detection for smart home networks under advers...
CSITiaesprime
 
PDF
Transfer learning: classifying balanced and imbalanced fungus images using in...
CSITiaesprime
 
PDF
Implementation of automation configuration of enterprise networks as software...
CSITiaesprime
 
Vector space model, term frequency-inverse document frequency with linear sea...
CSITiaesprime
 
Electro-capacitive cancer therapy using wearable electric field detector: a r...
CSITiaesprime
 
Technology adoption model for smart urban farming-a proposed conceptual model
CSITiaesprime
 
Optimizing development and operations from the project success perspective us...
CSITiaesprime
 
Unraveling Indonesian heritage through pattern recognition using YOLOv5
CSITiaesprime
 
Capabilities of cellebrite universal forensics extraction device in mobile de...
CSITiaesprime
 
Company clustering based on financial report data using k-means
CSITiaesprime
 
Securing DNS over HTTPS traffic: a real-time analysis tool
CSITiaesprime
 
Adversarial attacks in signature verification: a deep learning approach
CSITiaesprime
 
Optimizing classification models for medical image diagnosis: a comparative a...
CSITiaesprime
 
Acoustic echo cancellation system based on Laguerre method and neural network
CSITiaesprime
 
Clustering man in the middle attack on chain and graph-based blockchain in in...
CSITiaesprime
 
Smart irrigation system using node microcontroller unit ESP8266 and Ubidots c...
CSITiaesprime
 
Development of learning videos for natural science subjects in junior high sc...
CSITiaesprime
 
Clustering of uninhabitable houses using the optimized apriori algorithm
CSITiaesprime
 
Improving support vector machine and backpropagation performance for diabetes...
CSITiaesprime
 
Video shot boundary detection based on frames objects comparison and scale-in...
CSITiaesprime
 
Machine learning-based anomaly detection for smart home networks under advers...
CSITiaesprime
 
Transfer learning: classifying balanced and imbalanced fungus images using in...
CSITiaesprime
 
Implementation of automation configuration of enterprise networks as software...
CSITiaesprime
 
Ad

Recently uploaded (20)

PDF
UiPath DevConnect 2025: Agentic Automation Community User Group Meeting
DianaGray10
 
PDF
Staying Human in a Machine- Accelerated World
Catalin Jora
 
PDF
Transcript: Book industry state of the nation 2025 - Tech Forum 2025
BookNet Canada
 
PDF
CIFDAQ Market Wrap for the week of 4th July 2025
CIFDAQ
 
PDF
🚀 Let’s Build Our First Slack Workflow! 🔧.pdf
SanjeetMishra29
 
PPTX
Seamless Tech Experiences Showcasing Cross-Platform App Design.pptx
presentifyai
 
PPTX
New ThousandEyes Product Innovations: Cisco Live June 2025
ThousandEyes
 
PDF
NASA A Researcher’s Guide to International Space Station : Physical Sciences ...
Dr. PANKAJ DHUSSA
 
PPTX
The Project Compass - GDG on Campus MSIT
dscmsitkol
 
PPTX
Mastering ODC + Okta Configuration - Chennai OSUG
HathiMaryA
 
PDF
[Newgen] NewgenONE Marvin Brochure 1.pdf
darshakparmar
 
PPT
Ericsson LTE presentation SEMINAR 2010.ppt
npat3
 
PDF
Peak of Data & AI Encore AI-Enhanced Workflows for the Real World
Safe Software
 
PDF
ICONIQ State of AI Report 2025 - The Builder's Playbook
Razin Mustafiz
 
PDF
The Rise of AI and IoT in Mobile App Tech.pdf
IMG Global Infotech
 
PPTX
AI Penetration Testing Essentials: A Cybersecurity Guide for 2025
defencerabbit
 
PDF
LOOPS in C Programming Language - Technology
RishabhDwivedi43
 
PDF
Kit-Works Team Study_20250627_한달만에만든사내서비스키링(양다윗).pdf
Wonjun Hwang
 
PDF
Newgen Beyond Frankenstein_Build vs Buy_Digital_version.pdf
darshakparmar
 
PPTX
COMPARISON OF RASTER ANALYSIS TOOLS OF QGIS AND ARCGIS
Sharanya Sarkar
 
UiPath DevConnect 2025: Agentic Automation Community User Group Meeting
DianaGray10
 
Staying Human in a Machine- Accelerated World
Catalin Jora
 
Transcript: Book industry state of the nation 2025 - Tech Forum 2025
BookNet Canada
 
CIFDAQ Market Wrap for the week of 4th July 2025
CIFDAQ
 
🚀 Let’s Build Our First Slack Workflow! 🔧.pdf
SanjeetMishra29
 
Seamless Tech Experiences Showcasing Cross-Platform App Design.pptx
presentifyai
 
New ThousandEyes Product Innovations: Cisco Live June 2025
ThousandEyes
 
NASA A Researcher’s Guide to International Space Station : Physical Sciences ...
Dr. PANKAJ DHUSSA
 
The Project Compass - GDG on Campus MSIT
dscmsitkol
 
Mastering ODC + Okta Configuration - Chennai OSUG
HathiMaryA
 
[Newgen] NewgenONE Marvin Brochure 1.pdf
darshakparmar
 
Ericsson LTE presentation SEMINAR 2010.ppt
npat3
 
Peak of Data & AI Encore AI-Enhanced Workflows for the Real World
Safe Software
 
ICONIQ State of AI Report 2025 - The Builder's Playbook
Razin Mustafiz
 
The Rise of AI and IoT in Mobile App Tech.pdf
IMG Global Infotech
 
AI Penetration Testing Essentials: A Cybersecurity Guide for 2025
defencerabbit
 
LOOPS in C Programming Language - Technology
RishabhDwivedi43
 
Kit-Works Team Study_20250627_한달만에만든사내서비스키링(양다윗).pdf
Wonjun Hwang
 
Newgen Beyond Frankenstein_Build vs Buy_Digital_version.pdf
darshakparmar
 
COMPARISON OF RASTER ANALYSIS TOOLS OF QGIS AND ARCGIS
Sharanya Sarkar
 
Ad

Hand gesture recognition using machine learning algorithms

  • 1. Computer Science and Information Technologies Vol. 1, No. 3, November 2020, pp. 116~120 ISSN: 2722-3221, DOI: 10.11591/csit.v1i3.p116-120  116 Journal homepage: http://iaesprime.com/index.php/csit Hand gesture recognition using machine learning algorithms Abhishek B1 , Kanya Krishi2 , Meghana M3 , Mohammed Daaniyaal4 , Anupama H S5 1,2,3,4 B. E, Computer Science and Engineering, BMS Institute of Technology, Bangalore, India 5 BMS Institute of Technology, Bangalore, India Article Info ABSTRACT Article history: Received Apr 24, 2020 Revised Jun 14, 2020 Accepted Jun 29, 2020 Gesture recognition is an emerging topic in today’s technologies. The main focus of this is to recognize the human gestures using mathematical algorithms for human computer interaction. Only a few modes of human-computer interaction exist, they are: through keyboard, mouse, touch screens etc. Each of these devices has their own limitations when it comes to adapting more versatile hardware in computers. Gesture recognition is one of the essential techniques to build user-friendly interfaces. Usually, gestures can be originated from any bodily motion or state, but commonly originate from the face or hand. Gesture recognition enables users to interact with the devices without physically touching them. This paper describes how hand gestures are trained to perform certain actions like switching pages, scrolling up or down in a page. Keywords: Gesture recognition Human–computer interaction User-friendly interface This is an open access article under the CC BY-SA license. Corresponding Author: Abhishek B, B.E, Computer Science and Engineering, BMS Institute of Technology, Bangalore, India. Email: [email protected] 1. INTRODUCTION Gesture recognition is a technique which is used to understand and analyze the human body language and interact with the user accordingly. This in turn helps in building a bridge between the machine and the user to communicate with each other. Gesture recognition is useful in processing the information which cannot be conveyed through speech or text. Gestures are the simplest means of communicating something that is meaningful. This paper involves implementation of the system that aims to design a vision-based hand gesture recognition system with a high correct detection rate along with a high-performance criterion, which can work in a real time human–computer interaction (HCI) system without having any of the limitations (gloves, uniform background etc.) on the user environment. The system can be defined using a flowchart that contains three main steps, they are: learning, detection, recognition as shown in Figure 1. Learning involves two aspects such as: − Training dataset: This is the dataset that consists of different types of hand gestures that are used to train the system based on which the system performs the actions. − Feature extraction: It involves determining the centroid that divides the image into two halves at its geometric centre. Detection involves three aspects: − Capture scene: Captures the images through a web camera, which is used as an input to the system. − Preprocessing: Images that are captured through the webcam are compared with the dataset to recognize the valid hand movements that are needed to perform the required actions. − Hand detection: The requirements for hand detection involve the input image from the webcam. The image should be fetched with a speed of 20 frames per second. Distance should also be maintained
  • 2. Comput. Sci. Inf. Technol.  Hand gesture recognition using machine learning algorithms… (Abhishek B) 117 between the hand and the camera. Approximate distance that should be between hand the camera is around 30 to 100 cm. The video input is stored frame by frame into a matrix after preprocessing. Recognition consists of: − Gesture recognition: The number of fingers present in the hand gesture is determined by making use of defect points present in the gesture. The resultant gesture obtained is fed through a 3-dimensional convolutional neural network consecutively to recognize the current gesture. − Performing action: The recognized gesture is used as an input to perform the actions required by the user. Figure 1. Flowchart of HCI 2. LITERATURE SURVEY The implementation is divided into four main steps: 1) image enhancement and segmentation, 2) orientation detection, 3) feature extraction, and 4) classification [1]. This work was focused on above four categories but main limitation was change of color was happening very rapidly by the change in the different lighting condition, which may cause error or even failures. For example, due to insufficient light condition, the existence of hand area is not detected but the non-skin regions are mistaken for the hand area because of same color [2]. Involves three main steps for hand gesture recognition system: 1) segmentation, 2) feature representation 3) recognition techniques. The system is based on hand gesture recognition by modeling of the hand in spatial domain. The system uses various 2D and 3D geometric and non-geometric models for modeling. It has used fuzzy c-means clustering algorithm which resulted in an accuracy of 85.83%. The main drawback of the system is it does not consider gesture recognition of temporal space, i.e., motion of gestures and it is unable to classify images with complex background i.e., where there are other objects in the scene with the hand objects [3]. This survey focuses on the hand gesture recognition using different steps like data acquisition, pre-processing, segmentation and so on. Suitable input device should be selected for the data acquisition. There are a number of input devices for data acquisition. Some of them are data gloves, marker, and hand images (from webcam/Kinect 3D Sensor). But the limitation with this work was change in the illumination, rotation and orientation, scaling problem and special hardware which is pretty costlier [4]. The system implementation is divided into three phases: 1) Hand gesture recognition using kinetic camera, 2) algorithms for hand detection recognition, 3) hand gesture recognition. The limitation here is that the edge detection and segmentation algorithms used here are not very efficient when compared to neural networks. The dataset being considered here is very small and can be used to detect very few sign gestures. The system architecture consists of: 1) image acquisition, 2) segmentation of hand region, 3) distance transforms method for gesture recognition [5]. The limitations of this system involve 1) the numbers of gestures
  • 3.  ISSN: 2722-3221 Comput. Sci. Inf. Technol., Vol. 1, No. 3, November 2020: 117 – 120 118 that are recognized are less and 2) the gestures recognized were not used to control any applications [6]. In this implementation there are three main algorithms that are used: 1) Viola–Jones algorithm. 2) convex hull algorithm, 3) the AdaBoost based learning algorithm. The work was accomplished by training a set of feature set which is local contour sequence. The limitations of this system are that it requires two sets of images for classification. One is the positive set that contains the required images, the other is the negative set that contains contradicting images [7]. The system implementation consists of three components: 1) hand detection 2) gesture recognition, and 3) HCI. It has implemented the following methodology: 1) the input image is preprocessed and the hand detector tries to filter out the hand from the input image, 2) a CNN classifier is employed to recognize gestures from the processed image, while a Kalman Filter is used to estimate the position of the mouse cursor, and 3) the recognition and estimation results are submitted to a control centre which decides the action to be taken. One of the limitations of this system is that it recognizes only the static images [8]. This implementation focuses on detection of hand gestures using java and neural networks. It is divided into two phases: 1) Detection module using java where in the hand is detected using background subtraction and conversion of video feed into HSB video feed thus detecting skin pixels; 2) The second module is the prediction module; a convolutional neural network is used. The input feed image is gained from Java. The input image is fed into the neural network and is analyzed with respect to the dataset images. One of the limitations of this system is that it requires socket programming in order to connect java and python modules. 3. IMPLEMENTATION A hand gesture recognition system was developed to capture the hand gestures being performed by the user and to control a computer system based on the incoming information. Many of the existing systems in literature have implemented gesture recognition using only spatial modelling, i.e., recognition of a single gesture and not temporal modelling i.e., recognition of motion of gestures. Also, the existing systems have not been implemented in real time, they use a pre captured image as an input for gesture recognition. To overcome these existing problems a new architecture has been developed which aims to design a vision-based hand gesture recognition system with a high correct detection rate along with a high-performance criterion, which can work in a real time HCI system without having any of the mentioned strict limitations (gloves, uniform background, etc.) on the user environment. The design is composed of a HCI system which uses hand gestures as input for communication as show in Figure 2. Figure 2. Design of the proposed HCI system
  • 4. Comput. Sci. Inf. Technol.  Hand gesture recognition using machine learning algorithms… (Abhishek B) 119 Input to the system is from the web camera or a prerecorded video sequence. Later it detects the skin color by using an adaptive algorithm in the beginning of the frames. For the current user skin color has to be fixed based on the lighting and camera parameter and condition. Once it has been fixed, hand is localized with a histogram clustering method. Then a machine learning algorithm has been used to detect the hand gestures in consecutive frames to distinguish the current gesture. These gestures are used as an input for a computer application as shown in Figure 3. The system is divided into 3 subsystems: 3.1. Hand and motion detection The Web-camera captures the hand movement and provides it as input to OpenCV and TensorFlow Object detector. Edge detection and skin detection are performed to obtain the boundary of the hand. This is then sent to the 3D CNN. 3.2. Dataset Dataset is used for training the 3D CNN. Two types of datasets are being used–one for the hand detection and the other for the motion or gesture detection. Hand detection uses EGO dataset, Motion or Gesture Recognition uses Jester dataset. 3.3. 3D CNN CNN’s are a class of deep learning neural networks used for analyzing videos and images. It consists of several layers–input layer, hidden layers, and output layer. It performs back propagation for better accuracy and efficiency. It performs training and verification of the recognized gestures and HCIs take place–turning of the pages, zooming in, and zooming out. The interactions with the computer take place with the help of PyAutoGUI or System Calls. Figure 3. System recognized hand gestures 4. CONCLUSION The importance of gesture recognition lies in building efficient human-machine interaction. This paper describes how the implementation of the system is done based upon the images captured. Hand detection is done using OpenCV and TensorFlow object detector. And further it is enhanced for interpretation of gestures by the computer to perform actions like switching the pages, scrolling up or down the page.
  • 5.  ISSN: 2722-3221 Comput. Sci. Inf. Technol., Vol. 1, No. 3, November 2020: 117 – 120 120 ACKNOWLEDGEMENT This work is done, supervised, and supported by the students and faculty members of the Department of Computer Science and Engineering, BMS Institute of Technology, Bangalore. REFERENCES [1] M. Panwar and P. S. Mehra, “Hand gesture recognition for human computer interaction,” 2011 International Conference on Image Information Processing, Shimla, pp. 1-7, 2011. [2] R. Zaman Khan and N. A. Ibraheem, “Comparative Study of Hand Gesture Recognition System,” International Conference of Advanced Computer Science & Information Technology, 2012. [3] A. R. Sarkar, G. Sanyal, and S. Majumder, “Hand Gesture Recognition Systems: A Survey,” International Journal of Computer Applications, vol. 71, no.15, pp. 25-37, May 2013. [4] A. E. Manjunath, B. P. V. Kumar, H. Rajesh, “Comparative Study of Hand Gesture Recognition Algorithms,” International Journal of Research in Computer and Communication Technology, vol 3, no. 4, April 2014. [5] D. R. Jadhav and L. M. R. J. Lobo, “Navigation of PowerPoint Using Hand Gestures,” International Journal of Science and Research (IJSR), vol. 4, no. 1, pp. 833-837, 2015. [6] R. M. Gurav and P. K. Kadbe, "Real time finger tracking and contour detection for gesture recognition using OpenCV," 2015 International Conference on Industrial Instrumentation and Control (ICIC), pp. 974-977, 2015. [7] P. Xu, “A Real-time Hand Gesture Recognition and Human-Computer Interaction System,” 2017, [Online]. Available: https://arxiv.org/abs/1704.07296 [8] P. Suganya, R. Sathya, and K. Vijayalakshmi, “Detection and Recognition of Gestures to Control the System Applications by Neural Networks,” International Journal of Pure and Applied Mathematics, vol. 118, no. 10, pp. 399-405, January 2018.