SlideShare a Scribd company logo
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3130
Chatbot with Gesture based User Input
Rajkumar G1, Gayathri M2
1Student/CSE, SRM Institute of Science and Technology, Chennai-603203
2Assistant Professor/CSE, SRM institute of Science and Technology, Chennai
---------------------------------------------------------------------***----------------------------------------------------------------------
Abstract: The paper describes a approach towards
building a chatbot with gesture based input. A gesture table
will be described here which is a mapping sequence of the
appropriate gestures to the appropriate commands they
mean. After successful gesture recognition, the classified or
recognized gesture is looked up in the table and is
converted accordingly into text and then that text is fed
into the chatbot program as input and appropriate
command is executed. The paper selects existing known
algorithms and aims to explores the overall effects it has
and if such a application is suitable for areas such as deaf
and dumb sign language or in public galleries, to quickly
get required information
1. Introduction
Gesture recognition is a topic in computer science with the
goal of interpreting human gestures algorithms. Gestures
can be created from any bodily motion but commonly
originate from the face or hand. Quite a few approaches
have been suggested using cameras and computer vision
algorithms to interpret sign language.
Gesture recognition applications in various fields such as
automotive, consumer electronics, Transit, gaming,
smartphones, defence, sign language translation sectors. Its
used to provide a more natural interaction with the UI to
save time and make things more intuitive. In
entertainment, mainly gaming ,it can be used to add a new
way of interaction to attract players. In smartphones, its
used for faster unlock. For deaf and dumb, its used to for
translating sign language into text.
Gesture recognition has spawned several correlated
technology such as touchless interface. Touchless user
interface (TUI) is a interface through which a computer can
be given commands or operations in form of body motion
and gestures without touching a keyboard, mouse, or
screen. For example, Microsoft's Kinect is a touchless game
interface; This is basically another major application of
gesture recognition. Many companies are invested in
development of this technology.
Intel corporation is researching how touch multi-factor
authentication can help healthcare organizations minimize
security risks and at the same time improve clinical
efficiency.Here Multi-factor authentication means multiple
layers/levels to authorise a particular transaction.
Microsoft corporation is researching the use of touchless
interaction within surgical settings, allowing images to be
viewed and manipulated without any physical contact with
traditional computer hardware through gesture
recognition.
Gesture recognition is linked to the field of digital image
processing. It uses many algorithms and concepts of digital
image processing along with concepts and algorithms of AI.
Some of these include thresholding, otsu algorithm,
anisotropic diffusion, Hidden Markov model, image editing,
image restoration etc(belonging to digital image
processing), machine learning ,computer vision involving
automatic inspection, Assisting humans in identification
tasks, Controlling processes, detecting events, interaction,
modeling objects, navigation, organizing information.
2. Literature Survey
Rupesh Prajapati, Vedant Pandey, Nupur Jamindar, Neeraj
Yadav, Prof. Neelam Phadnis have published a research
paper titled “Hand Gesture Recognition and Voice
Conversion for Deaf and Dumb”. In it they accept the input
in form of Video feed through webcam, they have proposed
a method of Create a database of images for training and
use PCA for creating a relationship out of its linearly
uncorrelated characterstics and then use classfication
algorithm like KNN and SVN.[1]
Anchal Sood and Anju Mishra have published a research
paper titled “AAWAAZ: A communication system for deaf
and dumb-”.They have proposed a sign recognition system
based on Harris algorithm for feature extraction and then it
is stored in a matrix. This matrix is further used to match
the image from the database. The system does have a few
limitations. Binary imaging technique doesn’t perfectly
seperates discrete objects from backgrounds of similar
colors due to inaccurate thresholding algorithm. But the
results are efficient.[2]
Ms R. Vinitha and Ms A. Theerthana have published a
research paper titled “Design And Development of Hand
Gesture Recognition System For Speech Impaired People.”
In it they accept input in form of Data glove(consisting of
sensors,accelometer and PIC-microcontroller). Gesture
recognition is based on hardware approach[3]
Shangeetha, R. K., V. Valliammai, and S. Padmavathi have
published a research paper titled “Computer vision based
approach for Indian Sign Language character recognition”.
In it they accept input in form of Live video feed from
webcam. They have also proposed a method of using HIS
model and then the feature extraction is done by distance
transform method. The results are as follows,The gestures
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3131
stored and recognized and coverted to appropriate voice
output.[4]
R. Aravind, D.Anbasaran, K.Alice have published a research
paper titled “GRS-Gesture Based Recognition System for
Indian Sign Language Recognition System for Deaf and
Dumb people”. In it they accept input in form of webcam.
They have proposed a method of gesture recognition using
color intensity variation and edge detection and
background subtraction. The results were as follows,
gesture symbols extracted from the image alone The gaps
identified were problems identifying concurrent
symbols[5]
Tripathi K. and Nandi have published a research paper
titled “Continuous indian sign language gesture recognition
and sentence”. In they accept input in form of webcam.
They have proposed a method using Orientation Histogram
and PCA. The results were that highly accurate gesture
recognition by reducing dimension features after
extraction.[6]
Daniel Simões Lopes have developed a gesture based
recognition system as a basis for touchless user interface.
In it they accept input in form of Hand gestures and body
postures. They have proposed a method of Developed a
gesture recognition system where each hand is can make
gestures and body postures can also be used. They found
out that that using touchless interface is indeed better than
relying on 2-D devices as it allows for faster retrieval and
greater flexibility in the sensitive surgical enviornment.
The gaps that were identified were Precision tasks such as
clipping plane and visualization and tagging are better
performed with mouse due to not good gesture
recognition,[7]
Another interesting method proposed was relying on
passive props.3D medical visualization relied on passive
interface props. Hardware interface placed in a position
where the user needs manipulate the control devices. The
result was that the system facilitates a natural interaction
between UI and user. The gaps that were identified were
that It was not easy to stop or use in several situations, the
proposed mechanism is not flexible enough[8]
-Chatbot -Gonzalas Cenelia Article Internet
Gonzalas Cenelia, gives a excellent approach to building a
chatbot based on keyword concept. It involves finding
keywords in the given user input and match it with a
already predefine knowledge base which has keywords
and assoiciated responses to it. Revolving around the
concept of keyword, things like context of the conversation
,learning from feedback to add knowledge base, keyword
ranking etc are also introduced [9]
3. Proposed System:-
Note the high level view of the proposed system in the
given diagram.
Input: Imagestream from webcam
Output: Appropriate textual output or execution of a
command like search
A. Gesture Recognition Module:-
Input: Imagestream from webcam representing correct
gestures
Output:Successfully recognized gesture label(See gesture
mapping sequence)
The recognized gestures can be divided into two types.
Counting gestures
-Counting gestures refer to the gestures that indicate
numbers like using only your index finger to count 1 ,or
any finger for that matter.
Implementation used:Convexity Defect
General gestures
-General gestures are the set of gestures that can be
classified by a AI model which is trained on a database of
images.
Implementation used:CNN model
For each of these types, A separate implementation is
provided which is considered more optimum for
recognizing those gestures. Using the convexity defect
approach, we don’t have to train the AI using large number
of images for each number representational gesture.
Besides the convexity defect approach is very flexible ,it
simply counts the number of fingers so we don’t have to
exact representational gesture like for example using the
pinky finger to represent 1 rather than the index finger.(To
its recommended to use the normal ones).
I. Convexity Defect Implementation:-
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3132
1)Main GUI Control Window
-This window is the main navigation window used to
activate all the modules /algorithms for sucessful gesture
recognition
2)GUI Input window
-This window activates the webcam and gives you input
video feed to manipulate on
3)Thresholding(Background substraction)
As the video feed continues ,every frame is continously
getting converted to grayscale image.
A grayscale image is a image without saturation aka color
.The only information its pixels have is on the intensity of
light falling on that pixel. So basically the image appears in
shades of grey.
After this process, thresholding is applied. Thresholding
converts the grayscale image to a binary color image which
typically are black and white. A value called the threshold
point can be manually provided or using an algorithm can
be automated . The threshold function will make the pixels
whose intensity are greater than the threshold value white
and vice versa for black.
In our case,we have use otsu algorithm to calculate
threshold point.
4)Finding contours and convex hull and defects
Contours are the areas of the discrete objects that we
capture from the image. We assume that the contour which
has the maximum area in the binary image will be of our
hand, thats why a white screen should be used to hide the
background objects.
We calculate then the convex hull which is the smallest
polygon that encloses a set of points. Then the convexity
defect is defined as the deepest point in the said binary
image between two vertices of the convex hull. The actual
definition of convexity defect in terms of physical
boundaries of the object is given below
Convexity defect is a cavity in an object (blob, contour)
segmented out from an image. That means an area that do
not belong to the object but located inside of its outer
boundary -convex hull.
II. Deep Learning Implementation:
(Contd)
The input consists of 20,000 images . The first Module
Image URL loader and reader is a simple algorithm which
reads all the appropriate pathnames of the files and
converts them into arrays using opencv functions .It
extracts all the images required for training and stores the
labels for each image in a separate list. One list all the
images and the other list contains all the corresponding
labels of each image. During the process the images are
converted to grayscale and resized to smaller size so that
the training phase of CNN model will be shorter. Then its
fed into the CNN model. It splits its data into 70% as
training and 30% as testing.
(In the below table ,the gestures can be mapped to the
appropriate text according to the usage, for example thumb
down to what or L to how etc, say if the application is to be
used in a deaf/dumb language or a custom gesture system
required etc.)
It takes 30 mins approximately for five iterations over the
entire database of images for training as well as the
validating(testing) to complete. After that model is ready to
predict furthur images.
(A sample tableis given if this application is to be used in a
museum)
Gesture Label
Thumb down How many
artifacts are
here?
palm(Horizontal) What is the name
of the place
L Where is it
located
Fist(Horizontal) When was it
founded
Fist(Vertical) What is the most
famous artifact
here.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3133
Tools used:
-Python(Anaconda package)
-Spyder(IDE)
-Keras,tensorflow and opencv as python libraries
B.Chatbot
Template Response:- We put up ready-made templates as
replies and those templates will have parameters to which
will be replaced by the appropriate data relevent to the
conversion
Keywords:-The concept on the basis of which the chatbot
will be implemented .The chatbot will scan the user input
and see if a keyword exists in it. Theses keywords are
stored in the knowledge base. And each of the keyword
have a set of responses linked to it. So a response is chosen
and given as reply
Knowledge base:-The knowledge base is a 2-D array
storing a set of keywords in the first column and its
associated responses in the next columns.
Keyword Ranking:-This is used when there are multiple
keyword match and based on the furthur corresponding
keyword match and repetitions, the appropriate keyword
is selected then the corresponding reply is given
The chatbot will be written in java based module which
will accept the input as text which was converted
appropriately from the gesture input.
Screenshots:
Deep learning implementation
(Note:Predict prints a command line out put.
The image is first captured and then saved
As png file and then is fed into the deep
Learning model)
Convexity defect to count fingers to display numbers
The output is planned to be written to
Text file which will be fed into a java based
chatbot
Chatbot Output:(Separate module)
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072
© 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3134
Example sentences are given as input, it can be customised
according to the table created by simple adding those
responses to the knowledge base of the chatbot.
References
[1] Rupesh Prajapati, Vedant Pandey, Nupur Jamindar,
Neeraj Yadav, Prof. Neelam Phadnis,”Hand Gesture
Recognition and Voice Conversion for Deaf and
Dumb”,IRJET,2018
[2] Anchal Sood and Anju Mishra, “AAWAAZ:
communication system for deaf and dumb-”.IEEE,2016
[3]Ms R. Vinitha and Ms A. Theerthana,"Design And
Development Of Hand Gesture Recognition System For
Speech Impaired People.",2016
[4]Shangeetha, R. K., V. Valliammai, and S.
Padmavathi,"Computer vision based approach for Indian
Sign Language character recognition." Machine Vision and
Image Processing (MVIP), IEEE,2012
[5] R.Aravind, D.Anbasaran, K.Alice,”GRS-Gesture Based
Recognition System for Indian Sign Language Recognition
System for Deaf and Dumb people”, International Journal of
Trend in Scientific Research and Development -
IJTSRD,2018
[6]Tripathi K. and Nandi, “Continuous indian sign language
gesture recognition and sentence”,Procedia
Computer,2015
[7]Daniel Simões Lopes “On the utility of 3D hand cursors
to explore medical volume datasets with a touchless
interface”, Journal of biomedical informatics,2017 Passive
real-world interface props for neurosurgical visualization,
Hinckley et al.CHI '94,
[9]Otsu Algorithm-(Algorithm used for Thresholding,
grayscale to binary),See OpenCv 3.4.2 documentation for
algorithm implementation
[10]Peter-Draukr algorithm(For calculation of contours)
[11]Satoshi Suzuki and others. Topological structural
analysis of digitized binary images by border following.
Computer Vision, Graphics, and Image Processing,
30(1):32–46, 1985.(Algorithm used for finding contours)
[12]CNN(For training the AI model)

More Related Content

What's hot (20)

PDF
IRJET - Hand Gestures Recognition using Deep Learning
IRJET Journal
 
PDF
Gesture recognition using artificial neural network,a technology for identify...
NidhinRaj Saikripa
 
PDF
Paper id 25201413
IJRAT
 
PDF
IRJET- Real-Time Object Detection System using Caffe Model
IRJET Journal
 
PDF
IRJET - NETRA: Android Application for Visually Challenged People to Dete...
IRJET Journal
 
PDF
Development of Real Time Face Recognition System using OpenCV
IRJET Journal
 
PDF
Real Time Vision Hand Gesture Recognition Based Media Control via LAN & Wirel...
IJMER
 
PDF
Smart Door Access using Facial Recognition
ijtsrd
 
PDF
IRJET- Applications of Object Detection System
IRJET Journal
 
PDF
IRJET- Vision Based Sign Language by using Matlab
IRJET Journal
 
PDF
Profile Identification through Face Recognition
ijtsrd
 
PDF
Paper id 21201494
IJRAT
 
PDF
Iris Encryption using (2, 2) Visual cryptography & Average Orientation Circul...
AM Publications
 
PDF
IRJET-Human Face Detection and Identification using Deep Metric Learning
IRJET Journal
 
PDF
Implementation of video tagging identifying characters in video
eSAT Publishing House
 
PDF
Final Year Project-Gesture Based Interaction and Image Processing
Sabnam Pandey, MBA
 
PDF
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET Journal
 
PDF
Authentication Schemes for Session Passwords using Color and Images
IJNSA Journal
 
PDF
IRJET - Mutecom using Tensorflow-Keras Model
IRJET Journal
 
IRJET - Hand Gestures Recognition using Deep Learning
IRJET Journal
 
Gesture recognition using artificial neural network,a technology for identify...
NidhinRaj Saikripa
 
Paper id 25201413
IJRAT
 
IRJET- Real-Time Object Detection System using Caffe Model
IRJET Journal
 
IRJET - NETRA: Android Application for Visually Challenged People to Dete...
IRJET Journal
 
Development of Real Time Face Recognition System using OpenCV
IRJET Journal
 
Real Time Vision Hand Gesture Recognition Based Media Control via LAN & Wirel...
IJMER
 
Smart Door Access using Facial Recognition
ijtsrd
 
IRJET- Applications of Object Detection System
IRJET Journal
 
IRJET- Vision Based Sign Language by using Matlab
IRJET Journal
 
Profile Identification through Face Recognition
ijtsrd
 
Paper id 21201494
IJRAT
 
Iris Encryption using (2, 2) Visual cryptography & Average Orientation Circul...
AM Publications
 
IRJET-Human Face Detection and Identification using Deep Metric Learning
IRJET Journal
 
Implementation of video tagging identifying characters in video
eSAT Publishing House
 
Final Year Project-Gesture Based Interaction and Image Processing
Sabnam Pandey, MBA
 
IRJET- Review on Human Action Detection in Stored Videos using Support Vector...
IRJET Journal
 
Authentication Schemes for Session Passwords using Color and Images
IJNSA Journal
 
IRJET - Mutecom using Tensorflow-Keras Model
IRJET Journal
 

Similar to IRJET - Chatbot with Gesture based User Input (20)

PDF
Sign Language Identification based on Hand Gestures
IRJET Journal
 
PDF
SLIDE PRESENTATION BY HAND GESTURE RECOGNITION USING MACHINE LEARNING
IRJET Journal
 
PDF
Sign Language Detection using Action Recognition
IRJET Journal
 
PDF
VIRTUAL PAINT APPLICATION USING HAND GESTURES
IRJET Journal
 
PDF
Hand Gesture Recognition System Using Holistic Mediapipe
IRJET Journal
 
PDF
Mouse Cursor Control Hands Free Using Deep Learning
IRJET Journal
 
PDF
Mouse Cursor Control Hands Free Using Deep Learning
IRJET Journal
 
PDF
A Survey Paper on Controlling Computer using Hand Gestures
IRJET Journal
 
PDF
Smart Presentation Control by Hand Gestures Using Computer Vision and Google’...
IRJET Journal
 
PDF
Controlling Computer using Hand Gestures
IRJET Journal
 
PDF
Dynamic Hand Gesture Recognition for Indian Sign Language: A Review
IRJET Journal
 
PDF
IRJET- Human Activity Recognition using Flex Sensors
IRJET Journal
 
PDF
IRJET- Secure Online Payment with Facial Recognition using CNN
IRJET Journal
 
PDF
Attendance System using Android Integrated Biometric Fingerprint Recognition
IRJET Journal
 
PDF
Sign Language Recognition using Machine Learning
IRJET Journal
 
PDF
Accessing Operating System using Finger Gesture
IRJET Journal
 
PDF
IRJET- Finger Gesture Recognition Using Linear Camera
IRJET Journal
 
PDF
IRJET- Convenience Improvement for Graphical Interface using Gesture Dete...
IRJET Journal
 
PDF
IRJET- Survey on Development of Fingerprint Biometric Attendance Management S...
IRJET Journal
 
PDF
A Survey on Virtual Whiteboard-A Gesture Controlled Pen-free Tool
vivatechijri
 
Sign Language Identification based on Hand Gestures
IRJET Journal
 
SLIDE PRESENTATION BY HAND GESTURE RECOGNITION USING MACHINE LEARNING
IRJET Journal
 
Sign Language Detection using Action Recognition
IRJET Journal
 
VIRTUAL PAINT APPLICATION USING HAND GESTURES
IRJET Journal
 
Hand Gesture Recognition System Using Holistic Mediapipe
IRJET Journal
 
Mouse Cursor Control Hands Free Using Deep Learning
IRJET Journal
 
Mouse Cursor Control Hands Free Using Deep Learning
IRJET Journal
 
A Survey Paper on Controlling Computer using Hand Gestures
IRJET Journal
 
Smart Presentation Control by Hand Gestures Using Computer Vision and Google’...
IRJET Journal
 
Controlling Computer using Hand Gestures
IRJET Journal
 
Dynamic Hand Gesture Recognition for Indian Sign Language: A Review
IRJET Journal
 
IRJET- Human Activity Recognition using Flex Sensors
IRJET Journal
 
IRJET- Secure Online Payment with Facial Recognition using CNN
IRJET Journal
 
Attendance System using Android Integrated Biometric Fingerprint Recognition
IRJET Journal
 
Sign Language Recognition using Machine Learning
IRJET Journal
 
Accessing Operating System using Finger Gesture
IRJET Journal
 
IRJET- Finger Gesture Recognition Using Linear Camera
IRJET Journal
 
IRJET- Convenience Improvement for Graphical Interface using Gesture Dete...
IRJET Journal
 
IRJET- Survey on Development of Fingerprint Biometric Attendance Management S...
IRJET Journal
 
A Survey on Virtual Whiteboard-A Gesture Controlled Pen-free Tool
vivatechijri
 
Ad

More from IRJET Journal (20)

PDF
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
IRJET Journal
 
PDF
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
IRJET Journal
 
PDF
Kiona – A Smart Society Automation Project
IRJET Journal
 
PDF
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
IRJET Journal
 
PDF
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
IRJET Journal
 
PDF
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
IRJET Journal
 
PDF
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
IRJET Journal
 
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
PDF
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
IRJET Journal
 
PDF
BRAIN TUMOUR DETECTION AND CLASSIFICATION
IRJET Journal
 
PDF
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
IRJET Journal
 
PDF
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
IRJET Journal
 
PDF
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
IRJET Journal
 
PDF
Breast Cancer Detection using Computer Vision
IRJET Journal
 
PDF
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
PDF
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
PDF
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
IRJET Journal
 
PDF
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
PDF
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
PDF
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Enhanced heart disease prediction using SKNDGR ensemble Machine Learning Model
IRJET Journal
 
Utilizing Biomedical Waste for Sustainable Brick Manufacturing: A Novel Appro...
IRJET Journal
 
Kiona – A Smart Society Automation Project
IRJET Journal
 
DESIGN AND DEVELOPMENT OF BATTERY THERMAL MANAGEMENT SYSTEM USING PHASE CHANG...
IRJET Journal
 
Invest in Innovation: Empowering Ideas through Blockchain Based Crowdfunding
IRJET Journal
 
SPACE WATCH YOUR REAL-TIME SPACE INFORMATION HUB
IRJET Journal
 
A Review on Influence of Fluid Viscous Damper on The Behaviour of Multi-store...
IRJET Journal
 
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Explainable AI(XAI) using LIME and Disease Detection in Mango Leaf by Transfe...
IRJET Journal
 
BRAIN TUMOUR DETECTION AND CLASSIFICATION
IRJET Journal
 
The Project Manager as an ambassador of the contract. The case of NEC4 ECC co...
IRJET Journal
 
"Enhanced Heat Transfer Performance in Shell and Tube Heat Exchangers: A CFD ...
IRJET Journal
 
Advancements in CFD Analysis of Shell and Tube Heat Exchangers with Nanofluid...
IRJET Journal
 
Breast Cancer Detection using Computer Vision
IRJET Journal
 
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
A Novel System for Recommending Agricultural Crops Using Machine Learning App...
IRJET Journal
 
Auto-Charging E-Vehicle with its battery Management.
IRJET Journal
 
Analysis of high energy charge particle in the Heliosphere
IRJET Journal
 
Wireless Arduino Control via Mobile: Eliminating the Need for a Dedicated Wir...
IRJET Journal
 
Ad

Recently uploaded (20)

PPTX
DATA BASE MANAGEMENT AND RELATIONAL DATA
gomathisankariv2
 
PDF
3rd International Conference on Machine Learning and IoT (MLIoT 2025)
ClaraZara1
 
PDF
mbse_An_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
PPT
Testing and final inspection of a solar PV system
MuhammadSanni2
 
PPTX
UNIT 1 - INTRODUCTION TO AI and AI tools and basic concept
gokuld13012005
 
PDF
Electrical Machines and Their Protection.pdf
Nabajyoti Banik
 
PPTX
How Industrial Project Management Differs From Construction.pptx
jamespit799
 
PDF
Submit Your Papers-International Journal on Cybernetics & Informatics ( IJCI)
IJCI JOURNAL
 
PDF
AI TECHNIQUES FOR IDENTIFYING ALTERATIONS IN THE HUMAN GUT MICROBIOME IN MULT...
vidyalalltv1
 
PDF
aAn_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
PDF
Digital water marking system project report
Kamal Acharya
 
PPT
Footbinding.pptmnmkjkjkknmnnjkkkkkkkkkkkkkk
mamadoundiaye42742
 
PDF
Electrical Engineer operation Supervisor
ssaruntatapower143
 
PDF
Viol_Alessandro_Presentazione_prelaurea.pdf
dsecqyvhbowrzxshhf
 
PDF
MODULE-5 notes [BCG402-CG&V] PART-B.pdf
Alvas Institute of Engineering and technology, Moodabidri
 
PPTX
darshai cross section and river section analysis
muk7971
 
PPTX
MODULE 04 - CLOUD COMPUTING AND SECURITY.pptx
Alvas Institute of Engineering and technology, Moodabidri
 
PDF
20ES1152 Programming for Problem Solving Lab Manual VRSEC.pdf
Ashutosh Satapathy
 
PPTX
Distribution reservoir and service storage pptx
dhanashree78
 
PDF
Reasons for the succes of MENARD PRESSUREMETER.pdf
majdiamz
 
DATA BASE MANAGEMENT AND RELATIONAL DATA
gomathisankariv2
 
3rd International Conference on Machine Learning and IoT (MLIoT 2025)
ClaraZara1
 
mbse_An_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
Testing and final inspection of a solar PV system
MuhammadSanni2
 
UNIT 1 - INTRODUCTION TO AI and AI tools and basic concept
gokuld13012005
 
Electrical Machines and Their Protection.pdf
Nabajyoti Banik
 
How Industrial Project Management Differs From Construction.pptx
jamespit799
 
Submit Your Papers-International Journal on Cybernetics & Informatics ( IJCI)
IJCI JOURNAL
 
AI TECHNIQUES FOR IDENTIFYING ALTERATIONS IN THE HUMAN GUT MICROBIOME IN MULT...
vidyalalltv1
 
aAn_Introduction_to_Arcadia_20150115.pdf
henriqueltorres1
 
Digital water marking system project report
Kamal Acharya
 
Footbinding.pptmnmkjkjkknmnnjkkkkkkkkkkkkkk
mamadoundiaye42742
 
Electrical Engineer operation Supervisor
ssaruntatapower143
 
Viol_Alessandro_Presentazione_prelaurea.pdf
dsecqyvhbowrzxshhf
 
MODULE-5 notes [BCG402-CG&V] PART-B.pdf
Alvas Institute of Engineering and technology, Moodabidri
 
darshai cross section and river section analysis
muk7971
 
MODULE 04 - CLOUD COMPUTING AND SECURITY.pptx
Alvas Institute of Engineering and technology, Moodabidri
 
20ES1152 Programming for Problem Solving Lab Manual VRSEC.pdf
Ashutosh Satapathy
 
Distribution reservoir and service storage pptx
dhanashree78
 
Reasons for the succes of MENARD PRESSUREMETER.pdf
majdiamz
 

IRJET - Chatbot with Gesture based User Input

  • 1. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3130 Chatbot with Gesture based User Input Rajkumar G1, Gayathri M2 1Student/CSE, SRM Institute of Science and Technology, Chennai-603203 2Assistant Professor/CSE, SRM institute of Science and Technology, Chennai ---------------------------------------------------------------------***---------------------------------------------------------------------- Abstract: The paper describes a approach towards building a chatbot with gesture based input. A gesture table will be described here which is a mapping sequence of the appropriate gestures to the appropriate commands they mean. After successful gesture recognition, the classified or recognized gesture is looked up in the table and is converted accordingly into text and then that text is fed into the chatbot program as input and appropriate command is executed. The paper selects existing known algorithms and aims to explores the overall effects it has and if such a application is suitable for areas such as deaf and dumb sign language or in public galleries, to quickly get required information 1. Introduction Gesture recognition is a topic in computer science with the goal of interpreting human gestures algorithms. Gestures can be created from any bodily motion but commonly originate from the face or hand. Quite a few approaches have been suggested using cameras and computer vision algorithms to interpret sign language. Gesture recognition applications in various fields such as automotive, consumer electronics, Transit, gaming, smartphones, defence, sign language translation sectors. Its used to provide a more natural interaction with the UI to save time and make things more intuitive. In entertainment, mainly gaming ,it can be used to add a new way of interaction to attract players. In smartphones, its used for faster unlock. For deaf and dumb, its used to for translating sign language into text. Gesture recognition has spawned several correlated technology such as touchless interface. Touchless user interface (TUI) is a interface through which a computer can be given commands or operations in form of body motion and gestures without touching a keyboard, mouse, or screen. For example, Microsoft's Kinect is a touchless game interface; This is basically another major application of gesture recognition. Many companies are invested in development of this technology. Intel corporation is researching how touch multi-factor authentication can help healthcare organizations minimize security risks and at the same time improve clinical efficiency.Here Multi-factor authentication means multiple layers/levels to authorise a particular transaction. Microsoft corporation is researching the use of touchless interaction within surgical settings, allowing images to be viewed and manipulated without any physical contact with traditional computer hardware through gesture recognition. Gesture recognition is linked to the field of digital image processing. It uses many algorithms and concepts of digital image processing along with concepts and algorithms of AI. Some of these include thresholding, otsu algorithm, anisotropic diffusion, Hidden Markov model, image editing, image restoration etc(belonging to digital image processing), machine learning ,computer vision involving automatic inspection, Assisting humans in identification tasks, Controlling processes, detecting events, interaction, modeling objects, navigation, organizing information. 2. Literature Survey Rupesh Prajapati, Vedant Pandey, Nupur Jamindar, Neeraj Yadav, Prof. Neelam Phadnis have published a research paper titled “Hand Gesture Recognition and Voice Conversion for Deaf and Dumb”. In it they accept the input in form of Video feed through webcam, they have proposed a method of Create a database of images for training and use PCA for creating a relationship out of its linearly uncorrelated characterstics and then use classfication algorithm like KNN and SVN.[1] Anchal Sood and Anju Mishra have published a research paper titled “AAWAAZ: A communication system for deaf and dumb-”.They have proposed a sign recognition system based on Harris algorithm for feature extraction and then it is stored in a matrix. This matrix is further used to match the image from the database. The system does have a few limitations. Binary imaging technique doesn’t perfectly seperates discrete objects from backgrounds of similar colors due to inaccurate thresholding algorithm. But the results are efficient.[2] Ms R. Vinitha and Ms A. Theerthana have published a research paper titled “Design And Development of Hand Gesture Recognition System For Speech Impaired People.” In it they accept input in form of Data glove(consisting of sensors,accelometer and PIC-microcontroller). Gesture recognition is based on hardware approach[3] Shangeetha, R. K., V. Valliammai, and S. Padmavathi have published a research paper titled “Computer vision based approach for Indian Sign Language character recognition”. In it they accept input in form of Live video feed from webcam. They have also proposed a method of using HIS model and then the feature extraction is done by distance transform method. The results are as follows,The gestures
  • 2. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3131 stored and recognized and coverted to appropriate voice output.[4] R. Aravind, D.Anbasaran, K.Alice have published a research paper titled “GRS-Gesture Based Recognition System for Indian Sign Language Recognition System for Deaf and Dumb people”. In it they accept input in form of webcam. They have proposed a method of gesture recognition using color intensity variation and edge detection and background subtraction. The results were as follows, gesture symbols extracted from the image alone The gaps identified were problems identifying concurrent symbols[5] Tripathi K. and Nandi have published a research paper titled “Continuous indian sign language gesture recognition and sentence”. In they accept input in form of webcam. They have proposed a method using Orientation Histogram and PCA. The results were that highly accurate gesture recognition by reducing dimension features after extraction.[6] Daniel Simões Lopes have developed a gesture based recognition system as a basis for touchless user interface. In it they accept input in form of Hand gestures and body postures. They have proposed a method of Developed a gesture recognition system where each hand is can make gestures and body postures can also be used. They found out that that using touchless interface is indeed better than relying on 2-D devices as it allows for faster retrieval and greater flexibility in the sensitive surgical enviornment. The gaps that were identified were Precision tasks such as clipping plane and visualization and tagging are better performed with mouse due to not good gesture recognition,[7] Another interesting method proposed was relying on passive props.3D medical visualization relied on passive interface props. Hardware interface placed in a position where the user needs manipulate the control devices. The result was that the system facilitates a natural interaction between UI and user. The gaps that were identified were that It was not easy to stop or use in several situations, the proposed mechanism is not flexible enough[8] -Chatbot -Gonzalas Cenelia Article Internet Gonzalas Cenelia, gives a excellent approach to building a chatbot based on keyword concept. It involves finding keywords in the given user input and match it with a already predefine knowledge base which has keywords and assoiciated responses to it. Revolving around the concept of keyword, things like context of the conversation ,learning from feedback to add knowledge base, keyword ranking etc are also introduced [9] 3. Proposed System:- Note the high level view of the proposed system in the given diagram. Input: Imagestream from webcam Output: Appropriate textual output or execution of a command like search A. Gesture Recognition Module:- Input: Imagestream from webcam representing correct gestures Output:Successfully recognized gesture label(See gesture mapping sequence) The recognized gestures can be divided into two types. Counting gestures -Counting gestures refer to the gestures that indicate numbers like using only your index finger to count 1 ,or any finger for that matter. Implementation used:Convexity Defect General gestures -General gestures are the set of gestures that can be classified by a AI model which is trained on a database of images. Implementation used:CNN model For each of these types, A separate implementation is provided which is considered more optimum for recognizing those gestures. Using the convexity defect approach, we don’t have to train the AI using large number of images for each number representational gesture. Besides the convexity defect approach is very flexible ,it simply counts the number of fingers so we don’t have to exact representational gesture like for example using the pinky finger to represent 1 rather than the index finger.(To its recommended to use the normal ones). I. Convexity Defect Implementation:-
  • 3. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3132 1)Main GUI Control Window -This window is the main navigation window used to activate all the modules /algorithms for sucessful gesture recognition 2)GUI Input window -This window activates the webcam and gives you input video feed to manipulate on 3)Thresholding(Background substraction) As the video feed continues ,every frame is continously getting converted to grayscale image. A grayscale image is a image without saturation aka color .The only information its pixels have is on the intensity of light falling on that pixel. So basically the image appears in shades of grey. After this process, thresholding is applied. Thresholding converts the grayscale image to a binary color image which typically are black and white. A value called the threshold point can be manually provided or using an algorithm can be automated . The threshold function will make the pixels whose intensity are greater than the threshold value white and vice versa for black. In our case,we have use otsu algorithm to calculate threshold point. 4)Finding contours and convex hull and defects Contours are the areas of the discrete objects that we capture from the image. We assume that the contour which has the maximum area in the binary image will be of our hand, thats why a white screen should be used to hide the background objects. We calculate then the convex hull which is the smallest polygon that encloses a set of points. Then the convexity defect is defined as the deepest point in the said binary image between two vertices of the convex hull. The actual definition of convexity defect in terms of physical boundaries of the object is given below Convexity defect is a cavity in an object (blob, contour) segmented out from an image. That means an area that do not belong to the object but located inside of its outer boundary -convex hull. II. Deep Learning Implementation: (Contd) The input consists of 20,000 images . The first Module Image URL loader and reader is a simple algorithm which reads all the appropriate pathnames of the files and converts them into arrays using opencv functions .It extracts all the images required for training and stores the labels for each image in a separate list. One list all the images and the other list contains all the corresponding labels of each image. During the process the images are converted to grayscale and resized to smaller size so that the training phase of CNN model will be shorter. Then its fed into the CNN model. It splits its data into 70% as training and 30% as testing. (In the below table ,the gestures can be mapped to the appropriate text according to the usage, for example thumb down to what or L to how etc, say if the application is to be used in a deaf/dumb language or a custom gesture system required etc.) It takes 30 mins approximately for five iterations over the entire database of images for training as well as the validating(testing) to complete. After that model is ready to predict furthur images. (A sample tableis given if this application is to be used in a museum) Gesture Label Thumb down How many artifacts are here? palm(Horizontal) What is the name of the place L Where is it located Fist(Horizontal) When was it founded Fist(Vertical) What is the most famous artifact here.
  • 4. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3133 Tools used: -Python(Anaconda package) -Spyder(IDE) -Keras,tensorflow and opencv as python libraries B.Chatbot Template Response:- We put up ready-made templates as replies and those templates will have parameters to which will be replaced by the appropriate data relevent to the conversion Keywords:-The concept on the basis of which the chatbot will be implemented .The chatbot will scan the user input and see if a keyword exists in it. Theses keywords are stored in the knowledge base. And each of the keyword have a set of responses linked to it. So a response is chosen and given as reply Knowledge base:-The knowledge base is a 2-D array storing a set of keywords in the first column and its associated responses in the next columns. Keyword Ranking:-This is used when there are multiple keyword match and based on the furthur corresponding keyword match and repetitions, the appropriate keyword is selected then the corresponding reply is given The chatbot will be written in java based module which will accept the input as text which was converted appropriately from the gesture input. Screenshots: Deep learning implementation (Note:Predict prints a command line out put. The image is first captured and then saved As png file and then is fed into the deep Learning model) Convexity defect to count fingers to display numbers The output is planned to be written to Text file which will be fed into a java based chatbot Chatbot Output:(Separate module)
  • 5. International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056 Volume: 07 Issue: 02 | Feb 2020 www.irjet.net p-ISSN: 2395-0072 © 2020, IRJET | Impact Factor value: 7.34 | ISO 9001:2008 Certified Journal | Page 3134 Example sentences are given as input, it can be customised according to the table created by simple adding those responses to the knowledge base of the chatbot. References [1] Rupesh Prajapati, Vedant Pandey, Nupur Jamindar, Neeraj Yadav, Prof. Neelam Phadnis,”Hand Gesture Recognition and Voice Conversion for Deaf and Dumb”,IRJET,2018 [2] Anchal Sood and Anju Mishra, “AAWAAZ: communication system for deaf and dumb-”.IEEE,2016 [3]Ms R. Vinitha and Ms A. Theerthana,"Design And Development Of Hand Gesture Recognition System For Speech Impaired People.",2016 [4]Shangeetha, R. K., V. Valliammai, and S. Padmavathi,"Computer vision based approach for Indian Sign Language character recognition." Machine Vision and Image Processing (MVIP), IEEE,2012 [5] R.Aravind, D.Anbasaran, K.Alice,”GRS-Gesture Based Recognition System for Indian Sign Language Recognition System for Deaf and Dumb people”, International Journal of Trend in Scientific Research and Development - IJTSRD,2018 [6]Tripathi K. and Nandi, “Continuous indian sign language gesture recognition and sentence”,Procedia Computer,2015 [7]Daniel Simões Lopes “On the utility of 3D hand cursors to explore medical volume datasets with a touchless interface”, Journal of biomedical informatics,2017 Passive real-world interface props for neurosurgical visualization, Hinckley et al.CHI '94, [9]Otsu Algorithm-(Algorithm used for Thresholding, grayscale to binary),See OpenCv 3.4.2 documentation for algorithm implementation [10]Peter-Draukr algorithm(For calculation of contours) [11]Satoshi Suzuki and others. Topological structural analysis of digitized binary images by border following. Computer Vision, Graphics, and Image Processing, 30(1):32–46, 1985.(Algorithm used for finding contours) [12]CNN(For training the AI model)