0% found this document useful (0 votes)
10 views

Project Report to EDITED

Uploaded by

gupta1803yashi
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
10 views

Project Report to EDITED

Uploaded by

gupta1803yashi
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 17

Gesture-Controlled PowerPoint Tool

A PROJECT REPORT

Submitted by

Pratik Pawar 202101103063


Yashi Gupta 202101103060
Rutuja Tarte 202101103053
Sanskruti Pawar 202101103004

in partial fulfillment for the award of the degree

of

BACHELOR OF TECHNOLOGY

in

COMPUTER SCIENCE & ENGINEERING

Guided By

MR. SUJEET MORE

Department of Computer Science & Engineering

JAWAHARLAL NEHRU ENGINEERING COLLEGE


MGM UNIVERSITY CHHATRAPATI SAMBHAJINAGAR (M.S.),

YEAR 2024-25

APPENDIX 2

CERTIFICATE

This is to certify that the project report


“Gesture-Controlled PowerPoint Tool”
Submitted by
Pratik Pawar 202101103063
Yashi Gupta 202101103060
Rutuja Tarte 202101103053
Sanskruti Pawar 202101103004
is a bonafide work carried out by them under the supervision of <<Name of Guide>>
and it is approved for the partial fulfillment for award of Bachelor of Technology
(Computer Science & Engineering ), degree of MGM University, Chhtrapati
Sambhajinagar, MS, India
Date:

Mr. Sujeet More Dr. D. S. Deshpande


Guide Head of
Department
Dept. of Computer Sci. & Engineering Dept. of Computer Sci. &
Engineering

Dr. H. H. Shinde
Principal
Jawaharlal Nehru Engineering College
MGM University Chhatrapati Sambhajinagar (M.S.)
APPENDIX 3
CONTENTS

List of Abbreviations i
List of Figure ii
List of Table iii
Abstract iv

1. INTRODUCTION
1
1.1 Introduction 1
1.2 Necessity
2

2. LITERATURE SURVEY

3. PROBLEM DEFINITION AND SRS


3.1 Problem Definition

Problem Statement

Objectives
Major Inputs

Major Outputs
Major Constraints
Hardware Resources Required

Software Resources Required

Area of Project

3.2 Software Requirements Specification

3.2.1 Introduction

Purpose
Scope of the Project
Intended Audience & Reading Suggestions
3.2.2 Data Model and Description

Data Objects described in detail

3.2.3 Functional Model and Description

4. SYSTEM DESIGN AND IMPLEMENTATION


4.1 Proposed System Architecture
4.1.1. System Components

4.2 Design ER , DFD or suitable diagrams


4.3 Implementation Details
4.4 Timeline Chart
4.5 Cost Estimation

5 PERFORMANCE ANALYSIS
5.1 Different Modules and their working, Output Screens
5.2 Analysis
5.3 Testing

6 CONCLUSIONS
6.1 Conclusions
6.2 Future Scope
References
Acknowledgement

*******
List of Abbreviations

Abbreviation Full Form


AI Artificial Intelligence
GUI Graphical User Interface
HCI Human-Computer Interaction
JPG Joint Photographic Experts Group
ML Machine Learning
NLP Natural Language Processing
OpenCV Open Source Computer Vision Library
PRN Permanent Registration Number
SRS Software Requirements Specification

List of Figure
Sr No. Figure Name Page No.
Fig No. 3.2.1 Data Flow Pg no. 8
Fig No. 3.2.2 Mediapipe Hand Landmark Pg no. 8
Fig No. 4.2.1 Use Case Diagram Pg no. 10
Fig No. 4.2.2 System Architecture Pg no. 10
Fig No. 5.1.1 Pointer Gesture Pg no. 13
Fig No. 5.1.2 Eraser Gesture Pg no. 14
Fig No. 5.1.3 Next Slides Gesture Pg no. 14
Fig No. 5.1.4 Previous Slides Gesture Pg no. 14
Fig No. 5.1.5 Gesture Writing Pg no. 15
Fig No. 5.1.6 Recognition of Gestures Pg no. 15
Fig No. 5.1.7 Next Slide Demo Pg no. 16

List of Table

Table No. Table Name Page No.


Table no. 3.2.3.1 Functional Model and Description Pg no. 7
Table no. 4.3.1 Project Phases and Tasks Overview Pg no. 11
Table no. 4.4.1 Timeline Chart Pg no. 11
Table no. 4.5.1 Cost Estimation Pg no. 12

ABSTRACT

In today's digital world Presentation using a slideshow is an effective and attractive way that
helps speakers to convey information and convince the audience. There are ways to control
slides with devices like mouse, keyboard, or laser pointer, etc. The drawback is one should have
previous knowledge about the devices in order to manage them. Gesture recognition has acquired
importance a couple of years prior and are utilized to control applications like media players,
robot control, gaming. The hand gesture recognition system builds the use of gloves, markers
and so on However, the utilization of such gloves or markers expands the expense of the system.
In this proposed system, Artificial intelligence-based hand gesture detection methodology is
proposed. Users will be able to change the slides of the presentation in both forward and
backward directions by just doing hand gestures. Use of hand gestures cause connection simple,
helpful, and doesn't need any additional gadget. The suggested method is to help speakers for a
productive presentation with natural improved communication with the computer. Specifically,
the proposed system is more viable than utilizing a laser pointer since the hand is more apparent
and thus can better grab the attention of the audience
The project encompasses various stages, including data collection, preprocessing, model training,
and evaluation, culminating in the deployment of a functional sign language recognition system.
Through rigorous testing and evaluation, the system's accuracy, speed, and robustness to
variations in gestures and environmental factors are assessed to ensure its effectiveness and
reliability in real-world scenarios.

1. INTRODUCTION

1.1 INTRODUCTION
In today's fast-paced and increasingly digital world, effective communication is essential for
success in various domains, ranging from business presentations to educational seminars.
Traditional methods of delivering presentations often rely on static slides and conventional input
devices, limiting the presenter's ability to engage with the audience and convey information
dynamically. The emergence of gesture-based interaction technology has paved the way for a
paradigm shift in presentation delivery, offering presenters a more intuitive and immersive
means of controlling and interacting with their slides.

The "Gesture-Controlled PowerPoint Tool" represents a groundbreaking project at the forefront


of this technological revolution, aiming to transform the way presentations are delivered by
introducing gesture-based control for PowerPoint slides. By leveraging advanced computer
vision and machine learning techniques, the project seeks to empower presenters with the ability
to navigate through PowerPoint slides using natural hand gestures, thereby enhancing
interactivity, engagement, and overall presentation effectiveness.

A comprehensive literature survey forms the foundation of the project, exploring existing
research and technologies in the fields of gesture recognition, computer vision, and presentation
control. Insights gleaned from the literature survey inform the design and implementation of the
Gesture-Controlled PowerPoint Tool, ensuring alignment with state-of-the-art techniques and
best practices.

The implementation of the Gesture-Controlled PowerPoint Tool involves the utilization of


cutting-edge technologies and programming languages, such as Python, Flask and OpenCV. The
application is modularized into components for gesture detection, interpretation, and integration
with PowerPoint, with code snippets and GUI design screenshots offering insights into the
implementation details.

1.2 NECESSITY

Traditional methods of controlling presentations can be restrictive and cumbersome, often


requiring users to handle separate devices like laser pointers, keyboards, or remotes. These
methods may hinder fluid interactions, especially in high-stakes or interactive environments.

Moreover, individuals with physical challenges may find these tools less accessible, creating a
barrier to effective communication. There is a pressing need for a more inclusive and intuitive
system that accommodates diverse user requirements while maintaining simplicity and
efficiency.

The Gesture-Controlled PowerPoint Tool addresses these challenges by:

1) Introducing a device-free interaction mechanism for presentations.


2) Eliminating the learning curve associated with traditional tools.
3) Reducing reliance on additional hardware, ensuring cost-efficiency.
4) Enhancing accessibility for users with physical disabilities.
This innovation is not only practical but also aligns with the modern trend toward natural and
immersive user interfaces, making presentations more interactive and engaging.

2.LITERATURE SURVEY

In recent years, the intersection of human-computer interaction (HCI) and gesture recognition
has witnessed substantial progress, leading to diverse applications across various domains.
Lawrence and Ashleigh (2019) conducted a comprehensive study focusing on the impact of HCI
within the educational context of the University of Southampton. Their findings not only
highlighted the positive influence of HCI on literacy and efficacy but also underscored its
potential to transform educational environments [1]. Ren et al. (2013) made significant strides by
developing a robust hand gesture recognition system that harnessed the power of Kinect sensors.
Their system boasted impressive accuracy and speed, showcasing the feasibility of gesture
recognition technologies in practical applications [2]. Additionally, Dhall, Vashisth, and
Aggarwal (2020) delved into the realm of automated hand gesture recognition, leveraging deep
convolutional neural networks. Their research not only advanced theoretical understanding but
also provided valuable insights into deploying such systems in real-world scenarios, thereby
bridging the gap between theory and practice [3]. Meanwhile, Talele, Patil, and Barse (2019)
introduced an innovative real-time object detection approach using TensorFlow and OpenCV,
tailored specifically for mobile technology domains. This demonstrated the versatility and
adaptability of gesture recognition technologies in addressing contemporary technological
challenges [4]. Moreover, AlSaedi and Al Asadi (2020) proposed an economical hand gesture
recognition system, highlighting the potential for achieving remarkable recognition rates with
minimal hardware requirements. This research represents a significant step towards
democratizing access to gesture recognition technologies, making them more accessible and
practical for a wider range of applications [5]. Collectively, these selected studies not only
contribute substantially to the advancement of HCI and gesture recognition technologies but also
underscore their diverse applications across various domains, ranging from education to mobile
technology.
3. PROBLEM DEFINITION AND SRS

3.1 PROBLEM DEFINITION

Problem Statement
The current presentation tools rely on physical input devices such as keyboards, mice, or laser
pointers, which can hinder dynamic interaction between the presenter and the audience. These
methods are not inclusive for individuals with physical challenges and limit the natural flow of
communication. The Gesture-Controlled PowerPoint Tool aims to resolve these issues by
enabling hands-free slide navigation through intuitive hand gestures.

Objectives
• To develop a gesture-based tool for seamless navigation of PowerPoint slides.
• To enhance presentation interactivity and engagement using natural hand gestures.
• To eliminate reliance on traditional input devices, making the tool more inclusive.
• To integrate a web-based feature for slide-to-JPG conversion, ensuring compatibility.
• To maintain affordability and simplicity without requiring expensive hardware.

Major Inputs
1. Live video stream from a webcam for gesture recognition.
2. PowerPoint slides converted into JPG format for processing.
3. Predefined gestures recognized by the system (e.g., swipe left/right).

Major Outputs
1. Navigation commands (Next slide, Previous slide).
2. Dynamic feedback displayed during gesture recognition.
3. A seamless presentation flow controlled by gestures.

Major Constraints
1. Lighting Conditions: The system’s accuracy may vary under poor or inconsistent lighting.
2. Hardware Dependency: Requires a functioning webcam with sufficient resolution for
hand tracking.
3. Gesture Complexity: Limited to simple gestures to ensure system responsiveness.
4. Processing Speed: Real-time processing is necessary to minimize latency.

Hardware Resources Required


1. Personal Computer or Laptop with the following specifications:
1. Processor: Intel i3 or higher
2. RAM: 4 GB or higher
3. Camera: Standard webcam with at least 720p resolution
2. Internet connection for accessing the slide conversion web feature.

Software Resources Required


1. Python 3.x
2. Libraries:
o OpenCV
o MediaPipe
o NumPy
3. PowerPoint (for creating presentations).
4. Web-based slide-to-JPG converter.

Area of Project
• HCI: Leveraging natural hand gestures to interact with systems.
• Computer Vision: Using image processing techniques to detect and interpret gestures.
• Presentation Technology: Revolutionizing slide navigation tools to enhance user
experience.
3.2 Software Requirements Specification

3.2.1 Introduction
Purpose
The purpose of this document is to define the functional, performance, and user interface
requirements for the Gesture-Controlled PowerPoint Tool. This tool allows presenters to
navigate slides using hand gestures, eliminating the need for conventional input devices and
enhancing inclusivity and interactivity.

Scope of the Project


The Gesture-Controlled PowerPoint Tool utilizes computer vision to detect and interpret hand
gestures for seamless slide navigation. It incorporates a web-based utility for converting
PowerPoint slides to JPG format for compatibility with the gesture recognition system. This
project focuses on accessibility, enabling users to present dynamically without relying on
physical input tools.
Intended Audience & Reading Suggestions
• Audience: Presenters, educators, business professionals, and individuals with physical
disabilities.
• Suggestions: This document is structured to provide clear guidance for technical
implementation, user interaction, and system requirements. Stakeholders are advised to
review sections relevant to their roles (e.g., developers for functional descriptions, users
for interface details).

3.2.2 Data Model and Description


Data Objects
• Input Data:
o Live video feed from a webcam.
o JPG images of slides processed for gesture mapping.
• Processed Data:
o Detected hand landmarks and gestures.
o Navigation commands (e.g., next/previous slide).
• Output Data:
o Slide navigation actions executed in real-time.
o Visual feedback confirming gesture recognition.

3.2.3 Functional Model and Description

Sr No. Function ID Name of Function Description


1 F-1 Login Allows users to access and configure the
tool.
2 F-2 Gesture Handler Detects hand gestures for slide navigation
3 F-3 Slide Conversion Converts slides to JPG format for processing
4 F-4 Feedback Display Provides visual cues for recognized gestures

Table no. 3.2.3.1 Functional Model and Description Table.

Fig 3.2.1 Data flow

Fig 3.2.2 Mediapipe Hand Landmark


4. SYSTEM DESIGN AND IMPLEMENTATION

4.1 Proposed System Architecture

The Gesture-Controlled PowerPoint Tool is designed with a modular architecture to ensure


flexibility, scalability, and ease of integration with existing systems. The architecture comprises
the following components:

4.1.1 System Components

Input Module:
Captures video input from a webcam and sends it for processing.
Key technology: OpenCV for real-time image capture.

Processing Module:
Detects and interprets hand gestures using the MediaPipe framework.
Converts input gestures into meaningful navigation commands.

Slide Control Module:


Translates navigation commands into actions like moving to the next/previous slide in
PowerPoint.
Utilizes the slide-to-JPG conversion tool for compatibility with gesture recognition.

Feedback Module:
Provides real-time visual feedback on recognized gestures to the user.

4.2 Diagrams

Fig 4.2.1 Use-Case Diagram


Fig 4.2.2 System Architecture

4.3 Implementation Details


Programming Language: Python 3.x.
• Libraries Used:
o OpenCV: For image processing and video input handling.
o MediaPipe: For hand gesture detection and landmark recognition.
o NumPy: For numerical computations in gesture mapping.
• Integration:
o The tool integrates with PowerPoint slides converted into JPG format.
o Real-time gesture data triggers navigation commands via a Python script.
• Code Highlights: Includes gesture detection, recognition logic, and integration with
PowerPoint commands.

Phase Duration Tasks


Research and Planning Research and Planning Research and Planning

Design and Architecture Design and Architecture Design and Architecture

Testing and Evaluation Testing and Evaluation Testing and Evaluation

Table no. 4.3.1 Project Phases and Tasks Overview

4.4 Timeline Chart


Phase Duration Tasks
Research and Planning 2 weeks Literature survey, system requirements gathering.
Design and Architecture Designing system components and workflow
3 weeks
diagrams.
Implementation 6 weeks Developing modules for gesture recognition and
control.
Testing and Evaluation 3 weeks Testing for accuracy, responsiveness, and
stability.
Documentation and Finalization 2 weeks Preparing project report and user guide.

Table no. 4.4.1 Timeline Chart Table

4.5 Cost Estimation


Component Estimated Cost (INR)
Hardware (Webcam) 1,500
Software Tools (Free) 0 (OpenCV, MediaPipe, Python are open-source)
Development Resources N/A (student project)
Miscellaneous (Printing, Documentation, etc.) 500
Total Estimated Cost 2,000

Table no. 4.5.1 Cost Estimation Table.

5. PERFORMANCE ANALYSIS

5.1 Different Modules and Their Working, Output Screens

Modules and Functions:


1. Gesture Handler Module:
o Function: Detects and interprets gestures using the webcam.
o Output Screen: Visual feedback confirming recognized gestures (e.g., swipe
left/right).
2. Slide Conversion Module:
o Function: Converts PowerPoint slides into JPG format for compatibility.
o Output Screen: Confirmation of successful slide conversion.
3. Feedback Display Module:
o Function: Displays feedback for gestures, such as successful recognition.
o Output Screen: Real-time visualization of gestures and corresponding actions.

4. Output Screens:
Fig.5.1.1 Pointer Gesture

Fig 5.1.2 Eraser Gesture

Fig 5.1.3 Next Slide Gesture

Fig 5.1.4 Previous Slide Gesture


Outputs

Fig 5.1.5 Gesture Writing

Fig 5.1.6 Recognition Of Gesture

Fig 5.1.7 Next Slide Demo


5.2 Analysis

The system stands out for its impressive responsiveness, delivering real-time gesture recognition
with almost no noticeable delay. This makes it ideal for presentations, ensuring smooth and
natural interactions. With gesture-based controls, users can navigate their slides without the need
for a mouse or keyboard, keeping their focus on the audience and the content.

The accuracy of gesture recognition is a key feature of the tool. Testing shows a success rate of
85% to 95%, depending on factors like lighting and the clarity of hand movements. To improve
reliability, the system uses smart algorithms that adapt to slight variations in gestures, making it
work well in different environments and conditions.

Accessibility is one of the biggest strengths of this tool. It has been designed to include everyone,
especially those with physical disabilities. By removing the need for traditional input devices, the
tool ensures that people with limited mobility can still use it easily. This thoughtful design makes
the system accessible to a wider range of users.

Another advantage is its efficiency in using resources. The tool works perfectly on regular
hardware and doesn’t require any fancy or expensive add-ons. This makes it not only easy to set
up but also affordable for users, whether they are individuals or organizations. By leveraging
existing devices, it ensures flexibility and widespread usability.

Users have shared highly positive feedback about the tool. Many have praised its user-friendly
interface, which is straightforward and easy to understand. The system is quick to learn, even for
first-time users. Its interactive features and smooth performance have been especially appreciated
for making presentations more engaging and effortless.
Overall, the system combines advanced technology with a simple, practical design. It’s a reliable
and accessible tool that makes giving presentations easier and more enjoyable for everyone.

5.3 Testing
Testing of Gesture-Controlled PowerPoint Tool
Objective of Testing
The purpose of testing the Gesture-Controlled PowerPoint Tool is to evaluate its functionality,
usability, and performance to ensure that it meets the intended design goals and user
requirements. This section details the testing methodology, results, and conclusions.

Testing Methodology
Testing was conducted in two phases: Functional Testing and Usability Testing.

Functional Testing:
This phase tested whether the system functions correctly according to the specified requirements,
such as gesture recognition, PowerPoint control features, and system response time. Various
gestures (e.g., swipe left, swipe right, and gestures for transitions) were tested for accuracy and
responsiveness.

Usability Testing:
In this phase, the focus was on the ease of use and user experience. Participants (including team
members and external users) were asked to interact with the tool during a PowerPoint
presentation. Feedback was collected on aspects like comfort, intuitiveness, and any difficulties
experienced during the interaction.

Test Cases and Results


Test Case Description Expected Outcome Actual Outcome Status
Gesture Recognition (Swipe Left): Test swipe left gesture to move to previous slide Tool
should move to the previous slide Tool correctly moves to previous slide Pass
Gesture Recognition (Swipe Right): Test swipe right gesture to move to next slide
Tool should move to the next slide Tool correctly moves to next slide Pass
Gesture for Transition (Fist): Test fist gesture to start a transition Tool should start
the defined transition effect Tool correctly initiates transition effect Pass Accuracy of
Gesture Recognition Evaluate system’s accuracy in recognizing gestures under different
lighting conditions Tool should accurately recognize gestures in all environments. Gesture
recognition was accurate under controlled lighting, but recognition was slower in low light
Pass/Fail
Usability (User Feedback): Evaluate user satisfaction with the tool's ease of use Users
should feel the tool is intuitive and easy to use Majority of users reported ease of use and
intuitive design, though some had trouble with complex gestures Pass

Issues and Challenges


Gesture Misrecognition: In some cases, gestures were misinterpreted, especially in low-light
conditions or when gestures were performed too quickly.
Latency: There was a slight delay in response time when transitioning between slides. Further
optimization is needed for real-time interaction.
Learning Curve: While most users adapted quickly, a few needed assistance in performing
complex gestures, such as the fist gesture for transitions.

Conclusion:
The testing phase revealed that the Gesture-Controlled PowerPoint Tool works well in most
conditions and provides an innovative way of interacting with presentations. However, there are
areas for improvement, particularly in gesture recognition under varied lighting conditions and
response times. With further fine-tuning and user feedback, the tool has the potential to enhance
presentation delivery.
6.CONCLUSION

6.1 Conclusions
The Gesture-Controlled PowerPoint Tool makes presentations easier and more interactive by
letting users navigate slides without using a mouse or keyboard. This hands-free approach
adds a fun and engaging element, making it perfect for keeping the audience’s attention. It’s
also designed with accessibility in mind, making sure people with physical challenges can
use it comfortably, so everyone gets a fair chance to benefit from it.

We built this tool to be affordable by using open-source software and regular hardware, so
anyone can set it up without spending too much. It’s practical and fits right into today’s
needs, whether you’re using it in a classroom, office, or business meeting. By combining
simple tech with modern trends, the tool works well in a variety of settings and adds real
value for its users.

6.2 Future Scope


The Gesture-Controlled PowerPoint Tool has the potential to evolve into a completely
hardware-free, cloud-based system with the help of IoT technology. By using IoT-enabled
devices like smart projectors and cameras, we can remove the need for PCs or laptops
entirely. This would make the tool easy to set up and use, offering a simple plug-and-play
experience. Presenters could control their slides directly from these smart devices, making
the system more portable, compact, and user-friendly.

Adding cloud integration would take things even further. Users could store their
presentations online and access them from anywhere, eliminating the hassle of carrying files
around. This upgrade would make the tool perfect for classrooms, conference halls, and even
remote meetings. It would offer a flexible, convenient solution that adapts to modern
presentation needs while keeping everything smooth and interactive.

REFERENCES

1] D. Jadhav, Prof. L.M.R.J. Lobo, Hand Gesture Recognition System To Control Slide Show
Navigation, IJAIEM, Vol. 3, No. 4 (2014)
2] M. Harika, A. Setijadi P, H. Hindersah, Finger-Pointing Gesture Analysis for Slide
Presentation, Bong-Kee Sin Journal Of Korea Multimedia Society, Vol. 19, No. 8, August
(2016)
3] Ahmed Kadem Hamed AlSaedi, Abbas H. Hassin Al Asadi, A New Hand Gestures
Recognition System, Indonesian Journal of Electrical Engineering and Computer Science,
Vol 18, (2020)
4] D.O. Lawrence, and Dr. M.J. Ashleigh, Impact Of Human-Computer Interaction (HCI) on
Users in Higher Educational System: Southampton University As A Case Study, Vol.6, No 3,
pp. 1-12, September (2019)
5] I. Dhall, S. Vashisth, G. Aggarwal, Automated Hand Gesture Recognition using a Deep
Convolutional Neural Network, 10th International Conference on Cloud Computing, Data
Science & Engineering (Confluence), (2020)
ACKNOWLEDGEMENT

We extend our sincere gratitude to the Principal, Vice Principal, and the Head of the
Department of Computer Science at JNEC College, MGM University, for their unwavering
support and encouragement throughout the duration of our project, Gesture-Controlled
PowerPoint Tool. Their guidance and motivation played a pivotal role in fostering our
development and enabling us to accomplish this milestone.

A heartfelt thank you to our project mentor, whose advice and feedback helped us tackle
challenges and improve our work. We are also grateful to our faculty members, whose
knowledge and insights guided us throughout the process and contributed to the success of
this project.

Lastly, we want to thank our team members for their dedication and teamwork. This project
was a journey of learning, problem-solving, and creativity, and it wouldn’t have been
possible without everyone’s effort and enthusiasm.

You might also like