0% found this document useful (0 votes)
36 views

Design Project 2

Copyright
© © All Rights Reserved
Available Formats
Download as ODT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
36 views

Design Project 2

Copyright
© © All Rights Reserved
Available Formats
Download as ODT, PDF, TXT or read online on Scribd
You are on page 1/ 9

ABSTRACT

Communicating with a specially challenged person who can’t speak or hear is quite difficult,
especially when you don’t know sign language.
So to ease this problem, today you will learn to make a sign Language Translator device that
converts sign language into a spoken language. This device will be based on an ML model
that can recognize the different sign language gestures for accurate translation
INTRODUCTION

There are different sign languages practiced in different countries. For example, India uses
Indian Sign Language (ISL) while the USA uses American sign language (ASL). So, you first
need to decide which type you wish to implement. here are more than 70 million people who
are mute or hard of hearing. These people can communicate via postures, body movements,
eyes, eyebrows, and hand gestures thanks to sign languages. People with hearing and/or
speech impairments use sign language as their natural mode of communication. Signs are
gestures made with one or both hands accompanied by facial expressions with specific
meanings. Although deaf, hard-of-hearing, and mute people can easily communicate with
each other, integration in educational, social and work environments is a significant barrier
for the differently abled. There is a communication barrier between an unimpaired person
who is not aware of the sign language system and an impaired person who wishes to
communicate. The use of computers to identify and render sign language has progressed
significantly in recent years. The way the world works is rapidly changing and improving due
to technological?advancements. As programs over the last two decades have progressed,
barriers for the differently abled are dissipating. Researchers are working to build gear and
software that will help these people interact and learn using image processing, artificial
intelligence, and pattern matching approaches. The project's goal is to help people overcome
these obstacles by developing a vision-based technology for recognising and translating sign
language into text. We want to create a Raspberry Pi application for real-time motion gesture
recognition using webcam input in Python. This project combines real-time motion detection
and gesture recognition. The user has to perform a specific gesture. The webcam captures and
recognizes the gesture(from a set of known gestures) and displays the accurate representation.
LITERATURE SURVEY

“SVBiComm: Sign Voice Bidirectional Communication System based on Machine


Learning” Nada Gamal Mohammed; Rania Ahmed Abdel Azeem Abul Seoud, 2018
1st International Conference on Computer Applications & Information Security
(ICCAIS). The purpose of this study is to create a desktop human computer interface
application that will let persons who are normal, "deaf/dumb," or blind communicate
more easily. The SVBiComm system allows blind people to hear a voice stating the
word gestured by the "deaf/dumb," while the deaf receive a gesture representing the
blind's word. SVBiComm operates in two directions. The first is to convert video to
speech. The animated word movements are mapped into text using a language
knowledge base. The audio is then created using the Text-to-Speech (TTS) API.
Processing from speech to video is the second direction. The voice of the blind person
is translated into text using the Speech-to-Text (STT) API. Using a 3D graphical
model, the natural language is mapped from the database to "deaf/dumb" in a
corresponding sign language form. A total of 113 sentences with 244 signals were
used to test the system. In speech recognition, the system correctly identified words
from 19 different people with a 90% accuracy rate. For picture recognition, the
system correctly identified photographs for 21 different people with an 84 percent
accuracy rate. The SVBiComm system provides a variety of low-cost facilities that
can be utilised in a variety of applications.
 “Real Time Sign Language Recognition Using a Consumer Depth Camera” Alina
Kuznetsova,Laura Leal-Taixe, Bodo Rosenhahn, Proceedings of the IEEE
International Conference on Computer Vision (ICCV) Workshops, 2018. Recent
improvements in sensing technologies, such as time-of-flight and structured light
cameras make hand gesture detection practical.We offer a highly precise approach for
recognising static gestures from depth data provided by one of the sensors specified
above in this paper. Rotation, translation, and scale invariant characteristics are
derived from depth pictures. After that, a multi-layered random forest (MLRF) is
trained to categorise the feature vectors, resulting in hand sign recognition.
 “American Sign Language Alphabet Recognition using Convolutional Neural
Networks with Multiview Augmentation and Inference Fusion” Wenjin Tao, Ming C
Leu Engineering Applications of Artificial Intelligence, 2018. From depth images
recorded by Microsoft Kinect, this study offers a method for ASL alphabet
identification utilising Convolutional Neural Networks (CNN) with multiview
augmentation and inference fusion. Our method adds to the original data by producing
extra viewpoints, making the training more successful and reducing the risk of
overfitting. Our approach comprehends information from many viewpoints during the
inference stage for the final prediction to handle the perplexing circumstances
produced by orientational fluctuations and partial occlusions. Our technique
outperforms the state-of-the-art on two available benchmark datasets.
 “Vision based sign language recognition”Sadhana Bhimrao Bhagat, Dinesh Rojarkar
Journal of Emerging Technologies and Innovative Research (JETIR), 2020. This
study suggests an algorithm or approach for an application that will aid in the
recognition of Indian Sign Language's various signs. The approach has been designed
with a single user in mind. The real-time images will be captured first, and then they
will be saved in the directory. By using the SIFT (scale invariance Fourier transform)
algorithm, it will be possible to determine which sign has been articulated by the user.
The comparison will be done in reverse, and the result will be generated based on
matched key points from the input image to the image that has already been saved for
a certain letter. In Indian Sign Language, there are twenty-six signs that match to each
letter of the alphabet and the proposed algorithm delivers 95% accuracy.
 “Human-Computer Interaction with Hand Gesture Recognition using ResNet and
MobileNet” Abeer Alnuaim,Vikas Tripathi, Hussam Tarazi, Comput Intell Neurosci,
2022. A framework was developed that consists of two CNN models, each of which
is trained separately on the training data. To attain better outcomes, the final forecasts
of the two models were combined.The main contribution of this study is resizing the
images to 64 x 64 pixels, converting them from grayscale to three-channel images,
and then applying the median filter, which acts as lowpass filtering to smo
OBJECTIVE

 The primary purpose of gesture language translator is to create a system that


can recognise different human gestures and translate into text. Vision-based
hand gesture interfaces require real-time hand detection and gesture
recognition to do this.
 Promotes greater sensitivity and awareness of the deaf and hard of hearing
communities.
 The outcomes allow signers and non-signers to communicate successfully,
effortlessly, and instantly in their respective languages.
 The capacity of this technology to open up discussions between signers and
speakers in the marketplace, employment, schools, health care, and civic
institutions is what makes it intriguing.

COMPONENTS

HARDWARE:

 Raspberry Pi 3 Model B+
 Zebion Webcam 20 Mega Pixel
 SD Card 32 GB
 Desktop or Laptop
SOFTWARE:

 Python IDLE
 Packets:

a. CV2

b. Numpy

c. Mediapipe

d. Tensorflow
e. Tf.keras

DESCRIPTION

 Raspberry PI 3 Model B+

1. Windows and Android operating systems are supported by this computing device.
2. RPi is essentially a low-cost system programming and administration test machine.
3. The BCM2837B0 system-on-chip (SoC) features a 1.4 GHz quad-core ARMv8 64bit
processor and a powerful VideoCore IV GPU.
4. Snappy Ubuntu Core, Raspbian, Fedora, and Arch Linux, as well as Microsoft
Windows 10 IoT Core, are among the ARM GNU/Linux distributions that can be run
on the Raspberry Pi.

 ZEBION Webcam 20 MP

This full functionality 20 MP camera can deliver smooth and detailed high-quality

Specifications :

1. Image Sensor – CMOS


2. Interface – USB
3. Frame Rate – 30 FPS
BLOCK DIAGRAM
FUTURE SCOPE

 It removes the need for an interpreter.

 Application that is both mobile and web-based.

 Image processing should be enhanced so that the system can communicate in both
directions, i.e. transform conventional language to sign language and vice versa.

 Recognize motion-related indicators.

 Concentrate on turning the gesture sequence into speech.

 The system may be improved by allowing for multi-language display and speech
conversion.

You might also like