Capstone Project Proposal
Capstone Project Proposal
EasyCom
Submitted by:
101603276 Rishab
101603278 Rishabh Gupta
101610077 Sabhan Goyal
Assistant Professor
Jan-Dec 2019
2
TABLE OF CONTENTS
SIGNATURE of Mentor: . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
4
Project Overview
Communication is one of the fundamental step for the existence and survival of humans.
Moreover, it is mandatory for good and sustainable of an organisation. Everyone needs to
communicate in order to live a good and healthy life. But it is often very difficult for the
hearing-impaired community to communicate their ideas and creativity to the normal
humans. Our project “EasyCom” aims to aid these special group of people who have
difficulty in communicating in verbal form. There is several software to recognize
gestures but there is no dedicated device which can convert gestures to speech.
We aim to build a device which consists of a camera which will record the signs
performed by a differently abled person and then convert the signs into text and speech.
The text will be displayed on a screen attached to the device and an audio output will be
produced using a speaker. We will train a model on a dataset of a standard sign language
and the model will be installed in our device.
5
Need Analysis
Most people don’t understand sign language, therefore it is observed in social interactions
that impaired people face a lot of difficulty while communicating to normal people. This
can lead to miscommunication and conflict between these people.
This project aims to bridge this communication gap by providing a device which can
convert sign gestures to speech and text. This can be a huge step as it can help in
communication between speech impaired and normal people. The device will contain a
screen so that people can read the text on it and understand what the user of the device is
trying to say using gestures. It will also contain a speaker so that people can also hear the
audio translation of the gestures.
● More Generalist Approach: It focuses on the broad classes of all actions, not
specifically concentrating on the hand gestures.
● No Hardware Implementation: There is no hardware implementation of the above
mentioned project.
6
Literature Survey
A literature survey provides a quick summary/overview of recent research in a topic. In
our project we will talk about the existing research papers related to this.
Real Time Gesture Recognition for Easy Load Transport: Motion of the
body is interacting with other things for communication is called gesture recognition. For
more accurate results, we need to know the gesture information which is taken is possibly
same to provide to receiver. In this paper we are usually focus on control machine using
movement of the user body. Kinect is used to track the Body skeleton of a user or
operator. For easy to carry a load we used the system contains Kinect, Arduino & matlab.
We integrate the Kinect sensor and MATLAB image processing mechanism to process
video input which is captured by Kinect. The machine capable of carrying out a complex
series of action, and control device which is used to control the motor drivers which
implementing human skeleton tracking method. Real time moving position of an
Operator is captured using Kinect sensor having three cameras one is RGB, IR emitter,
IR depth measurement camera then captured body movements are providing for
processing and instruction of body movements are used to perform vehicle according to
the user actions. Proposed system will consist of process to provide a facility to user or
operator to ease their field work to drive a vehicle, with the help of Kinect sensor and
processing mechanism to follow the movement of body.
8
Assumptions
1. The user should know different gestures that can be recognised and successfully
decoded by the device.
2. The user must know about the device like battery life, range of the speaker,
position of the camera relative to the user and position of the screen.
3. The user must know optimal position of hands while performing the gestures for
best recognition by the camera.
4. The user cannot communicate with a person who is both deaf and blind using this
device.
Constraints
Standards
• IEEE
• Industry 4.0
Objectives
1. To build a Deep learning model which can accurately recognise the limited set of
gestures of a standard sign language.
2. To design and develop a portable hardware device which can run the above
mentioned model.
3. To install the model on the hardware device and then test the final product.
Methodology
1. Successfully build an efficient and accurate deep learning model which can
convert the gestures to text and speech.
2. To train this model we need an ample dataset to increase the efficiency and the
accuracy of the model. So, we need to record the dataset of the gestures. There are
a lot of gestures but we will only concentrate on a limited set of these gestures.
3. Assemble all the hardware components to create the final hardware device and
install the trained model on it.
4. Test the final product.
10
Work Plan
Rishab Garg
Rishabh Gupta
Ravi Singla
Sabhan Goyal
11
Course Subjects
1. Python
2. Keras
3. OpenCV
4. Numpy
5. Scipy
6. Computer Architecture
7. Electronics
8. Machine Learning
9. Deep Learning
References
https://arxiv.org/abs/1608.00859
https://www.talkinghands.co.in/
https://projects.raspberrypi.org/en/projects/raspberry-pi-setting-up
https://www.researchgate.net/publication/319406874_Text_-_To_-
_Speech_Synthesis_TTS