Synopsis
Synopsis
BACHELOR OF ENGINEERING
IN
COMPUTER SCIENCE WITH SPECIALIZATION IN
ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING
Submitted by:
Keywords:
Sign language detection, Sign language processing, Deaf-dumb hearing
impaired, intelligent systems
Table of Contents
Abstract i
1. Introduction
1.1 Problem Definition 1
1.2 Project Overview 2
1.3 Hardware Specification 3
1.4 Software Specification 3
2. Literature Survey 4
2.1 Existing System 6
2.2 Proposed System 7
2.3 Literature Review Summary 8
3. Problem Formulation 9
4. Research Objective 10
5. Methodologies 11
6. Conclusion 13
7. Tentative Chapter Plan for the proposed work 14
8. Reference 15
1. INTRODUCTION
1.1 Problem Definition
This project falls within the HCI (Human Computer Interface) sector and seeks
to recognise multiple alphabets (a-z), digits (0-9) and several typical ISL family
hand motions such as Thank you, Hello, and so on. Hand-gesture recognition is
a difficult problem, and ISL recognition is particularly difficult owing to the use
of both hands. Many studies have been done in the past employing sensors (such
as glove sensors) and various image processing techniques (such as edge
detection, Hough Transform, and so on), but they are quite costly, and many
people cannot afford them.
1
1.2 Problem Overview
Many people in India are deaf or hard of hearing, thus they communicate with
others using hand gestures. However, aside from a small group of people, not
everyone is familiar with sign language, and they may need an interpreter, which
may be complex and costly. The goal of this research is to build software that
can anticipate ISL alphanumeric hand movements in real time, bridging the
communication gap.
Sign language is largely used by the disabled, and there are few others who
understand it, such as relatives, activists, and teachers at SekolahLuarBiasa
(SLB). Natural gestures and formal cues are the two types of sign language[1].
The natural cue is a manual (hand-handed) expression agreed upon by the user
(conventionally), recognised to be limited in a particular group (esoteric), and a
substitute for words used by a deaf person (as opposed to body language). A
formal gesture is a cue that is established deliberately and has the same language
structure as the community's spoken language.
2
1.3 Hardware Specification
1.Interface: Jupyter notebook for inserting python libraries in a notebook format,
it is typically a python code where we can easily estimate our data sets model in
one single notebook.
3
2. LITERATURE SURVEY
"Sign Language Recognition Using Spatial-Temporal Graph Convolutional
Networks" by Wei Liu et al. [1] (2021) – This paper proposes a spatial-temporal
graph convolutional network (ST-GCN) for sign language recognition. The
method achieves state-of-the-art performance on benchmark datasets.
4
language recognition. The method achieves state-of-the-art performance on
benchmark datasets.
"Real-time Sign Language Recognition with Dynamic Hand Gesture and Facial
Expression Extraction" by Yurui Zhou et al. (2022) - This paper [8] proposes a
real-time sign language recognition system with dynamic hand gesture and facial
expression extraction. The proposed method achieves high accuracy and low
latency.
"A Robust and Efficient Sign Language Recognition System Using Graph
Convolutional Networks" by Rupam Das et al. (2022) - This paper [9] proposes
a robust and efficient sign language recognition system using graph
convolutional networks (GCNs). The method achieves state-of-the-art
performance on benchmark datasets.
5
2.1 Existing System
In existing system the module was developed for dumb person using flex sensor,
there user hand is attached with the flex sensors. On this module the flex sensor
reacts on bend of each finger individually. By taking that value controller starts
to react with speech, each flex sensor holds unique voice stored in APR Kit and
for each sign it will play unique voice. And in other existing system, the work is
done only for some alphabets and not for the words or sentences, and accuracy
obtained is very low.
In the proposed system the unable or dumb person should provide a gesture or
sign image to the system. The system evaluates the sign input with matlab image
processing technique and classifies the input to the recognized identification.
Later it initiates the voice media through the system when the input image
6
matches with the given dataset. And the output will be shown in the text format
too. This is a prototype to develop the concept of converting the sign language
to speech and text. The aim of this paper is to provide an application to the
society to establish the ease of communication between the deaf and mute people
by making use of image processing algorithm.
▪ When comparing with existing system user can give more signs
▪ The module provides two way communications which helps in easy
interaction between the normal people and disables
▪ Easy to Interface
▪ Flexible
7
2.3 Literature Review Summary
Year
Article/ Tools/ Evaluation
and Technique Source
Author Software Parameter
Citation
Real Time
Sign
Language
Recognition Python
Accuracy of
2022 System for Machine CNN http://surl.li/fhxbe
80%
Hearing and Learning
Speech
Impaired
People
Sign
Python
Language https://rb.gy/tqutsj Precision of
2019 Machine RNN
Detection “in 83%
Learning
the Wild”
ML Based
Sign Python
https://rb.gy/qejolc 65%
2021 Language Machine KNN
Accuracy.
Recognition Learning
System
Machine
Learning
Techniques Python
ML https://rb.gy/gakute
2017 for Indian Machine 90%
Algorithms
Sign Learning
Language
Recognition
8
3. PROBLEM FORMULATION
Conversing with people having a hearing disability is a major challenge.
Deaf and Mute people use hand gesture sign language to communicate,
hence normal people face problems in recognizing their language by signs
made. Hence there is a need for systems that recognize the different signs
and conveys the information to normal people.
The solution is to develop a translator that can detect sign language used by
a disabled person, and then feed that sign into a machine-learning algorithm
called transfer learning, which is then detected by the neural network and
translated on the screen so that a normal person can understand what the sign
is saying.
It's a lot easier now, thanks to speech to text and translators. But what about
individuals who are unable to speak or hear? The main goal of this project
is to create an application that can assist persons being unable to speak or
hear. The language barrier is also a very significant issue. Hand signals and
gestures are used by people who are unable to speak. Ordinary people have
trouble comprehending their own language. As a result, a system that
identifies various signals and gestures and relays information to ordinary
people is required. It connects persons who are physically handicapped with
others who are not.
Many firms are creating solutions for deaf and hard of hearing persons, but
not everyone can afford them. Some are very pricey for ordinary middle-
class individuals to bring.
9
4. OBJECTIVES
The proposed work is aimed to carry out work leading to the
development of an approach for SIGN LANGUAGE DETECTION
MODEL. More than 360 million of world population suffers from
hearing and speech impairments [3]. Sign language detection is a project
implementation for designing a model in which web camera is used for
capturing images of hand gestures which is done by open cv.
After capturing images, labelling of images are required and then pre
trained model SSD Mobile net v2 is used for sign recognition. Thus, an
effective path of communication can be developed between deaf and
normal audience. Three steps must be completed in real time to solve our
problem:
10
5. METHODOLOGY
12
6.CONCLUSION
Sign languages are kinds of visual languages that employ movements of
hands, body, and facial expression as a means of communication. Sign
languages are important for specially-abled people to have a means of
communication. Through it, they can communicate and express and share
their feelings with others. The drawback is that not everyone possesses the
knowledge of sign languages which limits communication. This limitation
can be overcome by the use of automated Sign Language Recognition
systems which will be able to easily translate the sign language gestures into
commonly spoken language. In this paper, it has been done by TensorFlow
object detection API. The system has been trained on the Indian Sign
Language alphabet dataset. The system detects sign language in real-time.
For data acquisition, images have been captured by a webcam using Python
and OpenCV which makes the cost cheaper. The developed system is
showing an average confidence rate of 85.45%. Though the system has
achieved a high average confidence rate, the dataset it has been trained on is
small in size and limited. In the future, the dataset can be enlarged so that the
system can recognize more gestures. The TensorFlow model that has been
used can be interchanged with another model as well. The system can be
implemented for different sign languages by changing the dataset.
13
7. TENTATIVE CHAPTER PLAN FOR THE PROPOSED
WORK
CHAPTER 1: INTRODUCTION
This chapter will cover the overview of the existing models and the need of better model
to detect sign language.
This chapter include the literature available for SIGN LANGUAGE DETECTION MODEL.
The findings of the researchers will be highlighted which will become basis of current
implementation.
CHAPTER 3: OBJECTIVE
This chapter will provide introduction to the concepts which are necessary to understand the
proposed system.
CHAPTER 4: METHODOLOGIES
This chapter will cover the technical details of the proposed approach.
This chapter will provide information about the subject system and tools used for evaluation
of proposed method.
14
REFERENCES
[1] Liu, W., Chen, L., Chen, J., & Zhang, J. (2021). Sign Language Recognition
Using Spatial-Temporal Graph Convolutional Networks. IEEE Transactions on
Neural Networks and Learning Systems, 32(2), 436-448.
[2] Kociolek, M., Kozakowski, K., & Kraft, M. (2021). Sign Language Recognition
Using Deep Convolutional Networks and Temporal Squeeze-and-Excitation. Applied
Sciences, 11(4), 1697.
[3] Zhang, L., Yin, L., & Zhang, Y. (2021). A Transformer-Based Method for
Continuous Sign Language Recognition. IEEE Signal Processing Letters, 28, 162-
166.
[4] Wang, Y., Yuan, Y., & Huang, H. (2021). Sign Language Recognition with Point
Clouds and a Graph Convolutional Network. Pattern Recognition Letters, 146, 10-17.
[5] Xu, H., Xu, X., & Xu, Y. (2021). Sign Language Recognition Using a Hybrid
CNN-LSTM Model with Channel Attention Mechanism. Sensors, 21(7), 2331.
[6] Dipietro, L., Pourhashemi, D., & Radke, R. J. (2021). Handshape Recognition in
American Sign Language Using Convolutional Neural Networks and a Novel
Balanced Dataset. IEEE Transactions on Human-Machine Systems, 51(6), 551-562.
[7] Wang, H., Zhang, Z., & Huang, X. (2021). Sign Language Recognition Using a
Spatio-Temporal Attention-Based Network. Neurocomputing, 460, 349-361.
[8] Zhou, Y., Liu, W., Xie, J., & Xu, Y. (2022). Real-time Sign Language Recognition
with Dynamic Hand Gesture and Facial Expression Extraction. Journal of Ambient
Intelligence and Humanized Computing.
15
[9] Das, R., Das, S., & Dey, N. (2022). A Robust and Efficient Sign Language
Recognition System Using Graph Convolutional Networks. IEEE Access, 10, 28090-
28105.
[10] http://timesofindia.indiatimes.com/home/education/news/realising-the-
importance-of-indian-sign-languagedictionary/articleshow/86221166.cms
16