Project Synopsis
Project Synopsis
ON
Real Time Conversion of American Sign Language to text
with Emotion using Machine Learning
Submitted by
Project Supervisor:
Dr. Rachna Jain
Department of Computer Science and Engineering
JSS Academy of Technical Education, Noida
October 2023
TABLE OF CONTENTS
REFERENCES
15
INTRODUCTION
American sign language is a predominant sign language Since the only disability
Deaf and Dumb (hereby referred to as D&M) people have is communication
related and since they cannot use spoken languages, the only way for them to
communicate is through sign language. Communication is the process of
exchange of thoughts and messages in various ways such as speech, signals,
behavior and visuals. D&M people make use of their hands to express different
gestures to express their ideas with other people. Gestures are the non-verbally
exchanged messages and these gestures are understood with vision. This
nonverbal communication of deaf and dumb people is called sign language. A
sign language is a language which uses gestures instead of sound to convey
meaning combining hand-shapes, orientation and movement of the hands, arms
or body, facial expressions and lip-patterns. Contrary to popular belief, sign
language is not international. These vary from region to region.
Sign language is a visual language and consists of 3 major components.
Minimizing the verbal exchange gap among D&M and non-D&M people turns
into a want to make certain effective conversation among all. Sign language
translation is among one of the most growing lines of research and it enables the
maximum natural manner of communication for those with hearing impairments.
A hand gesture recognition system offers an opportunity for deaf people to talk
with vocal humans without the need of an interpreter. The system is built for the
automated conversion of ASL into textual content and speech.
In our project we primarily focus on producing a model which can recognize
Fingerspelling based hand gestures in order to form a complete word by
combining each gesture. The gestures we aim to train are as given in the image
below.
In the recent years there has been tremendous research done on the hand gesture
recognition.
With the help of literature survey, we realized that the basic steps in hand gesture
recognition are: -
Data acquisition
Data pre-processing
Feature extraction
Gesture classification
The different approaches to acquire data about the hand gesture can be
done in the following ways:
● In [1] the approach for hand detection combines threshold-based colour detection
with background subtraction. We can use AdaBoost face detector to differentiate
between faces and hands as they both involve similar skin-color.
The goal is to recognize two classes of gestures: deictic and symbolic. The
image is filtered using a fast look–up indexing table. After filtering, skin
colour pixels are gathered into blobs. Blobs are statistical objects based on
the location (x, y) and the colorimetry (Y, U, V) of the skin color pixels in
order to determine homogeneous areas. In [2] Naïve Bayes Classifier is used
which is an effective and fast method for static hand gesture recognition. It
is based on classifying the different gestures according to geometric based
invariants which are obtained from image data after segmentation.
Thus, unlike many other recognition methods, this method is not dependent
on skin colour. The gestures are extracted from each frame of the video,
with a static background. The first step is to segment and label the objects
of interest and to extract geometric invariants from them. Next step is the
classification of gestures by using a K nearest neighbor algorithm aided
with distance weighting algorithm (KNNDW) to provide suitable data for a
locally weighted Naïve Bayes‟ classifier.
Since sign language is distinct from regular writing, a language barrier is created
to prevent communication between regular people and D&M people. They
therefore rely on communication based on visuals for interaction.
People who are not familiar with D&M may be able to understand the
movements if there is a common interface that converts sign language to text.
Therefore, efforts have been made to develop a visual interface that would enable
D&M people to communicate without truly understanding each other's
languages.
The goal is to create an intuitive human computer interface (HCI) where the
computer understands the human sign language.
There are numerous sign languages used around the world, including American
Sign Language (ASL), French Sign Language, British Sign Language (BSL),
Indian, and others.
SCOPE
Additionally, the project will be designed with scalability in mind, allowing for
future adaptations to accommodate emerging sign language variants and
advancements in technology. This forward-looking approach ensures that the
Sign-Language-To-Text-Conversion system remains relevant and effective in an
ever-evolving landscape.
Ultimately, this project is poised to revolutionize communication for the deaf and
hard of hearing community, setting a new standard for inclusive and accessible
language technologies that have the potential to benefit individuals worldwide.
LITERATURE SURVEY
TIMELINE CHART
CONCLUSION
In this report, a functional real time vision based American Sign Language
recognition for D&M people have been developed for asl alphabets.
We achieved final accuracy of 98.0% on our data set. We have improved our
prediction after implementing two layers of algorithms wherein we have verified
and predicted symbols which are more similar to each other.
This gives us the ability to detect almost all the symbols provided that they are
shown properly, there is no noise in the background and lighting is adequate.
Future Scope:
We are planning to achieve higher accuracy even in case of complex
backgrounds by trying out various background subtraction algorithms.
We are also thinking of improving the Pre Processing to predict gestures in low
light conditions with a higher accuracy.
This project can be enhanced by being built as a web/mobile application for the
users to conveniently access the project. Also, the existing project only works for
ASL; it can be extended to work for other native sign languages with the right
amount of data set and training. This project implements a finger spelling
translator; however, sign languages are also spoken in a contextual basis where
each gesture could represent an object, or verb. So, identifying this kind of a
contextual signing would require a higher degree of processing and natural
language processing (NLP).
REFERENCES
[1] T. Yang, Y. Xu, and “A., Hidden Markov Model for Gesture Recognition”,
CMU-RI-TR-94 10, Robotics Institute, Carnegie Mellon Univ., Pittsburgh, PA,
May 1994.
[2] Pujan Ziaie, Thomas M uller, Mary Ellen Foster, and Alois Knoll “A Na ̈ıve
Bayes Munich, Dept. of Informatics VI, Robotics and Embedded Systems,
Boltzmannstr. 3, DE-85748 Garching, Germany.
[3]https://docs.opencv.org/2.4/doc/tutorials/imgproc/
gausian_median_blur_bilateral_filter/gausian_median_blur_bilateral_filter.html
[5]aeshpande3.github.io/A-Beginner%27s-Guide-To-Understanding-
Convolutional-Neural Networks-Part-2/
[6] http://www-i6.informatik.rwth-aachen.de/~dreuw/database.php
[7] Pigou L., Dieleman S., Kindermans PJ., Schrauwen B. (2015) Sign Language
Recognition Using Convolutional Neural Networks. In: Agapito L., Bronstein
M., Rother C. (eds) Computer Vision - ECCV 2014 Workshops. ECCV 2014.
Lecture Notes in Computer Science, vol 8925. Springer, Cham
[8] Zaki, M.M., Shaheen, S.I.: Sign language recognition using a combination of
new vision-based features. Pattern Recognition Letters 32(4), 572–577 (2011).
[12] https://opencv.org/
[13] https://en.wikipedia.org/wiki/TensorFlow
[14] https://en.wikipedia.org/wiki/Convolutional_neural_nework
[15] http://hunspell.github.io/