Sign Language Recognition
Sign Language Recognition
RECOGNITION SYSTEM
BASED TEXT TO
ANIMATION CONVERSION
Team Members
Rimjhim
(2300520140053)
Manishka Singh
(2300520140032)
Payal Singh
(2300520140044)
Rashi Chaudhary Under the
(2300520140052) auspices of
Dr Upendra Kumar
Mr. Sandeep Yadav
INTRODUCTION
Sign language bridges communication gaps by
incorporating gestures, expressions, and lip movements,
adding meaning and context to conversations. It is vital
for nearly 6 million people in India who rely on Indian
Sign Language.
Our application leverages Natural Language Processing
(NLP) to convert spoken English into Indian Sign
Language animations. It acts as both a translator for
seamless communication and a learning tool for those
interested in ISL.
Special focus is given to capturing the unique phonetic
characteristics and linguistic nuances of Indian languages,
ensuring precise speech recognition and accurate
conversion into sign language.
MOTIVATION
Challenges Faced by the Hearing-Impaired
Individuals with hearing impairment face significant
communication barriers in daily life.
Lack of Accessible Sign Language Education
Only a small percentage of the deaf population in India
receive formal education in sign language, and sign language
training centers are often costly.
Prevalence of Hearing Impairment
Around 63 million people in India suffer from significant
auditory impairment.
Need for Better Communication Solutions
An estimated 5 million deaf and hard-of-hearing people in
India use sign language, highlighting the need for more
accessible communication tools.
RELATED WORK
METHODOLOGY
The methodology involves converting speech to text using JavaScript's
Speech Recognition library. The text is then tokenized, parsed, and
transformed into Indian Sign Language (ISL) syntax, following specific
grammatical rules. Stemming, lemmatizing, and removal of stop words
ensure accuracy in translation. Finally, the system scrapes an ISL
dictionary for sign language videos, which are merged into a cohesive
video sequence to visually represent the spoken sentence in ISL.
Merging Videos
Once the sign language videos are obtained, they are merged or concatenated into
a single video. This video is essentially a stream of videos that represents the
spoken sentence in ISL.
Data Flow Diagram
0 level DFD:
1 level DFD:
Use Case Diagram
Class Diagram
Sequence Diagram
Speech Recognition Sign Language
Service Database
User
Speech Input
Speech Input
Text Output
Gestures Output
Send to Animation
Animation Output
Animated Output
Animation
Rendering
Engine
Tech Stack Used:
Speech to text: IBM Watson speech-to-text
Text Preprocessing: Spacy
Text to ISL Glass mapping: OpenNMT(Python)
ISL Gesture Animation: Unity 3D
Optional Enhancement: Power BI for visual
analysis
Database: Indian Sign Language Research &
Training Center
CHALLENGES
Speech Recognition Accuracy: Achieving high accuracy
in converting spoken language to text can be difficult
due to accents, background noise, and speech
variations.
ISL Syntax Complexity: Translating text into ISL
involves complex grammar rules, which can be hard to
map accurately from spoken language.
Tokenization Issues: Handling ambiguous words and
context-dependent phrases during tokenization can lead
to errors in processing.
Video Synchronization: Merging multiple sign language
videos into a cohesive stream without losing context or
meaning is a technical challenge.
Future Potential and Innovation
The system's future potential includes improving speech recognition
accuracy and expanding sign language video databases. With AI-
driven personalized learning and multilingual support, it can enhance
communication, offering greater independence and inclusion for the
deaf and speech-impaired community.