Project - 01
Project - 01
DARBHANGA
PROJECT:- 01
PROJECT - 01
ABOUT 1
.
PROJECT TITLE
US SIGN LANGUAGE
DETECTOR
TEAM MEMBER
2
.
SUB-TITLE
1.SONU KUMAR JHA - REAL-TIME SIGN LANGUAGE
RECOGNITION WITH MEDIAPIPE AND
21105111025 DEEP LEARNING
2.VIDYAPATI KUMAR -
21105111035 3
.
TAGLINE
3.CHHAVINATH KR. CHY- "BRIDGING COMMUNICATION GAPS
21105111025 THROUGH AI-POWERED SIGN
4.SAURAV KUMAR- LANGUAGE RECOGNITION."
21105111038
INTRODUCTION
This project focuses on developing an AI-powered model that
interprets sign language and translates it into text .
Leveraging Recurrent Neural Networks (RNNs) and Long
Short-Term Memory (LSTM) architectures, the model captures
and processes sequential data from sign language gestures.
By analyzing motion patterns, hand positions, and contextual
cues, it ensures accurate recognition and real-time
translation. The system bridges communication gaps for the
hearing-impaired community, enabling seamless interaction
with non-signing individuals. This innovative approach
combines deep learning and natural language processing,
promising accessibility and inclusivity through cutting-edge
technology.
WHAT IS SIGN LANGUAGE?
• A VISUAL LANGUAGE USED BY THE DEAF AND HARD-
OF-HEARING COMMUNITY.
• Importance of accurate and real-time recognition.
PROBLEM STATEMENT:
• COMMUNICATION BARRIERS BETWEEN SIGN
LANGUAGE USERS AND NON-USERS.
• Need for an automated system to translate sign
language into actionable outputs.
SOLUTION
• A REAL-TIME SIGN LANGUAGE DETECTOR USING
DEEP LEARNING (LSTM) AND COMPUTER VISION
(MEDIAPIPE).
METHODOLOG SOFTWARE/TOOLS
YPipeline
Input:
Overview:
Preprocessing and Feature Extraction:
Real-time video feed from a webcam. OpenCV,
MediaPipe
Keypoint Extraction:
Mediapipe Holistic model extracts pose, Deep Learning Architectures :
face, and hand landmarks. RNN ,
LSTM
Data Preprocessing:
Keypoints are flattened and normalized
Data Handling :
for LSTM input.
TensorFlow , NUMPY
Model Training:
LSTM neural network trained on labeled
sign language data. Platform for execution :
Jupyter
Output: Notebook
Predicted sign language gesture
displayed in real-time.
FLOWCHART
IMPLEMENTATION
AND RESULTS
• Data Collection:
⚬ Collected 30 sequences per gesture (e.g.,
"hello," "thanks," "I love you").
⚬ Each sequence consists of 30 frames of
keypoint data.
• Model Architecture:
⚬ 3 LSTM layers with 64, 128, and 64 units
respectively.
⚬ Dense layers for classification.
⚬ Activation: ReLU for hidden layers,
Softmax for output.
• Training:
⚬ 2000 epochs with Adam optimizer.
⚬ Achieved 100% accuracy on test data.
• Real-Time Testing:
⚬ Successfully recognized gestures in real-
time with a confidence threshold of 0.5.
RESULT
THAN
YOU!
K
PROJECT GROUP DETAILS