New RM AAT
New RM AAT
Impact Analysis
Dr. Lakshmi Β N
Assistant Ρrοfessοr
Deρt. Cοmρuter Science and Εngineering
ΒΜS Institute οf Τechnοlοgy and Μanagement
Βangalοre, Κarnataka- 560094
Pranjal Tyagi
Raghvendra Sharma
Computer Science & Engineering
Computer Science & Engineering
BMS Institute Of Technology & Management
BMS Institute Of Technology & Management
Bangalore,India
Bangalore,India
text using MediaPipe and Streamlit. The integration of these population does not know sign language, they can not
models creates a unified platform that bridges communication communicate effectively.Also, traditional ways to interact with
gaps and promotes accessible learning through innovative to gather that information are not as simple.icant barriers in
technology.
communication, often relying on sign language as their primary
Keywords — AI, Gesture Recognition, Sign Language mode of interaction. However, the limited understanding of
Detection, MediaPipe, Streamlit, Educational Technology sign language among the general population restricts their
ability to communicate effectively.
This makes hand gestures a less than ideal vehicle for
I. INTRODUCTION mathematical content, and thus underscores the desire for
simpler, more accessible interfaces. In this context, the most
Revolutioning due to the rapid advance in artificial important issues are high accuracy for gesture and sign
intelligence (AI) and machine learning (ML), several fields language detection, real-time processing to have immediate
have come up with new solutions to improve human–computer feedback, and to make the best user interfacing. These
interaction. Assistive solutions for the disabled are one challenges present the need for making effective assistive
important application of these technologies. The potential of technologies that meet the hearing impaired users' needs and
AI- driven systems, particularly those in sign language solution seekers in interactive math support.
recognition and gesture based interactions in bridging the The main purpose of the project is to create two new advanced
communication gap for deaf and hard of hear community has AI models and combine them into a unified platform in order to
been demonstrated. In addition to allowing communication, improve communication and problem solving abilities. We start
these technologies also enable intuitive interfaces for with constructing a gesture based math problem solving model
educational and problem solving applications. that can accurately predict and solve mathematical problems by
Yet, though progress has been made, current solutions suffer using hand gestures.
from failure to achieve accuracy, real time performance, or A second, is we’re going to work on a sign language detection
ease of use. AI models that can interpret sign language and system where using MediaPipe and Streamlit, it will read sign
mathematical gestures held promise to settle these issues and language signs and make it readable in real time! By integrating
deliver better and more accessible assistive technologies. The these models together into a single platform, users can interact
purpose of this project was to create an integrated system built with mathematics while they communicate using sign language.
around gesture based math problem solving and sign language
detection to improve communication and learning for users.
The objective of this project is to make communication easier robust way to perform real time gesture recognition
for the deaf and hard of hearing community, and to do it which is useful for the project of sign language
through an interactive and an optimal way of learning and detection.
solving mathematical problems.
This project contributes in very important ways to the field of [2] Kumar, Thankachan, and Dominic (2023) aim to
AI driven assistive technologies. First, we develop two develop a system to ease communication for those
innovative AI models: It is then applied to two problems one with hearing and vocal disabilities. The system
for gesture based math problem solving, and another for sign recognizes static and dynamic gestures of American
language detection. Second, we make seamless integration Sign Language (ASL), specifically the letters a-z. It
between these models within the same platform possible, with uses skin color segmentation and Viola-Jones face
'ease of transition between functionalities. Thirdly, we ensure detection to focus on hands and face, eliminating
real time processing capabilities for both models so that we facial regions for static gestures. Zernike moments
can get user experience and immediate feedback. Fourth, we are used for static gesture recognition, and curve
create a user friendly interface that accommodates people at feature vectors for dynamic gestures, classified with
different levels of technical expertise, thus making the system SVM classifiers. The Sphinx module maps spoken
more accessible to a wider audience. Finally, we implement alphabet to text, and subsequently to a picture or
the communications and learning accessibility for the hearing video. The system performs in dynamic
impaired as well as for those who need interactive math backgrounds with minimal clutter, contributing to
support, using innovative technology to enhance educational real-time hand detection and gesture recognition.
and communicative accessibility.
[3] M. J, Krishna, S. N. S and S. K use machinelearning
II. RELATED WORK
to address communication barriers for deaf and
dumb individuals.. For acquiring image they utilize
Previous research has shown the potential of using AI in
OpenCV, for training they use CNN and the output
gesture recognition and sign language detection. Studies like
is the text. Unlike other studies of partial sign
[1] have demonstrated the effectiveness of convolutional
language identification, this study focuses on full
neural networks (CNNs) in recognizing hand gestures, while
recognition of American Sign Language (ASL)
[2] have explored the use of MediaPipe for real-time hand
using 26 letters and 10 numbers. Finger and hand
tracking. However, integrating these technologies into a
motions are extracted to differentiate static and
single platform that addresses both mathematical problem-
dynamic gestures. It also emphasizes the necessity
solving and sign language translation remains relatively
of accessible technology, and not depend on
unexplored. This paper builds on existing work by
external sensors.
combining these approaches to create a unified system that
[4] Which, in the paper, explores how this might be
enhances accessibility and learning experiences.
used in video games by using gestures captured
through a user’s video camera to performactions in
[1] Khubchandani and Karthick (2023) build a sign
the game. Hand gestures on Hill Climb Racing and
language recognition system using Convolutional
body gestures in Subway Surfers are used to
Neural Networks (CNNs) on the Microsoft Kinect
investigate. The practicability of this approach is
platform led to GPU acceleration.. They then
indicated through promising real time results using
preprocess the hand gesture data before passing
a webcam, and the approach is shown to be useful
video samples to the neural network to achieve
as a learning and rehabilitation tool using gesture
95% recognition accuracy. Most errors came from
based controls.
poor lighting an image quality. Suggested future
research directions to improve the system efficacy [4] Focuses on Human Computer Interaction and
and reduce errors were presented. This represents a Gesture Recognition through the use of vision
based, sound wave based and Radio Frequency them into text or speech.
(RF) methods with a focus on vision based.. The Integration: We integrate the Streamlit interface
paper proposes the use of deep neural network to with the sign language detection model to allow
improve the gesture recognition since this relates for a seamless user experience.
to interference from factors like palm size, skin
color and occlusion, although it is susceptible to
such interference. The study then uses MediaPipe
to detect 21 hand landmarks and get their
coordinates and trains RNN, CNN, hybrid (CNN
and RNN), and Transformer Encoder models.
With a 99.28% accuracy, the optimized RNN
model was achieved. Practical contactless
interaction application is envisaged by this
research.
III. METHODOLOGY
A. System Architecture
Convolutional Neural Network (CNN): They are to be able Python: Is the main language when we are
to recognise hand gestures and interpret mathematical implementing the algorithms and also integrating
model to classily differentiate between different hand Convolutional Neural Networks (CNNs): Used for
gestures. PatchNet for recognition is multiple layers, processing images, e.g. feature extraction from hand
convolutional layers, pooling and fully connected layers, to gestures, sign language and facial images.
extract features and predict on input images. Long Short-Term Memory (LSTM) Networks: Used
for sequence prediction, a critical task for reading
C. Hand Detection Landmark
continuous sign language gestures in time.
In order to improve hand detection and gesture recognition
Experimental Procedures:
accuracy as well as decrease false positives. It comes down
Gesture Recognition: Users make predefined
to processing hand landmarks detected by MediaPipe with
mathematical gestures in front of the webcam. We
advanced techniques. Using machine learning models
capture these gestures, these process them with the
trained on gesture data, the algorithm filters and refines
CNN and the LSTM models, and provide the
these landmarks to identify correctly gestures.
equation that gives the mathematical solution.
D. Experimental Setup
V. RESULTS
translated sign language into readable text, making it very We would also like to thank the CSE department of BMS
easy to communicate for the deaf and mute community. Institute of Technology for the facilities and support
Hand tracking module of MediaPipe captured robust hand provided to us to carry out this work.
landmarks which helped the system decode a variety of
REFERENCES
sign language gestures. Evaluation of the system included a
series of tests on users practicing different sign language
[1] H. Khubchandani and K. T, "Sign Language Recognition," 2023 9th
gestures. The system translated gestures into text without
International Conference on Information Technology Trends (ITT),
error, quite reliably. Real time interaction was enabled
Dubai, United Arab Emirates, 2023, pp. 199-203, doi:
through integration with a user friendly interface giving 10.1109/ITT59889.2023.10184252.
immediate visual feedback for users as they gestured. In
communication and educational contexts, particularly, [2] A. Kumar, K. Thankachan and M. M. Dominic, "Sign language
having such a feature was very suitable. Finally, we recognition," 2016 3rd International Conference on Recent Advances
in Information Technology (RAIT), Dhanbad, India, 2016, pp. 422-
ensured that the sign language detection system is a useful
428, doi: 10.1109/RAIT.2016.7507939.
tool for increasing communication accessibility and
inclusivity for the deaf and mute community.
[3] M. J, B. V. Krishna, S. N. S and S. K, "Sign Language Recognition
VI. CONCLUSION using Machine Learning," 2022 International Conference on
With the Gesture Learn system, previously unheard of promise Innovative Computing, Intelligent Communication and Smart
Electrical Systems (ICSES), Chennai, India, 2022, pp. 1-5, doi:
has been demonstrated in the capacity of math instruction and
10.1109/ICSES55317.2022.9914155.
communication for the deaf and mute community. By using AI
driven gesture recognition, sign language detection, the
[4] U. Patel, S. Rupani, V. Saini and X. Tan, "Gesture Recognition
Using MediaPipe for Online Realtime Gameplay," 2022
IEEE/WIC/ACM International Joint Conference on Web
Intelligence and Intelligent Agent Technology (WI-IAT), Niagara
Falls, ON, Canada, 2022, pp. 223-229, doi: 10.1109/WI-
IAT55865.2022.00039.