Final Paper IEEE
Final Paper IEEE
Abstract— By employing the latest technologies such as the this study could lead to improvements in assistive devices for
ESP32 camera, TFT display, TTL serial converter, and people with hearing loss not just in Bangladesh but
speaker, this project seeks to create a device for real-time throughout the world. To encourage inclusion and
interpretation of Bangla Sign Language (BSL) movements. By accessibility on a larger scale, comparable devices could be
addressing communication barriers, the device aims to implemented in other areas where sign languages have been
improve accessibility and inclusivity within the deaf used by improving and expanding the technology established
community in Bangladesh for those with hearing impairments. in this research.
The device analyzes BSL movements collected by the camera
and sensor using machine learning techniques and image
processing algorithms, providing clear output over a speaker II. LITERATURE REVIEW
and display.
A. Previous work summary(in tabular form)
Keywords— Bangla Sign Language (BSL), ESP32 camera, Table 1. Literature Review
TTL serial converter, real-time interpretation, accessibility,
Title and Applications Methodology Performances
inclusivity, image processing, machine learning.
References
1. Deep Assistive This System The YOLOv7
I. INTRODUCTION Learning- technology is trained on model
Sign language is a form of interaction method that is based Bangla for the achieved
usually associated with individuals who are deaf or have Sign individuals Okkhornama [email protected]
other communication challenges. Sign language enables Language with hearing category with accuracies
people to communicate with others more easily and more Detection disabilities. 49 BdSL ranging from
efficiently, reducing the need for assistance devices such as with an categories. 85% to 97%
hearing aids or cochlear implants.[1] It also allows them to EdgeDevice and
participate more fully in interactions with others, such as Siddique et [email protected]
discussions, social events, and performances and
al. (2023) [7] from 41% to
presentations[1]. Bangla Sign Language (BdSL) is critical
53%.
for encouraging communication and promoting social
integration among Bangladesh's deaf people[2]. BdSL 2. Bangla YOLOv5 for 24.2GB The prototype
recognition and detection is a difficult problem in computer Sign BdSL BdSL video implementati
vision and deep learning research because sign language Language recognition from SignBD on of the
recognition accuracy changes with skin tone, hand position, Recognition aids converted to project
and background[3]. Minor changes in hand placement, facial using communicati images using achieves an
expressions, and movement are common in BdSL signs, and YOLOv5 on in FFmpeg. average
they can drastically change the meaning. Those who are not M. Karim et Bangladesh/I accuracy of
familiar with the language may easily overlook the al. (2023) [8] ndia. 91.62%.
complexity, which could result in misunderstandings. 3. Sign A The model is The
Resources for BdSL are frequently more limited than those Language TensorFlow- trained to developed
for spoken languages, which makes it difficult for people to Recognition powered sign recognize system is
learn and use the language outside of specialized classes or System using language different showing an
contacts with the deaf community[4]. Large datasets of TensorFlow interpreter gestures of average
BdSL movements can be used to train the system, giving it Object translates sign language confidence
the capacity to detect small variations and complexities that Detection hand gestures and translate rate of
are essential for accurate interpretation[5]. Researchers in API captured in them into 85.45%
artificial intelligence and natural language processing are at
Mishra et al. real-time into English.
present attempting to create a system that can automatically
recognize sign language. ASL, BSL, and JSL belong to the (2022) [9] English text.
sign languages whose automatic recognition has improved to
a high degree of accuracy (more than 99%)[6]. In the future,
4. Deep Recognition Deep learning- Highest allows real-time engagement via mobile networks[11]. In
learning- of Indian based training addition, in order to improve accessibility and inclusivity in
based sign Sign convolutional accuracy sign language interaction, GestureTek, a company that
language Language neural networks of 99.72% focuses on gesture recognition technology, has developed
recognition (ISL) static (CNN) using and interactive systems that employ motion-tracking algorithms
system for signs, robust modeling 99.90% on and depth-sensing cameras to translate and interpret sign
static signs including of static signs colored language gestures into spoken language or text[12].
digits, and
Kumar et alphabets, grayscale C. Survey Analysis
al. (2020) and words images. In order to understand the needs of the user base for this
[13] used in day- project, an initial survey was administered to people in the
to-day life. age range of 5 to 55 and older. The surveyed demographic
5. A CNN Real-time, Convolutional Average was primarily made up of friends and family members of
sign web-camera- Neural Network recognitio people who are Deaf or hard of hearing. When asked how
language based (CNN) based n accuracy they wanted to benefit from the initiative, they indicated that
recognition British Sign system of 89%. they wanted to improve educational opportunities and
system with Language learning as well as communication accessibility with an
single & recognition emphasis on understanding. When responding to questions
double- system about what aspects of the project they would like to see
handed improved, participants emphasized how critical it is to
gestures achieve high accuracy and speed of interpretation in
addition to being able to identify and accommodate different
N. Buckley signers and dialects. Moreover, there was a clear preference
et al. (2021) for the use of intelligent robotic elements in the proposed
[14] system.
D. Low-Cost Device
6. Machine The Hand Gesture Works
application recognition well in Table 2. Detail Cost
learning
based sign of the using HOG black Component Price
language research is (Histogram of backgroun Arduino Uno R3 ৳789
recognition: to convert Oriented d and ESP-32 Cam Module ৳ 848
a review sign Gradients)SVM white
language to (Support Vector foreground
1.8-inch TFT Display ৳480
and its
research speech. Machine) as a HC-05 Bluetooth Module ৳315
frontier. classifier, PL2303 USB to TTL ৳98
Conversion of Serial Converter
sign language to
Hasan et al. Mini Speaker ৳30
(2016) audio using
[15] TTS (Text to Breadboard ৳88
Speech) Jumper Wires ৳112
converter Total ৳2760