Smart Glove for Sign Language Translation Using Arduino
Smart Glove for Sign Language Translation Using Arduino
Article history: Sign language is a vital mode of communication for deaf people, yet it presents a
Received 11 August 2023 significant barrier when interacting with those who do not understand it. The advent
Received in revised form 13 October 2023 of technology has paved the way for innovative solutions to bridge this communication
Accepted 29 October 2023 gap. This abstract explores the development and implications of a smart glove
Available online 7 December 2023
designed for sign language translation (SLT). The primary aim of this study is to create
a wearable device, the Smart Glove, capable of recognizing and translating sign
language gestures into text or speech. Key objectives include designing a lightweight
and ergonomic glove prototype, developing machine learning algorithms for sign
language recognition, implementing real-time translation capabilities, evaluating the
glove's accuracy and usability, and assessing the potential impact on facilitating
communication for deaf people. The Smart Glove utilizes only one sensor, flex sensors,
to capture hand movements and gestures. These data inputs are processed through a
custom-built machine learning model trained on a comprehensive sign language
dataset. Preliminary results indicate a high accuracy rate in recognizing sign language
gestures, with an average recognition rate of over 90% across a diverse set of signs.
While challenges such as expanding gesture recognition and refining translation
Keywords: algorithms remain, this technology offers a promising solution to break down
Sign Language Translation (SLT); Smart communication barriers and enhance the quality of life for those who rely on sign
Glove; Wearable Device language.
1. Introduction
Sign languages are widely used in deaf communities as a primary mode of communication [1].
They rely on hand forms, hand orientation, movement, and facial expressions to convey thoughts
and messages [2]. It is important to acknowledge that only a minority of the overall population
possesses proficiency in sign language [3]. Therefore, individual dependent on sign language may
face considerable difficulties engaging in conversations or effectively expressing their ideas. To bridge
this communication gap, several methods have been developed to aid individuals who are deaf or
*
Corresponding author.
E-mail address: [email protected]
https://doi.org/10.37934/aram.112.1.8087
80
Journal of Advanced Research in Applied Mechanics
Volume 112, Issue 1 (2023) 80-87
have communication impairments [4]. The methods include various types of hearing aids, such as
behind-the-ear, in-the-ear, and canal devices, as well as advanced technologies that convert sign
language gestures into written or spoken language [5]. While hearing aids provide valuable
assistance, they can occasionally cause discomfort or introduce background noise to the user's
auditory experience. Translating sign language gestures into text and speech for individuals
unfamiliar with sign language shows great potential as a technological advancement. Wearable
technology, such as a smart glove, can detect and translate the movements of the user's arms and
fingers into an easily understandable format [6].
The objective of this project is to create a smart glove device that can translate American Sign
Language (ASL) gestures into text and voice [7]. By utilizing flex sensors and advanced algorithms,
this technology has the potential to facilitate communication between sign language users and
speakers of different languages [8]. Preliminary studies have confirmed the efficacy of this smart
glove in translating sign language motions into text and audio, thereby improving communication
between sign language users and individuals who primarily use spoken language. Individuals who are
unable to speak often rely on body language and sign language as their main modes of
communication. Their limited knowledge of sign language hinders their communication with a wider
audience. The main objective of this project is to develop a wearable smart glove system capable of
accurately converting sign language gestures into comprehensible text and speech. This technology
aims to facilitate inclusive communication [9].
Our current project phase is centered on the development of a wearable smart glove. The glove
will utilize flex sensors to precisely capture the intricate hand and finger movements of the user [10].
Additionally, we will ensure that the glove is equipped with the necessary hardware components for
efficient sensor data capture and wireless connectivity with the companion mobile application
[11,12]. Our project focuses on capturing a diverse range of hand and finger movements to improve
the accuracy of algorithms. We aim to achieve high precision in recognizing a specific set of sign
language gestures, specifically those that correspond to the ASL alphabet, common words, and
phrases [13]. Our system will utilize effective translation techniques to facilitate communication
between sign language and spoken language. These processes facilitate the conversion of recognized
sign language motions into coherent text, enhancing comprehension for users unfamiliar with sign
language [14]. To enhance natural communication, we will integrate a robust speech synthesis engine
that can generate realistic spoken language using received gestures [15].
During the forthcoming testing phase, our smart glove solution will be thoroughly evaluated with
a diverse range of participants, including individuals proficient in sign language as well as those with
no prior knowledge of sign language [16]. Our evaluation will cover the performance of the smart
glove and mobile app, including the accuracy of gesture recognition and translation. The experiments
will provide valuable insights for optimizing the system's functionality. This project has the potential
to revolutionize communication for deaf communities and promote inclusivity and understanding
among users of different languages.
2. Methodology
2.1 Flowchart of The Project
During the planning phase, the flowchart of the project has been constructed to ensure it will run
smoothly. Figure 1 shows the flowchart for this project. The project will start with software
development when the research has been completed. The process of developing software also
involves GUIs and code. Because Arduino will be used in this project, C++ will be the coding language
81
Journal of Advanced Research in Applied Mechanics
Volume 112, Issue 1 (2023) 80-87
used [17]. The next stage, once software development is complete, is to test and run the code to
identify any coding and GUI errors. Before moving on to the next phase, any faults must be fixed first.
Hardware development is the next phase. Hardware development involves putting together the
Flex and Arduino sensors [6,9]. These two pieces of gear will assist in gathering data from tool
vibration. The software component must be linked to the hardware components. The system will be
tested once again to identify any faults that may have occurred after integrating the software and
hardware. Before doing the experiment, any mistakes must be corrected.
A flex sensor has been selected to collect the vibration signal from three separate axes in order
to gather the vibration signal from the tool [10]. One signal-detecting axis is all that the fundamental
flex sensor has. This project will have a difficult time since the sensor needs to be rotated in order to
collect data on several axes. Table 1 shows all the properties of the flex sensor in this paper.
Table 1
Flex sensor’s properties
Parameters Units
Straight Resistance 25K ohms
Resistance Tolerance ±30%
Bending Resistance Change 10K ohm
Rated Power 0.5 watt
Peak Power 1 watt
Bending Life >1 million times
Working Temperature -35 to 80 °C
Length x Width 73.666 x 6.35mm
82
Journal of Advanced Research in Applied Mechanics
Volume 112, Issue 1 (2023) 80-87
Creating a circuit diagram for five flex sensors demands a thorough knowledge of the sensor's
characteristics and the electrical principles governing their operation [10]. When developing a circuit
diagram for flex sensors, it is important to consider the following factors: First of all, sensor
specifications The sensor specifications, such as the voltage and current requirements, the operating
frequency, and the output signal, should be carefully evaluated before building the circuit schematic
[18]. It is crucial to ensure that the sensor's specifications are compatible with the rest of the circuit
components. Next, the manner in which the components are linked will be carefully planned to
ensure that the sensors are properly connected and that the circuit operates effectively [19]. Lastly,
signal conditioning Before it can be used by an amplifier or a microcontroller in the circuit, the output
signal of the flex sensors may need to be first checked. To filter or amplify the signal, the circuit may
need to be supplemented with a resistor or a capacitor [20].
Based on the above criteria, the flex sensor circuit design has been built as shown in Figure 2. The
components of the circuit, their connections, and the direction of current flow are all shown visually
in the circuit diagram. The sensor's features, such as its voltage and current needs, operating
frequency, and output signal compatibility with other circuit components, were carefully taken into
account when creating the circuit layout. This initial step made sure that the flex sensors were
seamlessly included in the circuit [21]. The precise design of the circuit topology, which specifies how
the components are linked, ensures that the sensors are positioned correctly and that the circuit
functions as a whole. Additionally, signal conditioning elements were added to the design to improve
signal quality, such as an integrated potentiometer coupled to the sensor [22]. In order to ensure the
security and efficiency of the circuit's operation, grounding was also stressed. Enough grounding
precautions were taken to guard against any electrical dangers. In order to provide adaptation for
different use cases, provisions were included to link the circuit's output to either an indicator or a
microcontroller, taking into account the intended application of the circuit. Rigid testing processes
were used to confirm the circuit's integrity and performance. The circuit's dependability and
conformity with the set parameters were validated throughout this thorough testing phase.
83
Journal of Advanced Research in Applied Mechanics
Volume 112, Issue 1 (2023) 80-87
3. Results
3.1 Data Measurement
To properly capture the complex motions of sign language gestures for our research, exact
measurement is essential. The smart glove's sensors are carefully chosen and calibrated to guarantee
that they produce reliable results [23]. An important part of the measuring process is sensor
calibration, which involves making changes to remove biases and guarantee that sensor outputs
correspond to the actual physical movements of the hand and fingers. In order to attain the highest
accuracy possible, calibration methods entail meticulously monitoring the reference motions and
fine-tuning the sensor settings [24,25].
Participants expressed letters, words, and sentences in American Sign Language (ASL) using a
wide variety of sign language movements. Flex sensors in the smart glove recorded complex hand
and finger motions in addition to non-manual cues like head and facial gestures, which are essential
for communicating grammatical complexity and context [26]. The resultant dataset is a rich supply of
raw data points that can be used to train and improve our gesture recognition system. It includes
sensor readings, timestamps, and other contextual data. The results of some of the American Sign
Language (ASL) letters are listed in Figure 3.
Fig. 3. The result of some of the American Sign Language (ASL) letters
Complex technological hurdles must be overcome in the creation of a smart glove for sign
language translation. The precision and accuracy of gesture recognition are two important
challenges. It is a challenging effort to ensure that the glove can accurately discern between a large
variety of sign language motions, even those with small changes [27]. Optimizing sensor location,
reducing latency, and developing a smooth user experience that is practical and comfortable for
regular usage are additional technical hurdles [28]. We collected a total of 10 samples of each letter
84
Journal of Advanced Research in Applied Mechanics
Volume 112, Issue 1 (2023) 80-87
to evaluate the system's accuracy. Examining each sign's reproducibility was the goal. The accuracy
graph in Figure 4 allows for evaluation of the results attained. Through the aforementioned
investigations, this device had a mean success rate of 94.23%. Specifically, when altering the hand
signals, sensor faults are blamed for the failures.
120
100
80
60
40
20
0
A B C D E F G H I J K L M N O P Q R S T U V W X Y Z
In conclusion, this study documents the creation of a glove-based sign language translator, a
technological breakthrough in accessibility. By harnessing a single Arduino UNO and five flex sensors,
this device adeptly captures the intricate movements of each finger and arm. The presented
experimental results are compelling, with a remarkable success rate of 94.23% in translating finger
movements into spoken and written English letters, signifying a significant leap in bridging
communication barriers for the Deaf community. Beyond its technical achievements, the project
aspires to be a transformative solution, poised to dramatically enhance the lives of Deaf individuals
while promoting inclusivity in communication. The project's ability to surmount intricate challenges,
respect the cultural and linguistic diversity inherent in sign languages, and enable Deaf individuals to
communicate naturally across diverse settings is pivotal to its ultimate success. In doing so, it holds
the potential to redefine the way Deaf individuals interact with the world and contribute to a more
inclusive society.
Acknowledgement
This research was funded by a grant from Ministry of Higher Education of Malaysia with project ID
RDU232707 and UIC231511.
References
[1] Cerruti, Massimo, and Riccardo Regis. "Partitive determiners in Piedmontese: A case of language variation and
change in a contact setting." Linguistics 58, no. 3 (2020): 651-677. https://doi.org/10.1515/ling-2020-0080
[2] Jayanath, T. "A Comparative Study on Bilingualism and Multilingualism." KnowEx Social Sciences 1, no. 02 (2021):
79-91. https://doi.org/10.17501/27059901.2020.1207
[3] Millán Jiménez, Ana, and Juan José García Escribano. "Entrepreneurship in people with disabilities. Cultural and
social aspects." Suma de Negocios 10, no. SPE22 (2019): 27-34. https://doi.org/10.14349/sumneg/2019.V10.N22.A4
[4] Zumalt, L. Eugene, Sallie Silver, and Lynne C. Kramer. "Evaluation of a communication device for deaf-blind persons."
Journal of Visual Impairment & Blindness 66, no. 1 (1972): 20-25. https://doi.org/10.1177/0145482X7206600105
[5] Quinteros Baumgart, Cibel, and Stephen Bates Billick. "Positive cognitive effects of bilingualism and multilingualism
on cerebral function: A review." Psychiatric Quarterly 89, no. 2 (2018): 273-283. https://doi.org/10.1007/s11126-
017-9532-9
85
Journal of Advanced Research in Applied Mechanics
Volume 112, Issue 1 (2023) 80-87
[6] Abougarair, A., and W. Arebi. "Smart glove for sign language translation." Int Rob Auto J 8, no. 3 (2022): 109-117.
https://doi.org/10.15406/iratj.2022.08.00253
[7] Caselli, Naomi, Corrine Occhino, Bruno Artacho, Andreas Savakis, and Matthew Dye. "Perceptual optimization of
language: evidence from American sign language." Cognition 224 (2022): 105040.
https://doi.org/10.1016/j.cognition.2022.105040
[8] Lakho, Rafique Ahmed, Zamir Ahmed Abro, Jun Chen, and Rui Min. "Smart Insole Based on Flexi Force and Flex
Sensor for Monitoring Different Body Postures." Sensors 22, no. 15 (2022): 5469.
https://doi.org/10.3390/s22155469
[9] Setiawan, Joga Dharma, Mochammad Ariyanto, M. Munadi, Muhammad Mutoha, Adam Glowacz, and Wahyu
Caesarendra. "Grasp posture control of wearable extra robotic fingers with flex sensors based on neural network."
Electronics 9, no. 6 (2020): 905. https://doi.org/10.3390/electronics9060905
[10] Zakri, Azriyenni Azhari, Amir Hamzah Arfianti Arfianti, M. Iqbal, Hamdy Madjid, and Naufal Fikri Aulia. "Designing
Flex Sensor Gloves with Temperature Sensor & Pulse Sensor to Help Stroke Patients."
[11] Neto, Antônio José Alves, José Aprígio Carneiro Neto, and Edward David Moreno. "The development of a low-cost
big data cluster using Apache Hadoop and Raspberry Pi. A complete guide." Computers and Electrical Engineering
104 (2022): 108403. https://doi.org/10.1016/j.compeleceng.2022.108403
[12] Urrea, Claudio, and John Kern. "Design and implementation of a wireless control system applied to a 3-DoF
redundant robot using Raspberry Pi interface and User Datagram Protocol." Computers and Electrical Engineering
95 (2021): 107424. https://doi.org/10.1016/j.compeleceng.2021.107424
[13] Gurbuz, Sevgi Z., Ali Cafer Gurbuz, Evie A. Malaia, Darrin J. Griffin, Chris S. Crawford, Mohammad Mahbubur
Rahman, Emre Kurtoglu, Ridvan Aksu, Trevor Macks, and Robiulhossain Mdrafi. "American sign language recognition
using rf sensing." IEEE Sensors Journal 21, no. 3 (2020): 3763-3775. https://doi.org/10.1109/JSEN.2020.3022376
[14] Rizwan, Shaheer Bin, Muhammad Saad Zahid Khan, and Muhammad Imran. "American sign language translation via
smart wearable glove technology." In 2019 International Symposium on Recent Advances in Electrical Engineering
(RAEE), vol. 4, pp. 1-6. IEEE, 2019. https://doi.org/10.1109/RAEE.2019.8886931
[15] Torroja, Yago, Alejandro López, Jorge Portilla, and Teresa Riesgo. "A serial port based debugging tool to improve
learning with arduino." In 2015 Conference on Design of Circuits and Integrated Systems (DCIS), pp. 1-4. IEEE, 2015.
https://doi.org/10.1109/DCIS.2015.7388612
[16] Verpoorten, Eve, Giulia Massaglia, Candido Fabrizio Pirri, and Marzia Quaglio. "Electrospun peo/pedot: Pss
nanofibers for wearable physiological flex sensors." Sensors 21, no. 12 (2021): 4110.
https://doi.org/10.3390/s21124110
[17] Berbesi, J. Martheyn, K. Saumeth, and F. Pinilla. "Parallel control firmware for CNC milling machine based in
Arduino." In 2017 12th International Microsystems, Packaging, Assembly and Circuits Technology Conference
(IMPACT), pp. 319-322. IEEE, 2017. https://doi.org/10.1109/IMPACT.2017.8255919
[18] Ibrahim, Haziqatulhanis, Norazlianie Sazali, Ahmad Syahiman Mohd Shah, Mohamad Shaiful Abdul Karim, Farhana
Aziz, and Wan Norharyati Wan Salleh. "A review on factors affecting heat transfer efficiency of nanofluids for
application in plate heat exchanger." Journal of Advanced Research in Fluid Mechanics and Thermal Sciences 60, no.
1 (2019): 144-154.
[19] Noh, Sun-Kuk, Kuk-Se Kim, and Yoo-Kang Ji. "Design of a room monitoring system for wireless sensor networks."
International Journal of Distributed Sensor Networks 9, no. 8 (2013): 189840. https://doi.org/10.1155/2013/189840
[20] Krestovnikov, Konstantin, Anton Saveliev, and Ekaterina Cherskikh. "Development of a circuit design for a capacitive
pressure sensor, applied in walking robot foot." In 2020 IEEE 20th Mediterranean Electrotechnical Conference
(MELECON), pp. 243-247. IEEE, 2020. https://doi.org/10.1109/MELECON48756.2020.9140509
[21] Bhanuteja, G., J. Sowmya, Veda Sandeep Nagaraja, and S. L. Pinjare. "Design and simulation of readout circuit for
mems sensor." In 2015 2nd International Symposium on Physics and Technology of Sensors (ISPTS), pp. 264-267.
IEEE, 2015. https://doi.org/10.1109/ISPTS.2015.7220126
[22] Yan, Mengting, Haoran Wei, and Marvin Onabajo. "Modeling of thermal coupling and temperature sensor circuit
design considerations for hardware Trojan detection." In 2018 IEEE 61st International Midwest Symposium on
Circuits and Systems (MWSCAS), pp. 857-860. IEEE, 2018. https://doi.org/10.1109/MWSCAS.2018.8623865
[23] Saggio, Giovanni, and Giancarlo Orengo. "Flex sensor characterization against shape and curvature changes."
Sensors and Actuators A: Physical 273 (2018): 221-231. https://doi.org/10.1016/j.sna.2018.02.035
[24] Ghani, J. A., M. Rizal, M. Z. Nuawi, M. J. Ghazali, and C. H. C. Haron. "Monitoring online cutting tool wear using low-
cost technique and user-friendly GUI." Wear 271, no. 9-10 (2011): 2619-2624.
https://doi.org/10.1016/j.wear.2011.01.038
[25] Sazali, Norazlianie. "A comprehensive review of carbon molecular sieve membranes for hydrogen production and
purification." The International Journal of Advanced Manufacturing Technology 107 (2020): 2465-2483.
https://doi.org/10.1007/s00170-020-05196-y
86
Journal of Advanced Research in Applied Mechanics
Volume 112, Issue 1 (2023) 80-87
[26] Saggio, Giovanni, Francesco Riillo, Laura Sbernini, and Lucia Rita Quitadamo. "Resistive flex sensors: a survey." Smart
Materials and Structures 25, no. 1 (2015): 013001. https://doi.org/10.1088/0964-1726/25/1/013001
[27] Kumar, Davinder, and Aman Ganesh. "A Critical Review on Hand Gesture Recognition using sEMG: Challenges,
Application, Process and Techniques." In Journal of Physics: Conference Series, vol. 2327, no. 1, p. 012075. IOP
Publishing, 2022. https://doi.org/10.1088/1742-6596/2327/1/012075
[28] Mujahid, Abdullah, Mazhar Javed Awan, Awais Yasin, Mazin Abed Mohammed, Robertas Damaševičius, Rytis
Maskeliūnas, and Karrar Hameed Abdulkareem. "Real-time hand gesture recognition based on deep learning
YOLOv3 model." Applied Sciences 11, no. 9 (2021): 4164. https://doi.org/10.3390/app11094164
87