SignTone - Full
SignTone - Full
Communication methods such as sign language may not be universally understood, limiting
interaction with those who do not know sign language.
Deaf individuals may face challenges in accessing real-time communication tools, making it difficult to
engage spontaneously in conversations.
While technology has advanced, some deaf individuals may still lack access to affordable and effective
assistive devices or software.
Objectives
To develop and build the deaf companion System.
To enable two-way communication between deaf and normal people.
To design and develop a high-performance sign recognition module.
To generate high-quality speech from the text.
To transform the text from a normal individual into SL through an Avatar.
Abstract
Deaf individuals face significant difficulties in communicating with others in society, as only a small
number of them possess knowledge of and utilize sign language for communication.
In general, deaf individuals use sign language or text to interact or communicate with others. While
these methods are effective within the deaf community, they face significant limitations when trying to
communicate with the hearing community.
This can lead to isolation, frustration, and discrimination. The main contribution of this project is to
develop and build the deaf companion System (DCS) to enable two-way communication between non-
deaf and normal people in Indian Sign Language using Temporal Convolutional Network(TCN).
The proposed system has three modules; the sign recognition module (SRM) that recognizes the signs of
a deaf individual which were integrated into the sign translation with Multilingual Interpreter System,
the speech recognition using Hidden Markov Model and synthesis module (SRSM) that processes the
speech of a non-deaf individual and converts it to text, and an Avatar module (AM) to generate and
perform the corresponding sign of the non-deaf speech.
Introduction
Effective communication is a fundamental aspect of human interaction, enabling the exchange of
information, ideas, and emotions.
However, individuals with hearing and mute disabilities often face significant challenges in
expressing themselves and understanding others, leading to communication barriers.
Sign language has traditionally served as a crucial means of communication for the deaf community,
but its interpretation remains challenging for non-signers.
In response to these challenges, this project introduces the development of a Deaf Companion
System, leveraging advanced technologies such as Temporal Convolutional Networks (TCNs) to
enhance communication between deaf individuals and the wider community.
The system aims to bridge the gap by recognizing sign language gestures, converting spoken
language to text, and generating realistic sign language avatars.
Existing System
The traditional systems for communication and education for the deaf community have often
relied on established methods, but these methods may have limitations in addressing the
diverse needs of individuals with hearing and mute disabilities.
Assistive Devices
Hearing Aids and Cochlear Implants: Traditional assistive devices include hearing aids and
cochlear implants, which aim to improve auditory perception.
Manual Communication Tools
Pen and Paper: Traditional tools such as writing or using pen and paper are often employed for
basic communication.
Interpreters
Sign Language Interpreters: In various situations, sign language interpreters are employed to
bridge communication gaps between deaf and hearing individuals.
Existing Algorithms
Support Vector Machines (SVMs)
SVMs have been employed for classification tasks in SLR. They work well with high-dimensional
feature vectors extracted from sign language images.
Ensemble Learning
Ensemble methods, such as Random Forests and Gradient Boosting, can be applied to combine
multiple SLR models to improve overall accuracy and robustness.
Two-Way Communication
Implementation of a comprehensive two-way communication system, fostering seamless
interaction between deaf individuals and the broader community.
System Maintenance
SL Recognition
User Management
DCS Admin ML Interpretation
Avatar Generation
Input Sign
Input Sign
he User Authentication module establishes secure access, while the Dashboard serves as a
central hub for seamless navigation.
The Sign Language Recognition interface allows users to input gestures for real-time
interpretation, complemented by the Speech-to-Text interface for comprehensive
communication.
The Avatar Customization module enables personalization, and language, cultural, and
accessibility settings enhance adaptability.
The Admin Dashboard provides tools for training model, monitoring and system management.
This design ensures a holistic and accessible platform for individuals with hearing and mute
disabilities.
2. System User Dashboard
2.1. Admin
Login
Admin authentication for secure access to the system's administrative functions.
User Management
Admin-exclusive capabilities to manage user profiles, permissions, and system maintenance
tasks.
2.2. Deaf User
Non-Deaf Receive
View text representations and hear voice outputs corresponding to the predicted sign language
gesture.
3. Deaf Companion System
In the process of building and training the SignNet Model for the Deaf Companion System,
several essential steps are followed to ensure accuracy and effectiveness.
3.2. Preprocessing
This step involves a series of preprocessing tasks to enhance the dataset quality. It includes
standardizing image dimensions, converting images to grayscale for simplicity, applying a Gabor
Filter for noise reduction, and binarizing images to facilitate effective feature extraction.
3.3. Segmentation
Employing a Region Proposal Network (RPN), this step focuses on identifying and isolating
distinct regions within the images. Segmentation is crucial for improving the model's ability to
accurately recognize and interpret individual sign language gestures.
3.4. Feature Extraction
Implementing a Fully Connected Layer, this phase captures key features and nuances of the sign
gestures.
3.5. Classification
Utilizing a Pooling Layer, this step categorizes the extracted features, aiding in the identification
of specific sign language gestures based on the captured features.
Synchronization
Intricately synchronizes with the rhythm and content of non-deaf speech, creating a seamless
and harmonious connection between spoken words and visual representation.
System Requirements
Language : Python 3.7.4(64-bit) or (32-bit)
SN Design : HTML, CSS, Bootstrap
IDE : IDLE
Web Framework : Flask 1.1.1
Database : MySQL
Local Server : Wampserver 2i
OS : Windows 10 64 –bit
Packages : Tensor Flow, Pandas, Sickit Learn, Matplotlib, Mediapipe
Dataset
The Sign Language MNIST data came from greatly extending the small number (1704) of the
color images .
Each training and test case represents a label (0-25) as a one-to-one map for each alphabetic
letter A-Z (and no cases for 9=J or 25=Z )
The MNIST dataset for SL is collected from UCI Repository or Kaggle or github
References
1. F. Wen, Z. Zhang, T. He and C. Lee, "AI enabled sign language recognition and VR space bidirectional
communication using triboelectric smart glove", Nature Commun., vol. 12, no. 1, pp. 1-13, Sep. 2021
2. R. Gupta and A. Kumar, "Indian sign language recognition using wearable sensors and multi-label classification",
Comput. Electr. Eng., vol. 90, Mar. 2021.
3. A. Wadhawan and P. Kumar, "Deep learning-based sign language recognition system for static signs", Neural
Comput. Appl., vol. 32, no. 12, pp. 7957-7968, Jun. 2020.
4. R. Cui, H. Liu and C. Zhang, "A deep neural framework for continuous sign language recognition by iterative
training", IEEE Trans. Multimedia, vol. 21, no. 7, pp. 1880-1891, Jul. 2019.
5. G. A. Rao, K. Syamala, P. V. V. Kishore and A. S. C. S. Sastry, "Deep convolutional neural networks for sign
language recognition", Proc. Conf. Signal Process. Commun. Eng. Syst. (SPACES), pp. 194-197, Jan. 2018.
6. A. Sadeghzadeh and M. B. Islam, "Triplet loss-based convolutional neural network for static sign language
recognition", Proc. Innov. Intell. Syst. Appl. Conf. (ASYU), pp. 1-6, Sep. 2022.
7. C. Li, Y. Hou, P. Wang and W. Li, "Joint distance maps based action recognition with convolutional neural
networks", IEEE Signal Process. Lett., vol. 24, no. 5, pp. 624-628, May 2017.