Madhuri 2013
Madhuri 2013
[email protected],[email protected],[email protected]
Abstract_ This report presents a mobile VISION-BASED application program for automatic translation of Indian sign
SIGN LANGUAGE TRANSLATION DEVICE for automatic language into speech in English to assist the hearing and/or
translation of Indian sign language into speech in English to speech impaired people to communicate with hearing people.
assist the hearing and/or speech impaired people to
This sign language translator should be able to translate
communicate with hearing people. It could be used as a
translator for people that do not understand sign language, alphabets (A-Z) and numbers (0-9).
avoiding by this way the intervention of an intermediate Although facial expressions add important information to
person and allow communication using their natural way of the emotional aspect of the sign; but in this project work they
speaking. The proposed system is an interactive application are excluded from the area of interest, since its analysis
program developed using LABVIEW software and complicates the already difficult problem. Our system aims at
incorporated into a mobile phone. The sign language gesture listening to deaf, which means that it could be used as a
images are acquired using the inbuilt camera of the mobile
translator between deaf and people that do not understand sign
phone; vision analysis functions are performed in the
operating system and provide speech output through the language, avoiding by this way the intervention of an
inbuilt audio device thereby minimizing hardware intermediate person. Both deaf and people that do not have
requirements and expense. The experienced lag time hearing problems would communicate using their natural way
between the sign language and the translation is little of speaking.
because of parallel processing. This allows for almost
instantaneous recognition from finger and hand movements
to translation. This is able to recognize one handed sign II.METHODOLOGY
representations of alphabets (A-Z) and numbers (0-9).The
results are found to be highly consistent, reproducible, with We present a real-time vision-based system for
fairly high precision and accuracy. recognizing finger spelling continuous Sign Language (ASL)
Keywords: using a single camera to track the user's unadorned hands.
Image processing, LabVIEW, pattern matching, sign This system is broken down into three main parts starting with
language.
the image acquisition followed by image processing to extract
features for recognition and last comes the recognition stage
where signs are identified and audio output is given. This
I. INTRODUCTION section explains the details of each process. The program
starts with image acquisition, i.e. sign images capturing by the
Sign Language (SL) is the natural way of communication camera. The acquired images are pre-processed to differentiate
of hearing and/or speech -impaired people. A sign is a static and dynamic signs, and also the start and end of a sign.
movement of one or both hands, accompanied with facial The images are processed to identify the region of interest.
expression, which corresponds to a specific meaning. The unique features of each sign in the region of interest are
Although the deaf, hard of hearing and hearing signers can extracted to be used in the recognition stage. In the recognition
communicate without problems amongst themselves, there is a stage, the features extracted are compared with the available
serious challenge for the deaf community trying to integrate database of pattern matching templates. A threshold value is
into educational, social and work environments. The overall set for the maximum difference between the input sign and the
goal of this project is to develop a new vision based database, if the difference is below the maximum limit, a
technology for recognizing and translating continuous sign match is found and the sign is recognized. Corresponding
language to text. To paper presents a mobile interactive audio file is played on audio device. The program can be
implemented in a laptop, desktop or an IOS mobile phone to
operate with its inbuilt camera, processor and audio device.
The experiments use a 36 word lexicon.
4. CONCLUSION