Advanced Blind Helper Android Application Using Text-To-speech Synthesis
Advanced Blind Helper Android Application Using Text-To-speech Synthesis
https://doi.org/10.22214/ijraset.2023.49861
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue III Mar 2023- Available at www.ijraset.com
Abstract: This project aims to develop an Android application that enables blind people to detect currency denominations and
make phone calls using handwritten text recognition and voice commands. The application uses image processing techniques
and a trained TensorFlow model to recognize the denomination of banknotes, and an ML model to recognize handwritten text
and convert it to a phone number. The user can initiate a phone call by saying a voice command, and the application provides
audio feedback throughout the process.
Keywords: Money detection, Calling, text-to speech
I. INTRODUCTION
Blind people face several challenges in their daily lives, such as recognizing currency denominations and making phone calls. These
tasks often require visual cues or touchscreen interactions, which are not accessible to blind people. To address these challenges, we
propose an Android application that uses machine learning and image processing techniques to detect currency denominations and
make phone calls using handwritten text recognition and voice commands.
The currency detection module uses a trained TensorFlow model to recognize the different features and patterns of banknotes, such
as edges, corners, colors, and symbols. The user can capture an image of a banknote using the smartphone camera, and the
application provides audio feedback on the denomination of the currency.
The calling activity module uses an ML model to recognize the user's handwritten text and convert it to a phone number. The user
can initiate a phone call by saying a voice command, such as "call" or "dial," and the application provides audio feedback
throughout the process.
The application aims to provide a user-friendly and accessible interface for blind people to detect currency denominations and make
phone calls using handwritten text recognition and voice commands. It can help improve the independence and quality of life of
blind people, enabling them to perform tasks that were previously inaccessible or challenging.
III. METHODOLOGY
A. Methodology
a. First I have added the required dependencies that allows us to include external library or local jar files or other library modules
in our Android project. Then in the xml I have designed the user interface of the application.
In MainActivity java I have created all the methods that will help the user to open certain tasks by simple voice command.
We have also implemented a swiping touch event as given in [4] so that we have left and right swipe.
By left swiping on the screen the user will read the feature or operations of the app.
By right swiping on the screen voice input will start. After the user gives the voice command it will automatically be redirected to
that particular activity. Let's say If user say “ I want to detect the currency” then it will automatically open currency activity. So that
user will just tap on the screen and take the picture and know the value of currency.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1951
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue III Mar 2023- Available at www.ijraset.com
b. b) Methods Used: i. Text to Speech (TTS):- TTS is a method that converts speech from text. TTS is important for voice output
for voice feedback for users. TTS is implemented in software where audio capability is required. When a user enters a voice
command, TTS will convert that voice into text format and perform specific action. ii.[3] Speech to Text(STT):- Android has a
inbuilt feature that is speech-to-text through which the user can provide speech input to the software. In the background speech
input will be converted to text and perform action in the form of TTS.
B. System Architecture
The system proposes the following applications: w
C. Project Requirements
The requirements were arranged in three groups: user interface, functional requirements.
1) User interface
a) Easily accessible
b) Flexibility of voice control (Set speed, pause speech)
2) Functional Requirements
Switching among the different voices
Text to speech
Voice assistant
(x) Exit - close the app.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1952
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue III Mar 2023- Available at www.ijraset.com
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1953
International Journal for Research in Applied Science & Engineering Technology (IJRASET)
ISSN: 2321-9653; IC Value: 45.98; SJ Impact Factor: 7.538
Volume 11 Issue III Mar 2023- Available at www.ijraset.com
VI. CONCLUSION
At present, mobile apps on smartphones are used to perform most of our daily activities. But the people with vision impairment
require assistance to access these mobile apps through handheld devices like mobile and tablets. Google, Android applications has
been developing various mobile apps for visually impaired people Still it needs to provide more effective facilities in app by
adopting and synergizing suitable techniques from Artificial Intelligence.[5] This report introduced two environmentally-friendly
designs for a blind people. We presented information about the Blind people application. This application will be more effective
for blind people. It is important to develop this application for the future. The system is used by Blind people but normal people
also can use it.
REFERENCES
[1] H. Nguyen, M. Nguyen, Q. Nguyen, S. Yang and H. Le, "Web-based object detection and sound feedback system for visually impaired people," 2020
International Conference on Multimedia Analysis and Pattern Recognition (MAPR), 2020, pp. 1-6, doi: 10.1109/MAPR49794.2020.9237770.
[2] H. Jiang, T. Gonnot, W. Yi and J. Saniie, "Computer vision and text recognition for assisting visually impaired people using Android smartphone," 2017 IEEE
International Conference on Electro Information Technology (EIT), 2017, pp. 350-353, doi: 10.1109/EIT.2017.8053384.
[3] Nwakanma, Ifeanyi & Oluigbo, Ikenna & Izunna, Okpala. (2014). Text – To – Speech Synthesis (TTS). 2. 154-163.
[4] Wu, Xiangyu & Jiang, Yanyan & Xu, Chang & Cao, Chun & Ma, Xiaoxing & Lu, Jian. (2016). Testing Android Apps via Guided Gesture Event Generation.
201-208. 10.1109/APSEC.2016.037.
[5] S. M. Felix, S. Kumar and A. Veeramuthu, "A Smart Personal AI Assistant for Visually Impaired People," 2018 2nd International Conference on Trends in
Electronics and Informatics (ICOEI), 2018, pp. 1245-1250, doi: 10.1109/ICOEI.2018.8553750.
.
©IJRASET: All Rights are Reserved | SJ Impact Factor 7.538 | ISRA Journal Impact Factor 7.894 | 1954