Report
Report
Sujeet K Gupta
Aman kumar
Fense N Raju
AIT 2024-2025
1
LIST OF FIGURES
Fig 4 Components 12
Specification
AIT 2024-2025
2
ABSTRACT
People deal with problems on a daily basis. Even walking without assistance is
impossible for them. They frequently depend on others to assist them. Several
technologies have been created to help those who are visually impaired. A
variety of technologies are being used.
Help the blind, Solutions based on computer vision are becoming one of the
most promising choices because they are accessible and reasonably priced. For
those who are blind or visually handicapped, this research suggests a system.
For visually challenged individuals, the suggested system seeks to provide a
wearable visual assistance that can respond to verbal commands. The capability
it offers deals with item and signboard identification.
This will make it easier for the person with vision impairments to go about their
daily lives and navigate their environment. The Open CV platform uses the
Raspberry Pi to develop artificial vision using the Python programming
language. Keywords: Raspberry Pi, Python, OpenCV, speech commands, video
recording, and navigation.
AIT 2024-2025
3
INTRODUCTION
This project's goal is to assist blind individuals in their daily lives by utilizing
the listen. There are a few key components to this visual-based endeavor,
including a camera, Vibrators, sensors, microphones, and a Raspberry Pi
mounted together with extra Internet-based working technologies are
interconnected.
The project's input will be an image or video (many frames), the picture was
taken and examined using the camera with an IoT and AI interface for the
Raspberry Pi. Thus, the object is identified. and the blind person receives
information using earphones that are transmitted via audio. Blind people's lives
are improved by this method since it will use the most recent technology and is
designed to help people with vision impairments live unrestricted lives.
AIT 2024-2025
4
LITERATURE SURVEY
• Highlights:
• Highlights:
• Highlights:
AIT 2024-2025
5
• Uses deep learning for real-time object detection.
PROBLEM STATEMENT
AIT 2024-2025
6
AIM
The aim of this project is to develop a portable, real-time visual assistant for
visually impaired individuals that enhances their independence by providing
audio feedback on surrounding objects and obstacles. Using a Raspberry Pi,
camera module, and speaker, the device detects objects in the user’s
environment and communicates relevant information via voice prompts,
allowing for safer and more informed navigation in both indoor and outdoor
settings. This project seeks to address the daily navigation challenges faced by
visually impaired individuals, offering a practical, affordable, and user-friendly
solution to improve their quality of life.
OBJECTIVE
The primary objective of this project is to design and develop a portable visual
assistant device that uses a Raspberry Pi, camera, and speaker to provide
visually impaired individuals with real-time object detection and audio
feedback. The system aims to reliably identify common objects and obstacles,
enabling users to navigate safely in various environments, including both indoor
and outdoor settings. Through a text-to-speech (TTS) system, the device will
deliver immediate audio feedback, giving users clear information about their
surroundings and enhancing their situational awareness. Additionally, we aim to
optimize the device’s performance by minimizing processing latency and
ensuring a long-lasting battery for extended usability. User-centered testing with
visually impaired individuals will be conducted to refine the system based on
real-world feedback, ensuring it is user-friendly, accurate, and effective.
Ultimately, this project strives to create an affordable, accessible assistive tool
that empowers visually impaired individuals, supporting greater independence
and improving their quality of life through enhanced mobility.
AIT 2024-2025
7
Block Diagram
AIT 2024-2025
8
Methodology
1. Requirements Analysis
2. Hardware Setup:-
9
processes to identify nearby objects. The speaker is used to communicate
detected objects and obstacles to the user in real-time. By choosing these
compact components, we aim to keep the device lightweight and portable.
Raspberry Pi
AIT 2024-2025
10
5. USB Ports: It features two USB 3.0 and two USB 2.0 ports,
supporting faster data transfer rates for peripherals.
6. GPIO Pins: The 40-pin GPIO (General Purpose Input/Output) allows
users to connect external components, enabling a wide range of
electronics and IoT applications.
7. Power Supply: Requires a USB-C 5V 3A power source, providing
more stable and reliable power than previous models.
The Raspberry Pi Camera Module is an accessory designed specifically for the Raspberry Pi
boards, enabling high-quality image and video capture. Here’s an overview of its theoretical
aspects:
1. Sensor: The camera module typically features a high-resolution sensor (like the Sony
IMX219 for older versions or Sony IMX477 for newer high-quality versions) that
allows capturing images and videos with high clarity and detail.
2. Resolution: The module offers different resolutions depending on the version:
Camera Module v2: 8 MP resolution, capable of capturing still images and
Full HD (1080p) video at 30 fps.
High Quality (HQ) Camera Module: 12.3 MP resolution, with support for
various lens mounts (e.g., CS and C-mount), providing greater flexibility and
improved image quality for advanced applications.
3. Lens and Field of View (FoV): The standard modules have a fixed-focus lens with a
broad field of view (around 62.2° horizontally for the v2 module). The HQ Camera
AIT 2024-2025
11
allows for interchangeable lenses, which helps customize the field of view and depth
of focus.
4. Interface: The camera connects to the Raspberry Pi via the CSI (Camera Serial
Interface) port using a ribbon cable. The CSI interface allows high-speed data
transmission, enabling real-time image and video processing.
5. Image Processing: With built-in support in the Raspberry Pi OS, the camera module
works seamlessly with software libraries like OpenCV, Python’s PiCamera library,
and other imaging tools, making it easy to integrate into applications requiring
computer vision.
6. Applications: Due to its compact design and good performance, the camera module is
ideal for projects involving surveillance, facial recognition, object detection, and, in
your case, creating a visual assistant for blind individuals.
3. Software Design:-
System Integration: The object detection and TTS modules work together in real
time. When an object is detected, the TTS module converts the object’s name to
speech, which the speaker then relays to the user, ensuring timely and accurate
feedback.
1. Model Selection
Chose a pre-trained model (e.g., MobileNet or YOLO) that balances high accuracy
and efficient processing, suitable for real-time use on the Raspberry Pi.
2. Dataset Selection
Utilized standard datasets like COCO (Common Objects in Context) to ensure the
model recognizes a wide range of objects relevant to daily environments.
AIT 2024-2025
12
3. Data Preprocessing
Applied preprocessing steps such as resizing images, normalizing pixel values, and
data augmentation to improve the model’s robustness to different conditions.
Performed data augmentation techniques like rotating, flipping, and adjusting
brightness to make the model more adaptable to real-world conditions.
After confirming that each component worked effectively on its own, we conducted
integrated system tests. We placed the device in controlled environments with
common obstacles like chairs, tables, and doors to assess its ability to detect and
describe objects. During these tests, we adjusted the sensitivity of the detection
AIT 2024-2025
13
algorithm, ensuring it accurately identified objects at practical distances without
overwhelming the user with unnecessary details about objects that were too far away.
Field testing with blind individuals provided crucial insights into the practical
usability of the device. We tested the system both indoors (in rooms, hallways, and
cluttered spaces) and outdoors (on streets and open spaces), noting how the device
responded in real-world settings. The feedback from users helped us fine-tune the
timing of audio feedback so that voice cues were delivered promptly without being
delayed or overly frequent. Additionally, we calibrated the text-to-speech volume and
speed to ensure that feedback was clear and easily understandable, even in noisy
environments.
Final calibration involved optimizing power usage to ensure a longer battery life,
allowing the device to be functional over extended periods. After several rounds of
testing and calibration, we achieved a setup where the device consistently delivers
accurate, timely information about the user’s surroundings, enhancing both its
usability and effectiveness for visually impaired individuals. This thorough testing
and calibration process ensured that the device would be practical, reliable, and
comfortable for real-world usage.
AIT 2024-2025
14
EXPERIMENTATION
What is YOLO:
YOLO (You Only Look Once): YOLO is a popular real-time object detection
system. It's known for its speed and accuracy, making it suitable for applications
like aerial surveillance. YOLOv5, the latest version, offers improved accuracy
and efficiency.
Steps:-
AIT 2024-2025
15
1. Component Testing
2. Environment-Based Testing
AIT 2024-2025
16
Measured real-time performance, including detection speed,
latency, and responsiveness, to ensure the device provided quick
feedback without noticeable delays.
CODE :-
AIT 2024-2025
17
Object Detection Pipeline:
OCR Pipeline:
AIT 2024-2025
18
Audio Feedback Using gTTS:
AIT 2024-2025
19
CHALLENGES DUE TO HARDWARE
1. OCR Performance:
2. Frame Skipping:
• Understanding signs and labels in public spaces (e.g., bus stops, restroom
signs), which are often inaccessible.
AIT 2024-2025
20
• Recognizing and identifying people, which can lead to social discomfort
and awkwardness.
AIT 2024-2025
21
• Understanding Signs and Labels in Public Spaces
Solution: Text-to-speech for reading signs and labels provides access to
critical information in public areas, enhancing independence.
AIT 2024-2025
22
Project Schedule
AIT 2024-2025
23
CONCLUSION
AIT 2024-2025
24
REFERENCES
AIT 2024-2025
25