Seminar Report2
Seminar Report2
KOZHIKODE-5
SEMINAR REPORT
ON
Submitted by
ASWATHI P
guided by
Mrs.ROJNA N
KERALA GOVERNMENT POLYTECHNIC
COLLEGE KOZHIKODE-5
CERTIFICATE
i
ACKNOWLEDGEMENT
It is with great enthusiasm and learning spirit that I am bringing out the
report of my Seminar. Here I would like to mark my token of gratitude to all those
who influenced me during the period of my work. I would like to express my
sincere thanks to Dr.Jawharali B SK, Principal, Kerala Government Polytechnic
College Kozhikode for the facilities provided here.
And above all I praise and thank the Almighty God, who showered His
abundant grace on me to make this work a success. I also express my special thanks
and gratitude to my family and all my friends for their support and encouragement.
ABSTRACT
While QR code reading finds its applications in many diverse fields like retail envi-
ronments, industries, product identification, marketing, education, warcraft etc., this work aims
to streamline the performance of the algorithm according to the requirements of a visually-
impaired user for assistance in product identification and or way-finding in both indoor and
outdoor environments. Vision is a gift and being able to make up for it for someone would be
a great cause to serve. Given the laborious hassle involved taking their route and the alarming
consequence of taking the stairs instead of an elevator, this work aims to contribute a robust and
effective QR code scanning system addressing the possible problems and challenges associated
with reading the QR code within such an environment using a deep learning based method-
ology. Especially in case of an emergency to identify a fire exit for instance, this work will
effectively contribute towards robust recognition of the QR Code even when it might be in a
condition not interpretable by standard QR Code readers. It was concluded that while the same
Computer Vision algorithm can be tweaked to execute targeted drone attacks, it is also the same
technology that can assist humanity hand in hand and help visually-impaired users in seeing the
missed and fulfill the purpose that technology was originally created for i.e. to serve humanity,
rather than the opposite.
iii
TABLE OF CONTENTS
TITLE PAGE i
BONAFIDE CERTIFICATE i
ABSTRACT iii
LIST OF FIGURES vi
1 INTRODUCTION 1
1.1 Background and Evolution of QR Codes 1
1.2 QR Code Capabilities 1
1.3 QR Code Structure 2
1.4 QR Codes and Visually-Impaired Users 3
1.5 Increasing Relevance of QR Codes 3
1.6 Existing Work and Limitations 3
1.7 Challenges and Requirements 3
2 RELATED WORK 4
2.1 Challenges with Existing QR Code Scanning Techniques 4
2.2 Image Morphological Operations and Probabilistic Methods 5
2.3 Deep Learning Approaches 5
2.4 Assisting Visually-Impaired Users 5
2.5 State-of-the-Art Research Comparisons 5
3 ALGORITHM 7
3.1 Overview 7
3.2 Performance Criteria 7
3.3 Methodology 7
iv
3.4 Detection and Localization 8
3.5 Bounding Box Prediction 9
3.6 Confidence Score and Classification Loss 9
3.7 Confidence Score and Loss 9
4 DATA SET 10
4.1 Training Data 10
4.2 Test Data 10
4.3 Synthetic Data Generation 11
4.4 Summary of Datasets 12
5 RESULTS 13
5.1 Performance Evaluation 13
5.2 Comparative Analysis 13
5.3 Dataset Evaluation 14
5.4 Summary of Results 15
REFERENCES 17
v
LIST OF FIGURES
vi
CHAPTER 1
INTRODUCTION
1
1.3 QR Code Structure
The architecture of a typical QR code consists of several components:
• Modules: Many small black and white squares each represent a single bit, organizing
themselves into rows and columns to form a matrix.
• Timing Patterns: A combination of light and dark modules that interlink the FPs, pro-
viding key information about the size, matrix dimensions, and distortion of the code.
• Alignment Patterns: Used to determine the perspective distortion of the QR code’s im-
age.
• Quiet Zone: A light area of minimum width equivalent to four modules forming a bound-
ary around the QR code.
These properties, when varied, determine the version of the QR code as one of 40 different
versions.
2
1.4 QR Codes and Visually-Impaired Users
A visually-impaired user may be defined as partly sighted or blind persons that are part
of a wider group of persons with reduced mobility. Globally, over 2.2 billion people are affected
by some kind of visual impairment, of which 79% use a smartphone. Several user requirements
support the use of a camera-based device to assist visually impaired users in daily life due to
the wide range of difficulties faced in day-to-day tasks
3
CHAPTER 2
RELATED WORK
4
2.2 Image Morphological Operations and Probabilistic Methods
QR code localization may be achieved through image morphological operations, as in
the QR code may be split into segments called ”tiles,” and a probabilistic method is applied to
classify each tile before the post-processing step where the whole QR code is framed together.
Neural networks have shown significant promise in addressing this problem.
5
bounding box in images with QR codes surrounded by text. When they used their own dataset,
the bounding box was too small if the QR code was rotated to about 45 degrees. They address
issues of perspective distortion in orientation, scale, image blur, and symbology invariance.
Their methodology involves using areas of high concentration of edge and corner structures to
construct a barcodeness map from the structure matrix and identifying peaks in the map to find
barcode borders. They also incorporated colored QR codes.
In the paper by Dubska, they developed their own open-source dataset of challenging QR
code images. Their methodology involved using the Hough Transform to detect vanishing
points with the Hough space as a 2D signal using parameterization. They then extracted the
grids and matrix codes. They compared their results with ZXing, finding that ZXing fails on
rotated and skewed images compared to their more robust algorithm.
6
CHAPTER 3
ALGORITHM
3.1 Overview
QR Code work proposes a streamlined approach for QR code detection and interpre-
tation from live video streams as well as static images. This will assist a visually impaired user
in indoor and outdoor navigation, for example, by identifying a fire-exit through a QR code.
The performance of state-of-the-art algorithms and frameworks has been reaffirmed. While QR
codes find applications in many fields, the performance requirements of this algorithm were
prioritized based on those most crucial to a visually-impaired user.
• Perspective Distortion: The image may not be aligned with the camera.
• Motion Blur:Addressed by iteratively capturing the image live from the camera as frames.
3.3 Methodology
The entire algorithm is split into three parts.
7
• Detection
• Localization
• Decoding
8
3.5 Bounding Box Prediction
The input image is divided into an SxS grid of cells. Each cell is responsible for
predicting an object lying in the center (x, y) of that cell in the image. The grid cell also predicts
the bounding box B with 5 components (x, y, w, h, confidence) and the class probabilities C,
giving a total of S x S x B x 5 outputs related to the bounding box predictions.
Bounding Box Loss: Equation (3.1) shows the loss in predicting the bounding box position.
The function computes a sum over each bounding box predictor of every cell.
S X
B2
X
1obj 2 2
λcoord ij (xi − x̂i ) + (yi − ŷi ) (3.1)
i=0 j=0
S 2
X X
1obj i c ∈ classes (pi (c) − p̂i (c))2 (3.2)
i=0
2
S X
B
X
λcoord 1obj
ij (3.3)
i=0 j=0
9
CHAPTER 4
DATA SET
• Soros Dataset: Consisting of real, sharp, defocused, and motion-blurred images. Figure
4.3 shows sample images from this dataset.
10
Figure 4.1: Dubska dataset samples
11
4.4 Summary of Datasets
• Dubska Dataset: 400 QR codes with various distortions for comprehensive evaluation.
Each dataset serves a specific purpose, ensuring the algorithm is robust, accurate, and reli-
able under various conditions.
12
CHAPTER 5
RESULTS
The results of my study on the intelligent QR code scanning system for visually-
impaired users are promising. The algorithm performed well, effectively handling various
challenges while maintaining real-time accuracy. This technology, when integrated into ETAs,
smartphone apps, or smart glasses, can significantly aid visually-impaired users in navigation
and product identification. Although there was a slight increase in decoding time, the improved
success rate justifies this trade-off, especially for critical tasks like identifying medications. En-
hancing the algorithm’s robustness and exploring additional applications could further expand
its potential and benefit users.
13
Algorithm Detection Rate
Ours 1.000
Gabour Soros 0.426
Hansen 0.890
• 130 Images Dataset: Out of 130 images, 70 were read successfully, while others were
not due to irrelevant challenges.
• Dubska Dataset: The model achieved a 100% detection rate and a 68% success rate in
decoding on the Dubska dataset consisting of 400 images of QR codes.
• Kaggle Dataset: On a test set from Kaggle with 9,999 QR codes, all were detected
and 9,651 were decoded. This method showed robustness in detecting QR codes with
perspective distortion, achieving a 99.9% success rate for detection and a 96.5% success
rate for decoding.
• Soros Dataset: A detection success of 76.7% and a decoding success of 30.83% were
achieved. The low decoding rate is due to various scales, orientations, and blurriness of
images in this dataset.
14
5.4 Summary of Results
The detection rate and decoding rates obtained on each of these datasets are listed in
Table2.
These results provide a generalized depiction of the proposed algorithm’s capability, demon-
strating its effectiveness in detecting and decoding QR codes under various conditions and chal-
lenges.
15
CHAPTER 6
16
REFERENCES
[1] A. Namane and M. Arezki, ”Fast Real Time 1D Barcode Detection FromWebeam Images
Using the Bars Detection Method”, Proceedings of the World Congress on Engineer-
ing,2017.
[2] S. Busaeed, R. Mehmood, I. Katib and J. Corchado, ”LidSonic for Visually Impaired:
Green Machine Learning-Based Assistive Smart Glasses with Smart App and Arduino”,
Electronics, 2022.
[3] L. B. Neto et al., ”A Kinect-Based Wearable Face Recognition System to Aid Visually
Impaired Users”, IEEE Transactions on Human-Machine Systems,2017.
[5] P. Theodorou and A. Meliones, ”Gaining insight for the design development deployment
and distribution of assistive navigation systems for blind and visually impaired people
through a detailed user requirements elicitation”, Universal Access in the Information
Society, 2022
17