Review Papers
Review Papers
UNIVERSITY
REVIEW PAPER
ON
SUBMITTED TO
(BATCH – 2021-2023)
(PROJECT GUIDE)
Submitted by
Shivani Enroll:(2102209028)
Face recognition with masks has become a critical area of research due to the
COVID-19 pandemic. The widespread use of masks has made traditional facial
recognition techniques ineffective. Machine learning algorithms have been
developed to identify faces with masks. These algorithms use deep learning
techniques, such as convolutional neural networks, to detect features on the upper
half of the face that are still visible despite the mask. Transfer learning is also used
to improve accuracy in the identification of faces with masks. Several datasets have
been created to train and test these algorithms. This abstract provides an overview
of the state-of-the-art techniques used in face recognition with masks in machine
learning.
Face recognition technology has become a vital tool in various applications, from
security to user authentication. However, the ongoing COVID-19 pandemic has
made face recognition with masks a crucial challenge to tackle. Traditional face
recognition algorithms rely on analyzing the entire face, including the nose and
mouth regions, which are covered by masks, making recognition accuracy difficult
to achieve. This paper presents a review of the recent advances in face recognition
with masks, including the development of innovative approaches that combine
machine learning algorithms with deep neural networks to accurately recognize
individuals wearing masks. These approaches involve capturing and analyzing the
unique facial features that are still visible when a mask is worn, such as the eyes and
forehead, and using them to identify individuals. The paper also discusses the
potential applications of face recognition with masks in various fields, including
public safety, security, and user authentication, highlighting the importance of this
technology in the current pandemic situation.
Introduction
The goal of this review paper is to provide an overview of recent research on face
recognition with masks in machine learning. This paper will discuss the impact of
masks on face recognition systems and describe various techniques used to improve
their performance. These techniques include traditional feature-based methods, deep
learning-based methods, and hybrid models. The review will also highlight the
importance of using diverse datasets to train face recognition systems to recognize
faces with masks. Furthermore, the paper will discuss the challenges and limitations
of existing approaches and propose future research directions to address these
challenges.
Face recognition technology has become a vital tool in various applications, from
security to user authentication. However, the ongoing COVID-19 pandemic has
made face recognition with masks a crucial challenge to tackle. Traditional face
recognition algorithms rely on analyzing the entire face, including the nose and
mouth regions, which are covered by masks, making recognition accuracy difficult
to achieve. This paper presents a review of the recent advances in face recognition
with masks, including the development of innovative approaches that combine
machine learning algorithms with deep neural networks to accurately recognize
individuals wearing masks. These approaches involve capturing and analyzing the
unique facial features that are still visible when a mask is worn, such as the eyes and
forehead, and using them to identify individuals. The paper also discusses the
potential applications of face recognition with masks in various fields, including
public safety, security, and user authentication, highlighting the importance of this
technology in the current pandemic situation.
Related Work
How many organizations rely on image processing and object recognition to identify
security issues and image classification, notably in health image processing .The
prime objective for this research is to create a face recognition algorithm that
combines real-time computer vision (CV2) with CNN (an advanced learning
algorithm) to recognition a person who is wearing a mask .Figure 1.1 shows that the
classification results are being shown independent of whether the individual is
wearing a mask or not.
The proposed model’s main principle is that in the first phase, a random image
dataset that is easy to access online is collected and processed even-before being
uploaded to the server. Following the generation of the dataset, the next step in this
field is to include image pre-processing to recover elements and improve the model’s
accuracy. Subsequent to preprocessing, the model must be trained to predict the final
output, which can take some time dependent on the hardware setup. The second step
is to recognize real-time computer vision data using an CV2-built library and,
finally, to present the results to determine whether or not a person is wearing a mask,
as shown in figure 1.2.
Image processing and data augmentation Pre-processing data assists in the cleaning
of data, the extraction of important characteristics from a dataset, and the
improvement of model performance in classification With an augmentation
algorithm, image pre-processing helps in transmitting information, improving
images, and making new data with smaller image datasets. The size, contrast,
position, and orientation of a classified data obtained may vary. Image pre-
processing is therefore the systematic, algorithm-based step done to ensure that
images are correctly edited and ready for training and inference. The image dataset
is pre-processed using augmentation methods. It may randomly crop, color filter,
position, flip, and add noise to the training data using Image Generator and CV2
built-in functions to minimize model over fitting. This study uses other pre-
processing methods and augmentation to create a verified dataset of various types.
As shown in Figure 1.2, the system will send preprocessed images to any advanced
learning algorithms to extract features. Several data augmentation methods are used
to ensure that the dataset increases in size and quality. This assisted in reducing over
fitting issues and improving the model’s overview ability during training. Table 4.1
shows the image augmentation settings that were employed. So, first and primarily,
the data is resized by a value of 1/ 255. This aids in the normalization process.
Following that, a zoom range of 20% is used to allow for some random cropping of
the image. The rotation range is set to 40 degrees, and the horizontal flip is enabled.
Literature Survey
2.1 Overview
The world is facing a huge health crisis due to the rapid transmission of the
coronavirus (COVID-19). According to WHO, the most effective preventive measure
against COVID-19 is wearing a mask in public places. It is very difficult to manually
monitor people in these areas. So in this paper, we propose a deep learning model
that detects people who are not wearing a mask, and a transfer learning model is
proposed to automate the process of identifying people who are not wearing a mask.
The proposed model is built by fine-tuning the pre-trained, state-of-the-art deep
learning model, InceptionV3. The proposed model is trained and tested on the Simulated
Masked Face Dataset (SMFD). An initiative was started by the French government to
identify passengers who are not wearing masks in public places with the help of artificial
intelligence (AI) techniques like machine learning (ML) and deep learning (DL), the
Internet of Things (IOT) and big data. These are being used for the faster diagnosis of
COVID-19.[2.]
Bosheng Qin and Dongxiao Li.[3.] designed a face mask identification method using
the SRCNet classification network and achieved an accuracy of 98.7 percent in
classifying the images into three categories, namely "correct face mask wearing,"
"incorrect face mask wearing," and "no face mask wearing."
K-nearest neighbours (KNN), support vector machine (SVM), and deep learning
are examples of classification methods. A popular DL algorithm for classification
tasks is the convolutional neural network (CNN). Because it is created of neurons that
can learn, CNN is equal to a convolutional network. Each neuron will receive data and
perform tasks, and they will all be connected in a network. Convolutional layers,
pooling layers, and fully connected layers are also the three stages of CNN’s operation.
CNN is designed to work with information that has a two-dimensional shape. Each
layer of the network is made up of various two-dimensional planes, each of which
includes multiple neurons. Neurons in two adjacent layers will connect with others.
CNN is the expert in image recognition.[6]
Face recognition using KNN, SVM, and CNN has been the subject of much research.
On a low-power processor, a reference used KNN for face recognition and achieved 91.5
%. For face recognition, the reference used SVM and achieved accuracy of 90.82 %. In
the YALE and ORL datasets, Reference used SVM for face recognition and achieved 96
% and 96.5 % accuracy, respectively. Reference employed CNN to obtain 99.5 % on the
AR database and 85.13 % accuracy on the FERET database for face recognition with
pose and lighting variation. SVM, KNN, and CNN are effective algorithms for
classifying face photos, according to previous research. As a result, we examine KNN,
SVM, and CNN in the same dataset for face mask detection. We expect similar
results.[7.]
The goal of this study is to examine and compare the performance of algorithms such
as SVM and KNN with a deep learning algorithm such as CNN. Pixel value is a feature
that is used to classify face masks. Accuracy, precision, recall, F1 score, and execution
time are all factors to consider. Researchers should be able to use the data to choose the
best and most suitable algorithm to identify face masks. Md. Sabbir Ejaz et al.[8.]
implemented the Principal Component Analysis (PCA) algorithm for masked and
unmasked facial recognition. The main aim of this work is to identify people who are
not wearing masks in public places to prevent the further transmission of COVID-19.
As per the research, wearing a mask can prevent and control the transmission of
COVID-19. Therefore, the governments of various countries have made it mandatory
to wear masks in public places and crowded areas. The same work can further be
improved by employing large volumes of data and can also be extended to classify the
type of mask and implement a facial recognition system, deployed at various
workplaces to support person identification while wearing the mask.
Roshan et. a l [9.] His paper proposes a python-based GUI tool for face mask
detection. Convolution Neural Network (CNN), PCA, and HAAR Cascade
algorithms are used in the development of the tool. Keras is the library used to create
the CNN architecture. To extract the image contents, PCA features are used. The
processing is first, the image is given to the model as input. Now the image has gone
through the preprocessing stage. At this stage, the algorithm detects the mask and eyes.
Face recognition will be done with the help of the HAAR algorithm in the next stage.
From this, the features are extracted and they move to the classification stage. In this
stage, principal component analysis (PCA) classifies the input images with various
factors. PCA has been trained with images. The next stage is the training stage. At
this stage, with the help of CNN, the qualified model from the feature extraction will
be saved to the prediction section. In the final stage, that is the prediction stage, the
prediction is By using CNN, the accuracy of the model is high. For the implementation
of the CNN algorithm, the Python Keras library is used. This paper also describes
the modules and the system architecture of the proposed model.
The COVID-19 virus spreads globally in 2020, and it can be spread through contact.
So, Shuang et. [ 10.] as mentioned in this paper, identity authentication systems based
on passwords or fingerprints are unsafe. Using facial biometrics seems a safe choice,
but most people are wearing masks. That’s why the performance of many algorithms
related to faces is challenged. This paper is to study the head pose classification problem,
which can help improve the performance of many algorithms related to faces. And this
research also has great value for many applications, such as human attention monitoring,
driver fatigue monitoring, and interaction-based face liveness detection. So, to solve
this issue, we use a method called HGL to deal with head pose classification by adopting
colour texture analysis of images and line portraits. This method uses the color
texture analysis of images and line portraits. This experiment was carried out on the
MAFA dataset because most face datasets with masks do not have head pose
information. The training data provided by the MAFA dataset is not sufficient, so
it is important for our experiment to expand the data. First, we use the information
about facial location provided by the MAFA data set to crop the image. In this
process, we increase the size of the facial frame to 1.2, 1.4, and 1.6 times the original
size for cropping, and normalize these facial images to 8080. Then, we extract the H-
channel in the HSV color space of the image and normalize the pixel values of the
image in the range of 0-255. by extracting and processing the pixel information of
the H channel in the HSV color space to distinguish the face and mask in the image.
The results show that the method proposed in this paper has better performance than
other algorithms in both front accuracy and side accuracy. For the front accuracy, this
algorithm is 0.8% better than the best accuracy of other algorithms. And this
algorithm is 2.28% higher than the best accuracy of other algorithms for side accuracy.
The reason for this phenomenon is that the images processed through the H-channel can
provide focused information to the net. The evaluation of the MAFA data set shows
that our method is superior to the other methods listed. The solution of this problem
also provides help for the study of multi-angle problems. In practical applications, we
can use a face recognition algorithm that distinguishes between wearing a mask and not
wearing a mask. If the facial image has masks, we can use the method proposed in the
paper for classification, and if it is a normal facial image, we can use FSA-Net. The line
portrait algorithms can also be used.
The recent COVID-19 epidemic has increased the amount of hygiene and contactless
identity confirmation. However, this epidemic has led to the widespread use of face
masks, which are important to maintaining this epidemic under control. The effect
of wearing a mask on facial recognition in the interaction area is a sensitive issue at
the moment but unread. Naser et.al.[11.] In his paper, he says that facial recognition
has become as important in our daily lives as it is simple and intangible how to ensure
accurate identity. Procedures such as authentication in the default border control of
gates or secure access to electrical equipment are increasingly dependent on such
technology. Face occlusion has been repeatedly addressed in the scope of face recognition
solutions. Moreover, developing occlusion-invariant face recognition solutions has been
a growing research challenge. How- ever, most of these works address general occlusion
that commonly appears in in-the-wild capture conditions, such as sunglasses and partial
captures. Our work aims at studying this effect to enable the future development of
solutions addressing accurate face recognition in such scenarios. To provide a wide view
of the effect of wearing a mask on face recognition performance, we analyzed the
performance of three face recognition algorithms. Two of these algorithms are among
the top performing academic approaches, namely the Arc Face and Sphere Fac. The
third algorithm is a COTS algorithm from the vendor Neurotechnology. In the
following, this section provides more details on these algorithms. I chose Sphere Face
as it achieved competitive verification accuracy on Labeled Face in the Wild (LFW)
of 99.42% and YouTube Faces (YTF) of 95.0% using 64-CNN layers trained on the
CASIA-Web Face dataset . However, Arc Face achieved state-of-the-art performance
on several face recognition benchmarks, such as LFW 99.83% and YTF 98.02%. In the
recent NIST report addressing the performance of vendor-face verification products,
Neurotechnogy achieved one of the best performances in the recent NIST report.
Matching the faces by COTS produces almost similar scores. For the Arc Face and
Sphere Face, the Multi-task Cascaded Convolutional Networks (MTCNN)solution is
used, as recommended in, to detect(crop) and align (affine transformation) the face.
Both networks process the input aligned and cropped image and produce a feature
vector of the size 512. To compare two faces, a distance is calculated between their
respective feature vectors. This is calculated as the Euclidean distance for arc-face
features, as recommended in, and as the cosine distance for sphere-face features, as
recommended in. Our analyses pointed out the significant effect of wearing a mask
on comparison scores and separability between genuine and imposter comparisons in
all the investigated systems. Moreover, we point out a large drop in the verification
performance of the academic face recognition solutions.
Lingxue et. [12.] It proposes two models, PDSN and FDM, based on the CNN
algorithm for face mask recognition. With this, a three-stage model is proposed. In
the initial stage, PDSN is used to capture the correspondence between the feature
elements and facial blocks. On the second stage, they create a dictionary which includes
the mask details that they learned from the generator. And in the final stage, using
FDM, perform the recognition. With the help of various mathematical formulas, train
the CNN algorithm to perform the learning process for the mask generators. In the next
stage, establishing the mask dictionary And finally, recognition. In the evaluation of the
proposed model, providing a decent output It is able to provide a maximum of 98.26%
accuracy in its algorithms. The analysis is conducted with various other data sets and
it is able to perform with the same accuracy all over the datasets. As a result, it can be
used as a good proposal.
We all are aware about the outbreak of Coronavirus Disease (COVID-19),Chavda et.
al.[13.] states As the world recovers from the pandemic and plans to return to a state of
normalcy, there is a wave of anxiety among all individuals, especially those who intend
to resume in- person activity. Studies have proved that wearing a face mask significantly
reduces the risk of viral transmission as well as provides a sense of protection. But it is
not possible to track the implementation of this policy manually. We introduce a
Deep Learning-based system that can detect instances where face masks are not used
properly. Our system consists of a dual-stage Convolutional Neural Network (CNN)
architecture capable of detecting masked and unmasked faces and can be integrated with
pre-installed CCTV cameras. We all know that technologies such as Machine Learning
and Artificial Intelligence have made our lives easier and provided solutions to a
variety of complex problems. Modern computer vision algorithms are approaching
human-level performance in visual perception tasks. In this paper, we propose a two-
stage CNN architecture where the first stage detects human faces, while the second stage
uses a lightweight image classifier to classify the faces detected in the first stage as either
‘Mask’or ‘No Mask’faces and draws bounding boxes around them along with the detected
class name. We are working with CNN in two categories: Multi-Stage Detectors and
Single-Stage Detectors. The face mask detection methods are further divided into three
methods: correct face mask wearing, incorrect face mask wearing, and no face mask
wearing. The system takes an image, detects and crops faces, and then uses SRCNe to
perform image super-resolution and classify them. Our architecture consists of two major
stages. The first stage of our architecture includes a Face Detector, and the second stage
of our architecture is a CNN based Face Mask Classifier. The results from the second
stage are decoded, and the final output is the image with all the faces shown. We also
use Dlib Deep Learning Face Detector, MTCNN-It uses a cascade architecture with
three stages, and Retina Face-It is a staggering design with pixel-wise localization that
uses a multi-task learning strategy to simultaneously predict face box. The second stage
of our system is a face mask classifier. This stage takes the processed ROI from the
Intermediate Processing Block and classifies it as either Mask or No Mask.
P. Gupta et. al.[14.] presented research on facial recognition that provides a novel
approach to using a deep neural network (another type of deep network). Rather than
sup- plying raw pixel values as input, this proposed approach simply provides the
derived facial features. The Haar cascade is used to extract facial features, which are
then fed instead of raw pixel data. The complexity of the neural network-based
recognition framework has been reduced as the number of redundant input features has
been reduced. By employing DNN instead of a convolutional network, the procedure
is also lighter and faster. The accuracy of the framework is not guaranteed by the given
method, as the average accuracy obtained is 97.05%.
For the face recognition system, K. J. Bhojane et. al.[15] used the embedded face
recognition and face tracking system methods available in MATLAB and the
Raspberry Pi B. It uses the Haarlike function, which was utilised to recognise and
authenticate the authenticated user’s face, to establish a safe environment for ignition
and access to the car. In a secure setting, the face of a person is the most important
part of the car’s ignition. Future work will be informed by the findings of the gesture
identification and control studies.
M. Rahman et. [16.] released a paper. aiming at establishing a system for recognising
whether or not a person wears a mask and It informs the appropriate authorities in the
smart city network. It captures facial photos by filming them in real time at numerous
public spots throughout the city. Masked faces are being identified using facial pictures
de- rived from this movie. After extracting features from images Using the
convolutional neural network (CNN) learning technique, the features are learned
through several hidden layers. When the architecture recognises people who aren’t
wearing a mask, the information is sent to the proper authorities via the municipal
network so that they can take the right action. Based on data gathered from a variety
of sources, the proposed method assessed promising findings. They also lay out a
framework for enforcing basic health guidelines in this pandemic situation in these
materials.
The goal of this review paper is to summarise and assess numerous well-known
techniques in the multiple steps of a pattern recognition system, as well as to
highlight the analytical and application techniques that are at the forefront of this
exciting and complicated topic. Closer machine learning methodologies are used in the
literature to approach pattern recognition frameworks. Pattern recognition techniques
are required for applications such as data processing, retrieval of multimedia
information, internet searching, face recognition, and cursive handwriting recognition.
Information gathering, segmentation, and feature ex- traction are just a few of the
obstacles that come with fully realising this. [17]
Vinitha et. al.[18.] released an article in which they suggested a method that em-
ploys a deep learning algorithm and computer vision to recognise a person with a
masked face in an image or video stream. Tensor Flow, Open CV, Keras, and PyTorch
are some of the libraries that have been used. In two stages, the project will be
implemented. The first phase involves training a deep learning model, which is followed
by the application of a mask detector to a live image or video stream in the second phase.
Open CV is the framework that was utilised to detect faces in real time from a live
webcam feed. A COVID-19 face mask detector has been built using Python and
computer vision using a data set.
The suggested system seeks to recognise the masked and faces are depicted using the
advanced YOLOv3 architecture, according to R. Bhuiyan et. al.[19] The learning
algorithm Convolution Neural Network is used by YOLO (You Only Look Once)
(CNN).YOLO connects to CNN via hidden layers, research, and quick algorithm
retrieval, and can recognise and locate any image. The process starts with the input
of 30 distinct photos from the dataset into the model, which is then combined to
produce action-level predictions. It produces outstanding imaging and detecting
performance. This model is tested on a live video to see if the model’s fps rate and
detection performance with masked and unmasked two layers are acceptable. With an
average frame rate of 17, our model produces amazing video inside. This system is more
efficient and quicker than methods that use their own data sources.
T. Meenpal et. al.[20.] proposed a semantic segmentation that categorises each pixel in
an image into face and non-face. It constructs a binary classifier quickly and then
recognises the fragments as chunks. From RGB photos, including localised objects, you
can generate accurate face masks for human objects. The findings of the Multi Human
Parsing were shown by the author. At the pixel level, this dataset has good average
accuracy. Further more, the issue of incorrect predictions is addressed, and the correct
bounding box is constructed around the segmented area. The suggested network is capable
of detecting both non-frontal and numerous faces in a single image. Complex tasks, such
as detecting parts of the face, can be accomplished using this technology.
Face recognition is performed by the paper. The face recognition network detects the
face and then estimates the pose. To process these regions for face identification and
posture estimation, a region proposal network is used to discover probable regions that
may include a face, and a single deep multitask CNN is used to process them. Because
the network is not capable of achieving real-time performance, it must be improved.
Face recognition is tested using the FDDB benchmark. Pose estimation is based on the
AFW benchmark. The head position can be used to determine a person’s motive, intent,
and gaze. According to recent research, convolutional neural networks (CNNs) have
made a series of advancements in the tasks of face recognition and position estimation.
Two CNN frameworks exist for performing these two integrated objectives at the same
time. One method is to first recognize faces using a face recognition network, and then
estimate each face’s pose using a pose estimation network. The other option is to utilise
a single deep multi-task CNN to analyse these regions for simultaneous face recognition
and posture estimation after using a region proposal technique to produce several
candidate regions that may contain faces. Because the two networks are different, the
performance of posture estimation is influenced by the performance of the face
recognition network. The latter generates a high number of candidate regions, which
takes CNN a great deal of time and resources to analyze.[21]
Facial expression recognition is presented in this study. Happy, surprised, fear, disgust,
sad- ness, anger, and a neutral state are the seven primary expressions that are identified
and classified. There are three major aspects to this system. To extract the face-centered
region, the initial step is to recognise the face and facial features. The next step is to
normalise and extract the edges of this interest zone. We have a face edge image at this
point, which we utilise to calculate the Euclidean distance between all pixels that form
edges. The SVM approach is used in the third stage to classify various emotional states.
Face Alignment for a Single Image is addressed in this study. Researchers show how an
ensemble of regression trees can be used to directly estimate the face’s landmark positions
from a sparse subset of pixel intensities, leading to super- real time performance and
high-quality predictions.[24.]
Using deep convolutional neural networks, this article presents a method for
automatically detecting faces, locating landmarks, predicting poses, and identifying
gender (CNN). Hyper is a suggested approach that uses a separate CNN to fuse the
intermediate layers of a deep network, then uses a multi-task learning algorithm to
operate on the fused features. It takes advantage of task synergy to improve individual
results.[25.]
In this paper, researchers proposed a dynamic neural network for human face
recognition. That outperforms other algorithms. A self-organizing map (SOM) neural
network, A convolutional neural network is used in the system. The SOM analysis shows
image samples into a topological space for which inputs that are near one another in the
original space are also near one another in the output space, which actually results in
dimensionality. reduction and invariance to minor changes in the image sample, although
the convolutional neural net- work provides partial invariance to translation, rotation,
scale, and deformation.[27.]
The standard face recognition pipeline consists of four stages: detect = > align
=>represent = > classify in modern face recognition To use explicit 3D face models to
apply a piece-wise affine transformation and extract a feature representation from a
nine-layer deep neural network, the paper revisits both the alignment and the
representation phases.[28.]
Due to various poses, illumination variations, and partial occlusion, face recognition
and matching in an unconstrained environment is a challenge. In this paper, we describe
a deep- casted multitask framework that increases productivity by using the intrinsic
association between detection and alignment. To identify the position of faces and
objects in a coarse- to-fine way, the framework uses a cascading approach with three
different levels of specially designed deep convolutional networks.[29.]
There has been a lot of discussion about sparse representation. The study of picture super-
resolution (SR) yielded significant results. In order to improve SR results, deep-
learning based SR approaches have also been proposed in the literature. For rapid and
high-quality image SR, the study employs a number of decision tree techniques. Our
proposed SR using the decision tree (SRDT) method employs a divide-and-conquer
strategy, in which an input low-resolution (LR) patch is classified into one of the leaf
nodes using a few simple binary tests, and then this LR patch is directly multiplied
with the regression model at that leaf node for regression. The classification and
regression processes both require very little time to complete.[32.]
RESULTS
Adam Optimizer with a learning rate of 0.15 is employed in the proposed model.
The model is trained over 15 epochs using all of the training images. Some data in
each epoch was lost during model training, namely accuracy, validation loss, and
validation accuracy. As a result, the decrease and accuracy are indicators of training
progress. It made a prediction about the classification of the training data and then
matched it to the new label. While accuracy is specified as the percentage of right
predictions, validation accuracy is defined as the measurement using data that has
not been used in training. Following training, graphs are displayed to show the
training accuracy, one for a loss and the other for accuracy. The overall training
accuracy is 1.01, while the validation accuracy is 0.47, with such a training loss of
0.73 as well as a validation loss of 0.463. The observed classification accuracy is
0.90, while the test loss is 0.15. Figures 5.1-5.3 show the training effectiveness.
CONCLUSION
[1.] Md Rafiuzzaman Bhuiyan, Sharun Akter Khushbu, and Md Sanzidul Islam. A deep
learning based assistive system to classify covid-19 face mask for human safety with
[2.] G Jignesh Chowdary, Narinder Singh Punn, Sanjay Kumar Sonbhadra, and
[3.] Bosheng Qin and Dongxiao Li. Identifying facemask-wearing condition using
Sensors,20(18):5236, 2020.
[6.] Anushree More, Harita Tilloo, Pradnya Kamble, Barkha Jagtap, and SN
[7.] Gargi Kale, Prashant Bhaware, Rohit Ingle, Sayali Sulbhewar, Yash
Gugaliya, Mayur Kaware, and Parag Thakare. Real time face mask
[8.] G Jignesh Chowdary, Narinder Singh Punn, Sanjay Kumar Sonbhadra, and
[9.] Gargi Kale, Prashant Bhaware, Rohit Ingle, Sayali Sulbhewar, Yash
Gugaliya, Mayur Kaware, and Parag Thakare. Real time face mask
[10.] Shuang Li, Xin Ning, Lina Yu, Liping Zhang, Xiaoli Dong, Yuan Shi, and
Wei He. Multi-angle head pose classification when wearing the mask for face
recognition under the covid-19 coronavirus epidemic. In 2020 international
[11.] Naser Damer, Jonas Henry Grebe, Cong Chen, Fadi Boutros, Florian Kirch-
buchner, and Arjan Kuijper. The effect of wearing a mask on face recognition
[12.] Lingxue Song, Dihong Gong, Zhifeng Li, Changsong Liu, and Wei Liu.
[13.] Amit Chavda, Jason Dsouza, Sumeet Badgujar, and Ankit Damani. Multi-
stage cnn architecture for face mask detection. In 2021 6th International Con-
ference
[14.] Priya Gupta, Nidhi Saxena, Meetika Sharma, and Jagriti Tripathi. Deep neural
[15.] Ketan J Bhojane and SS Thorat. Face recognition based car ignition and
security system. International Research Journal of Engineering and Technology
covid19 using facial mask detection in smart city network. In 2020 IEEE
[17.] Devyani Bhamare and Poonam Suryawanshi. Review on reliable pattern recog-
nition
2018.
[18.] Pierre Sermanet, Koray Kavukcuoglu, Soumith Chintala, and Yann LeCun.
[20.] Toshanlal Meenpal, Ashutosh Balakrishnan, and Amit Verma. Facial mask
2019.
[21.] Hao Wu, Ke Zhang, and Guohui Tian. Simultaneous face detection and
[22.] Lingxiao He, Haiqing Li, Qi Zhang, and Zhenan Sun. Dynamic feature match-
8(2):791–802, 2018.
[23.] Anagha S Dhavalikar and RK Kulkarni. Face detection and facial expres-
sion
78(12):1684316859, 2019.
. [25.] Steve Lawrence, C Lee Giles, Ah Chung Tsoi, and Andrew D Back. Face
[26.] Yaniv Taigman, Ming Yang, Marc’Aurelio Ranzato, and Lior Wolf.
[27] Kaipeng Zhang, Zhanpeng Zhang, Zhifeng Li, and Yu Qiao. Joint face
[28.] Amit Kumar, Azadeh Alavi, and Rama Chellappa. Kepler: Keypoint and pose
facemask detection with deep learning and computer vision. Int. J. Eng. Res.
[31.] Rajeev Ranjan, Vishal M Patel, and Rama Chellappa. Hyperface: A deep