VA Lecture 23
VA Lecture 23
using Face
Yashvi 2021UCD2117
Ritul 2021UCD2127
Meenal 2021UCD2128
Bhumika 2021UCD2167
Step 1: Face detection
The camera detects and locates the image of a face, either alone or in a crowd.
The image may show the person looking straight ahead or in profile.
Firstly, humans can be very different — skin color, haircuts, and tattoos change
their looks and affect the accuracy of the detection. Secondly, angles matter — a
face can be very different depending on the viewpoint of the camera. Thirdly,
different facial expressions can change the way a person looks. Finally, there are
matters of lighting and occlusion, as the face might be obstructed by something
or the picture can be taken in relative darkness.
Some of the Face Detection algorithms are :Yolo, SSD, HOG and HAAR
cascade, SIFT
Step 2: Face analysis
The software reads the geometry of your face. Key factors include the
distance between your eyes, the depth of your eye sockets, the distance
from forehead to chin, the shape of your cheekbones, and the contour of
the lips, ears, and chin. The aim is to identify the facial landmarks that
are key to distinguishing your face.
Step 3: Converting the image to data
The face capture process transforms analog information (a face) into a
set of digital information (data) based on the person's facial features.
Your face's analysis is essentially turned into a mathematical formula.
The numerical code is called a faceprint. This faceprint serves as a
unique identifier for the individual's face, similar to a fingerprint for
biometric identification. The faceprint captures the essential
characteristics of the face in a concise and standardized format.
Step 4: Finding a match
Your faceprint is then compared against a database of other
known faces. For example, the FBI has access to up to 650
million photos, drawn from various state databases. On
Facebook, any photo tagged with a person’s name becomes
a part of Facebook's database, which may also be used for
facial recognition. If your faceprint matches an image in a
facial recognition database, then a determination is made.
1. LBP is robust to changes in illumination, facial expressions, and small variations in pose.
2. It captures local texture details, which are important for distinguishing between different
facial regions.
3. LBP is computationally efficient and can be computed rapidly, making it suitable for real-
time applications.
Disadvantages of LBP:
1. LBP may not capture global facial features effectively, especially in cases of significant
pose variations or occlusions.
2. It may require additional techniques or feature fusion approaches to improve recognition
accuracy in complex scenarios.
Fisher Faces using LDA(linear
discrimnant analysis)
Linear Discriminant Analysis (LDA) is a dimensionality reduction technique commonly used in face
recognition to extract discriminative features from still images. The main goal of LDA in face recognition is to
project the high-dimensional face image data into a lower-dimensional subspace while maximizing the class
separability. Each face image in the dataset is typically represented as a high-dimensional feature vector,
where each element corresponds to a pixel or a feature extracted from the image. LDA method overcomes
the limitation of Principle Component Analysis method by applying the linear discriminant criterion.
Steps in LDA:
● LDA maximizes the class separability, making it effective for discriminating between
different individuals in face images.
● It reduces the dimensionality of the feature space, which can improve computational
efficiency and reduce the risk of overfitting.
● LDA is particularly useful when the number of training samples is limited compared to
the feature dimensionality.
Disadvantages of LDA:
● LDA assumes that the data follows a Gaussian distribution and that the classes have
equal covariance matrices, which may not always hold true in practice.
● It requires labeled training data to learn the discriminant features, making it a supervised
learning technique.
YOLO FOR FACE RECOGNITION
Face recognition in videos
Using OpenCV and Deep Learning (OpenFace)
1. Detect faces
2. Compute 128-d face embeddings to quantify a face
3. Train a “standard” machine learning classifier (SVM,
SGD classifier, Random Forest, etc.) on top of the face
embeddings, and therefore get a OpenCV face
recognition pipeline.
4. Recognize faces in images and video streams
To train a face recognition model with deep learning, each input batch of data includes three
images:
1. The anchor
2. The positive image
3. The negative image
The second image is our positive image — this image also contains a face of person A.
The negative image, on the other hand, does not have the same identity, and could belong
to person B, C, or even Y!
The point is that the anchor and positive image both belong to the same person/face while
the negative image does not contain the same face.
Knowledge-based methods
1. use information about the face, its features, shape, texture, or skin color.
2. set of rules is distinguished that a frame fragment must meet to be considered a human face.
Optical Flow Features:
1. The 128-d embeddings of the anchor and positive image lie closer together
2. While at the same time, pushing the embeddings for the negative image father away
And then train a “standard” machine learning classifier (SVM, SGD classifier, Random Forest,
etc.) on top of the face embeddings.In this manner, the network is able to learn to quantify faces
and return highly robust and discriminating embeddings suitable for face recognition.
Evaluation of face recognition techniques
Finding the best detector lies on the preference of the users and
what configurations give us the best balance of speed and
accuracy that a particular application will require.
Principal Component Method
recognition process to construction of a certain number of principal components of images for an input
image. However, in cases where there are significant changes in illumination or facial expression in the
face image, the effectiveness of the method is significantly reduced.
Convolutional Networks
CNN provides partial resistance to scale changes, shifts, rotations, changes in angle, and other distortions.
CNN was developed in DeepFace, which was acquired by Facebook to recognize the faces of its social
network users
LBP
Invariant to small changes in lighting conditions and small image rotations.. Among the disadvantages is
the need for high-quality image preprocessing due to high sensitivity to noise, as the number of false binary
codes increases in its presence.
Neural networks Based methods —- include addition of a new reference face to the database,
which requires complete retraining of the network on the entire available set, and this is a rather
lengthy procedure that, depending on the size of the sample, requires hours of work or even several
days.