0% found this document useful (0 votes)
264 views55 pages

Real Time Eye Detection (2) Edited Final

This document describes research on real-time eye detection. The objectives are to detect the eye from a face in a camera's field of view and to move the camera to align and capture the eye. It discusses digital image processing fundamentals like image acquisition, enhancement, restoration, and segmentation. It covers feature extraction, detection, and classification. The research involves phases of camera movement, face detection, locating the eye, and capturing the eye. Algorithms for face detection, color segmentation, and dark area extraction to find the eye are presented.

Uploaded by

Tushar Dhule
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
264 views55 pages

Real Time Eye Detection (2) Edited Final

This document describes research on real-time eye detection. The objectives are to detect the eye from a face in a camera's field of view and to move the camera to align and capture the eye. It discusses digital image processing fundamentals like image acquisition, enhancement, restoration, and segmentation. It covers feature extraction, detection, and classification. The research involves phases of camera movement, face detection, locating the eye, and capturing the eye. Algorithms for face detection, color segmentation, and dark area extraction to find the eye are presented.

Uploaded by

Tushar Dhule
Copyright
© Attribution Non-Commercial (BY-NC)
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 55

REAL TIME EYE DETECTION

BY- GUIDANCE BY-


MADHUR DHAMNE DR. S. P. NAROTE
ROHIT DHAYGUDE
TUSHAR DHULE

DEPT. OF ELECTRONICS & TELECOMMUNICATION ENGG.


SINHGAD COLLEGE OF ENGINEERING, PUNE
Objectives

 To detect the EYE from the face present in the field of


view of CAMERA

 To move the camera to align itself to capture eye


Introduction
Roadmap
 What is Digital Image Processing ?
 Fundamentals steps in DIP
 Image acquisition
 Image enhancement
 Image restoration
 Color image processing
 Compression
 Morphological processing
 Segmentation
 Representation and description
 Recognition
 Feature Extraction
 Feature Detection and Classification
Fundamental Steps in DIP
 Image acquisition- is the first process which involves preprocessing
such as scaling.
 Image enhancement- this is bringing out obscured detail or highlighting
certain features of interest in an image. This technique deals with a
number of mathematical functions such as the Fourier Transform.
 Image restoration- it improves the appearance of an image but is
objective in the sense that this technique tends to be based on
mathematical or probabilistic models of image degradation.
 Color image processing- this is used as a basis for extracting features of
interest in an image.
 Wavelets- are the foundation for representing images in various
degrees of resolution.
 Compression- deals with techniques for reducing the storage required to
save an image, or the bandwidth required to transmit it.
(continued…)

 Morphological processing- deals with tools for extracting image


components that are useful in the representation and description of
shape.
 Segmentation- partitions an image into its constituent parts or objects.
 Representation and description- representation is necessary for
transforming raw data into a form suitable for subsequent computer
processing. Description, also known as feature selection, deals with
extracting attributes that result in some quantitative information of
interest.
 Recognition- assigns a label to an object based on its descriptors.

 Feature Extraction- this is an area of image processing which involves


using algorithms to detect and isolate various desired portions of a
digitized image or video stream.
Image Enhancement
7

Histogram Example

Original
Histogram Example (cont… )
8

Poor contrast
Histogram Example (cont… )
9

Poor contrast
Histogram Example (cont… )
10

Enhanced contrast
Smoothing and Sharpening Examples
11

Smoothing Sharpening
Image analysis
 Image analysis is to identify and extract useful
information from an image or a video scene, typically
with the ultimate goal of forming a decision.
 Image analysis is the center piece of many
applications such as remote sensing, robotic vision
and medical imaging.
 Image analysis generally involves basic operations:
 Pre-processing,
 Object representation,
 Feature detection,
 Classification and interpretation.
Image Segmentation
 Image segmentation is an important pre-processing tool.
It produces a binary representation of the object with
features of interest such as shapes and edges.
 Common operations include:
 Thresholding: to segment an object from its background
through a simple pixel amplitude based decision.
Complicated thresholding methods may be used when the
background is not homogeneous.
 Edge detection: to identify edges of an object through a set
of high-pass filtering. Directional filters and adaptive filters
are frequently used to achieve reliable results.
Segmentation Examples

Thresholding Edge detection


14
Feature Extraction
15

 This is an area of image processing that uses algorithms to detect


and isolate various desired portions of a digitized image.
 What is a Feature?
A feature is a significant piece of information extracted from
an image which provides more detailed understanding of the image.
 Examples of Feature Detections
 Detecting of faces in an image filled with people and other

objects
 Detecting of facial features such as eyes, nose, mouth

 Detecting of edges, so that a feature can be extracted and

compared with another


Feature Detection and Classification

 Feature detection is to identify the presence of a certain type


of feature or object in an image.
 Feature detection is usually achieved by studying the statistic
variations of certain regions and their backgrounds to locate
unusual activities.
 Once an interesting feature has been detected, the
representation of this feature will be used to compare with all
possible features known to the processor. A statistical
classifier will produce a feature type that has the closest
similarity (or maximum likelihood) to the testing feature.
 Data collection and analysis (or the training process)have to
be performed at the classifier before any classification.
16
Block Diagram

INPUT
CAMERA COMPUTER

MECHANICAL
ASSEMBLY MICROPROSESSOR
CONTROLLING CIRCUIT
CAMERA FOCUS
Phases

 PHASE 1
 CAMERA MOVEMENT
 PHASE2
 FACE DETECTION
 PHASE3
 LOCATE EYE
 PHASE4
 CAPTURE EYE
Camera movement

 Mechanical Assembly
 Gears & links
 Motion control
 According to co-ordinates
 Servo motors
 Load consideration
 Controller/Processor
 Speed considerations
Face Detection

 Face Detection
 What is face detection?
 Importance of face detection
 Different approaches
 One example
What is Face Detection?
 Given an image, tell
whether there is any human
face, if there is, where is it.
Importance of Face Detection

 Required in any face recognition system

 Most important in any SURVEILLANCE system

 Extracting facial features


Different Approaches
Features based Approach

 Face color is a unique


feature
 This can be used to detect
a face.
Face Circle fitting Algorithm

 The continuous images as the input


 The the Skin Color pixel extracted
 Similar color pixel blocks identification
 Block grouping
 Final boundary of face region
 Face circle fitting
 (face is considered as the circular object)
 Face co-ordinates as the centre of the circle
Flow-chart
Details of the algorithm

 The algorithm uses HSI color modeling


 the H value of pixel decides the skin color pixel

 The value of H falling in the range HLo, HHi are selected


 Blocks of pixels are examined
 group of 4 blocks

 area more than 1/3rd then skin color

 The boundaries of face region are optimized


(continued…)

 The holes are filled


 Face circle is fitted
 centre location fixed

 Cvc= Ftop + R
 Chc is decided by vertical positions Pv[i]
 The maximum value of Pv[i] gives the Chc
 the radius is calculated using boundary values

 R= (Ftop - F bottom)/2
Color Segmentation Algorithm
 Color Segmentation
 Captured image => RGB color space
 Affected by lightning condition
 Remedy : use of YCbCr color space
 Y component => luminance
 Cb, Cr component => chrominance

 RGB to YCbCr conversion


 Y = 0.299R + 0.587G + 0.114B
 Cb = -0.169R - 0.332G + 0.500B
 Cr = 0.500R - 0.419G - 0.081B

 Result of color segmentation


(continued…)
 Image Segmentation
 Need : to obtain perfect binary image
 A three step process:
 Step 1: removal of small black and white regions

 Step 2: edge detection using Robert Cross Operator

 Step 3: integration of step 1 and step 2


(continued…)

 Step 1
 Elimination of small black holes on white region

 Removal of white region less than minimum face area

 Small region eliminated


image
(continued…)

 Step 2
 Robert cross algorithm

 Performs gradient measurement on binary image


 Highlights gradient of high magnitude
 Converts highlighted region into black lines by connecting
adjacent pixels
 Gradient magnitude |G | = |P1 – P4 | + |P2 – P3 |
(continued…)
 Edges detected by Robert cross operator
(continued…)
 Integration of step1 and 2
 Small black and white regions removed

 Clear view of skin region


(continued…)

 Image matching
 Template eigen image generation

 Correlating test image with eigen image template

 Drawbacks
 All the faces are vertical and have frontal view

 Images are captured under same lightning conditions


Eye Detection

 Importance of Eye Detection


 Challenges
 Different Approaches
Eye Detection(continued…)

 Importance of Eye Detection


 Man machine interaction technology

 Monitoring human vigilance

 Assisting people with disability

 Challenges
 Eye closure

 Eye occlusion

 Variability in scale and location, different

 Lighting conditions

 Face orientation
Different Approaches
Dark Area Extraction Algorithm

 The dark pixels extraction


 Intensity comparison
 Change the boundaries
 Detect the eye area
 Mark Eyes
 Send co-ordinates to processor
Flow Chart
Details of Algorithm

Eye Positioning
Eye positioning,
(a) source image,
(b) eye pixel extraction,
(c) eye pixel grouping,
(d) unreasonable pixel removal,
(e) horizontal position locating,
(f) nose pixel and eyebrow
pixel removal,
(g) vertical position location,
(h) eye marking.
(continued…)

Parallel Eye Detection


 Convert the captured
image in YCbCr color
space
 Build two separate
eye maps
 EyeMapC
 EyeMapL

 Above two eye maps


are combined to get a
single eye map.
(continued…)
 Eye Map C construction
 In YCbCr color space eye region have higher Cb value and lower Cr
value
 EyeMapC = 1/3 ((Cb)2 + (Cr)2 + (Cb/Cr))

 Eye Map L construction


 Design of dilation and erosion operator
 To emphasis on bright and dark area around eye

Eye map L = Y(x , y) + g(x , y)


Y(x , y) * g (x , y)
(continued…)

 Combining eye map C and


eye map L
 Table shows the
experiment results
Results(Comparison Chart)

METHOD APPROACH DETCTION ADVANTAGE DISADVANTAG


RATE E
(DATABASE)

COLOR BASED BACKGROUND RGB HSI


FACE CIRCLE (HSI) 85 -92 % FILLTERING IS FAST REQUIRED,
FITTING NO DIRECT
FUNCTION
FACE
DETECTION LIGHTNING EFFECT GENDER
COLOR 90% IS REDUCED, LESS RECOGNITION
SEGMENTATION FLASE COMPUTATION
ALGORITHM DETECTION 5% TIME
COLOR BASED
(YCbCr)
SPEED OF IMAGE NEEDS TO
PROCEESING BE
BUILDING EYE 85% INCREASE DUE TO FRONTAL,SSHOLU
MAP PARRALEL D NOT BE
EYE IMPLEMENTATION OCCLUDED BY
DETECTION GLASS ETC
COMPUTATIONAL EFFECT OF
DARK AREA FEATURE 85 -92 % TIME IS LESS LIGHTNING
EXTRACTION INVARIANT CONDITION
Mechanical Assembly

 Microprocessor system
 Motors systems
 Links and Gears
 Camera & Mountings
Assembly(how it will look/function)
Design Considerations

1. Camera Weight
 (around 1kg.)
2. Minimum displacement required
 Vertical & Horizontal
 Gear ratio calculation
3. Mobility with connecting cables
 The cables should be routed to achieve maximum
mobility for camera movement
Embedded system

 ARM
 Motor Drivers
 PC interfacing Circuit

(USB 2.0/RS-232)
 Manual Controls
ARM
 ARM 7 TDMI
 Designed to be small & reduce the power consumption
 High code density
 Hardware debug technology within the processor
 High speed Flash memory(32 to 128 kB)
 Large buffer size and high processing power
 Various 32-bit timer
 Supports USB 2.0 Full-speed device, multiple UARTs, SPI, SSP to
I2C-bus
 Supports In-System Programming/In-Application Programming
(ISP/IAP)
Proposed Work(Planning Sheet)

Activities Semester –I Semester -II

1. Basics of Image 
Processing

2. Literature Survey 
3. Software Study 
4.Development of  
algorithm

5.Implementation of 
prototype

6. Optimization analysis 
References

 [1] M. Yang, D.J. Kriegman, N. Ahuja, Detecting faces in images: a survey, IEEE
Transactions on Pattern Analysis and Machine Intelligence 24 (1) (2002) 34–58.
 [2] S. Zhang, Z. Liu, A robust, real-time ellipse detector, Pattern Recognition 38 (2) (2005)
[273–287].
 [3] K. Anderson, P.W. McOwan, Robust real-time face tracker for cluttered environments,
Computer Vision and Image Understanding 95 (2) (2004) [184–200].
 [4] F.Y. Shih, C. Chuang, Automatic extraction of head and face boundaries and facial
features, Information Sciences 158 (1) (2004) [117–130].
 [5] Z. Liu, J. Yang, N.S. Peng, An efficient face segmentation algorithm based on binary
partition tree, Signal Processing: Image Communication 20 (4) (2005) [295–314].
 [6] F. Tsalakanidou, S. Malassiotis, M.G. Strintzis, Face localization and authentication using
color and depth images, IEEE Transactions on Image Processing 14 (2) (2005) [152–168].
 [7] M. Soriano, B. Martinkauppi, S. Huovinen, M. Laaksonen, Adaptive skin color modeling
using the skin locus for selecting training pixels, Pattern Recognition 36 (3) (2003) [681–
690].
 [8] R.L. Hsu, M. Abdel-Mottaleb, A.K. Jain, Face detection in color images, IEEE
Transactions on Pattern Analysis and Machine Intelligence 24 (5) (2002) [696–706].
 [9] R. Xiao, M. Li, H. Zhang, Robust multipose face detection in images, IEEE Transactions on
Circuits and Systems for Video Technology 14 (1) (2004) [31–41].
 [10] C. Liu, A Bayesian discriminating features method for face detection, IEEE Transactions
on Pattern Analysis and Machine Intelligence 25 (6) (2003) [725–740].
 [11] Y. Li, S. Gong, J. Sherrah, H. Liddell, Support vector machine based multi-view face
detection and recognition, Image and Vision Computing 22 (5) (2004) [413–427].
 [12] P. Viola, M.J. Jones, Robust real-time face detection, International Journal of Computer
Vision 57 (2) (2004) [137–154].
 [13] L. Huang, A. Shimizu, Y. Hagihara, H. Kobatake, Gradient feature extraction for
classification-based face detection, Pattern Recognition 36 (11) (2003) [2501–2511].
 [14] M. Lie´vin, F. Luthon, Nonlinear color space and spatiotemporal MRF for hierarchical
segmentation of face features in video, IEEE Transactions on Image Processing 13 (1) (2004)
[63–71].
 [15] E. Loutas, I. Pitas, C. Nikou, Probabilistic multiple face detection and tracking using
entropy measures, IEEE Transactions on Circuits and Systems for Video Technology 14 (1)
(2004) [128–135].
 [16] S. Spors, R. Rabenstein, A real-time face tracker for color video, in: Proceedings of the
IEEE International Conference on Acoustics, Speech, and Signal Processing, 2001, pp. [1493–
1496].
 [17] S. Phimoltares, C. Lursinsap, K. Chamnongthai, Locating essential facial features using
neural visual model, in: Proceedings of the First IEEE International Conference on Machine
Learning and Cybernetics, 2002, pp. [1914–1919].
 [18] S. Phimoltares, C. Lursinsap, K. Chamnongthai, Facial feature extraction with rotational
invariance using neural visual model, in: Proceedings of the Third International Conference
on Intelligent Technologies, 2002, pp. [226–234].
 [19] S. Phimoltares, C. Lursinsap, K. Chamnongthai, Tight bounded localization of facial
features with color and rotational independence, in: Proceedings of the IEEE International
Symposium on Circuits and Systems,2003,pp.V-809–V-812.
 [20] M. Nixon, Eye spacing measurement for facial recognition, SPIE Proceedings, Vol. 575:
Applications of Digital Image Processing VIII, 1985, pp. [279-285]
 [21] P.W.Hallinan, Recognizing human eyes, SPIE Proceedings, Vol. 1570:
GeometricMethods in Computer Vision, 1991, pp. [214-226].
 [22] A. Pentland, B. Moghaddam, T. Starner, M. Turk, Viewbased and modular eigenspaces
for face recognition, Proceedings, IEEE Computer Society Conference on Computer Vision
and Pattern Recognition, 1994, pp. [84-91].
 [23] R. Herpers, M. Michaelis, K.H. Lichtenauer, G. Sommer, Edge and keypoint detection in
facial regions, Proceedings, Second International Conference on Automatic Face and Gesture
Recognition, 1996, pp. [212-217].
Thank You…!

You might also like