Refrence Image Captured Image: Figure 5.1 Sequence Diagram

Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 6

REFRENCE IMAGE CAPTURED IMAGE

RGB TO GRAY RGB TO GRAY


CONVERSION CONVERSION

IMAGE RESIZING IMAGE RESIZING

IMAGE IMAGE
ENHANCEMENT ENHANCEMENT

EDGE DETECTION EDGE DETECTION

IMAGE MATCHING

TIMING
ALLOCATION

Figure 5.1 Sequence Diagram

Above figure 5.1 shows the Sequence Diagram of Under sampled Face Recognition via Robust
Auxiliary Dictionary Learning (proposed algorithm)

The Block diagram above gives an overview of Under sampled Face Recognition via
Robust Auxiliary Dictionary Learning. Various boxes in Block diagram are explained below:
Fig.5.2: Robust auxiliary dictionary learning

5.2.1 IMAGE ACQUISITION

Generally an image is a two-dimensional function f(x,y)(here x and y are plane


coordinates).The amplitude of image at any point say f is called intensity of the image. It is
also called the gray level of image at that point. We need to convert these x and y values to
finite discrete values to form a digital image. The input image is a fundus taken from stare
data base and drive data base. The image of the retina is taken for processing and to check
the condition of the person. We need to convert the analog image to digital image to process
it through digital computer. Each digital image composed of a finite elements and each finite
element is called a pixel.

5.2.2 FORMATION OF IMAGE

We have some conditions for forming an image f(x, y) as values of image are
proportional to energy radiated by a physical source. So f(x, y) must be nonzero and finite.
i.e. 0< f(x, y) < .
5.2.3 IMAGE PRE-PROCESSING

5.2.3.1 IMAGE RESIZING/SCALING

Image scaling occurs in all digital photos at some stage whether this be in Bayer
demosaicing or in photo enlargement. It happens anytime you resize your image from one pixel
grid to another. Image resizing is necessary when you need to increase or decrease the total
number of pixels. Even if the same image resize is performed, the result can vary significantly
depending on the algorithm.

Images are resized because of number of reasons but one of them is very important in our
project. Every camera has its resolution, so when a system is designed for some camera
specifications it will not run correctly for any other camera depending on specification
similarities. so it is necessary to make the resolution constant for the application and hence
perform image resizing.

5.2.3.2 RGB TO GRAY CONVERSION

Humans perceive colour through wavelength-sensitive sensory cells called cones. There
are three different varieties of cones, each has a different sensitivity to electromagnetic radiation
(light) of different wavelength. One cone is mainly sensitive to green light, one to red light, and
one to blue light. By emitting a restricted combination of these three colours (red, green and
blue), and hence stimulate the three types of cones at will, we are able to generate almost any
detectable colour. This is the reason behind why colour images are often stored as three separate
image matrices; one storing the amount of red (R) in each pixel, one the amount of green (G) and
one the amount of blue (B). We call such colour images as stored in an RGB format.

In grayscale images, however, we do not differentiate how much we emit of different


colours, we emit the same amount in every channel. We will be able to differentiate the total
amount of emitted light for each pixel; little light gives dark pixels and much light is perceived
as bright pixels. When converting an RGB image to grayscale, we have to consider the RGB
values for each pixel and make as output a single value reflecting the brightness of that pixel.
One of the approaches is to take the average of the contribution from each channel: (R+B+C)/3.
However, since the perceived brightness is often dominated by the green component, a different,
more "human-oriented", method is to consider a weighted average, e.g.: 0.3R + 0.59G + 0.11B.

5.2.4 IMAGE ENHANCEMENT

Image enhancement is the process of adjusting digital images so that the results are more
suitable for display or further analysis. For example, we can eliminate noise, which will make it
more easier to identify the key characteristics.

In poor contrast images, the adjacent characters merge during binarization. We have to
reduce the spread of the characters before applying a threshold to the word image. Hence, we
introduce POWER- LAW TRANSFORMATION which increases the contrast of the
characters and helps in better segmentation. The basic form of power-law transformation is

s = cr ,

where r and s are the input and output intensities, respectively; c and are positive
constants.

A variety of devices used for image capture, printing, and display respond according to a
power-law. By convention, the exponent in the power-law equation is referred to as gamma.
Hence, the process used to correct these power-law response phenomena is called gamma
correction. Gamma correction is important, if displaying an image accurately on a computer
screen is of concern.

In our experimentation, is varied in the range of 1 to 5. If c is not equal to 1, then the


dynamic range of the pixel values will be significantly affected by scaling. Thus, to avoid
another stage of rescaling after power-law transformation, we fix the value of c = 1. With = 1,
if the power-law transformed image is passed through binarization, there will be no change in the
result compared to simple binarization. When > 1, there will be a change in the histogram plot,
since there is an increase of samples in the bins towards the gray value of zero. Gamma
correction is important if displaying an image accurately on computer screen is of concern.

5.2.5 EDGE DETECTION


Edge detection is the name for a set of mathematical methods which aim at identifying
points in a digital image at which the image brightness changes sharply or, more technically, has
discontinuities or noise. The points at which image brightness alters sharply are typically
organized into a set of curved line segments termed edges.

The same problem of detecting discontinuities in 1D signal is known as step detection


and the problem of finding signal discontinuities over time is known as change detection. Edge
detection is a basic tool in image processing, machine vision and computer envisage, particularly
in the areas of feature reveal and feature extraction.

5.2.5.1 EDGE DETECTION TECHNIQUES

Different colours have different brightness values of particular colour. Green image has
more bright than red and blue image or blue image is blurred image and red image is the high
noise image.

Following are list of various edge-detection methods:-

Sobel Edge Detection Technique

Perwitt Edge Detection

Roberts Edge Detection Technique

Zero cross Threshold Edge Detection Technique

Canny Edge Detection Technique

In our project we use CANNY EDGE DETECTION TECHNIQUE because of its various
advantages over other edge detection techniques

5.2.5.2 CANNY EDGE DETECTION

The Canny Edge Detector is one of the most commonly used image processing tools
detecting edges in a very robust manner. It is a multi-step process, which can be implemented
on the GPU as a sequence of filters. Canny edge detection technique is based on three basic
objectives.

Low error rate

All edges should be found, and there should be no spurious responses. That is, the edges must be
as close as possible to the true edges.

Edge point should be well localized:-

The edges located must be as close as possible to the true edges. That is, the distance between a
point marked as an edge by the detector and the centre of the true edge should be minimum.

Single edge point response


The detector should return only one point for each true edge point. That is, the number of local
maxima around the true edge should be minimum. This means that the detector should not
identify multiple edge pixels where only a single edge point exist.

The essence of Cannys work was in expressing the preceding three criteria
mathematically and then attempting to find optimal solution to these formulations, in general, it
is difficult to find a close form solution that satisfies all the preceding objectives. However,
using numerical optimization with 1-D step edges corrupted by additive while Gaussian noise
led to the conclusion that a good approximation to the optimal step edge detector is the first
derivative of Gaussian

Generalizing this result to 2-D involves recognizing that the 1-D approach still applies in
the direction of the edge normal. Because the direction of the normal is unknown beforehand,
this would require applying the 1-D edge detector in all possible directions. This task can be
approximated by first smoothing the image with circular 2-D Gaussian function, computing the
gradient of the result, and then using the gradient magnitude and direction to estimate edge
strength and direction at every point

You might also like