Image Processing Question Answer Bank
Image Processing Question Answer Bank
QUESTION BANK
(2 Marks and 16 Marks)
UNIT I
2 Marks
1. Define Image
An image may be defined as two dimensional light intensity function f(x, y)
where x and y denote spatial co-ordinate and the amplitude or value of f at
any point (x, y) is called intensity or grayscale or brightness of the image at
that point.
3. Define Brightness
Brightness of an object is the perceived luminance of the surround.
Two objects with different surroundings would have identical luminance
but different brightness.
The human being can resolve Several rods are connected to one
the fine details with these cones nerve end. So it gives the overall
because each one is connected picture of the image.
to its own nerve end.
27. Find the number of bits required to store a 256 X 256 image with 32 gray levels
32 gray levels = 25
= 5 bits
256 * 256 * 5 = 327680 bits
28. Write the expression to find the number of bits to store a digital image?
The number of bits required to store a digital
image is b=M X N X k
When M=N, this equation
becomes b=N^2k
16 Marks
be located thus limiting the search that has to be conducted in seeking that
information.
GB Sector(120o <= H < 240o).If the given value of H is in this ,we first
subtract 120o from it
H = H -120o
Then the RGB components are
B = I (1 – S)
G = I [1 + S cos H/cos(60o - H)]
B = 1 - (R + G)
BR Sector(240 <=H<=360o).Finally if H is in this range we subtract 240o from it
o
H = H - 240o
Then the RGB components are
G = I (1 - S)
B = I [1 + S cos H/cos(60o - H)]
R = 1 - (G + B)
UNIT II & III
2 Marks
5. Define histogram.
The histogram of a digital image with gray levels in the range [0, L-1] is a
discrete function h(rk)=nk.
rk-kth gray level
nk-number of pixels in the image having gray level rk.
k k
Sk= T(rk) = ∑ Pr(rj) = ∑ nj/n where k=0,1,2,….L-1
j=0 j=0
This transformation is called histogram equalization.
0 -1 0 -1 -1 -1
-1 A+4 -1 -1 A+8 -1
0 -1 0 -1 -1 -1
mean μ=a+√πb/4
standard deviation σ2=b(4-π)/4
16 Marks
1. Discuss different mean
filters
ƒ^(x,y)=1/mn Σ g(s,t)
(s,t)ЄSxy
Geometric mean filter
An image restored using a geometric mean filter is given by the
expression f^(x,y) = [ п g(s,t) ]
(s,t)ЄSxy
here ,each restored pixel is given by the product of the pixels in
the subimage window , raised to the power 1/mn
• Harmonic filters
The harmonic mean filtering operation is given by the expression
ƒ ^(x,y) = mn/∑(1/g(s,t))
• Contra harmonic mean filter
Contra harmonic mean filtering operation yields a restored image based on the
expression
Q+1 Q
f^(x,y)=∑g(x,t) /∑g(s,t)
where Q is called the order of the filter.This filter is well suited for
reducing or virtually eliminating the effect of salt and pepper noise.
H is a linear positive invariant process the the degraded image is given in the
spatial domain byg(x,y)=h(x,y)*f(x,y)+ŋ(x,y)
where h(x,y) is the spatial representation of the degradation
function and the symbol “*” indicates spatial convolution.
• The convolution in the spatial domain is equal to multiplication in the
frequency domain.
• The equivalent frequency domain representation is
G(u,v)=H(u,v)F(u,v)+N(u,v)
Where the terms in capital letters are the Fourier transforms of the
corresponding terms in the previous equation.
3. Write short notes on Median
Filters
Introduction:
-Median Filter is one of the part of the smoothing filter.
-No mask is used in the median filters.
-We choose 3x3 sub-image arranged in ascending order and
leave first four values.
3 5 7
2 10 20
30 9 4
2,3,4,5,7,9,10,20,30
-Take the median value.
-This median filter is the non-linear spatial
filtering. 1)median filtering smoothing
2)Max
filter
3)Min
filter
Max filter:
R=Max
-Max filter gives the brightest points
Min filter:
R=Min
-It helps to get the largest point in the image.
• It is assumed that the noise and the image are uncorrelated,that one or
other has zero mean:and that the gray levels in the estimate are a linear
function of the levels in the degradated image.
• Based on these conditions,the minimum of the error function in Eq is
given in the frequency domain by the expression
• This result is known as the wiener filter after N.Wiener,who proposed the
concepts in the year shown.the filter which consists of the term inside the
brackets also is commonly referred to as the minimum mean square error
filter or the least square error filter.
• We include references at the end of sources containing detailed
derivations of the wiener filter.
• The restored image in the spatial domain is given by the inverse Fourier
transform of the frequency domain estimate F(u,v).
• If the noise is zero,then the noise power spectrum vanishes and the
wiener filter reduces to the inverse filter.
• However the power spectrum of the undegraded image seldom is
known. Where k is a specified constant.
• Example illustrates the power spectrum of wiener filtering over direct
inverse filtering.the value of K was chosen interactively to yield the best
visual results.
• It illustrates the full inverse filtered result similarly is the radially
limited inverse filter .
• These images are duplicated here for convenience in making comparisons.
• As expected ,the inverse filter produced an unusable image.The \noise in the inring filter.
• The wiener filter result is by no means perfect,but it does give us a
hint as to image content.
• The noise is still quite visible, but the text can be seen through a
“curtain” of noise.
5. Explain Histogram processing
• The Histogram of the digital image with gray levels in the range [0,L-
1]is the discrete function p(rk)=nk/n where rk is the kth gray level, nk
is the number of pixel,n is the total number of pixel in the image and
k=0,1,2,…….L-1.
• P(rk) gives the an estimate probability of occurrence of gray-level rk..
Figure show the the histogram of four basic types of images.
Figure: Histogram corresponding to four basic image types
Histrogram Equalization
• Let the variable r represent the gray levels in the image to be
enhanced. The pixel value are continous quantities normalized that lie
in the interval [0,1] with r=0 represent black with r=1 represent
white.
• The transformation of the form
• S=T(r) ............................................ (1)
• Which produce a level s for every pixel value r in the original
image.it satisfy condition:
o T(r) is the single-valued and monotonically increasing in
the interval 0≤r≤1 and
o 0≤T(r)≤1 for 0≤r≤1
▪ Condition 1 preserves the order from black to white
in the gray scale.
▪ Condition 2 guarantees a mapping that is consistent
with the allowed range of pixel values.
R=T-¹(s) 0≤s≤1 ..................................... (2)
• The probability density function of the transformed graylevel is
Ps(s)=[pr(r)dr/ds] r=T-¹(s) ..................................(3)
• Consider the transformation function
S=T(r)= ∫Pr(w)dw 0≤r≤1 ............................... (4)
Where w is the dummy variable of integration .
From Eqn(4) the derivative of s with respect to r is
ds/dr=pr(r)
Substituting dr/ds into eqn(3)
yields
Ps(s)=[1] 0≤s≤1
Histogram Specfication
• Histogram equalization method does not lent itself to interactive application .
• Let Pr(r) and Pz(z) be the original and desired probability function.
Suppose the histogram equalization is utilized on the original image
S=T(r)=∫Pr(w) dw ............................................... (5)
W1 W2 W3
W4 W5 W6
W7 W8 W9
• Denoting the gray level of pixels under the mask at any location
by z1,z2,z3……,z9, the response of a linear mask is
R=w1z1+ w2z2 + ....... +w9z9
Smoothing Filters
• Lowpass Spatial filtering:
▪ The filter has to have positive coefficient.
▪ The response would be the sum of gray levels of nine
pixels which could cause R to be out of the gray level
range.
▪ The solution is to scale the sum by dividing R by 9.The
use of the form of mask are called neighborhood
averaging 1 1 1
1 1 1
1 1 1
1/9
• Median filtering:
▪ To achive noise reduction rather than blurring.
▪ The gray level of each pixel is replaced by the median
of the gray level in the neighbourhood of that pixel
Sharpening Filters
• Basic highpass spatial filtering:
▪ The filter should be positive ecoefficient near the center
and negative in the outer periphery.
▪ The sum of the coefficient are 0.
▪ This eliminate the zero- frequency term reducing
significantly the global contrast of the image
-1 -1 -1
-1 8 -1
• High_boost
filtering:
The definition is
High-boost=(A)(Original)-Lowpass
=(A-1) (Original)+ Original –Lowpass
=(A-1) (Original)+Hignpass
• Derivative Filters:
▪ Averaging is anlog to integration , differentiation can be
expected to have opposite effect and thus sharpen the
image
1. Spatial transformations:-
An image f of pixel coordinates(x,y) undergoes geometric
distortion to produce an image g with coordinates(x’,y’).this
transformation may be expressed as
x’=r(x,y)
y’=s(x,y)
• where r(x,y) and s(x’,y’) are the spatial transformations that
produced the geometrically distorted image g(x’,y’).
• If r(x,y) and s(x,y) were known analytically recovering f(x,y) from
the distorted image g(x’,y’) by applying the transformations in
reverse might possible theoretically.
• The method used most frequently to formulate the spatial relocation
of pixels by the use of tiepoints,which are a subset of pixels whose
location in the input and output image is known precisely.
• The vertices of the quadrilaterals are corresponding tiepoints.
▪ r(x,y)=c1x+c2y+c3xy+c4
▪ S(x,y)=c5x+c6y+c7xy+c8
▪ x’=c1x+c2y+c3xy+c4
▪ y’=c5x+c6y+c7xy+c8
• Since there are a total of eight known tiepoints these equations can
be solved for eight coefficients ci,i=1,2,…8.
• The coefficient constitute the geometric distortion model used to
transform all pixels within the quadrilateral region defined by the
tiepoints used to obtain the coefficients.
• Tiepoints are established by a number of different techniques
depending on the application.
2. Gray level Interpolation:-
• Depending on the values of coefficients ci equations can yield
noninteger values for x’ and y’.
• Because the distorted image g is digital ,its pixel values are defined
only at integer co ordinates .
• Thus using non integer values for x’, y’ causes a mapping into
locations of g for which no gray levels are defined.
• The technique is used to accomplish this is called gray level interpolation.
4. Describehomomorphic filtering
• The illumination – reflectance model can be used to develop a
frequency domain procedure for improving the appearance of an
image by simultaneous gray – level compression and contrast
enhancement.
• An image can be expressed as the product of illumination and
reflectance components.
f(x,y) = i(x,y) r(x,y)
F(f(x,y)) = F(i(x,y)
r(x,y))
mean μ=a+√πb/4
standard deviation σ2=b(4-π)/4
Gamma noise:
The PDF is
Uniform noise:
The PDF is
P(Z)=1/(b-a)if a<=Z<=b
0 otherwise
mean μ=a+b/2
standard deviation σ2=(b-a)2/12
Impulse noise:
The PDF is
P(Z) =Pa for z=a
Pb for z=b
0 Otherwise
UNIT IV
1. What is segmentation?
Segmentation subdivides on image in to its constitute regions or
objects. The level to which the subdivides is carried depends on the problem
being solved .That is segmentation should when the objects of interest in
application have been isolated.
6. What are the two properties used for establishing similarity of edge pixels?
(1) The strength of the response of the gradient operator used to
produce the edge pixel.
(2) The direction of the gradient.
7. What is edge?
An edge isa set of connected pixels that lie on the boundary between
two regions edges are more closely modeled as having a ramplike profile.
The slope of the ramp is inversely proportional to the degree of blurring in
the edge.
16 Marks
▼ƒ≡grad(ƒ)≡ gy = ∂ƒ/∂y
0 -1 0
-1 4 -1
0 -1 0
5. The real-time automatic images processing and pattern recognition are very
important for many problems in medicine, physics, geology, space research,
military applications and so on. For example, it is necessary for pilots and drivers
for immediate decision-making in poor visibility conditions. An approach to
image enhancement through artificial neural network’s (ANN) processing is
proposed.ANN is for images enhancement through approximation of image
transform function T. This function is approximated with use of ANN which is
trained evolutionary in the time of test images processing. Each ANN is
genetically encoded as the list of its connections. Truncation selection is used for
parental subpopulation formation. Original crossover and mutation operators,
which respect structures of the ANNs undergoing recombination and mutation,
are used. Nodes with Write Short notes on edge linking by local processing.
• One of the simplest approaches f or linking edge points is to
analyze the characteristics of the pixels in a small neighborhood
about every point in an image that has undergone edge detection.
• Two properties used for establishing similarity of edge pixels in the analysis are
▪ The strength of the response of the gradient
operator used to produce the edge pixel,
▪ The direction of the
gradient. The first property is given by the
value of ▼f.
Thus an edge pixel with coordinates (x’,y’) and in the predefined
neighborhood of (x,y) is similar in magnitude to the pixel at (x,y) if |▼f(x,y)
- ▼(x’,y’)|<=T where T is a nonnegative threshold.
The direction of the gradient vector is given by
α(x,y)=tanˉ¹ gy/gx
Then an edge pixel at (x’,y’) in the predefined neighborhood of (x,y) has an
angle similar to the pixel at (x,y) if | α(x,y)= α(x’,y’)|<A where A is an angle
threshold. Note that the direction of the edge at (x,y) in reality is
perpendicular to the direction of the gradient vector at that point.
A point in the predefined neighborhood of (x,y) is linked to the pixel at
(x’,y’) if both magnitude and direction criteria are satisfied. This process is
repeated for every location in the image. A record must be kept of linked
points as the center of the neighborhood is moved from pixel to pixel. A
simple bookkeeping procedure is to assign a different gray level to each set
of linked edge pixels.
The real-time automatic images processing and pattern recognition are very
important for many problems in medicine, physics, geology, space research,
military applications and so on. For example, it is necessary for pilots and
drivers for immediate decision-making in poor visibility conditions. An
approach to image enhancement through artificial neural network’s (ANN)
processing is proposed.ANN is for images enhancement through
approximation of image transform function T. This function is approximated
with use of ANN which is trained evolutionary in the time of test images
processing. Each ANN is genetically encoded as the list of its connections.
Truncation selection is used for parental subpopulation formation. Original
crossover and mutation operators, which respect structures of the ANNs
undergoing recombination and mutation, are used. Nodes with sigmoid
activation functions are considered. The population size adapts to the
properties of evolution during the algorithm run using simple resizing
strategy. In this application pixel-by-pixel brightness processing with use of
ANN paradigm is adopted. The topology of ANN is tuned simultaneously with
connections weights. The ANN approximating T function should have three
input nodes and one output node. During the training we evaluate each ANN
with respect to the visual quality of the processed images.
The artificial neural network training stage with use of single 128х128 pixels
image takes about 70 seconds on the Intel Pentium IV 3 GHz processor. After
completion of the learning process the obtained artificial neural network is
ready to process
arbitrary images that were not presented during the training. The processing
time for 512х512 pixels image is about 0.25 second. The ANN, as a rule,
included 3 input nodes, one or more hidden nodes and one output node.
16 Marks
Here, P(Ri) is a logical predicate defined over the points in set Ri and Ф is the null set.
➢ Condition (a) indicates that the segmentation must be complete that is
every pixel must be in a region.
➢ Condition (b) requires that points in a region must be connected in
some predefined sense.
➢ Condition(c) indicates that the regions must be disjoint.
➢ Condition(d) deals with the properties that must be satisfied by the pixels in
a segmented region.
Region Growing:
As its name implies region growing is a procedure that groups pixel
or subregions into larger regions based on predefined criteria. The basic
approach is to start with a set of “seed” points and from these grow regions.
➢ If the result of these computation shows clusters of values, the pixels
whose properties place them near the centroid of these clusters can be
used as seeds.
➢ Descriptors alone can yield misleading results if connectivity or adjacency
information is not used in the region growing process.
Region Splitting and Merging:
The procedure just discussed grows regions from a set of seed points.
An alternative into subdivided an image initially into a set of arbitrary,
disjointed regions and then merge and/or split the regions in an attempt to
satisfy the conditions.
R1 R
2
R3 R41 R42
R43 R44
1. Split into four disjoint quadrants any region Ri for which P(Ri)=FALSE.
2. Merge any adjacent regions Rj and Rk for which P(RjURk)=TRUE.
3. Stop when no further merging or splitting is possible.
Mean and standard deviation of pixels in a region to quantify the texture of region.
Role of thresholding:
We introduced a simple model in which an image f(x,y) is
formed as the product of a reflectance component r(x,y) and an
illumination components i(x,y). consider the computer generated
reflectance function.
➢ The histogram of this function is clearly bimodal and could be
portioned easily by placing a single global threshold, T, in the
histogram valley.
➢ Multiplying the reflectance function by the illumination function.
➢ Original valley was virtually eliminated, making segmentation
by a single threshold an impossible task.
➢ Although we seldom have the reflectance function by itself to work
with, this simple illustration shows that the reflective nature of
objects and background can be such that they are separable.
ƒ(x,y)=i(x,y)r(x,y)
Taking the natural logarithm of this equation
yields a sum: z(x,y)=ln ƒ(x,y)
=ln i(x,y)+ln r(x,y)
=i (x,y)+r (x,y)
➢ If i (x,y) and r (x,y) are independent random variable, the histogram
of z(x,y) is given by the convolution of the histogram of i (x,y) and r
(x,y).
➢ But if i (x,y) had a border histogram the convolution process would
smear the histogram of r (x,y), yielding a histogram for z(x,y)
whose shape could be quite different from that of the histogram of r
(x,y).
➢ The degree of distortion depends on the broadness of the histogram
of i (x,y), which in turn depends on the nonuniformity of the
illumination function.
➢ We have dealt with the logarithm of ƒ(x,y), instead of dealing
with the image function directly.
➢ When access to the illumination source is available, a solution
frequently used in practice to compensate for nonuniformity is to
project the illumination pattern onto a constant, white reflective
surface.
➢ This yields an image g(x,y)=ki(x,y), where k is a constant that
depends on the surface and i(x,y) is the illumination pattern.
➢ For any image ƒ(x,y)=i(x,y)r(x,y) obtained from the same
illumination function, simply dividing ƒ(x,y) by g(x,y) yields a
normalized function h(x,y)= ƒ(x,y)/g(x,y)= r(x,y)/k.
➢ Thus, if r(x,y) can be segmented by using a single threshold T, then
h(x,y) can be segmented by using single threshold of value T/k.
UNIT V
2 Marks
1. What is image compression?
Image compression refers to the process of redundancy amount of
data required to represent the given quantity of information for digital image.
The basis of reduction process is removal of redundant data.
A code word that is not a prefix of any other code word is called
instantaneous or prefix codeword.
31.Define I-frame
I-frame is Intraframe or Independent frame. An I-frame is
compressed independently of all frames. It resembles a JPEG encoded
image. It is the reference point for the motion estimation needed to generate
subsequent P and P-frame.
32.Define P-frame
P-frame is called predictive frame. A P-frame is the compressed
difference between the current frame and a prediction of it based on the
previous I or P-frame
33.Define B-frame
B-frame is the bidirectional frame. A B-frame is the compressed
difference between the current frame and a prediction of it based on the
previous I or P-frame or next P-frame. Accordingly the decoder must have
access to both past and future reference frames.
16 Marks
1. Define Compression and explain data Redundancy
in image compression
Data Redundancy:
The data or words that either provide no relevant information
or simply restate that which is already known .It is said to be data
redundancy.
Data Redundancy Rd = 1-
Compression Ratio.
` Cr=N1/N2.
Types of Redundancy
1. Coding Redundancy :
We developed this technique for image enhancement by histogram
processing on the assumption that the grey levels of an image are random
quantities. Here the grey level histogram of the image also can provide a
great deal of insight in the construction of codes to reduce the amount of
data used to represent it.
2. Interpixel Redundancy :
Inorder to reduce the interpixel redundancy in an image, the 2-D pixel array
normally used for human viewing and interpretation must be transformed in
to more efficient form.
3. Psychovisual Redundancy:
Certain information simply has less relative importance than other
information in the normal visual processing. This information is called
Psycovisual Redundant.
M
Average length p(ak)l(ak)
= k 1
I=3(0.2)+3(0.1)+3(0.2)+5(0.05)+1(0.3)+5(0.05)+4(0.1)
=2.7 bits/symbol
M
Entropy p(ak)log2p(ak)
=H(ak)= - k 1
=0.7667
Find Efficiency
There are two Structural model and they are broadly Classified as follows
1. An Encoder
2. A Decoder.
An Input image f(x,y) is fed in to encoder and create a set of symbols and
after transmission over the channel ,the encoded representation is fed in to
the decoder.
MAPPER:
It transforms the input data in to a format designed to reduce the interpixel
redundancy in the input image.
QUANTIZER:
It reduce the accuracy of the mapper’s output.
SYMBOL ENCODER:
It creates a fixed or variable length code to represent the
quantizer’s output and maps the output in accordance with the code.
Symbol Inverse
decoder mapper
SYMBOL DECODER:
The inverse operation of the source encoder’s symbol will be
performed and maps the blocks.
VI VII
I
This selection guarantees that the largest coefficient will lie in the
interval[T0,2T0].In each pass,the threshold Ti is reduced to half the value
it had in the previous pass:
Ti=1/2(Ti-1)
For given value of Ti,we assign one of four possible labels to the coefficients:
1. significance
positive(sp)
2.significant
negative(sn)
3.zerotree root(zr)
4.isolated zero(iz)
The coefficients labeled significant are simply those that fall in the outer
levels of the quantized and are assigned an initial reconstructed value of
1.5Ti or -1.5Ti,depending on whether the coefficient is positive or negative.
3. Discuss MPEGcompression
standard Introduction:
-The basic structure of the compression algorithm proposed by mpeg is
very similar to that of ITU-T H.261
-In mpeg the blocks are organized in macro blocks which are defined in the
same manner as that of H.261 algorithm
-The mpeg standard initially had applications that require digital storage and
retrieval as a major focus
Frames
I-Frames
-Mpeg includes some frames periodically that are coded without any
reference to the past Frames. These frames are called I-frames
-I-frames do not use temporal correlation for prediction.Thus the number
of frames between two consecutive Iframes is a trade-off between
compression efficiency and convenience.
P and B frames
-In order to improve the compression efficiency mpeg1 algorithm
contains two other types of frames: predictive coded and bidirectionally
predictive coded frames
-Generally the compression efficiency of P-frames is substantially higher than Iframes
Anchor frames
-The I and P frames are sometimes are anchor frames
-To compensate for the reduction in the amount of compression due to the
frequent use of Iframes the mpeg standard introduced Bframes
Group of pictures(GOP)
-GOP is a small random access unit in the video sequence
-The GOP structure is set up as a tradeoff between the high compression efficiency of
-Motion compensated coding and the coding and the fast picture acquisition
capability of periodic intra-only processing
-The format for mpeg is very flexible however the mpeg committee has
provided some suggested value for the various parameters
-For Mpeg 1 these suggested values are called the constraint parameter bitstream
MPEG2
-It takes a toolkit approach providing a number of subsets each
containing different options
-For a particular application the user can select from a set of profiles and levels
Types of profiles
-Simple
-Main
-Snr-scalable
-Spatially scalable
-High
-Simple profile uses the Bframes.but removal of the Bframes makes the
requirements simpler.
MPEG 4
-Provides a more abstract approach to the coding of multimedia.The
standard views the multimedia scene as a collection of objects.These
objects can be coded independently.
-Language called the binary format for scenes based on the virtual
reality modeling language has been developed by Mpeg.
-The protocol for managing the elementary streams and their multiplexed
version called the delivery multimedia integration framework is a part of
Mpeg4
-The different objects that makeup the scene are coded and sent to the multiplexer
-The information about the presence of these objects is also provided
to the motion compensator predictor
-It is also used in facial animation controlled by facial definition parameter
-It allows for object scalability.
MPEG7:
-Focus on the development of a multimedia content description interface
seems to be somewhat removed from the study of data compression
-These activities relate to the core principles of data compression
which is the development of compact descriptions of information
4. Discuss about Vector quantization procedure in detail
•Second
element 3 l2=0+(0.8-
0)0.82=0.656
u2=0+(0.8-0)0.1=0.8
•Third element 2
3=
l 0.312+(0.6-
0.312)0.8=0.5424
u3=0.312+(0.6-
0.312)0.82=0.54816
---------------------------------------------------------------