Leonid Yaroslavsky - Digital Holography and Digital Image Processing Principles, Methods, Algorithms PDF
Leonid Yaroslavsky - Digital Holography and Digital Image Processing Principles, Methods, Algorithms PDF
by
Leonid Yaroslavsky
Tel Aviv University, Israel
Title: Digital Holography and Digital Image Processing: Principles, Methods, Algorithms
Author (s): Leonid Yaroslavsky
ISBN 978-1-4419-5397-1 ISBN 978-1-4757-4988-5 (eBook)
DOI 10.1007/978-1-4757-4988-5
1. INTRODUCTION .....•..........................•................................. 1
REFERENCES .....•.....................•........•...........•..................... 10
References .............................................................................. 77
References ............................................................................312
References ............................................................................371
INTRODUCTION
§ Curiously enough that the initial meaning of digital holography is getting forgotten and, in
many recent publications, digital holography is associated only with numerical
reconstruction ofholograms .
4 Chapter 1
The theoretical base for digital holography and image processing is signal
theory. Ch. 2 introduces basic concepts of the signal theory and
mathematical models that are used for describing optical signals and imaging
systems. The most important models are integral transforms. All major
integral transforms, their properties and interrelations are reviewed:
convolution integral; Fourier transform and such its derivatives as Cosine,
Hartley, Hankel, Mellinn transforms; Fresnel transform; Hilbert transform;
Radon and Abel transforms; wavelet transforms; sliding window transforms.
In the last section of the chapter, stochastic signal transformations and
corresponding statistical models are introduced.
One of the most fundamental problems of digital holography and, more
generally, of integrating digital computers and analog optics is that of
adequate representation of optical signal and transformations in digital
computers. Solving this problem requires consistent account for computer-
to-optics interface and for computational complexity issues. These problems
are treated in Chs. 3 , for signals, and in Ch.4, for signal transforms.
In Ch. 3, general signal digitization principles and concepts of signal
discretization and element-wise quantization as a practical way to implement
the digitization are introduced. Signal discretization is treated as signal
expansion over a set of discretization basis functions and classes of shift,
scale and combined shift-scale basis function are introduced to describe
traditional signal sampling and other image discretization techniques
implemented in coded aperture imaging, synthetic aperture imaging,
computer and MRI tomography. Naturally, image sampling theory, including
I-D and 2-D sampling theorems and analysis of sampling artifacts, is treated
at the greatest length. Treatment of element-wise quantization is mostly
focused on the compander - expander optimal non-uniform quantization
method that is less covered in the literature then Max-Lloyd numerical
optimization of the quantization. Aseparate section is devoted to
peculiarities of signal quantization in digital holography. The chapter
concludes with a review of image compression methods. As this subject is
very weIl covered in the literature, only basic principles of data compression
and a classification and brief review of the methods are offered.
Ch. 4 provides a comprehensive exposure of discrete representation of
signal transforms in computers. In digital holography, computers should be
considered as an integral part of optical systems. This requires observing
mutual correspondence principle between discrete signal transforms in
computers and analog transforms in optical system that is formulated in Sect.
4.1. On the base of this principle, discrete representations of the convolution
integral, Fourier and Fresnel integral transforms are developed in Sects. 4.2 -
4.4. Conventional Discrete Fourier and Discrete Fresnel Transforms are
represented here in a more general form that takes into consideration that
6 Chapter 1
task. Methods for sensor signal perfection, image restoration and image
reconstruction are discussed in Ch. 8. The methods assurne a canonical
imaging system model, described in Sect. 8.1, that consist of a cascade of
units that perform signal linear transformation, nonlinear point-wise
transformation and stochastic transformation such as adding to the signal
signal independent noise. As a theoretical bench mark, the design of optimal
linear filters is used. The filters correct signal distortions in the linear
filtering units of the system model such as image blur, and suppress additive
signal independent noise in such a way as to minimize root mean square
restoration error measured over either a set of images in an image data base
or image ensemble or even over a particular individual image. In the latter
case, filters are adaptive as their parameters depend on the particular image
to which they are applied. In view of the implementation issues, the filters
are designed and work in the domain of orthogonal transforms that can be
computed with fast transform algorithms. The design of such filters is
discussed in Sect. 8.2. In Sect.8.3, this approach is extended to local adaptive
filters that work in transform domain of a window sliding over the image
and, in each window position, produce filtered value of the window central
pixel. Illustrative examples are presented that demonstrate edge preserving
noise suppression capability of local adaptive filters. In particular, it is
shown that such filters can be efficiently used for filtering signal dependent
noise, such as speckle noise. Sect. 8.4 extends this approach further to the
design of optimal linear filters for multi component images such as color,
multi spectral or multi modality images.
In real application, it very frequently happens . Very that noise randomly
replaces signal values that in this case are completely lost. In such situation,
the model of impulse noise is used. Filtering impulse noise, in general,
requires nonlinear filtering. Efficient and fast impulse noise filtering
algorithms that use simple linear filter and point wise threshold detector of
pixels distorted by impulse noise are described in Sect. 8.5. Other filters for
filtering impulse noise are described in Ch. 12.
All above mentioned filtering methods are aimed at correcting signal
distortions attributed to linear filtering and stochastic transformation units of
the imaging system model. Sect. 8.6 deals with processing methods for
correcting image and hologram gray scale distortions in the point-wise
nonlinear transformation unit.
In a broad sense, image restoration can be treated as applying to the
distorted signal a transformation inverse to that that caused the distortions
provided that the inverse transformation is appropriately modified to allow
for random interferences and other distortions that are always present in the
signals. This is how image reconstruction in holography and tomography is
usually carried out. The modification of the inverse transformation to allow
for random interferences and distortions is implemented as pre-processing of
holograms or, correspondingly, image projections before applying the
8 Chapter 1
REFERENCES
Scalar
------
Function value
Vectorial
-------
One-
Function arguments
MuJti-
(single (multi- dimensional dimensional
component) component) (l-D (2-D, 3-D,
signals signals signals) 4-D, etc.
--
signals)
/
Continuous
in
Continuous
in
~
-- Quantized
in values ""
Discrete in
arguments
vaIues arguments (quantized) (Discrete)
signals signals
Continoous signals
Signal
digitization •..1
~ /
....
Signal
.-.
I
Digital signals
I
reconstruction
one should convert analog signals into the corresponding digital ones. We
will refer to this conversion as signal digitization. Signal digitization always,
explicitly or implicitly, assumes that inverse conversion of the digital signal
into the corresponding analog signal is possible. We will refer to this digital-
to-analog conversion as signal reconstruction.
Throughout the book, we designate continuous signals by lower case
roman letters a,b, etc., and their arguments by x. Multi-component signals
will be designated by vectors of their components a = {ac J, c =1,2,••., C . For
multidimensional continuous signals, argument x is a vector variable with
components x = {x d J,d =1,2,••• ,D, where Dis the signal dimensionality.
Representative
signal
The signal spaee metries associates with each pair of signals in the space,
say, a(l)and a(2), a nonnegative real number d(a(l) ,a(2»). The method for
obtaining this number normally has the following properties:
(2.1.1)
(2.1.2)
(2.1.3)
2. Optical signals and trans/orms 15
(2.1.4)
where A V {.} designates the averaging over image data set. Local criteria
assume evaluation of signal distinctions "locally" around any particular point
k of argument of signal' s function. Applications of local criteria to the
design and substantiating adaptive linear and nonlinear filters will be given
in Chapts. 8 and 12 .
2. Optical signals and trans/arms 17
(2.1.5)
of the selected signals that are called basis signals (basis funetions).
Obviously, the set of basis functions should be selected from functions that
can't be represented as a linear combination ofthe rest of the functions from
18 Chapter2
the set, otherwise the set of basis functions will be redundant. The signal
space formed of signals generated as a linear combination (2.1.5) of N
linearly-independent basis functions is referred to as an N-dimensional
linear space.
Each signal a in N-dimensional signal space corresponds to a unique
linear combination of basis functions {cP k } , Le. is defined by a unique set of
scalar coefficients ~ k }. An (ordered) set of scalar coefficients of
decomposition of a given signalover the given basis is the signal's
representation with respect to this basis.
The "zero-signal" may be treated as a reference signal of the signal space.
The distinction of signal a from the "zero-signal" displays the individual
characteristic of this signal. Mathematically, this characteristic is described
through the concept of the signal' s norm Ilall :
(2.1.7)
This computation method satisfies the above conditions and, as we will see,
is a mathematical model of natural signal transformations such as those in
optical systems and signal sensors.
The concepts of signal scalar product and norm may be related by
defining the norm through the scalar product :
k:t;l
(2.1.10)
k=l
N-I N-I
(2.1.12)
(2.1.13)
where
Knowing the signal representation over the orthonormal basis, one can easily
ca1culate signal scalar product
andnonn
(2.1.16)
Representations tx i ,{ß ~t)} and , tx i1)} ,{ß ~l)} of signals a and b over
t )}
(2.1.17)
(2.1.18)
for arbitrary coefficients {A. r '# o} as a basis for representing any signal
vector:
N-t
aN = LCIr<Pr . (2.1.19)
r~O
(2.1.20)
2. Optical signals and transforms 21
(2.1.22)
IN = {ö k-'}, (2.1.23)
(2.1.24)
Such matrix treatment of discrete signals assumes, that signal scalar product
is defined as a matrix product
(2.1.25)
Transforms T N and T;' are called mutually orthogonal and signal vector
transformations described by Eqs. (2.1.20) and (2.1.23) are called direct and
inverse transforms, respectively. Transform matrix that is orthogonal to its
transpose
(2.1.26)
is called self conjugate matrix and the transform is called the unitary
transform. For unitary transforms, the Parceval's relationship holds:
(2.1.28)
22 Chapter 2
(2.1.29)
(2.1.31)
The function a{f) is called the integral transform of signal a{x), or its
spectrum, over continuous basis {<I> (f, x)} called transform kerne!. The
reciprocity condition for functions <p (x, f) and <I> (f, x) or the condition of
transform invertibility may be obtained by substituting (2.1.31) into (2.1.30):
(2.1.32)
where
where
(2.2.2)
Obviously, linear signal transformations are defined only for linear signal
space and linearly transformed signals also form a linear signal space.
Therefore for linear transforms, one may introduce the notion of the sum of
2. Optical signals and trans/orms 25
(2.2.4)
(2.2.5)
speetra ß(I) and a (I) of a signal and its linear transformation are also
related with an integral transform
(2.2.10)
1-- - - - - - - - - - - - - - - - - - - - - - -I
....
I ..
Input 1
Linear
transformation
...... Point-wise nonlinear
transformation
+-.
IOutput
Imaging system
~------------------------
Figure 2-3. Basie model of imaging and holographie systems
2. Optical signals and trans/orms 27
Figure 2-4. A woodcut by Albrecht Dürer showing the relationship between the light
distribution on an object and image plane (adopted form R. Bracewell, Two-dimensional
imaging, Prentice Hall Int. 1995)
28 Chapter2
Objective lens
----
------ Image
b(x. v)
------ --- ---
Objectplane (~,11)
where integration is assumed to be carried out over the object plane. Integral
ofEq. (2.3.1) implicitly shows image ofa point source located in coordinates
(~, Tl)·
In general, the point spread function is space variant: image of a point
source changes depending on its location in object plane. An important
special case is that of space invariant imaging systems in which image of a
point source is invariant to the point source location. For space invariant
systems, point spread function is a function of difference of coordinates in
object and image planes:
(2.3.2)
b(x,y)= JJa(~,Tl}h(x-~,y-Tl)d;dTJ.
'" '"
(2.3.3, a)
2. Optical signals and transforms 29
Note that integration limits in Eqs. (2.3 .3) are infinite. This is a direct
consequence of the assumption of space invariance.
Point spread function is the basic specification characteristic of imaging
systems. It is used for certification of imaging systems and characterizes the
system capability to resolve objects called imaging system resolving power.
Given noise level in the signal sensor, it detennines variance of error in
measuring coordinates of a point source ([5], see also Sect. 10.3). In
particular, one can show that, for separable PSF h{x,y)= hx{x}hy{Y),
variances cr; and cr: of the error in measuring (x,y) coordinates of a point
source are inversely proportional to the energy of the corresponding
derivatives of PSF:
(2.3.4)
------------~ ....
..
~
Coordioate
cooversion ~
Spau
invanant
~
Inverse
coordinate ...
processlug cooversion
Recording
Source of medium
Object beam
AObJ exp(i2n:<1> obi )
Reference beam
A~I exp{i2n:<1>,../ )
(2.3.5)
where A Obi exp(i<l> Obi) and Are! exp(i<l> Te!) are complex amplitudes of object
and reference beams, respectively.
32 Chapter 2
Scattered
Source of coherent
light
, /
I ~' I
J, ,
Object bellm /f
;' I I"
",,/ I I ,"
ADbjA"IAo exp(i<l> Dbj) ;' /,/' / J'
" I , t
--------T:."'-iIII ,/ " ", ' /
'/
Reconstruction
beam
Ao exp{i<l>"1 )
Consider 1-D model of free space wave propagation (Fig. 2-11). Let
complex amplitude of wave front of a monochromatic radiation in
coordinate x at the object ('input") plane is a(x). Let also hologram
("output") plane is situated at distance D from the object plane. At the
hologram ("output") plane at a point with coordinate J, this wave being
traveled a distance ~ D 2 + (x - f Ywill be accordingly attenuated and
x --_
a{x )exp[ i2rr [ ~D2 +{x- fY
A II
(2.3.6)
Obviously, 2-D point spread function offree space wave propagation from
input plane (x, y) to output plane Vx'
fy) is
34 Chapter2
(2.3.7)
distances ((x - Ix Y+ (y - I y Y t 2
involved in Kirchhoffs integral (2.3.8)
so as one can approximate
(2.3.9)
a Ix' I y
( )
= -LIa(x, y)exp[(x
00 00
in
- Ix Y+ (y - I )2]dxdy .
fJ) y (2.3.10)
that describes the so called near zone diffraction. Eq. 2.3.10 can
altemative1y be written as
2. Optical signals and trans/orms 35
-LIa{x, y}exp(Xl
00 00 + l)
irc d
[ xl + yJ ]dxdy.
exp - i2rc " ')J) y (2.3.11)
If, further, distance D is so large that one can neglect phase variations
( fl Jl J (l l)
rc +
x ')J) Y and rc x ; ; ,we arrive at Jar zone diffraetion, or
(2.3.12)
conclude that point spread function of optical systems whose input and
output planes are front and rear focal planes of a lens is
----------------------
-------------
--------
------ I
-- -----------.1
I
•
I
I
_____________ F_______ ~---------F-----------
I
Figure 2-12. Schematic diagram ofwave propagation between focal planes of a lens
(2.3.15)
It is more convenient to rewrite Eqs. 2.3.8, 2.3.10, 2.3.12, 2.3.14 and 2.3.15
in dimensionless coordinates normalized, correspondingly, by .J').J) or by
.JAF . In such normalized coordinates, obtain, neglecting irrelevant constant
factors:
uK(Jx,fy )=
-00 -00
(2.3.17)
2. Optical signals and transfarms 37
and
a F (2.3.18)
-00
00 00
Parallel beam of
monochromatic SpatiaJ Jight
light modulator
,, ~" ------- ----
,,
,-
, ... ~
~ -'
~
" ... . . .
<___ F__
~
~
~
From schematic diagram in Fig. 2-11 it follows also that lens itself can be
considered as a spatial light modulator that adds to the phase of the incident
light a component that is a quadratic function of a distance from the lens
optical axis. Indeed, a spherical wave at point ~ in lens plain from a point
source at the optical axis in the input plane has complex amplitude
(2.3 .20)
In paraxial approximation, when size of the lens is much less than its focal
distance, wave amplitude is approximately
(2.3.21)
Because this spherical wave is converted by the lens into a plain wave,
complex transparency of the lens is then
(2.3.22 )
t
: I
-r --------------- - ---------------
I I
I I
I I
I I
I I
I I
I F F I
where X" is transposed vector x. Eq. 2.4.1 can also be generally regarded
as a multi-dimensional integral transfonn assuming that X and f can be, in
general, multi component vectors.
Invertibilitv
Integral Fourier transfonn is invertible. According to the integral Fourier
theorem ([11]) signal a(x) can be found from its Fourier spectrum a(f) as
-s )]},
F
where
. ()
SIne x =sinx
-- (2.4.6)
x
of a rect-function:
1,0::;;x<1
reet ( x ) = { . (2.4.8)
0, otherwise
(2.4.9)
-SiRS] . (2.4.11)
cosS
42 Chapter 2
(2.4.12)
Changing variables Rot_s . f ~ fand by virtue ofthe fact that the Jacobian
of this coordinate conversion is
obtain finally:
(2.4.14)
Scaling theorem: Scaling signal coordinates S times causes I/S scaling ofits
spectrum magnitude and spectrum coordinates.
Proof
oe <Xl
(2.4.15)
is ~a(~ f).
2. Optical signals and transforms 43
fa(~}b(x - ~~
00
these signals.
Proof
Symmetry properties
Signal and their spectra symmetry properties are summarized in the
following equations in which symbol * means complex conjugation.
a(x)= ±a(- x)< FourierTransform )a (f)= ±a(- f); (2.4.18)
a(x)= -a(- X)= ±a' (X )< FourierTransform )a(f)= -a (- f)= ±a' (- f); (2.4.21)
44 Chapter 2
Proof:
(2.4.23)
(2.4.24)
Figure 2-15. Two types of image even symmetry observed in images reconstructed from
computer generated Fourier holograms.
(2.4.26)
a (j) = (2.4.27)
Similarly to the eosine transform, Hartley transform and its inverse are
identical. For real valued signals, the following relationships linking their
Fourier a F (I) and Hartley a H (j) spectra may be readily derived as
46 Chapter2
and inseparable
(2.4.30)
-00-00
a. (I) = 2x (2.4.32)
o
(2.4.33)
2. Optical signals and transforms 47
One can show this by replacing Cartesian coordinates (x,y) and Vx,f) in
2-D Fourier transform by polar coordinates (r,3o) and (/,8), respectively:
-00-00
"""
f fra{r )exp~2n (lx r cos3o + f y r sin 30 )}txdy =
0-"
(2.4.34)
"" "
a{/,8)= f fa{r )exp[i2n (cos8 cos3o + sin8 sin30 )r/lrdrd3o =
o -7t
"" 7t ""
(2.4.35)
From inverse 2-D Fourier transform one can derive that inverse Hankel
transform coincides with the direct one:
""
a{r)= 2n fa {/Yo{2n/r)/d/ . (2.4.36)
o
As one can see, Bessel function plays in Hankel transform role which
complex exponential function plays in Fourier transform. The role played by
sinc-function in Fourier transform, in Hankel transform is played by ajinc-
/unction ([17]) that, similarly to sinc-function as a Fourier transform of a
rect-function (Eq. 2.4.8), is Hankel transform ofthe rect-function:
""
jinc{r)= 2n frect{/Yo{2n/r)d/ . (2.4.38)
o
Sinc- andjinc -functions are compared in Fig. 2-16. As one can see from the
figure, jinc function oscillates much less and converges to zero much faster
then sinc function.
48 Chapter 2
sinc(x)
O.
,
O. .........................;............................................................. f·!~ ......·........·....·.................·............ · ........ ·<......•........ ·........ ·........ij
O. .. .... ·........+.. ·.... ·........ ·.... ·........ ·+......·..·.. ·............·.. t ..;...·......·.................... + ............................... ................................-iI
~
O.
0 ..........,.,.,..,."..,,--
-0.2.""===="",===="""===,,,,'9t'''''''"'===-========,",,,
-60 x 60
jinc(x)
0.8~1 ............................ !.................... + ............. ·.. ··· .... ·····IH··· .. ····· .... ······ .. ···!···················· ...... ·" .. ·· .. ·· .. ••·· .. · .... ··4
0.6~i····· .... ·.. ·.. ······..···· ..,·· .. ···················+· .... ····· .. ·· ....······l·H· .. ·.... ·..........·........ ·!.... ·· .... ·· .. ······ .. ··t .. ····· .. ·.. ·· .. ·.. ·· .. ···4
0.4 1;- ................................ ;.................................,........................... ·1 .. ·;·+· ..........·.... ··........ ·; ...... ·...... ·........ ·...... ·.. ".............. ·..·.............;1
0.21;-1 ................................ ;.................................;...........................1....;.. +...........................;................................"................................;1
o -------~""'~:(; ':(\T''''"~---------iI
j;-: .. ...
-0.1_ 60
-40 -20 o 20 40 x 60
The shift invariance of the Fourier transform may be tumed into ascale
invariance property if one introduces a logarithm scaling of the signal
coordinate:
x=lnz (2.4.39)
fa (Z)Zi2lt
00
(2.4.43)
2-D integral Fresnel transform was introduced in Sect. 2.3.2 (Eq. 2.3.17)
as a mathematical model that approximates Kirchhoff integral in near zone
diffraction. l-D Fresnel transform is defined respectively as
(2.4.44)
-00
Fresnel transform and Fourier transform are very c10sely connected and
may be reduced one to another by mean of phase modulation of signal and
its spectrum with chirp-functions functions chirp(x)= exp(- irrx 2 ) and
chirp(f)= exp(- i1tf2) :
(2.4.45)
-00
Therefore, the kerne 1 of the Fresnel transform is self conjugate and his
corresponding delta function is:
1 !i.
Fr± (z)= FrC(z)± FrS(z)= ~ fexp{± ix 2 ~ (2.4.49)
,,21t 0
fa
00
f fa
00 00
where
1
CiPr (J) = a(x )exp(- irr.x 2 )exP(i2rr.fx ~
-00
(2.4.55)
52 Chapter2
(2.4.56)
-00
(2.4.59)
1
a(x)=- f(lABCD(/)exp -i1t
<Xl [Ax 1 - 2.fx + Dfl ] d/. (2.4.60)
Bcl B
2. Optical signals and trans/orms 53
o
(2.4.61)
is called analytical signal for real valued signal a( x). Fourier spectrum of
analytical signals is nonzero only for positive frequencies.
54 Chapter 2
From Eq. 2.4.64 and by virtue of the convolution theorem it follows that
Hilbert transform is a signal convolution with PSF defined as
OOfsinmx dx =-slgn
1t . (m ) (2.4.66)
o x 2
1 -00 x
j
2 -00 x
1 2 -00 x
1
sin 2rrfx dx = sin 21t.fx dx =! sin 21t.fx dx +! COS 21t/x dx =
o x
~
'_
fsign(/)exp(- i21t.fx)d/ = - ~.
u
(2.4.68)
Note that these definitions are direcdy linked with the definitions of the
direct and inverse Fourier transform (Eqs. 2.4.1 and 2.4.3). If in these
definitions sign of exponents are opposite (see, for instance, [3]) , direct and
inverse Hilbert transforms also will also change their signs.
2. Optical signals and trans/orms 55
(2.4.71)
Such name is justified by the fact that A(x);;:: /a(x ~ and that functions A(x)
and a(x) never intersect and have common tangents wherever they touch
each other:
(2.4.74)
a(x) = ao(x )cos8 (x )cos 2rt/ox - ao(x )sin8 (x )sin 2rt/ox . (2.4.75)
Because Fourier spectra of functions cos 2rt/ox and sin 2rtfox are,
respectively, [8 (I - 10) + 8 (I + 10 )]/ 2 and [8 (I + 10) - 8 (I - 10 )]/ 2i ,
their Hilbert transforms are, respectively, - sin 2rt/ox and cos 2rt/ox. Then,
in the assumption that Fourier spectra of functions ao(x )cos8 (x) and
ao (x )sin8 (x) do not extend outside interval [- 10,/0]' Hilbert transform
of a(x) is:
56 Chapter 2
where A(x)= a(x )exp(i9 (x)), the form adopted for representation of
sinusoidal signals on a complex plain.
2. Optical signals and trans/arms 57
Parallel X-ny
beam y
sensitive sensors
s
Figure 2-17. Parallel beam projection tomography
1{f fa(x,
-00 XY
y)s (S - x cosS - y sin S )dxdY} exp(i21tA)/s =
(2.5.2)
-1(-00
(2.5.4)
2. Optical signals and transforms 59
obtain:
" ""
fd9 flw/a{w,9 )exp[- i21tw(xcos9 + ysin9 )}lw =
o
(2.5.6)
"
fFBproj(xcos9 + ysin9)1e,
o
where
FBproj(xcos9 + ysin9)=
oe
(2.5.7)
flw/a(w,9 )exp[- i21tw(xcos9 + ysin9 )}lw .
-00
a(w,e)
Figure 2-18. Cartesian and polar coordinate systems in Fourier transfonn domain
(2.5.8)
(2.5.9)
Let now signal a(x) is considered in a scaled coordinate system sx. Then
its convolution with PSF h() depends on both sand x :
00
In Eq. 2.6.2 signal is taken at different scales and convolution kernel is kept
the same. One can now translate signal scaling into convolution kernel
scaling by changing variables s~ ~ ~ and obtain:
s
J
a(s; x) =! a(~ )h( x - ~ )~ .
-00 s
(2.6.4)
(2.6.5)
with scale parameter .J').J) (see also Eq. 2.3.10). Haar Transform suggested
in 1910 ([35]) is yet another example of a wavelet trans form. Kernel of Haar
transform is abipolar step function
By the present time, wavelet transforms have become a very popular tool
for signal processing and quite a few of different wavelet transforms have
been suggested ([33]). Two dimensional and multi-dimensional wavelets are
usually defined as separable functions.
2. Optical signals and transforms 63
Fourier integral transform and its derivatives are all defined with respect
to signals defined on infinite intervals. In reality, however, only a finite
fraction of the signal selected by the signal sensor aperture is or should be
analyzed. This can mathematically be modeled by signal windowing. In this
way we arrive at windowed transforms:
J
00
where <p tf,~) is a transform kerne I and wt~) is a window function such
that it is elose to I within a certain vicinity to x =0 and then more or less
rapidly decays to zero:
w(O) = 1; limw(x)=O
x->oo
(2.7.2)
Windowed transforms of signal atx) are defined for every point x of signal
coordinates and usually are computed in a process of regular signal scanning
along its coordinate. In this sense they will be referred to as sliding window
transforms.
00
00
(2.7.6)
One ean give to Windowed Fourier Transform one more and very
instruetive interpretation. One ean regard Eq. 2.7.3 as a eonvolution integral
of a{x) with kernel w{1; )exp{i21t.ft ). Therefore, aeeording to the
eonvolution theorem, Fourier transform of the signal time-frequeney
representation a.{x,f) at frequeney j
A{p, f) = (2.7.7)
<Xl
Wf (2.7.10)
Because wavelet transforms are, for every scale, signal convolution they
can also be treated as signal sub-band decompositions formed by signal
filtering by filters with frequency response equal to Fourier Transform ofthe
mother function on the corresponding scale:
05
O~~O-I--~~----~"~
Figure 2-19. Arrangement ofsignal sub-bands for sliding window Fourier Transform
: :
b(x)=a(x)+n(x), (2.8.1)
where a(x) is input signal, b(x) is output signal and n(x )is a random
process statistically independent on signal a(x) and referred to as additive
noise. The assumption of statistical independence assumes that statistical
averaging applied to signal b(x) and to results of its transformations can be
carried out on statistical ensembles of signals {a{x)} and noise {n{x)}
separately:
(2.8.4)
where 0' n is its standard deviation and ii is its mean value, which, for
ASIN-model, is usually assumed to be zero, and by its correlation function
(2.8.6)
f
00
(2.8.8)
2. Optical signals and trans/orms 69
N(/)= No =a;;
(2.8.9)
Rn (x) = Noo (x)=a;o (x~
(2.8.12)
-q
p(q)= L exp(- q), (2.8.13)
q!
(2.8.14)
(2.8.15)
q-'>oo ) 0, (2.8.16)
which means that with growth of q fluctuations are becoming less and less
noticeable.
assurnes that noise is caused by two random factors: random binary process
e(x) that takes two values, zero and one, and an additive noise n(x). ImpN-
model is applicable to discrete signals. It is characteristic for signal
transmission systems with random faults. The process e(x) is specified by
the probability that e(x) = 1 (the probability of error, or the probability of
signal loss). Statistical properties of noise n(x) can be arbitrary. If random
process n(x) takes only two values equal to minimum and maximum of
a(x), it is, in image processingjargon, referred to as pepper and salt no;se.
2. Optical signals and trans/orms 71
Figure 2-21. Examples ofimages with additive and impulse noise. Observe different visibility
of different types of noise with the same standard deviation
72 Chapter2
(2.8.18)
Let the object has a uniformly painted surface such that a(~,,,) = A o.
Then:
(2.8.19)
where
If functions cos[e (I;, TI)] and sin~ (I;, TI)] are changing much faster then
point spread function h(x,y;I;,T1), or, in other words, many uncorrelated
values of exp[ie (I; ,TI)] are observed within aperture of the imaging system,
the central limit theorem of the probability theory can be used to assert that
b,e and b im are normally distributed with zero mean:
A; JJAVa {cos[e(~I)Jcos~(~z)]}h(Xd~l)h·(Xz;~z)~1 ~z =
<Xl <Xl
-co-co
A; J JR9(~I'~z}h(XI;~I)h·(Xz;~z)~1 ~2'
<Xl <Xl
(2.8.23)
a: =(breY
-- -- 1
=(b im )2 ="2 A ; JJlh(x,Y;~,l1Wd~d11'
0000
-00-00
(2.8.26)
In general, they are functions of coordinate (x, y) in image plane. For space
invariant imaging system (h(x, y;~, 11) = h(x - ~,y -11)) they are coordinate
independent:
(2.8.28)
Probability that b,e and b im take values within a rectangle d( b,e )d( b im )
is
d9 R
--exp
21t a b2
(R
---
2cr b2
2
JdR=---exp
d9 1
21t 2cr
--- JdR
(R
2cr 2
2
2
2 . (2.8.29)
b b
b 2 (x,y}=2cr; (2.8.31)
-00-00
Fluctuation of b 2 (x,y) around its mean value are called speckle noise. It is
customary to characterize intensity of speckle noise by ration of its standard
deviation to mean value called "speekle contras/':
0'2
Spekl eontrast = 6 • (2.8.33)
- b 2 (x,y)
It follows from Eqs. 2.8.31 and 32 that, for objects that scatter irradiation
almost uniformly in the space, speckle contrast of their image obtained in
coherent imaging systems is unity:
a) h)
Figure 2-22. Images with speckle noise for imaging systems with high (a) and low (h)
resolving power.
2. Optical signals and transforms 77
References
29. H. Barrett, The Radon Transform and its Applications, In: Progress in Optics, ed. By E.
Wolf, Elsevier, Amsterdam, vol. 21, pp. 219-286, 1984,.
30. S. R. Deans, Radon and Abel Transforms, in: The Transforms and Applications
Handbook, Ed. AD. Poularikas, CRC Press, 1966
31. Y. Meyer, Wavelets: Algorithms and Applications, SIAM, 1993
32. S. G. Mallat, A theoty for multiresolution signal decomposition: The wavelet
representation, IEEE Trans. Pattern Anal. Machine IntelI., PAMI-31, pp. 674-693,1989
33. S. Mallat, A Wavelet Tour ofSignal Processing, 2nd edition, Academic Press, 1999
34. M. VetterIi, J. Kovacevic, Wavelets and subband coding, Prentice Hall, Englewood
Cliffs, NJ., 1995
35. A Haar, Zur theorie der orthogonalen funktionensysteme, Math. Annal., Vol. 69,pp.
331-371, 1910
36. A Papoulis, Probability, Random Variables and Stochastic Processes, McGrow-Hill,
N.Y., 1984
37. J.W. Goodman. Statistical Properties of Laser Speckle Patterns, In: Laser Speckle and
related Phenomena, J. C. Dainty, Ed., Springer Verlag, Berlin, 1975
Chapter 3
Uk = fa{x}rid){x)c/x (3.2.1)
x
(3.2.2)
)
Functions ~!' (x)} characterize point spread functions of signal
reconstruction devices (such as, for instance, computer display), or their
reconstruction aperture functions. Approximation symbol in Eq. (3.2.2) is
used to indicate that discretization ~!d)(X)} and reconstruction ~!')(x)}
3. Digital representation 0/signals 81
The simplest method for generating set of functions from one "mother"
function is that oftranslation (coordinate shift):
(3.2.3)
Functions generated in this way are called shift basis funetions. Elementary
shift interval ~ is called the diseretization interval. It is usually, though
not always, the same for the entire range of signal coordinates x. Positions
{x =k~} are called sampling points (for images, raster points).
Two family of functions exemplify shift basis functions: reetangular
impulse funetions
where
82 Chapter 3
rect{~ }= { 1,
O~~ <1
(3.2.5)
0, otherwise
and sinc-functions
(3.2.6)
where
.
Slnc
{}
x =sin
--
x (3.2.7)
x
(3.2.8)
used as the discretization basis, signals are represented their mean values
over discretization intervals:
u k =-
1
fa{x}rect (x - kj},x) dx=-
1 (k+l)ll.x
Ja{x)1x=a(kj},x). (3.2.9)
j},x x j},x j},x kll.x
(3.2.10)
and signals are reconstructed with this basis in a form of piece-wise constant
functions (Figure 3-1):
(3.2.11)
1/2.ix 1 ( 1tX )
fexp(- i2rifx"pj = -sinc - , (3.2.12)
-1/2.ix Llx Llx
Figure 3-1. Continuous sinusoidal signal and its eopy reeonstrueted after diseretization with
rectangle impulse diseretization and reeonstruetion basis funetions
L1x
., ./
--"--------
I
I
\j.. :::~ ~inc(~ I ix) : x
I I I
basis:
-
1 oof.
SIßC
(1t (x - kAx)).SIßC(1t (x -IAx))dx =
Ax _00 Ax Ax
~ _!Sine(:: )sine( 1t (x - t-1)Ax )) dx =
l/2ar
-l/2Ax
fexp(- i21tf(k -1)Ax}tJf = sinc(1t (k -/))= Ö (k -/)= {IO,k*1
'
k=1
(3.2.14)
this basis:
(3.2.15)
in the range [-li 2Ax,1I 2Ax] (see Eq.3.2.12), basis composed of them
generates signals that belong to the class of band limited functions. It also
follows from Eq. (3.2.15) that coefficients ~ k} of discrete representation of
band limited functions are equal to functions' sampies at equidistant points
{x = kAx}:
(3.2.16)
3. Digital representation ofsignals 85
This is where the name "sampling functions" derives from. According with
the sampling theorem (see Section 3.3), the sampling functions are a gold
standard for the discrete representation of signals with shift basis functions.
Eqs. 3.2.11 and 3.2.15 can also be regarded as convolution conv(.,.) of
signals
(3.2.17a)
This justifies yet another name for shift basis functions, the eonvolution
basis funetions, and treatment of Eq. (3.2.17, b) for reconstruction of
continuous signals from their discrete representation as eonvolution based
interpolation.
(3.2.18)
Apparently, there are only a few examples of such sealing basis Junetions
also called Fourier kerneIs [1]. All of them are derivatives of sinusoidal
functions {cos(21tu / X)} and {sin(21tu / X)} . These functions are
orthogonal. Functions {cos(21tu / X)} when used as a reconstruction basis,
reconstruct signals
N-l
a{x)~ aCe) = La}e) cos{21tu / X) (3.2.19)
k=O
86 Chapter 3
that are periodical, with period X, and even ftmctions (a{x)= a{- x)).
Functions {sin(27tkx I X}}used as a reconstruction basis reconstruct signals
N-l
a(x}~ a(s) = La}s) sin(21tkx / X} (3.2.20)
k~O
that are also periodical, with period X, and odd ftmctions a(x) = -a(- x) .
Signal representation coefficients for these bases are found, respectively, as,
1 X/2
a~ = - fa(x}cos(21tkxl x)ix (3.2.21)
X -X/2
1 X/2
aZ = - fa(x}sin(21tkx I x)ix (3.2.22)
X -X/2
In reality, signals are, in general, neither even nor odd functions. There are
two ways to avoid this obstac1e in using sinusoidal basis functions. One way
is form, from signals taken at interval [0, X], auxiliary even or odd signals
defined on interval [O,2X] of double length as, respectively:
and
_( ) 1 2X_ 1 x
Ci/ =- fii(x)cos(1tkxl x)ix=- fa(x}cos(1tkxl x)ix (3.2.25)
2X 0 X 0
3. Digital representation 0/signals 87
_() 1 2X_ 1 X
a k
s = - fii(x)sin(1tkx/ X)ra=- fa(x)sin(1tkx/ X)ra (3.2.26)
2X 0 X 0
(3.2.27)
or
(3.2.28)
1 X/2
ak =- fa(x )exp(- i21tkx / X)ra (3.2.31)
X -X/2
88 Chapter3
.::::::::::::::::::,i::::::::::::::::::::l.·.·.·.·.·.·.·.·.·.·.·.·.·.·.·.·.·.·.·.·.·,1.·.·.·.·.·.·.·.·.·.·.·.·.·.· ... ~:t. . . . . . . . . L......... 'r v ..... i···· ..·.. j................· ·.. ~·
""1'" ...............~.
...................i..............""".! ............"" . . . !..
'. i .......................... '.':.......................................
, 't:"......................................... ;:",;.................
j : .. : .
.
.......... ),..... · · · i j · .. ········ .. j·.. ··+·. ·!....·..............
i
;.......!
.. ·.·.·.·.·.·.·.·.·.·..·.·.·.·.·.·.·.I,·.·.·.·.·.·.·.·.........................••...j"
k
exp(i2nu / X) = TI exp(i2nx / X) (3.2.32)
1=1
There exists yet another family of orthogonal functions buHt with the same
principle, Walsh functions. Walsh functions are distinguished by the fact
that they can assume only two values; either 1, or -1. They are generated by
multiplication of clipped sinusoidal functions called the Rademacher
functions
(3.2.33)
00 Ge
where k~C is the m-th digit of the so-called Gray code of number k. Gray
code digits are generated from digits {km} of binary representation of the
number
= L km 2 m;
00
k km = 0,1 (3.2.35)
m=O
(3.2.36)
90 Chapter3
is useful, where
= x / X = L~m 2- m; ~O = 0;
00
The Walsh functions are orthogonal on the interval [0, X]. Because of their
origin as a product of Rademacher functions, multiplication of two Walsh
functions results in another Walsh function with shifted an index or
argument. The shift, called dyadic shift, is determined through the bit-by-bit
modulo 2 addition offunctions' indices or, respectively, arguments:
(3.2.40)
00
of vectors ~~c} and {;m+l}' In this sense one can treat basis of Walsh
functions as ascale basis. Multiplicative nature of Walsh fimctions and
exponential ones justifies yet another name for these families of bases,
multiplicative bases.
Walsh function are akin to complex exponential functions
{exp(i21tkx / X)} in one more respect. This can be seen if (-1) in Eq.
(3.2.27) is replaced by exp(i1t):
(3.2.41)
'asO o
""'"'.- --
As one can see from the figure, an important parameter that unites both
families of functions is the number of zero crossing which coincides, for
these functions, with their index k. As it was l11entioned above, for
sinusoidal functions, this parameter has an association with the notion of
frequency. For Walsh functions, this parameter obtained name of
"sequency"([2]).
Sequency wise ordering of Walsh functions is not accidental. In
principle, ordering basis functions can be arbitrary. The only condition one
should obey is matching between signal representation coefficients {a. k } and
discretization and reconstruction basis functions ~ id)(x)} and ~ (x )}.ir)
However almost always natural ways of natural ordering exist. For instance,
for shift (convolution) bases the natural ordering is according to successive
coordinate shifts. Sinusoidal basis functions are naturally ordered in
frequency of signal Fourier transform. This why sequence (number of zero
crossing) wise ordering of Walsh functions can also be regarded natural.
However, the advantage of such an ordering is more fundamental. The
most natural requirement to ordering basis functions for signal representation
is the ordering for which signal approximation error for finite number N of
terms in signal expansion over the basis monotonically decreases with the
growth of N. For Walsh functions, it happens that, for conventional analog
signals, sequency wise ordering satisfies this requirement much better than
other methods of ordering. Figure 3-5 iIIustrates this feature of Walsh
function signal spectra. Graphs on the figure show averaged squared Walsh
spectral coefficients of image rows as functions of coefficients' indices for
sequence wise (Walsh) ordering (Ieft) and for Hadamard ordering (right). In
92 Chapter 3
oe .....!.......!.... ~ .. .. .: .. ... ",' (1.1 ..... ~ ... .. ! ... ... ~ ... .. ..' .. .. ..'
I I I I
001 ..... ;- ..... ';' .... ~ ........, ...... "', 01 .....:-...... ';' ... .. ~ ... ... ., .. .. -,
'1111 ..... I'" .. .. , ...... , ...... "'" ...... -. 1)0 ...... ,. ...... , .... , ..... ""I ...... -.
(JS
(1 4
.....
_ ..
.
~
-
,
.....
.. -
.
" ..............
...............
• ,
. J
,
..
,
......
... ..
.
_,
..
I
0' ...... Ir. ... ... , ... .. ~ .. ... J ... .. _,
:: t~_~~~_: ~ ~ : ~ ~ ~ ~ ~ ~
(I) .... 03 ..... -
Figure 3-5. Comparison ofimage Walsh function spectra in Walsh (Jeft) and Hadamard
(right) ordering
3.2.3 Wavelets
where fit is index of the most significant non-zero digit (bit) in binary
representation of k (Eq. (3.2.35)) and [k ]mod 2 m is residual from division
of k by 2 m.
The Haar functions are orthonormal on the interval [0, X]. Graphs of first
eight Haar functions are shown in Fig. 3-6.
haro(xF:::::: ..................... .
har6 (x).
'~:_i
The figure clearly reveals the way in which Haar fimctions are built. One
can see from the figure, that Haar functions form groups in scale and that
functions within each group are generated by shift in coordinates with shift
interval coordinated with the scale. One can also see that signals
reconstructed from their discrete representation over Haar basis functions
belong to the same family of piece-wise constant functions as signals
reconstructed in the base of rectangle impulse basis functions.
94 Chapter 3
(3.2.44)
Inverse:
I N-l
10FT = { r.; La k
kr)}
exp - i2rc-
(
...;N k=O N
Diserete eosine
Transforms (DCT) Direct: OCT ={.):N CO{rc (k +:: 2}r ) }
Inverse:
((k + 112))}
DCT-IV
J2
DCTIV = { JNcos rc + I/2Xr
N
IOCTIV = OCTIV
Diserete Sine
Transform (DST)
OST = {_I_
.)2N
Sin (2rc (k + lXr +
2N + 1
I))}
lOST = OST
DST-IV
OSTIV = {"';N Sin( 2rc (k + 11 ~~ + 1I2)J}
lOSTIV = OSTIV
Walsh
Transform WAL = {_l_
JN (_I)Ik~:m-,rm
.. } ., n =I
~O ogz N ; {rm }
are digits of binary representation of row index r;
~~:m-l} are digits of Gray code of bit reversed digits
{kn-m-1 } of binary representation {km} of column
index; m = O,I,••• ,n -1; {k;c = k s EB ks+ 1 }; EB - modulo
2 binary addition.
Haar Transform
HAAR = {2~ (_ly·-m-, TI 0(k EB rm )1;
...; N m=O
n- m
J
n = logz N; fit - index of the most significant
nonzero digit in binary representation of row index r;
96 Chapter 3
There are many ways to introduce and define wavelet transform as weIl
as there is unlimited number of modifications of wavelet bases. Fig. 3-6
represents one of the most simple and straightforward method, the
algorithmic one. As it is shown in the figure, signal wavelet decomposition
is made in several steps using two basic signal processing procedures:
decimation and interpolation. As a result, signal of N sampIes is decomposed
into log2 N sub-signals with descending resolution (number of sampIes)
each of which represents different sub-bands of the signal, from high to low
frequencies with the lowest level representing signal dc component The way
how low pass filtering for decimation and how signal interpolation from its
decimated copy are implemented defines type of the wavelets. This method
of signal representation is known also by names multiresolution, sub-band
representation and pyramidal coding (see, for instance, [3,7-9]).
N-l
a(x,ro) == a(x,ro) = LU k {ro)p k (x) (3.2.45)
k=O
l ll
lE(rof = la(x,ro)-a(x, ro f dx = a{x,ro)- ~Uk{ro)Pk{xf dx
(3.2.47)
98 Chapter3
Optimal values of {a. k (co)} that minimize MSE can be found by equaling to
zero derivatives of Eq. (3.2.47) over {a. k (ro )} :
(3.2.48)
where * stands for a complex conjugate. This means that, for basis {<P k (x)}
used as a reconstruction basis, the reciprocal discretization basis is composed
OffllOctions ~; (x)}.
From Eq. (3.2.48) it follows that minimal MSE is equal to
(3.2.50)
argmaX{AVn(~l
tpk (x)}
JJa{x,ro)a· {y,ro)p; {x)P
x k=O
k (Y)dxdY)} =
where AVn means averaging over the signal set n. Function Ra (x, y)
(3.2.55)
This means that optimal bases functions are eigen functions of integral
equation Eq.(3.2.56) with kerne I defined by the correlation function (3.2.52)
of the set of signals ([ 10, 11 D. Such bases are called Karhunen-Loeve bases.
Similar reasoning can be applied to integral representation of signals. Let
signals a(x,ro) are approximated by their integral representation over self-
conjugate basis <P (x, /):
(3.2.58)
(3.2.61)
In this case the basis optimality condition of Eq. 3.2.60 takes the form:
co
fRa (x - Y}P(y,j)ty = A(j)P * (x,j). (3.2.62)
-co
As one can see this condition is satisfied with exponential basis ftmctions
Ja {J,ro )P{x,J)lJ .
F
ä'{x,ro)= (3.2.64)
-F
( 3.2.66)
( 3.2.67)
( 3.2.68)
N-l }
a- m - {
- a k-- ~ar(ol)'
<Pk,r . ( 3.2.69)
T= argmaXAvn[rr!Olf] =
~',rJ r=O
~."J
Denote
102 Chapter3
(3.2.71)
T= argmax{I:1212 Ra (k,n}Pt)<i>n,J}
~k" J ,=0 k=O 1=0
=
art~rx{~(~(<i>n~J~Ra(k,n}Pt,,)} . (3.2.72)
( 3.2.73)
that become equality iff {at = 'Abk }, where 'A is an arbitrary constant, it
follows that the transfonn optimization equation is a transfonn is reduced to
the equation
( 3.2.74)
N-I
(3.2.76)
and
(3.2.77)
3. Digital representation 0/ signals 103
(3.2.78)
(3.2.79)
Eqs. (3.2.55, 73-76) determine MSE optimal bases and transforms via
signal correlation function, correlation matrix or outer product. One can also
deduce from the above reasoning that each particular orthogonal transform
can be regarded to be MSE optimal for a certain c1ass of signals whose
correlation function (matrix) satisfies these equations.
The theoretical value of Karhunen-Loeve, Hotelling and SVD transforms
in signal processing is that they provide the lower bound for compact (in
terms of number of terms in signal decomposition) signal representation for
MSE criterion of signal approximation. Their practical value is, however,
quite limited because of high complexity of generating transform basis
functions and matrices and computing signal representation coefficients.
= I IadP~~;(k,I)= I
N,-IN2-1 N, - IN2 - 1
Ur •• IadP~dl)(k }P~dl)(l)=
k=O /=0 k =O /=0
I I
N, - l N 2 -1
<i> ~dl)(l) ak./<i> ~dl)(k) (3.2.80)
k =0 /=0
As one can see from Eq. 3.2.80, direct computation of 2-D sum requires
N; N; multiplicationlsummation operations while two separable 1-0
summations require N1N2tN] + N 2) operations. For large 2_D arrays,
saving is very substantial. Note that in separable two- and multi-dimensional
bases multiplicand 1-0 basis functions are not necessarily should be
identical, though most frequently they are.
DFT DFT
(real) (imaginary)
DeT DST
Walsh
. ,..
!- ~-
~ bo
• I
I-
".
I I
,a J
r ~
Haar
"
F f-"' ;e ";}' l' .I
io;:o ~ ~ ~
" .
~ 30
Figure 3-7. Examples of8x8 sets of2-D basis functions for DFT, DST, Walsh and Haar
transforrns displayed as images
3. Digital representation 0/ signals 105
Cumsum(Spectrum).DFT:r(O.051);DCT-<:(0.037);WaISh.g(0.072);Haar-b(0.044
1~----~.------~
. ----~.~.~
--
~-. ~-.~ ..
0.9 __ . ___ t. . . . . . . . .. _ .. ' ........ - ~-::.;.: ........ • ' ..
0.3
./.
. . • •:;
.
. , • .
.
• • • '.' .
,
• . . . '.' • . • • • '.'
.
0.1 · · · · · · , ' · · · · · · , · · · · · · . · · · · · · · . ·
Figure 3·8. Comparison of image DFT,DCT, Walsh and Haar spectra. Graph illustrates the
speed with which magnitude wise sorted spectral coefficients converge to the entire image
signal energy.
a) b)
h)
Figure 3-9. Different types of vision organs in the animal kingdom. In all types of eyes one
can see devices for image discretization by sampling
3. Digital representation 0/ signals 107
Cl k = Ja {x }P Ax - kiU}dx, (3.3.la)
x
and its reconstruction over basis formed by the corresponding shifts of point
spread function of the reconstruction device is
C1!J
Cl k = Ja{x)PAx-kAx)dx (3.3.2a)
-co
00
obtained from signal a(x) by its convo1ution with a point spread function of
the discretization device. A model of the discretization device that
corresponds to such a representation is shown in Fig. 3-10.
t
Clock x=k8x
00
with a filter whose impulse response is equal to point spread function <P r (x)
of the signal reconstruction device:
00
er;
äl.. <Q
{ad/(kiU)}
I x=k
a,(x)
Clock
Figure 3-11. Mathematical model ofthe device for restoring continuous signals from their
sampIes
where
f
00
00
and
110 Chapter 3
k=-oo k=-oo
00
and
00
00
Then obtain:
(3.3.15)
(3.3.16)
Therefore,
(3.3.17)
3. Digital representation 0/ signals 111
which means that sampling signal adf (x) results, in Fourier domain, in
periodical replication of its spectrum with aperiod inverse to sampling
interval Llx . This phenomenon is iIIustrated in Fig. 3-12.
I
I1 11 11 11
11 11 11 11
"
11 rI
11 "
11 rI
11
I I 11 1 I I '
I I I I I I I I
" I I I I ,I
I I I I I I I I
I I I I I I I 1
I \ I \ I \ ,\
I I 1 I I I I 1
I I r \ I I I I
I I 1 1 I I I 1
I I 1 1 I I , 1
J I 1 1 1 1 J 1
I 1 I 1 I 1 I I
I \ I I I I 1 \
J .1 J I J I J I
1 1 1 1 1 I 1 I
1 I I I 1 I 1 I
I 1 J I J 1 1 I
I \ I \ I \ I \
I 'I \ I \ I \
~ 1/ ~x"'" /
Figure 3-12. Illustration ofEq. (3 .3. 17) ofperiodical replication ofsignal spectrum in signal's
sampled representation.
From Eq. (3.3 .18) one can easily see what type ofsignal distortions occur
in the case of signal discretization by sampling. One can also see that perfect
(distortion less) signal restoration from the result of its sampling is possible
only if signals are band-limited:
where
I, 0 ~ x ~ 1
reet ()
x ={ . (3.3.22)
0, otherwise
a(x)= IU
k=-oo
k sinc[n(x - k~)/~] (3.3.25)
provided
1 0 0 0 0
uk = - fa(x)sinc[n (x - kAx)/~]dx = fa(x)l (x - kAx}Lx = a(kAx)
Ax -00 -00
(3.3.26)
Eqs. (3.3.25 and 26) formulate the sampling theorem: band-limited signals
can exacdy reconstructed from their sampIes taken at the inter-sample
distance that is inversely proportional to the signal bandwidth.
The above analyses resulted in the sampling theorem establishes also a
link between the volume of the signal discrete representation, i.e., the
3. Digital representation of signals 113
N=X/Ax=XF (3.3.27)
(3.3.28)
If 2-D separable basis functions are build from 1-0 shift basis functions
~~I)(XI -kAxI);<p~2)(X2 -kAx2)} and ~~I)(XI -kAxI } <p~2)(X2 -kAx 2)}
2-D signal sampling is carried out in sampling points arranged in a
rectangular (Cartesian) raster which corresponds, in the domain of 2-D
Fourier transform, to periodical replication of signal spectrum in Cartesian
co-ordinates as it is shown in Fig. 3-12. Yet another example of a periodical
replication of image spectra in Cartesian coordinates is shown in Fig. 3-13
(b).
One can see from these figures that the Cartesian periodical replication of
2-D signal spectrum does not necessarily ensures efficient spectra dense
packing, and, correspondingly, least dense arrangement of signal sampling
points for minimization of the volume of signal discrete representation. In
general, tilted co-ordinate systems selected accordingly to the shape of the
figure that encompasses signal 2-D spectrum will secure better spectra
packing (Fig. 3-14 c,d). When sampling is carried out in a tilted co-ordinate
system, sampling functions are not anymore separable functions of two
Cartesian variables.
114 Chapter 3
Figure 3-13. 2-D separable image sampling in a rectangular raster and periodical replication
of its spectrum in a Cartesian co-ordinate system. Image at the bottom shows spectrum
aliasing (seen as dark squares around main spectral lobes) that takes place when sampling
interval is too large.
3. Digital representation 0/ signals 115
a) b)
"I. '.. ',. ',' ',,',.', .'1. '.' '.' ',I',. ';.1;..:,~
•; ••:. I:.•: .•: ••: ..:. "I.•: .••', .:••:. ':!t-';': "I. "••
• I . ' . ' . ' . ' I ' . ' . ' . ' . ' . ' . ' . ' . ' ....a;;• •
_ ...t
."1 ....
..' ~~ .... ' .. '
•• ~.',..J:
.. '
• ••.•••..•
.".~...,!~!.~ .......~...e-...a......
::::::::::~\~~~~:.:::::::::::::::::::::::::::::::::::::::::.::::.::::
;~..,;;.'.:.•..:....:•...:••..:..":'.'.:....:..•.:....:..":'.'.:.:..:.:....
., ."" " " " " ,',,',' ,',' .' .'.' ....
c) d)
Figure 3-14. An example ofan image spectrum (a), its periodical replication in rectangular
(b) and tilted (c) coordinate systems and tilted discretization raster (d)
b) c)
,/
"
...........................~I- --------
f.
d) e)
· ··.........
. . . . . . . . .....
...........
. . . . .. .
.... ..
· ·.....
. . . . .. ......
. . .. .. .. .
· ..........
... ........ .. . .. .
·· ............. . ..
, .............. .
··· ...............
.. . . . . . . . . . . .. .
· .......................
.... .
· · ........................... . .
· ·.. .....
. . . ................. . .
f)
Figure 3-15. 2-D sampling: a) figure in frequency domain that i1lustrates different sensitivity
of human visual system to different spatial frequencies : higher for horizontal and vertical and
lower for diagonal ones; b), c) - periodical replication ofthe figure a) in a conventional and
45°-rotated rectangle co-ordinate systems, respectively; d), e) - corresponding image
sampling rasters; f) an example of an image taken from a conventional book printing
3. Digital representation 0/ signals 117
..'
• .' • •,. .' • .• •..
''.. •• .' .' .'• • •
'I'
....i.,.....'I.'..'.'
'.
~ -- --... -
e · . ' '. ' •
-
• . ' ,. ,'
.'.........'.'.'...,.'. .
'..
" '
..
' ' '
..,.•••1-'.'
'. ' i. '
. .
I I. • . ' . ' •• ," I
• I•. • '
••• '
,' . ' ,.
..• •• • •• •• •. ..• '
' ' '
'
b)
a)
·· · ..
... ......... . . . .. . ... .. .. .. . .
· . ....... . .. ... . . ......... . .
· · . ......... ...... . ....... .. . ..
·· ·· ..
... . .
.
.. ... . .. ..
. . ... . ...
. .
. . .. .... . ..
. ...
... .. ..
. . . ..
· ... . . .. . . . .. . . .. .
·· · ..
. .................. ....
....... ........
. .
·' .·'.... ................... ..... . .. .
'.'
c)
Figure 3-16. Image spectra replication for Cartesian (a) and hexagonal (b) sampling: the same
number of circular spectra occupy less area when they are packed in a hexagonal co-ordinate
system then in Cartesian co-ordinates; c) - hexagonal sampling raster; d) Trichromatic retinal
mosaics: hexagonal arrangement ofcones ofhuman eye retina (adopted from [14]).
Rotating and tilting discretization rasters are not the only methods for the
optimization of 2-D and multi dimensional sampling. Two other methods are
"d;scret;zat;on w;th gaps" and d;scret;zat;on w;th sub band
decompos;t;on.([ 15])
Method of the discretization with gaps is based on the use of signal
spectra symmetry. It is ilIustrated in Fig. 3-17. Let signal spectrum is
bounded as it is shown in Fig. 3-17 a). If one rotates the signal by 90° and
superimposes the rotated copy onto the initial signal, a composite signal is
obtained whose spectrum perfectly fills a square. Therefore, the composite
signal can be optimally sampled in Cartesian co-ordinates. One can see that
for all groups of four composite and initial signal sampies centered around
the rotation center, the following relationships hold:
(3 .3.29)
118 Chapter 3
from which it follows that, from four sampIes tZl' zp Z3' Z4 J of the
composite signal one sampIe, say, Z4' is redundant because it can be found
from the other three ones:
(3.3.30)
~+j
' _0
a. a,
d e
c
Figure 3-17. Discretization with gaps: a) - a figure bounding a signal spectrum (shaded); b) -
the spectrum pattern obtained by superimposing the signal rotated by 90° and periodically
extended in Cartesian co-ordinate system; c) the discretization raster corresponding to the
spectrum's pattern (b); d) - group offour initial signal sampIes around the rotation center and
directions of their translation after the rotation; e) - the corresponding group of four sampIes
ofthe initial signal plus its 90° rotated copy. Empty circIes in (c) denote redundant sampIes.
3. Digital representation ofsignals 119
Subband I Subband 2
Subband6
Signal
spectrum 11
~~
+ -~
~-
+0+ + I..t.:~
'
~~
V-
Signal spectra
Figure 3-19. Strobe effects in sampling l-D sinusoidal signals. Graphs on top show sinusoidal
signals of different frequencies lower and higher the frequency that corresponds to the chosen
discretization bandwidth. Image on bottom shows spectra (bright spots) of sampled sinusoidal
signals of frequencies that increase from top to bottom. One can see that frequency of
sampled sinusoids increases until it reaches the border of the bandwidth. After that the
increase ofsignal frequency causes the decrease ofthe sampled signal frequency.
3. Digital representation 0/ signals 121
Distortions caused by inappropriate sampled signal low pass filtering for the
reconstruction exhibit themselves in appearance of signal high frequencies
doubles from spectrum replicas not cleaned out by the reconstruction filter.
Für sinusoidal signals, their doubles cause low frequency beatings in the
reconstructed signal that appear as müire patterns. Owing tü this reason,
these aliasing effects are called moire effects. In images, moire effects
usually appear in form of "pixelation" illustrated in Fig. 3-21. Fig. 3-22 by
Salvador Daly illustrates artistic use of the "pixelation" effect.
baqe recovery U1d. more senec'8lly, ,Ign. . . . . -I~ .ud. iaoIe tcDcnUY. " .
'OblclM fhlll ue &mOng the most rundam.· ~"1Ift' 81lOIII tbt,CDQII f'aDcIam
bC)' Ülvolve overy known ~fTom 1M . , .....,... .....,. lc8o'MI taJe-rroa dIoI
:tamining ettc IInIcturc or unrc:s.oLved sr.,:ca " ... cbc lInIaarc or,....achal AG
ent or the tinie9t motecufes. Sialt.d in 116 I _ or,. liaiaI meIe kr Swc:d ia iu. I
1:2 sampled and reconstructed image 1:2 sampled and reconstructed image
with no low-pass pre-filtering and with "box" pre-filtering and "box"
"box" post-filtering post-filtering
Figure 3-23. Influence of sampling artifacts on image readability: image sampling and
reconstruction with an ideal low pass filter and with a "box" filter that computes signal
average over a square box of 3x3 sampies around each pixel.
124 Chapter 3
In this section, we will derive formulas that can be used for numerical
evaluation of discretization artifacts. For the sake of simplicity, I-D case will
be analyzed. For 2-D signals, one should regard signal and spectral
coordinate and indices as two-component vectors.
For discretization aperture <J> Ax), sampling interval Ax and signal
reconstruction aperture <J>r(x), signal sampies at points {(k + u)Ax}, where
u!Jx is a shift in positioning discretization aperture with respect to signal
coordinate system x, are measured as (Eq. 3.3.1a):
(3.3.31)
where
CL)
N
ar (x)= ~:ak<J> r [x - (k + v)!Jx] (3.3.33)
k=-N
where 2N + 1 is the number of sampies used for the reconstruction and v!Jx
is shift of the reconstruction aperture with respect to the signal coordinate
system.
Find an explicit relationship between a(x) and ar (x) that involves
discretization and reconstruction device parameters:
N
ar (x)= Lak<J>r [x - (k + v)!Jx] =
k=-N
3. Digital representation 0/ signals 125
1 00 N 00
(3.3.34)
where
(3.3.35)
-00 -00-00
00
Ja(p)H(J,p}/p, (3.3.36)
-00
00
1
J J L cP
00 00 N
LU d [~ - (k + u)LU}P r [x - (k + V)LU ]exp[i21t (jX - p~ )]dxd~ =
-00 -00 k=-N
126 Chapter3
1 N
L J Jep AI; - (k + u)LU}p, [x - (k + v}ßx]exp[i21t (jX - pI; )]dx~ =
00 00
LU
k=-N -00-00
1
L Jep AI; - (k + u)Ax]exp(- i21tpl;)tI; x
N 00
LU
k=-N -00
00
where <I> d (. ) and <1>, (. ) are frequency responses of signal discretization and
reconstruction devices. Consider the summation tenn in Eq. (3.3.37):
N
L exp[i21t (I - p)kLU] =
k=-N
exp[i21t (f - pXN + 1)LU] - exp[- i21t (I - p)NLU] _
exp[i21t (I - p)Ax] - 1 -
sin(n(f - pX2N + 1}ßx] =(2N + 1) sin[1t(f - pX2N + 1)LU] .(3.3.38)
sin[1t (f - p}ßx] (2N + 1)sin(1t (I - p)LU]
Using Eq. 3.3.38, overall frequency response of the signal discretization-
reconstruction procedure can be written as
Um a,(f)=
N~oo
and reconstruction devices. Its spectrum is described by the first term of Eq.
3.3.41. The second component is aliasing one. Its spectrum is described by
the second term of Eq. 3.3.41. It consists of periodical replicas
3.3.32).
In the numerical evaluation of signal distortions caused by the
discretizationJreconstruction, it is devisable to distinguish two types of
distortions that are described by the above two reconstructed signal
components: distortions owing to the signal blur by the discretization and
reconstruction apertures and aliasing distortions. The former are completely
specified by deviation of reconstructed signal spectrum from the initial one:
(3.3.44)
(note that shift factor exp[i21t (u - v}Ax] in the evaluation of this error can
be neglected as irrelevant). For the evaluation of the second type of
distortions, the most straightforward measure is that of its energy which, by
Parceval theorem, is
(3.3.45)
-<X)
where
3.3.47)
/
-1 -112 o 1/2
High pass
Low pass
filtering
filtering
-I -1/2 0 1/2 I
(\,. .....Demodulation... / )
V a H (/) V
/
o -1/2 o 1/2
n
-1/2 1/2
n Subsampling
~
Subsampling
~ U2 U2
Figure 3-24. Multi-resolution sampling explained for a I-D signal in Fourier domain
130 Chapter3
....
•
..... High pass
r. Demodulation Subsampling JJ 2 f-+
filter 1
Lowpass Channel2
filter 1
....
•
..... High pass
r. Demodulation Subsampling JJ 4 ~
filter 2
Lowpass Channel3
filter 2
.... High pass .... Demodulation
f+ Subsampling JJ 8 -.
filter 3
..."...............•.......•.•..•..•..........................
Channel n
Subsampling U2 n
On each step, bandwidths of low pass and high pass filters, and therefore
signal sampling rate are half of those on the previous step. On the last n-th
step, low pass component obtained on this stage has bandwidth equal to
1/ r -th fraction of the initial signal bandwidth and it is sampled
accordingly with 1 / 2 n -th of the rate that corresponds to this initial
bandwidth. Signal reconstruction from its multi resolution sampled
representation is also carried out component (channel) wise. Each sampled
high pass component is reconstructed individually accordingly to its
bandwidth and its spectrum is shifted correspondingly and than all
reconstructed components are added up to form the entire reconstructed
signal.
Obviously, the described multi resolution sampling does not reduce the
number of signal sampies with respect to that corresponding to the signal
bandwidth. It however enables applying different quantization to different
signal components and in this way to gain in the final volume of signal
digital representation. The method has found its application in wavelet image
coding ([9]).
3. Digital representation 0/ signals 131
(3.4.1)
cl miD A
a)
(3.6.1)
(3.6.4)
-00
(3.6.5)
and
~ Jplrv
Ia(q+l)
D d, =
\lA. \n L(q)Lrv . a
)IJd, ~ pv."
(1) =a •• a (a) =a
mm , max , (3.6.6)
q=1 a g)
3. Digital representation 0/signals 135
(3.6.7)
(9) = 'ci
a Opl (9- 1) + ci opt
~ opt
(9»)/2 • (3.6.8)
(3.6.9)
(3.6.10)
w(a)= argmin{Ddr =
w(,,)
"l p(a )D
"mln
dr ( t;
dwa / da
Jda} , (3.6.11)
3. Digital representation of signals 137
aJ---------------- I
-------2\:-------
---------------- --:------------
I I
4--:------------
--------------
I I I
-t~-~------------
I I I I
f-ri-1------------
I I I I
~-J-L..-I--+-----Jl
I I ::: :
QuantizatlOn interval
indices
a)
r----.-----..-----.,...----.----,,".5
I ___ _
_____ 1..
Restored
---+-----L--- 0 signal
---r-----~-----~----
: :
_____ L.. _____ ~-----;-----L..----
: : values
I I I I
I I I I
I I I I
I I I I
I I I I
I I I I
I I I I
I I I I
I I I : -0.5
va ues
Expanding transformation
Figure 3-27. Schematic diagram of the companding quantization (a) and expanding
restoration (b)
138 Chapter 3
where Ddr (.) is a quantization 10ss function formu1ated in terms of the range
of quantization errors for each particu1ar va1ue a . The solution is provided
by the Euler-Lagrange equation, which in this case is written as:
(3.6.12)
where ~ = dw(a )/ da .
The following special cases will provide an insight into how probability
distribution p(a) and the type of loss function Ddr (.) affect optimal
arrangement of quantization intervals within the quantized value dynamic
range. We will begin with threshold quantization quality criteria that assume
that quantization error is nil if it does not exceed a certain threshold. For
threshold criteria, optimal arrangement of quantization intervals does not
depend on probability distribution p(a) because, according to the criteria,
los ses owing to quantization should be kept zero for all quantization
intervals.
(3.6.13)
In this case,
(3.6.17)
Ihe constant decreases with the stimulus size. Its lowest value of 1% - 3%
corresponds to stimuli of large angular size. Iherefore it follows from
Weber-Fechner's law that optimal, for digital image display, quantization of
image sampies should be uniform in the logarithmic scale. Ihe number Q of
required quantization levels for logarithmic quantization can be found from
equation
standard for image, sound and other analog signal representation and why 8
bits (byte) had become a basic unit in the computer industry.
Figure 3-28. False contours in image quantization (128; 64; 32; 16; 8; 4 quantization levels)
3. Digital representation 0/signals 141
(3.6.21)
(3.6.22)
(3.6.23)
which implies that the size of quantization intervals for the various values
of a. are inversely proportional to their probability densities raised to the
corresponding power. For the widely used mean squared quantization error
criterion ( q = 1 ),
a
J(p(a ))113 da
w(a )- w(a min ) amin
(3.6.24)
w(a max ) - w(a min ) = -aj=•• -(p-(a-))-1I3-tJ,-a
(3.6.26)
where
(3.6.28)
-co
One can find that, for this case, the reduction in the required number of
quantization levels, as compared , for the same averaged quantization error,
to the case of uniform quantization, is equal to
where a maxi min = ~a max - amin) / cr (J. • F or large a maxi min' the gain tends to
a maxlmin /5.7 .
(3.6.31)
(3.6.33)
P-th law quantization proved to be very useful for quantizing Fourier and
DCT transform coefficients of images. The optimal value of the nonlinearity
index P for which standard deviation of errors in reconstructed images
because of quantization of their spectra is minimal is usually about 0.3 (see,
for instance, Fig. 3-29). Remarkable enough, this corresponds to the
optimum for the mean squared relative quantization error criterion defined
by Eq. 3.6.3 3. Another option in quantization image spectral coefficients is
inhomogeneous quantization in which different coefficients are quantized
in different way. Inhomogeneous quantization in for of zonal quantization
has found its application in image transform coding (see Sect. 3.7.2)
................... _.......
10 ,." '"'' '" " .• "" " ............. ",' .............", ....." ...,.. ,,,; ... ,,,,,,,, ..... ,, ...... ,,,,,,..... !.,,,,,,..,
Figure 3-29. Optimization of P-th law quantization of image spectrum (33 quantization levels
ofspectrum real and imaginary parts in the dynamic range of±25 oftheir standard deviation):
Initial image and images reconstructed from uniformly (P=I) and P-th (P=0.3 and 0.1) law
quantized of spectral coefficients. Bar diagram at the bottom shows how standard deviation of
quantization error depends on the nonlinearity index P. Note that while not much difference
between images for P=O.1 and P=0.3 is visible, standard deviations of the quantization error
differ substantially for these two cases.
3. Digital representation 0/ signals 145
Figure 3-31. P-th law and pseudo-random diffuser quantization for encoding computer
generated display Fourier holograms: a) , b) - simulated images reconstructed from a Fourier
hologram ofthe image ofFig. 13-30, real and imaginary parts ofwhich are, correspondingly,
uniformly and P-th law (P=0.3) quantized to 257 quantization levels (2x128 for positive and
negative values plus one level for zero value); c) image reconstructed from a Fourier
hologram ofthe same image computed with the use ofpseudo-random binary (0 and n) phase
component assigned to the image.
3. Digital representation 0/signals 147
(3.7.1)
(3.7.2)
where IAI 2
is variance of the signal ensemble. The lower bound defined by
Eq. (3.7.1) can, in principle, be achieved as a result of general digitization
process discussed in Sect. 2.1.2.
In practice, data compression is applied for digital signals obtained for
continuous signals by their discretization and element-wise quantization
discussed in this chapter. For digital signals, lower bound for the code length
is determined by the entropy of digital signal ensemble ([20]). If {An} is a
set of digital signals to be encoded, and P(A n ) is probability with which
signal A k may occur, entropy
148 Chapter3
(3.7.3)
n
of the signal ensemble defined by the signal set {A,,} and their probabilities
{P(A,,)} determines minimal number of bits per signal that are necessary to
specify any particular digital signal from the ensemble. When all signals are
equally probable, entropy, and therefore number ofbits required, is maximal
and equal to
(3.7.7)
N-I
p({a"})= TIp(a,,), (3.7.8)
,,=0
3. Digital representation of signals 149
then
(3.7.9)
(3.7.10)
(3.7.11)
Entropy H=I.9554
Average number ofbits per symbol: 1.959
There are two most popular varieties of methods for coding rare symbols:
run length coding and coding co-ordinates 0/ rare symbols. In run length
coding, it is assumed that symbols to be coded form a sequence in which
runs of the most frequent symbol occur. The encoder computes length of the
runs and generates a new sequence of symbols in which the runs of the most
frequent initial symbol are replaced with new auxiliary symbols that
designate the run lengths. This new sequence can then be subjected to VL-
coding if necessary. Run length coding is an essentially I-D procedure. For
coding 2-D data, it is applied after 2-D data are converted into I-D data by
means of a zigzag scanning. Principle of zigzag scanning is illustrated below
in Fig. 3-34.
Coding coordinates of rare symbols is an alternative to the run length
coding. In this method, positions of rare symbols (others than the most
frequent one) are found. Then a new sequence of symbols is generated from
the initial one in which all occurrences of the most frequent symbol are
removed and auxiliary symbols are added to each occurrence of rare symbols
that designate their position in the initial sequence. Coding coordinates of
rare symbols is less sensitive to errors in transmitting the binary code than
run length coding. For the former, transmission errors cause only localized
152 Chapter3
errors in the decoded symbol sequence while they result in shifts of entire
decoded symbol sequence after the erroneous one in the latter. This property
is of especial importance in image coding for transmission where
transmission channel errors, in case of run length coding, may cause
substantial deformation of object boundaries.
Digital data compression methods have been developed for and found
their major applications in audio and video data compression. In this section,
we will briefly review methods for image and video compression.
Classification diagram of image coding methods is presented in Fig. 3-32.
Coding methods are classified in it into two groups: methods in which data
decorrelation and quantization are carried out separately one after another
and methods in which one can not separate decorrelation from quantization.
The latter attempt to implement, though in a reduced form, the idea of
encoding of signals as a whole to achieve the theoretical minimum of bits
defined by Eq. 3.7.3.
In decorrelation-then-quantization methods, data decorrelation is carried
out with linear transforms. Two large families of decorrelating linear
transforms are in use: "predictive" and orthogonal transforms. In predictive
transforms, for each pixel is compared its "predicted" value is generated in a
certain way from other pixels and a prediction error is computed as a
difference between the pixel value and its "predicted" value. The prediction
error is then sent to scalar quantizer. The simplest method that implements
decorrelation by prediction is known as DPCM (from "Differential Pulse
Code Modulation). Block diagram ofDPCM coding and decoding are shown
in Fig. 3-33. DPCM assumes that images are scanned row-wise/column-wise
and, for each current pixel ak,l' prediction error is computed as (Fig. 3-34)
(3.7.13)
,
decomposition and
sub-sampling
~
Hybrid "predictive"
and orthogonal
.Al
transforms
Combioed decorreJationlquandzation:
• DPCM with feedback
- LZWcoding
- Adaptive dtscrotization
- Vector quantizaflon
. EB ...
Statistical
~ Quantizer ~ enroder
InPU!
sca nned
si gnal
i Binary coded
Predictor
signal
i a)
..... Statistical
decoder ~
De-quantizer
~ ~
Restored signal
.
Binary coded '$l>
t
signal
P'redidor
i
b)
Figure 3-33. Block digram ofDPCM coding (a) and decoding (b)
Current
Ct pixd (k,I)
0.014
#:f:i=,
0.006
0.002
0.3
~ :
0.25 "·!'''''"·"t,,. , . . . . !'',,·'' ' ' '·i''' ' ' ' ·,,':
0.2 """""""""".""".,,,,, "" """ , """"".
0.05
Figure 3-35. Dynamic range, standard deviation and entropy reduction by predictive
decorrelation in DPCM
Inp ut
E9
~
r- Nonuniform
scale quantizer
.... Binary
statistical
encoder
-----.
I "
sig nal
I
•
De-quaoti:rer
r
Predictor ..... .....
....
I $
(3.7.14)
where
(3.7.15)
(3.7.16)
,.=1
illustrated in Fig. 3-38.
:-
~
F.s 111
r>t
I 1'- I"
I 1-, I·~
I
I
I
In the image block transform coding, image is split up into blocks. DCT
of each block is then computed and spectral coefficients are quantized
3. Digital representation 0/signals 159
References
1. E.C. Titchmarsh, Introduction to the theory ofFourier integrals, Oxford, New York, 1948
2. H. F. Harrnuth, Transmission ofInformation by Orthogonal Functions, Springer Verlag,
Berlin, 1970
3. S. Mallat, A Wavelet Tour ofSignal Processing, 2 nd edition, Academic Press, 1999
4. N. Ahmed, K.R. Rao, Orthogonal Transforms for Digital Signal Processing, Springer
Verlag, Berlin, Heidelberg,New Y ork, 1975
5. L. Yaroslavsky, Digital Picture Processing, An Introduction, Springer Verlag, Berlin,
Heidelberg, New York, Tokyo, 1985
6. K. G. Beauchamp, Transforms for Engineers, A Guide to Signal Processing, Clarendon
Press,Oxford,1987.
7. O.J. Burt, E.A Adelson, The Laplacian Pyramid as a compact Image Code, IEEE Trans. on
Com., Vol. 31, No. 4, pp. 5320540, Apr., 1983
8. Y. Meyer, Wavelets, Aigorithms and Applications, SIAM, Philadelphia, 1993
9. M. Vetterli, J. Kovaeevic, Wavelets and Subband Coding, Prentice Hall PTR, Englewwod
Cliffs, New Jersey, 199
10. H. Karhunen, Über Lineare Methoden in der Wahscheinlichkeitsrechnung", Ann. Acad.
Science Finn., Sero A.I.37, Helsinki, 1947
11. M. Loeve, Fonctions Aleatoires de Seconde Ordre, in.: P. Levy, Processes Stochastiques
et Movement Brownien, Paris, France: Hermann, 1948
12. H. Hottelling, Analysis of a Complex of Statistical Variables into Principal Components,
J. Educ. Psychology 24 (1933), 417:441, 498-520
13. R. Dawkins, Climbing Mount Unprobable, W.W.Norton Co, N.Y.,1998
14. H. Hofer, D. R. Willams, The Eye's Mechanisms for Autocalibration, Optics & Photonics
News, Jan. 2002
15. L. Yaroslavsky, Digital Picture Processing, An Introduction, Springer Verlag, Berlin,
Heidelberg, New York, Tokyo, 1985
16. E. Caroli, J. B. Stephen, G. Di Cocco, L. Natalucci, and A Spizzichino, "Coded Aperture
Imaging in X- and Gamma-ray astronomy", Space Science Reviews, vol. 45, 1987, pp.
349-403
17. G. K. Skinner, "Imaging with coded aperture masks", Nuclear Instruments and Methods
in Physics Research, vol. 221,1984, pp. 33-40
18. W.S. Hinshaw, AH. Lent, An introduction to NMR imaging, Proc. IEEE, Special issue on
Computerized Tomography, 71, No. 3, March 1983
19. A K. Jain, Fundamentals of Digital Image Processing, Prentice Hall, Englewood Cliffs,
1989
20. C. Shannon, A Mathematical Theory of Communication, Bell System Techn. J., vol. 27,
No. 3, 379-423; No. 4, 623-656,194
21. R.M. Fano, Technical Report No. 65, The Research Laboratory of Electronics, MIT,
March 17, 1949
22. D. A. Huffman, A Method for the Construction of Minimum Redundancy Codes, Prioc.
IRE, vol. 40, No. 10, pp. 1098-1101
22. R. Gonzalez, R. E. Woods, Digital Image Processing, Prentice Hall, Upper Saddle River,
N.J., 2002
23. H. S. Malvar, Signal Processing with Lapped Transfonns, Artech House,
Norwood, MA, 1992
Chapter 4
+
Discretization
and
quantization
~
Digital
signal
transformations
... Reconstruction
of the continuous
signal
I
~
I
I
- _.... -- ----------- - -- -
I
I Equivalent continuous transformation
I"
Continuous input signal Continuous output signal
Figure 4-1. Mutual correspondence principle between continuous and digital signal
transformations
(4.1.2)
'k = Ib(xx
},~d)(x}Lx = ILa(l)W)(x}Lx =
x
(4.1.6)
B=HA' , (4.1.7)
(4.1.8)
(4.1.9)
where 6 k ,n is the Kronecker delta (Eq. 2.1.10), and the relationship of Eq.
4.1.5 between input and output representation coefficients is reduced to a
simple element-wise multiplication:
(4.1.10)
la~{~~ h".9'l')~)t)(X)]d~
from which it follows that PSF heq (x,~) of the continuous linear
transformation equivalent, according to the mutual correspondence principle,
to the discrete transformation with DPSF {hk,n} is
(4.2.2)
k- n
-00 -00
(4.2.4)
Nh-l
bk = Lhnak-n (4.2.5a)
n=O
assuming that N h is the number of sampies of {hk}' and that the number of
signal sampies is, in accordance with infinite limits in the convolution
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 167
(4.2.5b)
(4.2.7)
-00-00
Nb-IN,,-I
L L
00 00
f f hnfl(r)(x-nAx),(d)~-(k-n~]x
_00_00 k=O n=O
Nb-IN,,-I
pe)}ixde = L L hk f
00 00
exp[i211'(fx - ffl(d)(~),(')(x)x
k=O n=O -00-00
168 Chapter4
(4.2.8)
-00 -00
where
Nh-l
DFR(P)= L hn exp(i2npnllX) (4.2.10)
n=O
(4.2.11)
-00
and
(4.2.12)
-00
sin[n{!
- [1r{ - p}N)iX]
bllX ] r --'I
eXPL'''', - p
XN I\A-]
F . (4.2.13)
sm 1- p b -
lim SV(f,p)=
Nb .....oo
lim Nb
Nb .....oo
Sin[~[;(fp}N
Nb sm
bt:l
- p
exp[in-(f - p XN b - 1}\x] = 6(f - p)
(4.2.14)
x
Continuos signal
Signal spectrum
llf f
Figure 4-2. Geometry of dispositions of signal and its Fourier spectrum sampIes
hk,r =
-00 -00
-QO -00
00
00
00
00
00
00
(4.3.2)
where N is the number of signal samples within interval X , and Eq. 4.3.1 b
can be rewritten as
Then basic equation 4.1.5 of digital filtering takes, for the representation of
Fourier transform, the form:
(
a,u,v ) N-l [(k+u)(r+v)] .
cx: Lak exp i2ff (4.3.4)
k=O N
where superscripts (u, v) are kept to denote that signal and its Fourier
spectrum samples are assumed to be taken with displacements u and v with
repect to signal and spectrum co ordinate systems.
Discrete transformation (4.3.4) is orthogonal. In order to make it
orthonormal, introduce a normalizing multiplier 1/JN and obtain:
(u 1 ~l [ '2 (k + u)(r + v>]
a,' v) = r,:-; .t...ak exp I ff (4.3.5)
..;N k=O N
Eqs. 4.3.5 and 6 contain multipliers exp (i2. ';) and exp( - i2. ';) that
a, = 1
La ex
r;;;
N-l
k
{. kr)
,21&- (4.3.9).
"N k=O N
and
1 La, ex
a k = r;;;
N-l { kr)
- ;21&- (4.3.10)
"N ,=0 N
In all applications, the OFT is the basic transform that can be computed
with Fast Fourier Transform algorithm (see Chapt. 6). Shifted OFTs can be,
for any shift parameters, computed via the OFT:
.1"'); k.~-Ik~)%[a,.x{-Ik;)]ex{-12,.:)
(4.3.12)
(4.3.15)
(4.3.16)
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 175
k = 0,1,..., N -1. Compute the inverse SDFT with shift parameters (p,q)
from a subset of K spectral sampies ~:,v }:
a:/P,v/ q = _1_[~a~'v exJ - i21f rp)Jexp[- i2lf n(r +q)] =
JN r=O ~ N N
1 r.;
1
r.;
K-I{
L N-I
L (kV) [ (k + u)r]} x
ak exp i21f- exp i21f
"N r=O "N k=O N N
where
. ( f rK
sm -X )
sincd(K;N;x)= N (4.3.18)
NSin(fr ~)
is the discrete sinc-function, discrete analog of the continuous sinc-function
defined by Eq. 2.4.6. For integer x=k-n function sincd(N;N;k-nJ is
identical to Kronecker delta-function:
N-I
auf p,vfq
n
=~ a sincd(N'" N' k -
~ n
n)= a n (4.3.20)
k=O
o 0 0 • 0
......
• 0
• 0 0
0.8 --:- - - - - - - - ~ - - - - - - -~ - - - - - -- --~--_ ~-
o
--------!----- --:..-------.:.
o 0 0 0
. .. .,.
0.6 -~ .............. -~ .............. ~- . . .. .. . . .. .. .. .. ........ .... ..: ................ ~ .......... .. .. .. ~
~. 2
~~ : "~ IW ' : ""li W-
0
Figure 4-3. Discrete sincd- (solid line) and continuous sinc- (dotted line)
functions for N =33 (top plot) and N =34 (bottom plot)
iu,o)
k+gN
=a(u,o)
k
=iu{o)
(kJlllodN'
(4.3.22)
a(u,v)
r+gN
=a(u,v)
r
exp(i2tr
N
gu) '. a(o,v) = a(o,v) = a(o{v)
r+ gN r (r JIIIod N
(4 3 23)
..
For the DFT, both signal and its spectrum are regarded as periodical:
Owing to the separability of the DFTs, the same holds for 2-D and multi-
dimensional DFTs for aU indices.
The periodicity property of DFTs is their main distinction from the
integral Fourier transform whose discrete representation they are. It
originates from image sampling with shift basis functions and from the fact
that the number of sampies in signal definition is finite.
Shift theorem
Shift theorem for the DFTs is completely analogous to that for integral
Fourier transform:
Permutation theorem
As one can see from Eq. 4.3.26, multiplying signal sampie index results, in
distinction to the continuous case, not in the corresponding scaling of signal
and its spectrum but rather in permutations of signal and its spectrum
sampies provided the scaling factor and the number of signal sampies have
no divisors in common. The distinctive feature of the finite dimensional
discrete signal space in which the DFT operates is that sampie indices are
counted modulo N and no other indices are defined.
Convolution theorem
Let ~:",v" } be SDFT(ua,va) of a signal {ak} and {a:b,Vb} be SDFT(Ub,Vb)
of a signal {bk}' k,r =O,1,•••,N -1. Compute the ISDFT(u,",vc) of their
product:
u v
c ke• e = .JN
1 )~I u v u v
, ; !::ocrr ". "Pr b' b exp - ,21&
exp(- ,,21& kv [, r(k N+ U c >] =
1
r.;exp kv
( -i21&-c )N-I{
L 1
r.;exp ( i21&-1I )N-t [
ru Lanexp i21& n(r + 11 V)] X
"+J N N r=O "+J N N n=O N
_1
JN exp{- i2ff[!!..-(V _ N
N C
-1) _ -1
2N
N
2N
(u + u _ u )~} x
abc J
La
N-I
n~O
n
N
N-1)]
n ( va + - -
exp[ i2ff-
2N
X
Lb
N-I
m~O
m exp[i2ff-
m ( Vb +N-l)]
N
- - sincd[N;N;tr{n+m-k+ua +U b -UJ]
2N
(4.3.27)
N-l
Select Va =Vb =-V c = - - - toobtain:
2N
(4.3.28)
where
N-l
bn = Lbm sincd[N;N;.(m - n + U a + Ub - u c )] (4.3.29)
m=O
(4.3.30)
In order to distinguish it from the general case of Eq. 4.3.28, we will refer to
Eqs. 4.3.27 and 28 as to shifted cyclic convolution
Symmetry properties
U,v
{ aN -k
['2
exp - I ff
2v(k N-1/ 2)]} SDFT(u,v)
)
. 2u(r N-112)]} ;
{au,;~r exp[ 12ff (4.3.31)
u,v _ + u,v
{ a k - _a N -k
['2
exp - I ff
2v(k N-112 )J} SDFT(u,v)
)
{aU'v
r
=±au,v
N_
. 2u(r -1/
exp[ 12ff
N'
. 2)]} (4.3.32)
(4.3.33)
(4.3.34)
If 2v is integer,
If 2v is integer,
182 Chapter4
Eqs. 4.3.34 - 37 show that semi-integer shift parameters play an especial role
in symmetry properties of SDFTs.
For the DFT, above equations are reduced to
Ja
~ N-k } SDFT(u,v) ){aN-r }., (4.3.39)
J =+
l.ak _aN-k } SDFT(u,v) )JIV
l..... r = +IV }.
-U,N-r' (4.3.40)
(4.3.41)
{a k = ±(ak r} SDFT(U,v»
(4.3.44)
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 183
l-D signals:
EvenN OddN
Even N. k.=O
kz=O --+_. N.I2;O
f··· --+ f· ..
k.=O
-- ~
Even NI kz=O
~ I
~
Cl ~ -- -~
...··
O,N212 I N./2;
N z/2 I t ••• .. . -
~- .C2- ..J
--+ ....
k.=O
k:=t) ~
......I
k.=O
k1- O --+ N./2 öO ~.... ~ I
• .
~
......
-- - - -+ ....··
O.Nz/2 I
I
I
~- . - -
...... I
:
·
- - - - are complex congugate to
Figure 4-4. Types ofDFT spectra symmetries for real valued signals
For even numbered N, coefficient tZN 12 plays a role of the highest signal
1 N-I
tZN/2= ~L(-lrak' (4.3.45)
vN k=O
Nil
I I• N-}
I I I 1(N-I)Il-1 I
•
(N+I)12 N-I o I J
••••••••••••
,.... ~ t- ,.... ~ +-
- - - are complex congugate to ••••• ••••• • ••••
Figure 4-5. Centering DFT spectrum (numbers in boxes indicate spectral indIces)
I 1' t I
i I1 I I I
Figure 4-6. Spectrum of a sinusoidal signal plotted without (central plot) and with (bottom
plot) spectrum centering
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 185
(4.3.46)
where 60 is Kronecker delta (see Fig. 4-7 a" b). Let us compute OFT of
this signal:
(4.3.47)
a) ~)
Sincd-Interpolated signal
t)
Figure 4-7. Graphical illustration of the discrete sampling theorem: a) initial signal; b) OFT
spectrum of signal (a); c) initial signal with zeros placed between its sampies; d) its OFT
spectrum: periodical replication of the initial signal spectrum; e) removing all spectrum
replicas but one by the ideal discrete low pass filter; t) discrete sinc-interpolated signal
resulted from inverse OFT of spectrum (e)
--
;;k = ~ LI-reet r-(N+l)/2]tZ(r)mOdNexP(-i21&-
1 LN-l[ kr) =
..;LN r=O LN -N -1 LN
1 [(N.;.!)/2 k )+
~ ~ Qr exp( - i21&-r
..;LN r=O LN
LN-l (, kr )]
). tZ(r )mod N exp - ,21&- =
r=LN~-1)/2 LN
~1 {(N.;!l/2[
~ 1 Lanexp
~ N-l nr)] ex{ -i21&-
( i21&- kr ) +
..; LN r=O ..; N n=O N LN
LN-l
r=r=Lf(N-l)/2
[1-La
.JN
N-l
n=O n
nr)] exp( -i21&-
exp( i21&-
N
kr )} =
LN
, (k -Ln,)
1{N-l
Ir L exp(-1211'
LQn [(N.:,!)/2 +
NvL n=O n=O LN
LN-l ' (k -
ex{-1211' Ln)-)]}
r=LN~-1)/2
-
LN-
ex [_i21&(k-Ln)(N +1)]_1
1 N-l P 2LN
-- La +
N JL n=O
n
exp[ -,'21& (k - Ln >] - 1
LN
. [ N(k -Ln)]
1 sm • LN 1
I r La n SIncd[N,N,(k -nL)/L],
La n Nsin[2.(k _ Ln>] = vL
N-l N-l. •.
Ir
vL n=O n=O
LN
(4.3.49)
(4.3.50)
while its additional zero sampIes are interpolated with the discrete sinc-
function from the initial nonzero ones (Fig. 4-7, f).
For even N, term with index (N + 1)/2 required by Eq. (4.3.46) does not
exist and maintaining the spectrum symmetry is possible either with the
mask
(2) 1 r - NI 2-1
Pr = - reet----- (4.3.53)
LN-N-2
that leaves two copies of this component in the extended spectrum. In these
two cases one can, similarly to the derivation of Eq. 4.3.49, obtain that
inverse DFT ofthe masked extended spectrum results in signals:
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 189
i(O)
k
=1DFTJ.·(O)a(
lI'r r) }= modN
or, correspondingly,
(4.3.55)
_ = {a
ak
k, k = 0,1,•.., N - 1
(4.3.56)
a2N-k-l,k = N, ...,2N -1
and eompute its SDFT(l/2,0):
__ 1 2N - 1 _
SDFT1I2 •0 {a k } - ~
[.
L
a k exp 121&
(k+1I2)r]_-
",2N k=O 2N
~
1 {~l_ ['2
,,2N
~
k=O
a k exp , Ir (k+1I2}r] + 2~1_
2N k=N 2N
['2
~ a k exp , Ir
(k+1I2}r]} =
1 {N-l
~ Lak exp i21r
[ (k+1I2}r] + 2N-l
L
[(k+1I2)r]} =
a2N-k-l exp i21r
,,2N k=O 2N k=N 2N
~ {~la k exp[i21r (k + 1I2}r] + ~la k exp[i21r (2N - k -1I2)r]} =
,,2N k=O 2N k=O 2N
1 {N-l
~ Lak exp i21r
[ (k+1I2}r] + N-l
Lak exp -i21r
[ (k+1I2}r]} =
,,2N k=O 2N k=O 2N
2 ~l
~ ~ak
v2N k=O
[(k
cos I +r . 112']
N
(4.3.57)
a rDeT __
- ~
2_~ [(k+1I2}r] .
~ak cos Ir (4.3.58)
,,2N k=O N
DCT signal speetrum is, as one ean see from Eq.(4.3.58), always an odd
(anti-symmetrie) sequenee ifregarded outside its base interval [O,N -1]:
a rDCT =-aDCT'a
2N-r' N
=0 (4.3.59)
(4.3.60)
From the above derivation of the DCT it also follows that, for the DCT,
signals are regarded as periodieal with period 2N :
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 191
(4.3.61)
The inverse DCT ean easily be found as the inverse SDFT(1I2,O) of anti-
symmetrie speetrum ofEq.(4.3.59):
- ,,2N1 2~1
Qk= LJa,DCT [.2
,=0
r=-:;
2N
exp-Iw (k+1I2'J =
La?!CT exp[(k+1I2),.J
-1- { afCT + N-l - i2w +
.J2N ,=1 2N
2~1 [.2
LJ a,DCT exp - 1 W(k+1I2'J =
,=N+l 2N
La?!CT exp[(k+1I2),.J
-1- { afCT + N-l - i2w +
.J2N ,=1 2N
~1
LJa,DCT exp[.2
-I W
(k + 1/2 )(2N - r >J =
,=1 2N
La?!CT exp[(k+1I2),.J
-1- {afCT + N-l -i2w +
.J2N ,=1 2N
~1
LJa,Dcr exp[ 1·2W(k + 1/2)- ] =
,=1 2N
-1- (afCT +2 N-l
La?!CT cos[ W (k+1I2)-JJ (4.3.62)
.J2N ,=1 N
Coeffieient afCT is, similarly to the case of DFT, proportional to the signal
de-eomponent:
Coeffieient a!~i
a NDer
_1
_
-
2 ~ (Ir.sm(1f k+ 112)
~Qk -
r;;-;;; (4.3.64)
...;2N k=O N
represents signal's highest frequeney (compare with Eq. 4.3.45 for DFT).
192 Chapter4
_ ={a
ak
k , k = 0,1,••. , N -1
(4.3.66)
-a2N-k-l' k=N,N +1,...,2N-l
1
. r:;:;
{N-l
Lak exp[{k+1I2)rJ
i2fl -
2N-l
L a2N-k-l exp[{k+l/2)rJ}
i2fl =
'" 2N k=O 2N k=N 2N
1 {~1
r:;:; ~ak exp['2
I If (k+1I2}rJ -
~1
~ak exp['2
, If (2N-k-1I2)rJ} =
i,,2N k=O 2N k=O 2N
1
. r:;:;
{N-l
Lak exp[{k
i21f +1I2)rJ - N-l
Lak exp[ - i2fl {k +1I2)rJ} =
'" 2N k=O 2N k=O 2N
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 193
2 La
r;;;-;:;
N-I
k sm
• [
tf
(k+1I2,] . (4.3.67)
,,2N k=O N
a rDcST -_ a 2N-r·
DcST (4.3.68)
(4.3.69)
In distinetion from the DFT and the DCT, the DeST does not eontain
signal' s de-eomponent:
(4.3.70)
a
k
= -t- ( tzDcST + 2 N-l
.J2N
~ tz DcST sin[ tf
N ;:. r
(k+1I2'JJ
N
(4.3.71)
(Zr
DCTIV _
-
&~l
- LJak cos [(k + 112)(r + 112>]
tf-L-_--.l.C1.-_--L (4.3.72)
N k=O N
that obtained a name DCT-IV ([3]). One ean easily verify that the DCT-IV
speetrum also exhibits an odd symmetry:
LDCTlV _ __ DCTIV }
pr - -2N-r-1 . (4.3.73)
194 Chapter4
DCTIV _
ak -
&~1 [(k + 1I2Xr + 11 2)] .
- ~a, cos Il (4.3.74)
N ,=0 N
a,DSTIV =&~1
-
N
~
k=O
.[
a k sm (k+1I2 Xr+1I2)]
Il~-----"=-_--L
N
(4.3.75)
a k , O~k~N-l
-
ak =
0,
-a 2N - k ,
k=N
N~k~2N
(4.3.76)
0, k=2N+l
DST _
a,
1
-~ ~aksm Il
~1 • [ (k+1 Xr+l)] (4.3.77)
,,2N k=O N
known as Discrete Sine Transform (DST). DST speetrum is also and in the
same way anti-symmetrie:
J_DST _
Pr -
-a2N-r'
DST }. a DST _ a DST - 0
N - 2N+l- (4.3.78)
ak -
_ ~
1 ~l
~ar
DST • [
sm Ir
(k + 1)(r + 1)] . (4.3.79)
,,2N r=O N
Discrete orthogonal transforms such as DFT, DCT and DST can be applied
to signals either globally, i.e. to all available signal sampIes, or block wise or
in a sliding window. If they are applied block wise, signal is subdivided into
blocks and the corresponding transform is applied to individual blocks. In
block transform signal processing, independent treatment of individual
blocks may cause so called blocking effects or artifacts on the borders of
blocks. Radical solution of this problem is processing in transform domain in
a sliding window. We will discuss application of sliding window transform
domain processing in Chs. 8 and 9. Another option that allows to
substantially reduce blocking effects is signal processing in hopping
overlapping windows. Most popular and computationally efficient is half
window overlapping. Transforms that are used in this processing are called
"Lapped" transforms ([4]).
One of the most popular lapped transforms is Modijied DCT (MDCT) or
Modulated Lapped Transform (MLn. The MDCT of a signal sequence
{a k }, k = 0,1,.••,2N -1 is defined as
a~DCT = fN
{22N-l N -
~ akhkcos[II( k+-2
+1)( 21)/]
r+ N (4.3.81)
. ( Ir k
h k =sm +1/2J
N
. (4.3.82)
196 Chapter4
One can show that MLT is equivalent to the SDFT ( N + 1 ,!. ) of a signal
2 2
{ak = akhk} modified in a certain way:
(4.3.83)
obtain finaIly:
From Eq. 4.3.86 it follows that the inverse MDCT coincides with the direct
one:
äk 1
= .J2N N-I N -+
f:otz~DCT cos[Ir( k+- 2
1)( 1)/]
r+2" N (4.3.88)
(4.3.89)
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 199
-00 -00
-00 -00
(4.4.1)
(4.4.2)
Then
-00 -00
-00
200 Chapter4
-co
where
(4.4.4)
-00
(4.4.5)
(4.4.6)
and
w=up-v/p. (4.4.7)
With this discrete PSF, basic equation Eq. 4.1.5 of digital filtering takes, for
the representation of Fresnel Transform, the form:
I •• )
a~'w = - - ~ a
1 N-l
exp[ -itr (kP-rlu+w)2]
r (4.4.9)
r '-;;;N
~ k N
"1'1 k=O
1 N-l [ Ckp
aY',w)=--Lakexp-itr
JN k=O N
\l]
- r I p}. (4.4.10)
a~'w) 1 {N_l
=-- La k 1p1) [
exp(-itr--
p
exp i2tr k(r - w )]} x
JN k=O
k
N N
From this relationship one can conclude that SOFrT is invertible and that
inverse SOFrT is defined as
I..)
a~'w
1 N-l
=__ I..) [
~a~'w exp itr
(kp - r lu
r.
+W )2] (4.4.12)
k JNf::, r N
202 Chapter4
(4.4.13)
(4.4.14)
N
W=- (4.4.15)
2p
(",;,.) _
(Zr -
1 ~
r;:; 4.- ak exp -u&
vN k=O
{ . [kp-(r-N /2)/
N
pr} . (4.4.16)
a~'w) = --La
1 N-I k 1p1 ) exp[ i21& k(r-wp )] .
exp( -i1&-- (4.4.17)
JNk=O k
N N
(4.4.19)
a" = 1
r,;; (_l)k exp(k1p1)N-I_ (
- il&-- La~ exp -
krJ
ill&- . (4.4.20)
"\IN N k:O N
Fig. 4-8 illustrates using Discrete Fresnel and Focal Plane invariant
Discrete Fresnel Transfonns for reconstruction of optical holograms.
Figure 4-8. Reconstruction of a hologram recorded in a form of amplitude and phase. a) - c):
reconstruction on different depth using Discrete Fresnel Transform (Eq. 4.4.10); d) - f) -
reconstruction using Focal Plane Invariant Discrete Fresnel Transform (Eq. 4.4.16). (The
hologram was provided by Dr. G. Pedrini 3,[8])
204 Chapter4
( .. ±
ar'
w±)
=
1 ~1
r.; ~a'f'
, __+ w+) {. [kp_ - r I p_ + W
exp Itl
-1 2}
=
..,;N r=O N
-1 ~l{ - 1 ~l
~ ~a
[.
exp -Itl (np+ - r I P+ + W +)2 ]} x
JN r=O JN n=O n N
. (kp_ - r I p_ + w _ )2 ]
exp[ Itl =
N
N-l [ r2 2 _ 1/u 2 )j
(1/ur~ ( n-k+w+;+
lu -w-lur- ) .
~ exp itr r- exp - i2nr
(4.4.21)
1/P: -1/p~ ;
N _ N _
w+ =--+w+; w - =--+w
2 -,. q=
2p+ p-
w± = w+lp+ - w_Ip_ . (4.4.22)
1
a~± ,w±) = -exp
. k +W
-Itl p-
N
- [ ( N
)2] Ia
N-l
n=O
n
[(nP+ + w )2] x
exp Itr
•
N
+
206 Chapter 4
where
frined(N; q; x)= ~
N~
l: exp[if& qr(rN- N )] exp(- ilf& xr)
N
(4.4.24)
and that
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 207
abs(frincd)
q=O •
0.9 _ . . • • . , . • . . •. ••• . . • . .• • . . •• • . • . . •. .
0 .7
. . .
. • • • . . • • . . . . . ; . . . • . • ,. . • • • . ; • . . • . •
.
0.6 ................ .. ......... ~ ........................... : .... .. ..... ..
0.2
0 .1
Figure 4-10. Absolute values of function jrincd (N;q;x) for N=256 and different va lues of
"focussing" parameter q. (The function is shown shifted to be centered around the middle
point ofthe range ofits argument)
Periodicity:
F or integer hand g:
ak+ Nh
. 2(k +"2
=a k exp[12f1hp w)] .
Nh + p (4.4.31)
(4.4.32)
208 Chapter 4
(4.4.33)
. {r1'2
a r exp[11& - ro
lV
r] (
SDFrT
".w) a k exp[ -121&
• ro(k +lVw / 1')] .(4.4.34)
L 6 } SDFrT".w
l"k = k-k o (
--(
JN
1 ISDFrT",w
frinCd(lV; ~; k+ w _ ~J exp[i1& (k + lV/21'2
1'2 I' 21'2 lV
r]
(4.4.37)
4. DIGITAL REPRESENTATION OF SIGNAL TRANSFORMATIONS 209
References
1. N. Ahmed, K.R. Rao, Orthogonal Transforms for Digital Signal Proeessing, Springer
Verlag, Berlin, Heidelberg, New York, 1975
2. A. K. Jain, Fundamentals of Digital Image Proeessing, Prentiee Hall, Englewood Cliffs,
1989
3. Wang, Z., Fast Algorithms for the diserete W transform and for the diserete Fourier
Transform", IEEE Trans. Aeoust., Speech, Signal Proeessing, vol. ASSP-32, Aug. 1984,
pp. 803-816
4. H. S. Malvar, Signal Proeessing with Lapped Transforms, Artech House Ine., Norwood,
MA,1992
5. L. Yaroslavsky, M. Eden, Fundamentals ofDigital Opties, Birkhauser, Boston, 1996
6. L. Yaroslavsky, N. Merzlyakov, Methods of Digital Holography, Consultant Bureau,
N.Y., 1980
7. T. Kreis, Holographie Interferometry, Prineiples and Methods, Akademie Verlag, Berlin,
1996
8. G. Pedrini, H. J. Tiziani, "Short-Coherenee Digital Mieroseopy by Use of a Lensless
Holographie Imaging System", Applied Opties-OT, 41, (22),4489-4496, (2002).
Chapter 5
Nh-I
bk = Lhna k - n (5.1.1)
n=O
!I
\ I
\0 \: .I I Direction of signal .anning
~ ~ ••••••••• I
11+.1
Ili,r.,.II,I"I. Output signal sampies
(5.1.2)
where N!t) and N~r) are the numbers of addends in the transversal (first
sum) and recursive (second sum) parts, respectively, of the filter. In
hardware implementation, recursive filters correspond to feedback circuitry.
Eq. 5.1.2 assumes that, for computing each output signal sampie, N!t)
input signal sampies andN~r) previously computed output signal sampies are
used. To illustrate computational saving with recursive filtering consider the
simplest case, when N!t) = 1 and N~r) = 1 :
(5.1.3)
1 N
b - ~a (5.1.5)
k - 2N + 1 L.J k-n'
-N
(5.1.6)
...... .
~
•••••••••••
• • • • • • • • • • •
. 0000
"Past" sampIes
Figure 5-2. "Past" and "future" sampies on a rectangular sampling raster at 2-D row wise
filtering
N-IM-I
(5.1.8)
In this case filtering can be carried out separately row-wise and column-wise
with corresponding 1-0 filters
(5.1.9)
(5.1.10)
DFR(P)= rr DFR,(p).
R
,=1
(5.1.11)
R-inputs adder
1-.-_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _...... output signal
R
h,. =~ h(')
~,. (5.1.12)
,=1
(5.1.13)
and assume that filter PSF {h,.} can be approximated, with the desired
accuracy, by expanding it into aseries over a system of basis functions
~,(n} r=1,2, ...,R; RSNfh)+N~h)+l}:
R
h,. == L'1,f',{n}. (5.1.14)
,=1
N~') R R Nlh ) R
where
(5.1.16)
5. METHODS OF DIGITAL FILTERING 217
Eq. 5.1.15 describes signal filtering with sub-filters with PSFs ~r(n)} in R
parallel branches and adding up the sub-filtering results with weights Wr}.
Let us now find conditions under which coefficients fJ r (k) in branch
filters can be computed recursively. In the simplest first order recursive
filter, every given output signal sampie is computed from the output sampie
determined at the preceding filter window position. Let us establish a link
between {P,(k)} and {P,(k -I)}:
Then obtain:
Eq. 5.1.19 signifies that when basis functions {tp,(n)} satisfy Eq. 5.1.18
coefficients fJ r (k) can be computed recursively with the recursiveness on
one preceding step.
It follows from Eq. (5.1.18) that the class of the recursive basis functions
is that of exponential functions:
(5.1.20)
Two most important special cases are the basis of discrete exponential
functions:
218 Chapter5
(5.1.21)
(5.1.22)
which can be obtained from Eq. (5.1.21) when Po =0. The latter
corresponds to recursive computation of signallocal mean:
When the basis of Eq. 5.1.21 with P =1 is used each of the R parallel
recursive filters (5.1.16) performs signal discrete Fourier analysis in the
window of M = Nfh) + N!h) + 1 sampies at the frequency r. The resultant
spectral coefficients are multiplied element-wise by the spectral coefficients
of the filter pulse response, i.e. by sampies of its discrete frequency
response, and summed up in order to obtain the filtering result. The basis
of Eq. 5.1.21 with Po =112 corresponds to Iocal DCT/DcST analysis.
Corresponding recursive algorithms of sliding window DFT and DCT/DcST
are discussed in Sect. 5.1.3
When the basis (5.1.22) is used, the digital filter is composed of a set of
loeal mean filters with different window size. The impulse response of the
filter is a sum of R rectangular functions of various extents. Synthesis of
such a representation for a particular filter is therefore reduced to
quantization of values of its PSF to R quantization levels (Fig. 5-5). For
minimization of filter PSF the approximation error caused its quantization,
optimal quantization methods discussed in Sect. 3.5 may be used.
Figure 5-5. An example ofa filter PSF (bold bars) and its quantized approximation with 8
quantization levels (c1ear bars)
5. METHODS OF DIGITAL FILTERING 219
(5.1.24)
R-IS-I
hm,n ~ L L'I"sf'"s(m,n) (5.1.25)
,=os=o
(5.1.26)
would then be the simplest possible. The above results are obviously
applicable to the one-dimensional functions f'~I)(m) and f'~2)(n). Hence in
the two-dimensional case, "pure" bases of exponential and rectangular
functions may be supplemented by "mixed" bases, rectangular in one
coordinate and exponential in the other coordinate.
There are also inseparable recursive two-dimensional bases. One of
them is a basis of rectangular functions within an arbitrary shaped area on a
rectangular raster. An analogous separable basis is bounded by a rectangle.
The aperture support of an inseparable two-dimensional filter is shown in
Fig.5-6, b) and its separable analog is shown in Fig.5-6, a) in which the
boundary elements of the aperture are cross hatched.
a) b)
Figure 5-6. Pattern of support of separable (a) and inseparable (b) 2-D moving average bases
(pixels that are drop out the filter aperture are single line hatched, those that enter it are
double line hatched).
220 Chapter 5
ar
(k-l) _
-
1
r;r
'\IN",
L
N w -I
n=O
(
• nF
)
ak-t-flX«Nw-t)/2)+n exp ,21&-
N",
(5.1.27)
and
(5.1.28)
Eg. 5.1.29 describes the recursive algorithm for computing DFT signal
spectrum in sliding window of N", sampIes that reguires, for signals whose
sampies are real numbers, 2(N", -1) multiplication operations and (N", + 1)
addition operations.
5. METHODS OF DIGITAL FILTERING 221
a,
(k-t) __I_~
- ~ -'..Jak-t-ftx«N",-t}i2)+-n exp ur
(0
(n+1I2)r) ( 5.1.30)
"N", n=O N",
and
(k) _ _1_~
a, - ~ -'..Jak-Jh{(N",-t}i2)+-n exp
(0,tr (n+1I2)r) (5.1.31)
"N", n=O N",
(5.1.34)
222 Chapter5
and
(k)ocST
Ir,
= Imagl~(k)}=
pr (k-l)ocSTC _
Ir, I
(k-I)DCTS
Ir, I
_ A
L.1
a k S 7.' (5 • 1•35)
where
1).
eomputing the DCT ean be eliminated in an algorithm that uses DCT
spectrum found on two previous window positions ([4 To derive this
algorithm , compute DFTs A 5DCT and A SDcST of p;
Llk)DcT and p ;,
LJk)ocST}
(5.1.37)
Z =ex{i2A" ;.) - the spectra1 sbift factor that corresponds to signal sbift
one sampIe back. Express, using Eq. 5.1.38, A: CST through A:CT and
substitute it into Eq. 5.1.37. Then obtain:
A •DCT ;::2CIADCTZ_ADCTZZ+C
. s z(M. -Ms
Z)· (5.1.41)
Finally, taking inverse OFT ofboth part ofEq. 5.1.42 and using the fact that
spectral multiplicands Z and ZZ correspond to signal shifts one and two
sampIes back, respectively, obtain for the sliding window OCT:
~k)oCT
r
;:: 2~k-I)DCT COs(._r_) _~k-l)oCT + (da - da )cos(._r_).
r N r k k-I 2N
w w
(5.1.42)
Similar recursive relationship can be derived for the sliding window OcST:
~k)oCST
r
;:: ~k-l)ocST cos(._r_) _~k-l)oCST -(da + da )Sin(.-r_)
r N r k k-I 2N
w w
(5.1.43)
(5.1.44)
(5.1.45)
5. METHODS OF DIGITAL FILTERING 225
(5 .1.46)
Original image
Figure 5-8. Three methods for extrapolating images without discontinuities on its borders
226 Chapter 5
(5.2.1)
(5.2.3)
_ { ak , = 0,1,..., N -1;
k
a(k)mod2N = a - - , k = N,N = 1,•••,2N -1
. (5.2.4)
2N k 1
N-l
ck = Lh"ii(k-..
..=0
-+(N/2])mod2N (5.2.5)
where
N-I_
C(k)mOd2N =L
n=0
"<n)mOd2NQ(k-IH{N /2Dmod2N (5.2.7)
for kernel
~ 1",1'1,1"
000000'
0000001
~.i;on ' :
11
r: 0000000
""'rn,vOIUtiOn tenn
11 1 0000000
Convolution kernel ~.} in its position for the middle convolution term
0000001
1 I 11 1 I1 11 I 1 1 0000000
Convolution kerne I ~.} in its position for t~e last convolution term
Figure 5-9. Cyclic convolution of a signal extended by its mirror reflection from its borders
As it follows from Eq. 4.2.27 for shifted cyc1ic convolution, for such signals
of double length 2N , one can retain advantages of computing convolution
with the use of a FFT if one DFT in the Fourier domain convolution
algorithm described by Eq. 5.2.2 is replaced by SDFT(u, v) with shift
parameters u = 11 2 and v = 0 :
5. METHODS OF DIGITAL FILTERING 229
It was shown in Sect. 4.3.3, that for signals extended to their double
length by mirror reflection, SOFT (1/2,0) is reduced to the Discrete Cosine
Transform. Therefore Eq. 5.2.9 is equivalent to the equation:
c=
k
1 2~1 a,DCT-'I, exp['2tr (k + 11 2}rJ =
~ ~ -I
~2N ~ 2N
~
~2N
1 {~DCT-'I, ['2
~a,
,=0
exp -I tr (k+1I2}r] +
2N
La:CTii,exp[(k+1I2)r]}
2N-I
-i2tr =
,=N 2N
,=1 N
- 2~
N-I
a:CT'I;m sin[(k+1I2)]}
.
N
tr r , (5.2.12)
where ~;e} and ~~m} are real and imaginary parts of OFT coefficients {'1,}
ofinterpolation kernel {hk}:
230 Chapter5
'Ir = 1
t.; L
2N-1 - ( • krJ
h" exp ,21&- , (5.2.13)
'\IN k=O 2N
As it was already mentioned, for computing the DFT and SDFT, FFT
algorithms are used that substantially reduce the computational burden of
digital filtering. The use of FFT assumes that data are complex numbers (see
Sect. 6.1). However in many image processing and digital holography
applications data are represented by real numbers. For direct application of
FFT algorithms to such data one should treat them as complex numbers with
zero imaginary part. This leads to additional computation and memory
expenditures. The DCT deals with real numbers and there exist special fast
DCT algorithms that are described in Sect. 6.3.6. However the DCT can also
be computed with FFT algorithms in virtue of the link between the DCT and
SDFT(1I2,O). Direct use of a FFT for computing the DCT exhibits four-fold
redundancy: two-fold redundancy owing to the fact that data are real
numbers and additional two-fold redundancy owing to the signal
symmetrization. Combined algorithms described in this section enable
avoiding these unnecessary expenditures.
Let "'k= a:} is signal with N real sampies and {a,} is a set of its
complex DFT coefficients. According to Eq. 4.3.42, {er,} exhibit complex
conjugate symmetry
(5.3.1 )
This means that one needs to compute only half of the coefficients {er,}, the
other half can then be found from Eq. 5.3.1. However FFT algorithms are
capable of efficient computing of only all transform coefficient vector as a
whole unless more complex and less widespread "pruned" transforms (See
Sect. 6.4) are available. One can avoid this limitation and efficiently exploit
the spectrum symmetry for reducing the number of operations needed for
computing signal DFT spectrum with conventional FFT algorithms. Two
ways are available to this goal: (i) by simultaneous computing spectra of two
signals with real sampies and (H) by splitting one signal with real sampies
into two signals, computing their spectra and then combining them into the
spectrum of the initial signal.
232 Chapter5
Consider the first method. Let ~k = a;} and {bk = b;} are two signals
with N real sampies and ~, =a~_,} and {P,
=fJ~-,} are their respective
DFT spectra. For an auxiliary signal with complex sampies:
(5 .3.2)
(5.3 .3)
Consider now
and obtain that spectra {a,} and {P,} can be found as:
(5.3.5)
(5.3.6)
{cA} !?FT
a)
Re
IFFT
Im
b)
Figure 5-10. Combined algorithms for simultaneous computing direct and inverse DFT of
two signals via FFT
5. METHODS OF DIGITAL FILTERING 233
a, = r:::-:;
...;2N
L a k exp( i21r-
1 ZN - I
k=O
kr)
=
2N
1 [N-I
r;:-;:;
...;2N k=O
Lau exp(2kr)
i21r- + Lazk+1 exp(i21r--r
2N
2 k + 1 )] =
2N
N-I
k=O
1r;;; { r.-;La
1 N-I
...;2 ...;N k=O
Zk exp(i2tr-
kr) +
N
Therefore computing the OFT ofthe signal can be reduced to computing two
OFTs of two auxiliary signals formed from its even and odd sampies. The
latter can be carried out by the above-described combined algorithm.
{exp(i2nr / N)}
Figure 5-11. Flow diagram ofthe combined algorithm for computing DFT ofa signal with
even number of real sampies
234 Chapter 5
2-D OFT is usually computed as two successive passes as raw wise and
column wise 1-0 OFTs. The above combined algorithms can be used at the
first pass, and the first algorithm is weB suited for simultaneous computing
1-0 OFT ofpairs ofimage rows. At the second (column-wise) pass, data are
complex. However, in virtue of their complex conjugate symmetry,
additional saving is possible as schematic diagram of Fig. 5-12 shows.
Because 1sI and (N/2+ 1)-th columns of 2-0 array of 1-0 row spectra
obtained at the first pass, where N is the number of image columns, are real
numbers, their column-wise spectra can be computed simultaneously with
the combined algorithm. Then one needs to additionaBy compute 1-0 OFTs
of N/2-1 columns beginning from the second one.
1-----------
I'· pass: I 2nd pass:
Combined DFTs of pairs of rows 1- Combined DFT ror I" and (N12+1)-th
,columns
.... 1 r DFTs for columns #2, .•.,N/2
.......
•
I . . •••••
...
...
...
• ••••••
........
•••••
1
~I
1
J.
I
••••
.,. II I
....
...
••••••
... r14.·····
I
I • ••••.
. .lex."
/.X'lt ....
.
. .
.....
~
I -""'-
... •••••• 1
~ X
I
-----------
I-D Spectra of image rows Final 2-D image spectrum
1 ______ - - - - -
•..
o
~.
-real valued sampies
Figure 5-12. Flow diagram of two pass 2-D DFT with the use of combined algorithms
Though special fast algorithms for computing the OCT are available
(Sec. 6.4), it can also be computed using FFT packages that are more
widespread. This option makes use of the fact that the OCT is SOFT(1I2,O)
of signals with even symmetry:
5. METHODS OF DIGITAL FILTERING 235
{a =a
k 2N - k -J, k = O,I,•..,N-l (5.3.8)
and
and
(5.3.12)
and compute SOFTe 1/2,0) of the composite signal {Zk}' Then obtain, with
an account for Eq. 5.2.10:
From this and the symmetry property (Eq. 5.2.9) two equations follow:
Therefore
j(I /2,0)
I'r
=1,(1/2,0)
\~r
_ ,(1 /2,0»)1 2 .
~2N-r , (5.3.15)
P-(1N-,/2,O) =1,(112,0)
\~,
+ ~2N-,
,(1 /2,0»)1 2 . (5.3.16)
j(I /2,O)
J'r
=1,(1/2,0)
\~ N+r
+ ,(112,0»)/2
~ N-r . (5.3.17)
Finally we obtain from Eqs. 5.3.15,16 and 11 that DCT spectra ofsignals
~I,k}' ~2,k}' ~3,1<}' ~4,k} can be found as:
a IDCT
,'
=DCT{a1,Je }= Ri,(1/2,O)
\-. r
_ ,(1 /2,0»)1 2 .
':t 2N-r ,
ReVt} ~I,,}
Imvrl
~2,k = a 2.2N_I_k }
ReV2N_, } k~)
Imfr2N_,}
RefrN+'}
~J.k =a J.2N-I-k }
ImVN+,} kJ
{a!1 = a~~~N_I_k} RefrN-' }
Imfr _,} ~4J
Figure 5-/3. Block diagram of the combined algorithm for computing DCT of four signals
5. METHODS OF DIGITAL FILTERING 237
References
1. C. de Boor, A Practical Guide to Splines, Springer Verlag, N.Y., 1978
2. M. Unser, A. Aldroubi, "Polynomial Splines and Wavelets - A Signal Processing
Perspective, In: Wavelets - A Tutorial in Theory and Applications, C. K. Chui, Ed., San
Diego, CA, Academic, 1992, pp. 91-122
3. R. Y. Vitkus and L.P. Yaroslavsky, "Recursive Algorithms for Local Adaptive Linear
Filtration", in Mathematical Research., L.P. Yaroslavskii, A. Rosenfeld and W.
Wilhelmi, Ed., Berlin: Academy Verlag, pp. 34-39,1987.
4. Jiangtao Xi, J. F. Chicharo, Computing Running DCT's and DST's based on Their
Second Order Shift Properties, IEEE Trans. On Circuits and Systems -I: Fundamental
Theory and Applications, vol. 47, No. 5, May 2000, pp. 779-783
5. Z. Wang, "Fast algorithms for the discrete W transform and for the Discrete Fourier
Transform", IEEE Trans. ASSP, vol. ASSP-32, pp. 803-816, Aug. 1984
6. H.S. Hou, "A fast recursive algorithm for computing the Discrete Cosine Transform",
IEEE Trans. on ASSP, vol. ASSP-35, No. 1O,0ct. 1987, p. 1455-1461
7. Z. Wang, "A simple structured algorithm for the DCT, in Proc 3rd Ann. Conj Signal
Process. Xi 'an, China, Nov 1988, pp.28-31 (in Chinese).
8. A. Gupta, K. R. Rao, "A fast recursive algorithm for the discrete sine transform", IEEE
Trans. on ASSP, vol. ASSP-38, No. 3, March 1990, pp. 553-557.
9. Z. Wang, "Pruning the Fast Discrete Cosine Transform", IEEE Transactions on
communications, Vol. 39, No. 5, May 1991.
10. Z. Cvetkovic, M. V. Popovic, "New fast recursive algorithms for the computation of
Discrete Cosine and Sine Transforms", IEEE Trans. Signal Processing, vol. 40, No. 8,
August 1992, pp. 2083-2086.
11. J. Ch. Yao, eh-V. Hsu, "Further results on " New fast recursive algorithms for the
diserete eosine and sine transforms", IEEE Trans. on Signal Proeessing, vol. 42, No. 11,
Nov. 1994, pp. 3254-3255
12. L. Yaroslavsky, A. Happonen, Y. Katyi, "Signal discrete sinc-interpolation in DCT
domain: fast algorithms", SMMSP 2002, Second International Workshop on Spectral
Methods and Multirate Signal Processing, Toulouse (France), 07.09.2002 - 08.09.2002.
Chapter 6
FAST TRANSFORMS
Fast Fourier Transforms (FFT) are algorithrns for fast computation of the
DFT. The principle of the FFT can easily be understood if one compares I-D
and 2-D DFT. Let ~!1)} and ~!~n be I-D and 2-D arrays with the same
number N=N\N 2 of sampIes: n=O,I,...,N\N2 -1; k=O,l, ...,N\-l;
1= O,1, •.. ,N 2 -1. Direct computing ofthe I-D DFT ofarray ~!\)}
1 N-\
a, =--La~\)exp i21r- (nr) (6.1.1)
IN n=O N
(6.1.2)
(Zr •• = 1 NI-I
ra;
"N k=O
L [
exp ;21f-
NI 1=0
La" exp ;21f-
kr ]N2-1 [ls ]
Nl
(6.1.3)
Then the 2-D OFT in Eq. 6.1.2 can be split into two successive 1-0 OFTs:
k, =; O;k, = 0
..._---tM " =0
DFTl , . DFT.,.
k, =J '1= 0
k, =0 ' Z= 0
k, =3
" = 3
k, =3 k, = 0
" =0
k, =J;k, =0 '1= 1;
DFT.,.
el{i2.:~)
k, = 1
'1 =I
........
k, =3 " =3
el{;2Jr~ )
DFTl , .
k, =1 DFT",.
k, = 3
k, =3 " =2
k, =3 --'L_--.Jr--;___.....J " =3
a,
a.
as
a,.__. . a , __- - - f
a) b)
Figure 6-2. Flow charts of an "in place" FFT (Ieft) and ofFFT without pennutations; numbers
in circles are values ofindex m in exponential twiddle factors exp(i2mn / 8)
242 Chapter6
One can see on this graph that the computations are split into logz 8 = 3
stages and are carried out in each stage "in-place", without a need to
allocate separate computer memory storage for intermediate arrays of values.
One can also note that output spectral data are arranged in this algorithm in a
bit-reversed order (see also inverse order of two-component representation
ofsignal and its spectrum indices in Eq. 6.1.4). The bit-reversed order is the
order sequential for binary codes of indices written in reversed order of
binary digits. Such an arrangement of output data may require an additional
computational stage for "bit reversa" data reshuffling into their natural
order. Algorithms with bit-reversed order of spectral coefficients are, in a
signal processing jargon, called "decimation-in-time" FFT algorithms. By
means of left-to-right mirroring the graph a "decimation-in-frequency"
FFT graph with bit reversed ordering of input data is obtained. The data or
spectra reshuffiing can be avoided by incorporating them into the graph as it
is illustrated in the graph of Fig. 6-2, b). The FFT algorithms with natural
ordering of input and output data do not enable "in-place" computation.
FFT algorithms are a special case of a broad class of fast transform
algorithms that have similar structure and whose computational complexity
is O{N log N J or lower.
6. FAST TRANSFORMS 243
o o 1
1 o
o o
Multiplied by a vector, it changes the order ofvector elements.
Hadamard matrix 0/ order 2:
h=[1 1].
2
1 -1
Matrix operations:
Matrix sum o/matrices:
N-I
~M(k)
~ JI,H
=
k=O
N-I (k)
LPJI-1,H-1
k=O
!:P~?p~~!
9=0
( ) .M(2)
MR,S -MI -
~ p(I,I!p~2,o)
L ..
- R,Q Q,S - 9=0
Matrix transposition:
~2) ].
"',H2 '
N-l
M8N = ®M
k=O
(6.2.1)
N-l
E9 M=I ®M; (6.2.2)
k=O N
(6.2.3)
6. FAST TRANSFORMS 247
(6.2.4)
(6.2.5)
(6.2.6)
(6.2.7)
(6.2.8)
(6.2.9)
(6.2.10)
(6.2.11)
(6.2.12)
( 6.2.13)
248 Chapter6
( 6.2.14)
( 6.2.15)
(6.2.17)
(6.2.18)
M(l) ®M(l)
v, .G, V2.G2
=[V~t(1 V ®A(k»)~.IM(l)
k=O 2 v, J ~ V2.G2
®M(I) ).[vgl(1
v, ,G, k=O G,
®A(k»)~.
G J'
2
(6.2.20)
(6.2.21)
(6.2.22)
6. FAST TRANSFORMS 249
(6.3.1)
s[
HAD 2" = [11fn ~:~llrn-r) ® (11]® h~(r-') = ll]
[11fn s[~:kllrn-r) ® (11]® HAD 2,-111] . (6.3.3)
(6.3.4)
(6.3.5)
250 Chapter6
ao t.ro ao (10
a\ (1\ a\ (1\
a2 (12 a2 ~
a3 (13 a3 (13
a4 (14 a4 (14
as (15 as (15
a6 (16 a6 (16
a7 (17 a7
a) b)
Figure 6-3. Flow diagrams of Fast Hadamard (a) and Fast Haar (b)
Transform algorithms.
One can observe the isomorphism between Eq. 6.3.5 and the graph
representation ofthe fast Walsh-Hadamard transform shown in Fig. 6-3, (a).
Term r = 0 in Eq. 6.3.5 corresponds to the right most stage in the graph.
Term r = n -1 corresponds to the left most stage in the graph. Term
described by matrix h 2 is represented in the graph by the "butterfly" uni!
that computes sum and difference of two inputs.
Eq. 6.3.5 explicitly shows also that Walsh-Hadamard transform of a
vector of N = 2" sampies can be computed in n = log2 N steps with
N additionlsubtraction operations in each step, or, totally, with
N log2 N additions/subtractions.
(6.3.6)
(6.3.7)
(6.3.8)
One ean eompare Eqs. 6.3.7 and 6.3.8 and Eqs. 6.3.2 and 6.3.3 to see that
bit reversal permutation that eonverts Walsh-Hadamard transform matrix
into Walsh-Paley matrix ean be attributed to permutation of their
eorresponding eonstituent elementary matriees.
Bit reversal permutation matrix, in its turn, ean be represented in layered
Kroneeker matrix and reeursive forms as
(6.3.9, a)
(6.3.9, b)
PAL 2n = BRP2n • TI {1
r=O
2n -,-t ® h 2 ® 12, }. (6.3.10, b)
(6.3.11)
To introduce Walsh transform matrix, we need a not ion of Gray code for
binary representation of numbers. m -th binary digit ~~c of Gray code of a
number written in its direct binary code representation as {-r-:BC },
m = O,I,••• ,n -1 is computed as
~GC
m
=~DBC
m
+ ~DBC
bin m+1
(6.3.12)
Table 6-1. Binary and Gray code representation of8 first natural numbers
Decimal index 0 1 2 3 4 5 6 7
Direct binary code 000 001 010 011 100 101 110 111
Bit reversal code 000 100 010 110 001 101 011 111
Gray code 000 001 011 010 110 111 101 100
(6.3.13)
(6.3.14)
Comparison of Eqs. 6.3.14 and 6.3.15 with corresponding Eqs. 6.3.7 and
6.3.8 shows that direct binary-to-Gray code permutation is associated with
reverse identity matrix 12 .
Matrix HADP" 2
= BGCP2" . BRP"2 that connects Walsh transfonn and
Walsh-Hadamard Transform is called Hadamard permutation matrix ([4]).
It is isomorphie to the Walsh transform matrix with substitutions
[11]~[10] and [lr]~[10]. With these substitutions obtain from Eq.
6.3.14:
(6.3.16, a)
(6.3.16, b)
By inversion ofthe order ofKronecker products in Eq. 6.3.16, b), one can
separate from the product HADPZ" = BGCPZ" • BRPZ" the direct binary-to-
Gray code pennutation matrix:
254 Chapter6
(6.3.21)
(6.3.21)
(6.3.22)
=(
HAR I" lEB !l(r-,,)tIII')O {(lI' ®[ll]}E(II, ®[ll])EBII"_I,+t} .
(6.3.23)
L
,,-1
r = rs 2 s (6.3.25)
s=O
Then matrix FOUR I " can be written in a layered Kronecker matrix form as:
(6.3.26)
256 Chapter 6
(6.3.27)
and rearrange vertical sum in Eq. 6.3.26 into n sums grouped according
{km}:
FOUR,
2
= 2'-I(n-l~
E ®
k=O .=0
(6.3.28)
Change now the order of summation over binary code digits {kJ of kin
Eq.6.3.27 to the reverse one while keeping indexing of {km} that
corresponds to Eq. 6.3.27. This will require introducing into Eq. 6.3.28 a bit
reversal matrix:
.
lJJJ}
I I ln-I 2' ~> 2"
FOUR ,
2
= BRP2" E E...
kO=O[ k, =0 [
E ® 1
k,_1 =0[ .=0[
Wn m.O
,-I
• (6.3.29)
{
(6.3.30)
(6.3.31)
Now, applying matrix identity Eq. 6.2.6, one can rewrite Eq. 6.3.31 as
258 Chapter6
{ [I (I
[11]® ~ ...
kt-O
E_ [n_z[
® 1 wn_~=o
kn_t-O.=o
2
s
~km+12m ]JJ] E
FOURzn = { [I (I
[11]® ~ ...
kt-O
E_ [n_z[
k._t-O
~ 1 wn_~=o
2$
.-0
~km+12'" ]JJ] E
{ [ 1[I (n_2[
[11]® E ...
kt =0
E ® 1 wn_r=o
k._ t =0 .=0
2$ ~km+12m ]JJ1 E
(6.3.35)
where
du =[1o exp(i2mt)J
0 1. (6.3.36)
6. FAST TRANSFORMS 259
Derive now a matrix formula for Fast Fourier Transform. Applying matrix
identities (6.2.1,6.2.7 and 6.2.18) to Eq. 6.3.34, obtain:
FOUR2n =
FOUR2o =
or, finally,
FOUR 2" =
(6.3.41)
{Ü(l y ® h, ®I,t..-J(I y ®[ I,(.~,) af~' d.~ )])}. BRP,..
( 12(0-,-1) ® [ 12, E9 (! d 2-(s+z) )]) . This adds, in eaeh r -th eomputation stage,
Matriees of the DCT and DeST defined in Seet. 4.3.3 are, for N =r,
real and imaginary parts of the matrix
(6.3.42)
(6.3.44)
(6.3.45)
where s = O,I, ... ,n, m = 0,1, .•. ,2 s -1, DCST~"" = 1, I~ is the conjugation
operator,
(6.3.47)
(6.3.48)
(6.3.49)
262 Chapter6
(6.3.50)
DCST.
2
·(i = DCST2.HADP.·(i
2
=DCT.2 ·a+iDCST.·a
2
(6.3.51)
Eq. 6.3.49 can be split up into formulas for separate computation of the
DCT and DcST as:
(6.3.52)
and
_
DST.
2
= DST2.HADP2• = n(2'.1
n-I
r=O
EI)
r=O
(, ))(
Rs 2:.: •.1 I,2 ® h 2 ® 12•., . 1 ) (6.3.53)
,
where
and
Rs(m)
2"-.1
=
[(0 E9 I 2n-,-1 )E(es~ E9 (- 12n -'-1 ))] m(I E9 02n->-1 )E(O E9 2d
s •m I 2n->-I)]'
(6.3.55)
Fig. 6-5 shows flowchart of the DCT and DcST for N=8.
To evaluate the computational complexity of the algorithrn, denote the
number of complex multiplications of matrix M by ,u{M). Then obtain:
Considering that the input vector for each stage consists of complex
numbers, the number of multiplications of real number can be found as
For matrix transform of size N = 2", there are n stages in the algorithm,
and the total number of muItiplications is then N IOg2 N .
ao ao ao ag
al al al (Z I
a2 (Z2 a2 (Z2
al (Z) a) (Zl
a4 (Z4 a4 (Z4
as as (Zs
a6 ~
a7 a7 (Z7
a) b)
Figure 6-5. Illustrative flow diagram ofthe DCT (a) and DcST (b) for N=8. Solid Iines denote
addition at the graph nodes; dotted lines denote subtraction; circle denotes multiplication by
respective coefficients {ds.m} and {es~} (numerical values are not shown)
264 Chapter 6
P=[1000]
2 (6.4.1)
(6.4.2)
(6.4.3)
FOUR(k,l)
2"
={"rr-
I
- k/I
~ 2'
® 501 ® pi2,,-I-r ).
,=0
6. FAST TRANSFORMS 265
and, for k + I ~ n,
FOUR(k,/)
n
= {nn-
2
I kll
- ® SO ® pi
n ~ 2' 2 2 - 1- r
)x
,=0
where
(6.4.6)
where
and left and right pruning matrices p~) and p~) containing ones in the
positions that correspond to nonzero input sampies and to transform
coefficients to be computed, pruned trans form may be written as
PFOUR2" =P(L).{B·rrn-lD
2" ,
·Hr }.p(R)=ö·rrn-l
2"
ö ·H
, T ,
(6.4.9)
r=O r=O
where
(6.4.10)
(6.4.11)
and
(6.4.12)
(6.4.13)
(6.4.14)
p(L)/D
\;0
)= RP(R){p(L)
~2."
B)·
P(L)(H o)= RP(R)~(L)(Do~Do);
p(L)(O"_I)= RP(R)~(L)(Hn-2~H"_2);
P(L)(H"_I)= RP(R)~(L)(On-I~O,,_J;
p(L)(O"-1)= RP(R)~(L)(Hn-2~H"_2);
P(L)(H"_I)= RP(R)~(L)(O"_I~O,,_J;
and
P(R)(H ) = peR) .
n-I 2'"
p(R)(0"_2)= P(L)(H"_2,P(R)(H"_2»);
P(R)(Hn-3)= p(L)(O"_l,P(R)(O"_l»);
p(R)(0"_3)= P(L)(H"_3,P(R)(H"_3»);
P(R)(Ho)= p(L)(OI,P(R)(OI»);
p(R)(OO)= P(L)(Ho,p(R)(Ho»);
268 Chapter6
Fig. 6-6 shows an example of a flow diagram of pruned FFT for computing 9
spectral coefficients of signals with 7 nonzero input sampies.
-
q -
q
-<3......
L()
L()
~
c
~
Figure 6-6. Illustrative flow diagram of a pruned DFT for N= 16. Solid
lines denote multiplication by 1, or j, doted lines denote multiplication by -1,
or -j and solid lines with black bubbles denote multiplication by "twiddle"
factors.
6. FAST TRANSFORMS 269
FOUR 2• =
BRP2' . fi[ 12(.-1-,) ® [1 2,Ea (! d 2-(s+2))]} (1 2
(.-,-1) ® h 2® 12,) .
(6.5.1)
(6.5.2)
is a diagonal matrix that consists of ones and twiddle factors exP(i8r ) in its
diagonal. Multiplication of eomplex numbers by these twiddle faetors may
be represented as multiplieation of a veetor eomposed of real and imaginary
parts ofthe numbers by a matrix:
three levels of quantization, qoS(lr ) and qin(lr ) take the values (-1, and °
I); with five levels of quantization, the values (-I, -112, 0, 112, I), and so on.
3 qUBQUZ8t1on 19"""5
o_S
o - - '. - I-l-"'r-r-.
-o.S
Figure 6-7. eosine function (dashed line) and corresponding qOS(.)functions for 3 and 5
quantization levels
QTFr =[ qO~'r
-qmlr
qinlr ]
qoslr
•
(6.5.4)
qin8r ] •
(6.5.5)
qos8r
":f1'I1~----UIL~h~~
0 .' -
·:/fI·:· ~Ir .'~I~:' Ilr': 1
--- I - ..1W ~w ..~I . JW
I ~ ,~ I'~~ 2~' 2'~ 3~ .0 .,
" ", " >0
Functions #4 Functions #5
~ ~J~N~~,~~~'~l
. .. .. ~!~W~~~I~~'~~
•0 '0
Functions #6
>0 :} 1 t') ':!J
Functions #7
2C) 2$ 30
Figure 6-8. Basis functions ofthe DFT (linearly interpolated, solid Iines) and the QDFT with
5 quantization levels (bars) for N=32
a) b)
Figure 6-9. Images optically reconstructed from digital holograms synthesized with the use of
the QOFT replacing the OFT: a) - the QOFT with three quantization levels; b) the QOFT with
5 quantization levels (adopted from [6])
0.8
0.6
, , .
.. , ...... r .... -,'" .... r ....
.............................
I I I
.
•
I • I ,
0,4 .. .. . . . . . . . . . . . . . . . . . . . . ,. .. ..
• I • • I
0.8 ..... .. .. r .... "," .... t ...... 0.8 "',- .... ,. ..... , ..... t' ....
. . . . .. . .. . . .. .. . . . .
• t t ,
...............................
. .
• • f t
0.•
. ... . . .... .. .. . .. . . . .. .. .
0-6 I I • , • • • I
I • I I • I t I
REFERENCES
1. L. Yaroslavsky, M. Eden, Fundamentals ofDigital Opties, Birkhauser, Boston, 1996
2. I. J. Good, J. Roy, Statist. Soe. Sero B 20, 362,1958
3. A. M. Trakhtman, V. A. Trakhtman, Fundamentals of the Theory of Diserete Signals
Defined ofFinite Intervals, Sov. Radio, Moseow, 1975, (In Russian)
4. N. Ahmed, K. R. Rao, Orthogonal Transforms for Digital Signal Proeessing, Springer
Verlag, 1975
5. L. Yaroslavsky, Digital Pieture Proeessing, An Introduetion, Springer Verlag,
Heidelberg, 1985
6. L. Yaroslavskii, N. Merzlyakov, Methods of Digital Holography, Plenum Press, N.Y.,
1980
7. L. Yaroslavsky, A. Happonen, Y.Katiyi, Signal Diserete Signal Sine-interpolation in
DCT Domain: Fast Algorithms, in : Proeeedings of International TISCP workshop on
Speetral Methods and Multirate Signal Proeessing, SMMSP'02, Toulouse ~ Franee,
07.09.2002 - 08.09.2002, Ed. T Saramaki, K. Egiazarian &1. Astola, p. 179-185, TICSP
Series #17, ISBN 952-15-08881-7, ISSN 1456-2774, TTKK Monistamo 2002
8. Y. Katiyi, L. Yaroslavsky, VIHS Strueture for Transforms and their Fast Algorithms, 3rd
Int. Symposium, Image and Signal Proeessing and Analysis, Sept. 18-20, 2003, Rome,
Italy
9. Y. Katiyi, L. Yaroslavsky, Regular Matrix Methods for Synthesis of Fast Transforms:
General Pruned and Integer-to-Integer Transforms, in Proe. Of International Workshop
on Speetral Methods and Multirate Signal Proeessing, ed. By T. Saramäki, K. Egiazarian
& J. Astola, SMMSP'2001, Pula, Croatia, June 16-18,2001
Chapter 7
(7.1.1)
(7.1.2)
(7.1.3)
pixels in the process of regular scanning image array {ai~)} row-wise and
then column-wise on the rectangular sampling grid. For each I-th row,
compute first histograms ~(k )(qm) of window columns:
(7.1.5)
(7.1.6)
(7.1.7)
(7.1.8)
It follows from this definition and from the definition of the histogram
(Eq.7 .1.1) that histogram moments can be directly computed from signal
sampies as:
Signal first moment is its mean value. Signal second moment is called
signal variance. Square root of the variance of signals centered around their
mean value
cr 1 1-- ~ii0(2) -
-
~ii(2) - 'a-(I)'\2
~
. a - a -ii(l) .
)'0- (7.1.10)
(7.1.11)
are local means of signal {(a k .,)"}. Therefore recursive algorithm for
computing signal local mean described in Sect. 5.1.1 can be efficiently used
for computing signallocal moments.
Digital signal statistical characteristic alternative to histograms and
moments are order statistics. Signal order statistics are defined throußh a
notion of the variational row. Variational row VR{lId of a signal tak},
k =0,1,.•., N -1 is obtained by ordering signal sampies according to their
values in a monotonically increasing sequence:
(7.1.12)
7. STATISTICAL COMPUTATION METHODS AND ALGORITHMS 279
If signal sampIes happen to have the same quantized value they are
positioned in the variational row one after another in an arbitrary order, for
instance, in the order they were accessed in the process of the variational
row formation. Note that coordinate indices {k} of signal sampIes are
ignored in the variational row.
Element a(R) that occupies R-th place in the variational row is called
signal R -th order statistics. Index R (R = 1,2,•••, N ) of the order statistics
is called its rank. The rank shows how many signal sampIes have value
lower or equal to that of the element. First order statistics is signal minimum:
(7.1.13)
(7.1.14)
(7.1.17)
_ _ 1 N~--<X (r)
aa. - - - k.Ja , (7.1.18)
N -Cl r=o.
280 Chapter 7
the arithmetic mean of the elements of the variational row which are
minimum (l elements apart from its ends.
Signal second moment, or its variance, is a parameter that characterizes
spread of the signal histogram. An alternative characteristics of signal values
spread based on order statistics is signal quasi-spread ([1]):
QSPR .. (a)=a
(-N+I) (N+I)
1+.. -a - 1- ..
(7.1.18)
with n as a parameter.
In image processing, order statistics and ranks can be measured both
globally over entire image or set of images and over a sliding window. In the
latter case they are called local. Local histograms, ranks and order statistics
are the base ofnonlinear filters discussed in Ch. 12.
Local ranks are usually computed for the window central pixel. Because
local ranks can be obtained from local histogram (see Eq. 7.1.17), they may
be computed in a recursive way. They mayaIso be calculated recursively not
through the local histogram, but in a direct manner as weIl. For this purpose,
it suffices to calculate the difference between the number of window
elements appearing at a given scanning step with a value smaller than that of
the window central pixel, and the number of such elements disappearing
from the fragment, and add this difference to the rank obtained at the
previous step.
(7.1.19)
7. STATISTICAL COMPUTATION METHODSANDALGORITHMS 281
(7.1.20)
00
(7.1.22)
(7.1.24)
(7.1.25)
282 Chapter 7
P" (r) = 1
r;;;
vN
L RII (n )exp(nrJ
N-l
n=O
i2n - ,
N
(7.1.26)
(7.1.27)
and
RI (n) =
1
r;;;
vN
L p" (r )exp(nr)
N-l
n=O
- i2n -
N
(7.1.28)
(7.1.29)
and
RII(n)= 1
r;;;
(nr) ,
LW}P)p,,(r)exp -i2n -
N-l
(7.1.30)
vN n=O N
where W}II) and W}p) are window functions that are equal to one in the
middle and the largest part of the interval 0 + N and decay more or less
rapidly to its ends. They are discrete analogs of optical apodization masks.
Among the different possible window functions the most known are
Hamming window:
W = 10(a~I-[(n-N/2)/N]2)
(7.1.32)
n lo(a)
0
where 1 0 is the modified zero order Bessel function of the first kind and
a is a parameter.
Hamming and Kaiser window functions do not decay to zero at ends of
interval 0 + N . They are applicable as spectral windows. As signal window
functions, they are applicable for signals with zero mean value. For images,
window functions that decay to zero are more appropriate. Two possible
examples are functions:
and
Wn =;atan
2 - 2cr /2)]
[ ß exp(n-N 2 (7.1.34)
(7.1.35)
AVrod ( -N1 t; a
N-l
ro.,k
a'ro.,k+r J+ AVro. (-N1 t; n
N-I
ro.,k
n'ro.,k+r
J+
286 Chapter 7
One can see from Eq. 7.1.36 that, owing to noise additivity and statistical
independence of the signal, the correlation function of the observed signal
Ra (r) is a sum of the correlation function the noise-Iess signal RAr), the
correlation function Rn (r) of the noise and of the realization of some
"random" process e n' The latter describes the error in measuring the noise
correlation function that are in practice carried out over its finite-size
realization rather then over the entire ensemble. The variance of the random
process e n is known to be inversely proportional to the volume of the
realization used for the measurement. Normally, in image processing
hundreds of thousands of pixels may be involved in the measurement.
Therefore, the variance of random error e n in (7.1.36) is usually small
enough and Rn{r) may be, with a good accuracy, estimated as
(7.1.37)
(7.1.38)
where (J; is the noise variance. In this case, the correlation function of the
observed image deviates from the correlation function of the noise-Iess
image in the coordinate origin only, and the difference is equal to the noise
variance:
(7.1.39)
For all other values r the magnitude of R ö (r) may serve as an estimate of
Ra (r). Assuming that Ra (r) is a monotonie function in the vicinity of
r = 0, one can use Rö (r) = Ra (r) for r > 0 to extrapolate its value Ra (o)
for r = 0 and then use the extrapolated value to compute noise variance
from Eq. 7.1.39.
A similar approach may be used to estimate the variance of a weakly
correlated (broad band) noise, i.e. a noise whose correlation function Rn (r)
7. STATISTICAL COMPUTATION METHODS AND ALGORITHMS 287
c)
nl) .•• •• •~
1!m .•. ~ •••• : .• • . :- .•• ~ •• --: .••. : ... . :- . . • ~ .... : .
• •
·
I
. . ..
• , I I
,..
I I •
---;----r---T---r---r---r-·-T--1----!-
I • • , . I I •
3D
2IXJ --.!----~.---:_---~-_.-!----~----:--
• • • , I • I
-- -. ,.
• • • • • • I
OL-~
1 0~~=-~3l~7.
40~50
=-~611~~70~OO~~90~
Frequency index
~ ~
Figure 7-2. Diagnostics ofperiodical interferences: a)· noisy image; b) averaged column
power spectrum ( top) and the resuIt of estimation of spectrum of the interference (bottom )
signal and that. The latter may be estimated by interpolation between those
image spectral components that are found not affected by the interference
assuming smoothness of noise-free image power spectrum. Power spectrum
ofthe interference may then be found as:
(7.1.40)
Figure 7-3. Noisy interferogram with a spatial carrier (Ieft) and its power spectrum displayed
as an image (right).Vertical and horizontal directions in this image correspond to vertical and
horizontal frequencies, respectively while brightness represents spectrum intensity.
.........................................
• I I I t
.. .. .. .. . . . . .
• I
... . . .. .. .. .. .. .. ..
• •
..
•
.. .. .. L ....... L ...................... J
......................................
• I • I
• • I I
.....
Figure 7-4. An interferogram quantized to 256 gray levels and its histogram
- ,'''
7. STATISTICAL COMPUTATION METHODSANDALGORITHMS 291
(7.2.1)
292 Chapter 7
where Cl' C 2 and C 3 are constant parameters of the algorithm [2]. For
generating 2-D pseudo-random arrays of numbers, the arrays are
successively filled in row-wise / column-wise.
The initial number, or seed, generally has a little effect upon the quality of
the sequence thus obtained. However the algorithm is very sensitive to the
parameter especially to parameter C 3 that plays a decisive role in
determining the period of repeatability of the generated numbers. Provided
appropriately selected parameters, the algorithm generates pseudo-random
numbers that do not exhibit noticeable correlations and have approximately
uniform distribution. We will refer to this algorithm as "primary pseudo-
random number generator". It is usually used to generate pseudo-random
numbers with any special statistical properties.
(7.2.3)
7. STATISTICAL COMPUTATION METHODS AND ALGORITHMS 293
Output pseudo-
Primary pseudo- . , Poiot-~e randorn nurn bers
randorn nurnber p
Doo)ineatity ...
r" .. , -"-
generator o:t .~Y· ~ .
Figure 7-5. Flow diagram of an algorithm for generating pseudo-random numbers with a
given distribution function
(7.2.5)
Output pseudo-
Primary pseudo- rando m numbers
random number ""-
Linear filter ""-
generator
Figure 7-6. Flow diagram of an algorithm for generating pseudo-random numbers with a
given correlation function
(7.2.7)
n
I I hmh/A ~ (~k-"'~k-/)= I I
m /
hmh/0':8 (m + n -I) = 0': I m
hmhm+n ,
(7.2.8)
where 8 (.) is the Kronecker delta-function. Linear filtering results also in
"normalization" of the distribution function of numbers {llk}' By virtue of
the central limit theorem of the probability theory, the higher is the number
of terms involved in weighted summation of initial pseudo-random numbers
for computing each of the numbers {ll k} according to Eq. 7.2.7, the c10ser is
their distribution function to the Gaussian (normal) one.
As it was described in Sect. 5.2.1, convolution is frequently more
efficiently implemented as cyc1ic convolution in the OFT domain. This
7. STATISTICAL COMPUTATION METHODS AND ALGORITHMS 295
Figure 7-7. Flow diagram ofthe algorithm for generation pseudo-random numbers with
normal distribution and given correlation function
. ,. . .. ' ., .
:
Figure 7-8. Empirical distribution function of 65000 pseudo-random numbers with Gaussian
distribution generated by the FFT method
In order to explain the algorithm, let ~}re)} and ~~Im)} be two arrays of
N non-correlated pseudo-random numbers. Form out of them an array of
complex numbers ~~re) + i~1Im)}, multiply these numbers by some weight
coefficients {wk } and compute DFT ofthe result:
(7.2.9)
296 Chapter 7
(7.2.10)
(7.2.11)
(7.2.12)
7. STATISTICAL COMPUTATION METHODS AND ALGORITHMS 297
and
(7.2.13)
Texture ima ge
Primary generator
of pseudo-random
Transformation ....
system
Dumbers
Texture images shown in Fig. 7-10 were generated by linear filtering ofa
primary array of uniformly distributed uncorrelated pseudo-random numbers
by the above-described FFT method. Image (a) was produced by linear
filtering with circular frequency response shown in the black box at the side
298 Chapter 7
c) d)
c)
Figure 7-11. Imitating natural textures: a), b) - images ofnatural textures; c), d) - generated
images
7. STATlSTlCAL COMPUTATION METHODS AND ALGORITHMS 299
of the image. The radius of the circle is equal to 0.1 of the highest spatial
frequency. Image (b) was produced using linear filter with frequency
response in a form of a ring shown in the black box. External radius of the
ring is 0.1 and internal radius is 0.075 of the highest spatial frequency of the
image.
Images (c) and (d) exemplify fractal images with power spectrum that
decays with spatial frequency index , as ,-1
and ,-I.S respectively. Such
images can be used to imitate clouds of different density.
Images shown in Fig. 7-11, (c) and (d) show textures that imitate natural
texture image of textile (a) and wood (b), correspondingly. They were
obtained by assigning power spectra of natural texture images to images
generated from primary independent random numbers.
a) b)
Computing rhase
I Primary array of pseudo-random numbers
....
~ !~) }
I
~(I)
k,/ ,
H.
- Quantization
- Computing A(,- I )(r, s)exp(i ~ (r, s)}) = DFT(exp~ ~1~/1) }D
Replacing A('-I)(r,s)
by
t ~k,/ }
[P{r,s)1'2;0 ~ P(r,s) ~ 1 ~ Continue iterations?
No
Yes
Figure 7-13. Flow diagram of an iterative algorithm for generating pseudo-random phase
mask with a specified power spectrum.
a) b)
~
0
0.3
~c 0 .25 -
0
0 .2
~
~
0 . 15
A 0. '
üi
0 .05
0
5 '0 '5 20 25
h:Ol'aUon ..
c) d)
Figure 7-14. Examples of images reconstructed from kinoforms: a) original image used for
the synthesis of kinoforms; b) and c) - reconstructed images for kinoforms with 255 and 5
phase quantization levels, respectively; d) plots of standard deviation of the reconstruction
error as a function of the number of iterations for no quantization, 255, 15 and 5 phase
quantization levels
302 Chapter 7
Flow diagram of the computation is shown in Fig. 7.16. Fig. 7-17 illustrates
results of reconstruction of a hologram synthesized in this way for a
unifonnly pained conic surface. As one can see, the hologram is not unifonn.
Since the object scatters most of radiation perpendicularly to its surface,
most of the hologram energy is concentrated along a circle. When fragments
of the hologram along the circle are reconstructed, that is when hologram is
viewed through windows placed along the circle, as in Fig. 7-17, c,
corresponding different aspects of the cone are seen. When viewing window
is situated out of the circle, practically nothing can be seen. Fig. 7-18
provides yet another example of reconstruction of a programmed diffuser
hologram synthesized for an object in a form of a hemisphere on which
numbers 0 and 2 are painted on its four different sides.
r----------I
3-0 I
Object I
object's
renectivity Object's wavetield complex
model
distribution ~ amplitude
-+ FIT
I
A(x,y)exp(i2lt(z(x,y)+PrDitT(x,y»]
Object 3-D ~
sbape z(x,y) I
I
I
t PrDitT(x,y) ~,
....
Generating 2-D array of
pseudo-random numbers
that specify the phase
---. Computing
object's wave
front
Simulating wave
front propagation
IDFT. DFrTl
component of the object
wave front ~~
-~
~ Introducing signal
1- - - - - - - - - - - distortions:
1 2-D array that specifies : - Array size
1 amplitude component of 1 limitation
1 the object wave front 1 - Dynamic
------'l---- range
limitation
Yes - Quantization
•
Comparing
reconstructed and initial
f , wave fronts; computing
( Continue \ and accumulating noise Simulating wave
iterations J.... statistical parameters ~ front
\. ) reconstruction
No .... •
Output data I ...
Figure 7-19. Flow diagram ofa computer Monte Carlo model for studying speckle noise
phenomena
Influence of three types of the wave front distortions on the speckle noise
variance will be illustrated on the base of simulation results ([8]):
limitation of the wave front area;
7. STATISTICAL COMPUTATION METHODS AND ALGORITHMS 307
(7.3.3)
where {y!:,im)}_ are wave front orthogonal components after their limitation
within the dynamic range [- 40" ,40" ]; roundO is an operation of rounding
to the nearest integer, Q -is the number of quantization levels assigned for
quantization within positive/negative halves of the dynamic range [0,40"].
Fig. 7-16 shows examples of images with speckle noise generated using the
described model.
Fig. 7-17 shows simulation results obtained for limitation of area of the
wave field area. The upper band in the image on Fig. 7-17 a) is the original
distribution of the object wave front magnitude generated in a form of a step
wedge function. Lower bands show magnitudes of images reconstructed
308 Chapter 7
from the wave front after limiting its area by portions of 1I8-th of the entire
area. They c1early demonstrate appearance of the speckle noise and the fact
that the higher is area limitation the more correlated is "~~;,", .....l~
c) d)
However, for moderate limitation, speckle contrast is lower then one and it
decreases with the increase of the fraction of the wave front used for object
reconstruction. Quite naturally, speckle disappears if image is reconstructed
from the whole wave field, which is the case of infinitely high resolving
power of the imaging system.
0.8
0.6
GrLv1l8
218
0.4 3'8
4/8
&8
0.2 &'8
7/8
ß/8
o'~~~~~~~--~~~
\
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9
Frajjcn dtdqJ;m .....
a) b)
Figure 7-21. Limitation ofwave front area: a) reconstructed images, b) - speckle contrast as a
function ofthe area limitation degree
(7.3.4)
n (n2cr"
P(n)=-z exp --z
2cr"
Z
) (7.3.5)
(7.3.6)
, . . '..... . : .
. ~:::::::::-r""
. -' -' ..... '..
',L: --~ -.. -_.- -
: f I , I •
\-, ........
~~:~~~~~:~~~:~
........... r- .. -. .. .. ,- -
<- . ; . .
~ • , I , •
~J._-="''___
'' ~''_''''''''''''''_'''''_''''''''''
!I) CI :m tSI 2m J)
..................
CI
_.&._
$I
....................210
:IX)
..
a) b) e)
Figure 7-22. Distribution density funetions of speekle noise eaused by the limitation of the
wave front area: a) - exponential distribution, high limitation (30%); b) - Rayleigh
distribution; "Medium" limitation (90%); c) - Normal distribution, very low limitation (97%).
Oscillating eurves show empirie al distribution funetions, plane curves - theoretieal models of
Eqs. 4-6, respeetively.
"
GrLv
7/8
"
6/8
5/8
4/8
3/8
218
1/8
0.04
0.02
10 20 30 40 50 60
Number of qualIliza U"" la""l$
that the model of additive Gaussian noise can be used with respect to the
magnitude ofthe reconstructed wave front (square root ofthe intensity):
(7.3.7)
1
cr oc- (7.3.8)
n Q'
0.2
0. 15
0.1
0.05
33 49 65 81 97 113 129
Number of quantization levels
Figure 7-24. Optimization ofnonlinearity index P in P-th low quantization ofwave front
orthogonal components.
312 Chapter 7
References
1. P. J. Huber, Robust Statistics, John Wiley and Sons, N.Y., 1981
2. W. H. Press, S. A. Teukolsky, W. T. Vetterling, B. P. Flannery, Numerical Receipes in
C. The Art of Scientific Computing, Second Edition, Cambridge University Press, 1995
3. L. Yaroslavsky, M. Eden, Fundamentals ofDigital Optics, Birkhauser, Boston, 1995
4. L. B. Lesern, P. M. Hirsch, J. A. Jordan, Kinoform, IBM Joum. Res. Dev., 13, 150, 1969
5. L. P. Yaroslavskii, N. S. Merzlyakov, Methods of Digital Holography, Consultant
Bureau, N.Y., 1980
6. N. C. Gallagher, B. Liu, Method for Computing Kinoforms that Reduces Image
Reconstruction Error, Applied Optics, v. 12, No. 10, Oct. 1973, p. 2328
7. Goodman, lW.,. Statistical Properties of Laser Speckle Patterns In Laser Speckle and
Related Phenomena, J.C. Dainty, ed. Springer Verlag, Berlin, 1975
8. L. Yaroslavsky, A. Shefler, Statistical characterization of speckle noise in coherent
imaging systems, in: Optical Measurement Systems for Industrial Inspection III, SPIE's
Int. Symposium on Optical Metrology, 23-25 June 2003, Munich, Germany, W. Osten,
K. Creath, M. Kujawinska, Eds., SPIE v. 5144, pp. 175-182
Chapter 8
SENSOR SIGNAL PERFECTING, IMAGE
RESTORATION, RECONSTRUCTION AND
ENHANCEMENT
(8.2.1)
A=H·B, (8.2.2)
(8.2.3)
(8.2.5)
gives a credit to Norbert Wiener for his pioneer works in the theory of
statistical methods of signal restoration ([ 1])1 .
In order to implement optimal scalar Wiener filter, one should therefore
know cross-corre1ation {AVnA AVnJa,ß;)} between filter input signal and
perfect signal spectral coefficients and power spectrum ~VnA AVnN ~ß, r)}
of the input signal in the selected basis. The statistical approach we adopted
that assumes averaging of the restoration error over statistical ensembles of
perfect signals and of filter input signals implies that these statistical
parameters should be measured in advance for these ensembles or over the
data bases.
(8.2.7)
In spectral domain, the same relationship holds for signal and noise spectral
coefficients:
ß,=a,+v" (8.2.8)
where ~,} = T {nk } • For this model one can obtain that
and
1 It will be just to give also a credit to Andrey N. Kolmogorov who developed the similar
theory for discrete signals ([2]).
318 Chapter8
(8.2.11)
One ean give to this formula a clear physieal interpretation. Define signal-
to-noise ratio as:
A V. I-v ,1 2
SNR = _....:O""A+jU'-'-'::-t (8.2.12)
, AVON IVk., 1
2
Then obtain:
SNR,
(8.2.13)
11, == 1 + SNR,
whieh means that seal ar Wiener filter weight eoefficients are defined, for
eaeh signal speetral eoefficient, by the signal-to-noise ratio for this
eoeffieient. The lower is signal-to-noise ratio for a partieular signal speetral
eomponent, the lower will be the eontribution of this eomponent to the filter
output signal.
In order to implement sealar Wiener filter one have to measure in advanee
power speetra A V0A (r, 2
1 ) and A VON (~ ,
2
1 ) of perfeet signals and of noise
in the seleeted basis. Noise power speetrum may be known from the
specifieation eertifieate of the imaging deviee. Otherwise it may be
measured in noisy input signals using methods deseribed in Seet 7.1.2. As
for the prefeet signal power speetrum, it is most frequently not known.
However, one ean, using Eq. 8.2.10, attempt to estimate it from the power
speetrum A V0A A VON (I ß, 1
2
) of input noisy signals as
(8.2.14)
The latter has to be estimated from the observed signal speetrum 1ß, 12 •
Denote this estimate as IßJ . This empirieal estimate made by averaging
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 319
RECONSTRUCTION AND ENHANCEMENT
(8.2.15)
(8.2.16)
11, = [
max 1-ß, 1
2
Iß ,1
-O'n
2
2
; ]
0 . (8.2.17)
11
,
= max[Iß,I Z -0';.
IßJ'
0]. (8.2.18)
Note that such an empirical Wiener filter is adaptive because its weight
coefficients depend on the spectrum of the image to which it will be applied.
Weight coefficients ofscalar Wiener filters assume values in the range
between zero and one. Aversion of the empirical Wiener filter of Eq. 8.2.18
with binary weight coefficients:
11 =
, 0,
if{t, Ißl
~ Thr
otherwise
(8.2.19)
320 Chapter8
Output
Input
Softlhard
thresholding
.L()w pass
flltering and
downsampling
SofUbard
thresholding
Low pass
fiJtering and
downsampling
Lowpass
illtering and
downsampling
narrow band noise filtering. Fig. 8-3 demonstrates filtering periodical noise
pattern in an image. Such interferences frequently appear in images digitized
by frame grabbers from analog video. Left column in Fig. 8.3 shows input
and filtered images. Right column shows averaged spectra of input and
output image rows. One can c1early see anomalous peaks of noise spectrum
in input image spectrum that are eliminated in the output image spectrum
after applying empirical Wiener filtering designed using methods of noise
spectrum estimation described in Sect. 7.1.3.2. Note that the fiItering is
carried out in this example as 1-0 row-wise filtering in OFT domain. This
type of interferences can also be successfully filtered in Walsh transform
domain ([4]).
+. . . . . . . ,. . . . . . . . ,. . .
Column wise avera ed ower spectrum alon TOWS
14 ................ !................... j.................. j................
12 ········..··.1·················1...··········.1 .. ·········
6 .. ~t~JCC==1\=J~A.( · ·'· ·.
20 40 60 80 100 120
"Filtered"
14 ...............,.................................. , ............... , ...............•................. , ..... .
12 ··························;······,,·,,·'·'·'1'···············r··············T····
6 '~~~=~~::J:::=~:::r""""""'T'
20 40 60 80 100 120
Figure 8-4. Periodical noise in a Fresnel hologram and a pattern of rejecting filter coefficients
for removing the noise. The noise manifests itself in horizontal and vertocal bands in OFT
spectrum of the hologram. (The hologram was kindly granted by Or. T. Kraus, BIAS,
Germany)
Upper right plot in Fig. 8-5 shows row-wise mean values, or row-wise
image Radon transform, (horizontal coordinate) as a function of the row
number (vertical coordinate). Taking as an es ti mate of perfect mean row-
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 323
RECONSTRUCTION AND ENHANCEMENT
wise values a mean value over all rows (shown with a straight line in the
bottom right plot in Fig. 8-5), one can estimate banding interference on
every row by subtracting this estimate from the observed mean values.
Subtracting the found values from all pixels in the corresponding row
eliminates the noise.
Wiener filtering of wide band noise and especially white noise is less
efficient. When filtering of white noise, Wiener filter tends to weaken low
energy signal spectral components. However usually these components are
exact1y the components that are the most important because they carry
information about signal changes such as at edges in images. Moreover,
Wiener filtering converts input white noise into output correlated noise
though with a reduced variance. As one can see from Eq. 8.2.11, in case of
the intensive input white noise, power spectrum of the residual noise is
roughly proportional to the signal power spectrum which means that the
residual noise becomes, statistically, signal alike. This may hamper
subsequent image analysis. In particular, it is well known that human vision
is more sensitive to correlated noise that to white noise ofthe same intensity.
Therefore Wiener filtering for image denoising may even worsen images.
Fig. 8-6 gives an example of computer simulation of Wiener filtering of a
test image with additive white noise. In the simulation, one can implement
the ideal Wiener filter because both signal and noise spectra are known. The
empirical Wiener filter was implemented in this example according to Eq.
8.2.18. As one can see from the figure, ideal Wiener filtering does improve
image quality. For the empirical Wiener filter, improvement is much less
appreciable. The difference image between initial noisy image and the result
of the empirical Wiener filtering (restoration error) shows that the filtering,
along with noise suppression, destroys image edges.
Consider now an imaging system model that accounts also for signal linear
transformations in imaging systems. Suppose that the linear transformation
unit can be modeled as a scalar filter with filter coefficients {A. r j in a
selected basis. For DFT basis, this assumption is just, to the accuracy of
boundary effects, for shift invariant linear filtering (see Ch. 5). In this case
coefficients {A. r j are sampies of imaging system frequency response. In
ideal imaging systems, they should all be equal to unity. In reality, they
decay with the frequency index r, which results, in particular, in image blur.
Processing aimed at correcting this type of distortions is frequently referred
to as image deblurring.
324 Chapter8
·i,~'iik~· )'~~:.\r~:~ ·t.f.~ :'$.c.r;~·; ).~ : . .~: , lrule reoo,~1)' uni. 111011: 8~'I1I), alt•
. . ~mit. 'i~(U, : .2.~': · jti~ ·( ~~~rp .· · ; 'oplerlLf lhw. an:: .!\Iang J be rnon Ilmulll'l\
IJl! . '}:~;:-i!~~ J:'1tt;~:. k.o\:~~!); ~\"J,P.ff~, ~ !' I.ey :iJtv'.lh' ~ evsY kJl~\jo1l SCAlt-ftom II
~ ~':iioi~ß CU! " !:~~'';•• <~(j;, ";" ~l'led ~. :lenniniog IM $'Inlt!tUJ'C of unroolved s:
~ . . ...... . t ;~~ .r..\ ~ ...... ''/ :'~.'" .,~ _" "', .
c.L I!l:: .,~$st ~~!I !t:(l .:_::il\\'t)!(1 i)\. ;11,
',.1 . . .
Figure 8-6. Ideal Wiener and Empirical Wiener filtering for image denoising
(8.2.20)
Using Eqs, 8.2.6, one can obtain that the scalar Wiener image restoration
filter is defined in this case by the equation:
SNR 1
(8,2.21)
11 = ~ l+S;Rr
r '
where SNR is signal-to-noise ratio at the output of the linear filter unit of
the imaging system model:
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 325
RECONSTRUCTION AND ENHANCEMENT
(8.2.22)
(8.2.23)
(8.2.24)
and
11 r _{~'ifIßJ
- Ar ~ Thr . (8.2.25)
0, otherwise
All these filters may be treated as two filters in cascade: the filter with
coefficients
in. 1
11 r = -:;: (8.2.26)
r
usually called the inverse filter and signal denoising filters described by Eqs.
8.2.17 - 19. Inverse filters compensate weakening signal frequency
components in the imaging system while denoising filters prevent from
excessive amphfication of noise and perform what is called
"regularization" of inverse filters. As one can see from Eq. 8.2.22, weight
coefficients of denoising filters for small {A r } fall faster than weight
coefficients ofthe inverse filter grow.
One of thc most immediate applications of inverse filters is correcting
distortions caused by finite size of apertures of image sensors, image
discretization devices and image displays. We will refer to this processing as
to aperture correction.
326 Chapter8
I I
: fiel) : I Ax I
~~. : n : n
Sensor's aperture
______ L.. ___ .1 ___ _
(8.2.29)
Parameters d(d) , d(r)and /).x are imaging system design parameters that
may be know system' s certificate. They can be used for correcting image
distortions by processing images in computer.
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 327
RECONSTRUCTION AND ENHANCEMENT
Figure 8-8. Aperture correction: initial (Ieft) and aperture corrected (right) images
Figure 8-9. Effect of spatial masking of reconstructed images due to the finite size of the
hologram recording aperture (a) and the result ofits compensation (b) (adopted from [5])
328 Chapter 8
In the derivation of the MSE optimal scalar linear filters, the size of input
signal vectors was a free parameter. In the filter implementation, one should
select the fashion in which signals are to be processed. Given an image to be
processed, filtering can be designed and carried out either over the entire set
of available image sampies or fragment-wise. We will refer to the former as
to global jiltering and to the latter as to local filtering.
There is a number of arguments in favor of "local" filtering versus
"global" one:
- "Global" filtering assumes signal "stationarity". For scalar linear
filtering, this means signal homogeneity in terms of signal power spectra:
spectra measured over different signal fragments should not substantially
deviate one from another. One hardly can regard such signals as images as
being spatially homogeneous. F ig. 8-10 illustrates image spatial
inhomogeneity on an example of a noisy "Lena" image. One can see that
while noise in the image is spatially homogeneous, individual fragments of
the image substantially differ one from another visually and, obviously, in
their power spectra as weIl.
- Adaptive filter design assumes empirical evaluation of signal spectra.
Global evaluation of image spectrum disregards spectra variations due to
image inhomogeneity. This obviously results in neglecting local image
information in favor of global one which usually contradicts processing
goals.
(8.3.1)
where W is it the filter window size and the size of the area over wh ich
filtering performance is evaluated, k is an index of the window position,
{ß ;k)} are spectral coefficients of the filter window sampIes, fx. !k)} are
spectral coefficients of filter window sampIes of a hypothetical perfect
signal. Solutions obtained in Sect. 8.2.2 and 8.2.3 for empirical Wiener
filters for image restoration may then be extended to local filtering by
replacing in Eqs. 8.2.16 - 19 and 8.2.23-26 signal spectra with corresponding
spectra ofthe signal window sampIes.
The most straightforward way to implement local filtering is to perform it
in a hopping window. This is exactly the way of processing implemented in
most popular audio and image coding methods such as JPEG and MPEG.
"Hopping window" processing being very attractive from the computational
complexity point of view suffers however from "blocking effects" - artifacts
in form of discontinuities at the edges of the hopping window. Obviously, an
ultimate solution of the "blocking effects" problem would be sliding window
processing. Sliding window filtering is local adaptive because, for empirical
330 Chapter8
Polnt-wln InVe1"se
multipUcadon
byÜte mter
Fllter
mask
spednlm
r=1 .=1
involves only signal spectrum coefficients with even indices. Therefore only
those spectral coefficients have to be computed and the computational
complexity of sliding window filtering in DCT domain is O[(W + 1)/2]
Y
operations for I-D filtering and ol(w + 1 /4 J operations for 2-D filtering
in a square window of W x W pixels.
Fig. 8-12 shows an example of local adaptive empirical Wiener rejecting
filtering in DCT domain according to Eq. (8.2.19) of a noisy
electrocardiogram in the window of 25 sampIes. Upper plot shows initial
electrocardiogram. Image in the second row ofthe figure shows local spectra
of the electrocardiogram for each window position. Image gray levels
represent here amplitudes of spectral coefficients, vertical coordinate
represents spectral component indices, horizontal coordinate represents
window position indices. Such a signal representation is its time-frequency
representation. Image in the third row displays the same spectra after
rejecting signal spectral components that do not exceed the selected
threshold. The bottom plot shows the filtering.
Fig. 8-13 illustrates local adaptive rejecting filtering of a piece-wise
constant test image. Upper left image is th initial test image. Upper right
image is the input image with noise added. Left bottom image is the result of
the filtering. Right bottom image shows a map of the rejecting filter
"transparency" for each window position; i.e., the ratio of the number of
ones in the array of the filter coefficients to the window size. One can see
from this image that, when filter window seats on an image patch where
image gray values are constant, the filter is almost completely opaque and
332 Chapter8
2500
2000
1500
1000
~. --~~-:Jt :-:::- .
~~~~-;!~.- ~-~~~.~~
Initial signal
::,:---~::..-m-= ~~
500 1000 1500 2000
local OCT speclrum;SzW=25
5
-:------.-------------
10
15
20
25
500 1000 1500 2000
Modified local speclrum: SZW=25: Thr-O.OOl
~ - -- ------ ~ -
5
10
15
20
25
500 1000 1500 2000
Reslo<ed signal
2500
2000
1500
200 F~"'''-'
158 .---- :::::: :: -------------~------I----~=='''''Iiuo L~~
~::::::::::::t:::::: :::: j-" - .:::.::::l::: -::::::::~
50 ~"·--·----------- -- f- --------- ~t·--' --· ........ ±
50 100 150 200 250
32-th row of the nolsy Image
200
100
~=--,"r-Wj~l' (· -------- --1--- -ur:'!~~:----,..~-·--t:·T--~'!:~rtl
0/---- -- - ------------- ,--- .-- -.. - W. __ .. ____ , ._. - .- .. ---,-
o !ui,•.
v. 1/1 ./' w. .. .J..J ~ I........, : i
50 100 150 200 250
32-th ro\IV of the filtered image
200
100 ~"" J'"''j----··---·-··I~-·-···{c·:tj·-·f------
:"·"·"·"-r··"·· . . . :
--- .. --- .. --l~~ -.--·-J··r
·-----·-::1:1
t . __ _ \ ____
Z
1
o _.~. - =-1 • \0-.. ,' •
Figure 8-14. A typical row ofthe initial, noisy and filtered images
334 Chapter8
Figure 8-15. Denoising MRI image. Left to right: initial noisy image, filtered image, and
differenee between initial and filtered images
Figure 8-16. An example of loeal adaptive filtering speekle noise in ultrasound images: initial
(Ieft) and filtered (right) images
In general, optimal window size depends on the image and may vary
even within the image. This requires filtering in multiple windows
with an appropriate combination of the filtering results.
One can avoid filtering in multiple windows by combining multi
resolution property of the wavelet shrinkage and local adaptivity of
DCT filtering in the hybrid filtering method ([10]). In the hybrid
filtering hard or soft thresholding in each scale level of the wave let
filtering shown in a flow diagram of Fig. 8.2 is replaced by local
adaptive filtering in DCT domain with a fixed window size. Owing to
multi scale image representation in wave let transforrn, this
replacement imitates parallel filtering of the input image with a set of
windows accordingly to the scales selected. Flow diagram of the
hybrid wave letlDCT filtering is shown in Fig. 8.17.
Output
Input
Sliding
window
filtering
Low pass
filtering and
downsampling
Sliding
window
ftItering
Low pass
tiltering and
down 'ampling
Low pass
filtcring and
downsampHng
Figure 8-17. Flow diagram of the hybrid sliding window DCT domain and wavelet denoising
336 Chapter8
1 1 1 -1 -1 -1
q> z,o =~ [ - 1
- 1 + 2 - 1]
+ 2 - 1 ; q> z,z = ~
[-1 - 2
-1]
- 2 + 4 -2 . (8.3.3)
-1 +2 -1 -1 -2 -1
The second and the third functions represent what is called vertical and
horizontal Laplacian operators. The last function can be decomposed into a
sum of diagonal Laplacians:
initial noise free image and filtered ones. For DCr moving window filtering,
window size and parameter Thr were optimized. For wavelet shrinkage and
hybrid filter, parameters Thr were optimized and the best of the following
types of wavelets was chosen: Haar wavelet ("haar"), Binomial coefficient
filter ("binom3", "binom9" and "binomI3), Daubechies wavelet ("daub2"
and "daub4") and Symmetrie Quadrature Mirror Filters ("qmf5", "qmf9"
and "qmfl3"). rhe table summarizes results obtained for four of 12 test
images (a test piece-wise constant image, "Lena" image, an MRI image and
an air photograph) of 256x256 pixels with 256 quantization level and
standard deviation of additive Gaussian noise of 13. For Dcr moving
window filtering, optimal window size is shown in brackets. For the wavelet
shrinkage, the best wavelet filter kerne I is indicated in brackets .
.,
1It:t
Figure 8-18. Effective basis functions in four scales ofthe hybrid DCT/wavelet filtering
Table 8-1. Standard deviation of the difference between initial noise free and filtered
images for different filtering methods Standard deviation ofadditive noise is 13)
Filter Piece-wise Lenna MRI Air photo
constant image
ima2e
DCT Hard 8.1 9.4 6.7 8.2
(3x 3) (3x 3) (5x 5) (3x 3)
Soft 7.5 8.6 6.3 7.7
(3 x 3) (5x 5) (7x 7) (3x 3)
Wave let Hard 8.6 10.1 8.5 9.3
shrinkage. (binom5) (binom5) (binom5) (binom5)
Soft 8.4 9.0 7.8 8.1
(binom5) (amf13) (binom5) (binom5)
Hybrid Hard 8.7 9.4 6.6 8.2
(binom5) (binom5) (binom5) (binom5)
Soft 7.9 8.6 6.2 7.5
(binom5) (binom5) (binom5) (binom5)
338 Chapter8
Fig. 8-19 compares filtering results for two of the test images. To ease
visual comparison, 64x64 pixel image fragments are shown.
Figure 8-19. Sliding window DCT domain, wave let and hybrid filtering for image denoising:
a) initial images; (b) fragments of the noisy images, (c, d) DCT filtering with hard and soft
threshold respectively. (e, f) wavelet shrinkage filtering with hard and soft threshold
respectively. (g, h) hybrid filtering with hard and soft threshold respectively.
2.S
1 .~
~
'... .,
'j
; .
! .
1
. : ·1 o.~ ~ ." . "
'!
., .:.
I
0 .4 0 .6 0 .8 1 0 .2 0 .4 0 .6 0 .8 1
i
u u .u ~:~ lf~" _ '
o~ ~: ..
'. '.]
j
::V\',~~ .
0 .2 0 .4 0 .6 0 .8 1 0 .2 0 .4 0 .6 0 .8 1
0 :3
/: :~'::.' ~
8~ ~ . u/ .. • u .
'
'<',.
~
~ • . .. :! . . . . . . .
... , . . .. ~l f·V~
/ ~. ". uj'-1'"'U'~"j
81_/· . : ..... · :-··~ . j . ' . _
0 .2 0 .4 0 .6 0 .8 1 0 .2 0 .4 0 .6 0 .8
u,.. u. u, ' '', ...
o+>;-~
8 ~ U' u ;~ u.
0 .2 0.4 0 .6 0 .8
!·· 1 1
~
0 .2 0 .4 0 .6 0 .8 1
Normalized 1requency NormaHzed frequency
Figure 8-20. Wave1et domain (Binom5) and hybrid sliding window ocr domain processing
as sub-band decompositions
340 Chapter8
(8.4.2)
(8.4.3)
one can obtain the following optimal restoration filter coefficients ([12]):
(8.4.4)
where
(8.4.5)
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 341
RECONSTRUCTION AND ENHANCEMENT
M
"b(m) /cr (m)
L... k n
(8.4.6)
where {b!m)} are sampies of image components and ~ ~m)} are standard
deviation of uncorrelated additive noise in image components, follows from
Eq. 8.4.2 and 8.4.44 in the assumption that tx~m)=ar}, {A~m)=l} and
AV 11 (m)1 2)oc{cr(m))2
nn \J" r ~ n •
It may very frequently happen that different images in the available set of
images contain the object in different position. In this case images should be
aligned before averaging. This conclusion follows in particular, from Eq.
8.4.4 in which when DFT basis is used
(8.4.7)
where, according to the shift theorem for DFT, {k~m,/)} are mutual
displacement between m-th anf I-th image components.
When mutual displacements of images are known, images can be aligned
through their resampling with an appropriate interpolation. Methods of
image resampling and other geometrical transformation are discussed in eh.
9. If displacements {k~m,/)} are known with a sub-pixel accuracy, in the
averaging sampies of images will be interlaced according to the mutual shifts
of the images. Therefore the resulting image will have accordingly higher
sampling rate than that of the original image set and, hence, will have higher
resolving power. This resolution increase is frequently referred to as super-
resolution from multiple images.
Mutual image displacements are a special case of general image
geometrie transformations. Obviously, image alignment before averaging
should allow for all transformations.
When mutual image displacements or corresponding parameters are not
known, they have to determined. Determination of image displacements or
342 Chapter8
(8.5.1)
"(I-I) ifl"(/-I)
ak,l
,,(tl _ , I ak,l "(H)I > De~,
- ak,l t'T''h r (I)
akr - --- { , (8.5.2)
'a(/-I)
k,l , otherwise
where {a~/,~} are filtered image sampIes on t-th iteration, t = 0, ••• , Nil '
{a(O)
k,l = bk,r'} a(/-I)
k,t are "predicted" sampIe values found by averaging of
neighbor pixels:
a"(H)
k,l
-_ -81 [~
L. ~L.a"(I-I) -a"(I-I)]
n=-lm=-] k-n,l-m k,l '
(8.5.3)
Figure 8-21. Filtering a printed text image corrupted by impulse noise with iterative and
recursive filters: (a) - noisy image with probability of impulse nooise 0.3; (b) - result of
iterative filtering according to Eq. 8.5.2 after 4 iterations; (c) - result of recursive filtering
according to Eq. 8,5.5
346 Chapter8
The described filters are not the only options for filtering impulse noise.
In general, for filtering impulse noise, non linear filters are required. Some
nonlinear filters for filtering impulse noise are described in Sect. 12.3 .
In conclusion of this section, consider ways of numerical evaluation of the
impulse noise filtering. MSE criterion used in the design of filters for
suppressing additive noise is not applicable for impulse noise. One can
easily see it from comparison of images distorted with additive and impulse
noise with the same MS-error with respect to the noiseless image (Fig. 8-22).
Figure 8-22. Images distorted by additive noise (a) and impulse noise (b) with the same MS-
error with respect to the noiseless image
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 347
RECONSTRUCTION AND ENHANCEMENT
The evaluation criterion for impulse noise filtering must take into account
the following three factors.
Distortions due to missing noisy sampies and the probability of
missing that determines contribution of these distortions to quality
corruption of the filtered image.
Distortions due to false noise detection and the probability of false
alarms that determines contribution of these distortions to quality
corruption of the filtered image Probability of false noise detection.
Estimation error for those pixels that were correctly detected as
replaced by noise.
Each of these errors can be evaluated in terms of their mean square values.
However, their contribution to the total evaluation is of more involved nature
and depends on application.
348 Chapter 8
In the imaging system model of Fig. 8-1, gray scale distortions are
attributed to the point-wise nonlinear transfonnation unit. Usually in
correcting gray scale distortions, noise described by the stochastic
transfonnations is ignored. In such an assumption, for correcting image
nonlinear transfonnations in imaging systems it is sufficient to apply to
images the corresponding inverse transfonnation. However, in digital
processing one should allow for quantization effects and for the fact that any
nonlinear transfonnation of quantized signals reduces the number of signal
quantization levels. The phenomenon of sticking of quantization levels
together as the result of their nonlinear transfonnation is illustrated in
Fig.8-23.
~l
Quantized I-""
output ,/
values
V
f-"'
Quantized input values
Figure 8-23. Nonlinear signal transfonnation and sticking quantization levels together. Thin
lines on the figure outline 16 unifonn quantization intervals. The curve shows transfer
function of a nonlinear transfonnation. Bold lines indicate 9 quantization levels left after the
nonlinear transfonnation of 16 quantization levels of input values
,-----------~---------
Compressing
transformation Distorted
for nonuniform signal values
quantization
imaging
system
Boundaries
ofuniform Boundaries of
quantization quantization
intervals
quantization
Perfeet
signal
values
Expanding
Figure 8-24. Principle of correcting nonlinear distortions with an account for effects of
quantization
350 Chap/er8
Contrast corrected image and false contours Destroying false contours by adding
noise to the corrected image
Figure 8-25. Image contrast correction with adding pseudo-random noise for randomization
of false contours
8. SENSOR SIGNAL PERFECTlNG, IMAGE RESTORATION, 351
RECONSTRUCTION AND ENHANCEMENT
~
Imerferogram with distorted gray scale Corrected"'=:''interferogram
,,~/- : ,:;.: ' w / ' - .. .
Perfecting Perfecting
sensor's signal: Inverse reconstructed image:
correcting
distortions and
--. transform
f-+ Correcting distortions,
denoising, and
denoising enhancement
Figure 8-27. Principle of image reconstruction by inverse filtering via inverse transform
speed. This is however, not the inverse transfonn required for the exact
reconstruction. Images reconstructed in this way are blurred and need
restoration. Left image in the second row of Fig. 8-28 illustrates such an
image restoration from parallel projections (right image in the first row) of a
test image shown in the left image in the first row. Inverse filtering applied
to images reeonstrueted by back projection can substantially improve image
quality and approximate exaet tomographie reconstruetion espeeially if it is
supplemented with a post-proeessing for removing inverse filtering artifacts.
To enable the inverse filtering, one should measure point spread funetion
of the tomographie maehine with back projeetion reeonstruetion. This ean be
done using specially phantom objects such as for instance, a point objeet.
The system PSF uniquely specify tomographie reeonstruction systems as
they can be treated as linear ones.
Let V be a veetor of observed image sampies, V be a result of its
"
reconstruetion in a given imperfect tomographie system, V be the eorreeted
result, Vc be a ealibrating test image for measuring the system point spread
funetion, and Vc is its eopy reconstructed by the system. The inverse
filtering proeessing can be deseribed, in a matrix fonn, as following:
(8.7.1)
where IF is a restoration linear operator obtained as:
(8.7.2)
v= IFFT{CD eFFT(V)} ,
IF (8.7.3)
where FFTO and IFFTO are direet and inverse Fast Fourier Transfonns
and CD IF is diserete frequency response of the inverse filter. It ean be found
from FFT(Vc ) and FFT(Vc ) as
(8.7.4)
(8.7.5)
with a spread parameter cr c was set to have magnitude of about pixel size.
Frequency response of the "inverse filter" was generated, similarly to
Wiener deblurring filter ofEq. 8.2.21, as
- IF·
IF = z ij
IIFI + E Z • var IFI z)'
(8.7.6)
Figure 8-28. Illustration of the inverse filtering for image restoration from back projection
reconstruction
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 357
RECONSTRUCTION AND ENHANCEMENT
Oprical hologram
--- ---- ---
--~~~~ - - -- -- --- - -- -- -- -- -
Analog-to-digital
conversion
-Preprocessing of
Computer I
digitized
hologram
,
\
\
,,
I
--
I
--
--
- - - - -----
~ "..~ - -- ,
- ------
,
I
I
I
I
I
,
,
I
Output image I
,,I
I
I
I
,
I
I
\
\
_ 1 N, N2
where {r"s} is an array of hologram sampies with (r, s) being the sampie
indices. As it is described in Sect. 5.1 .1, local mean in a sliding window can
be very efficiently computed recursively. In this application, window size is
usually 3 x 3 to 7 x 7 pixels. Fig. 8-30, center, shows result of
reconstruction of the hologram shown in Fig. 8-4 after removing zero-order
diffraction spot. Right image in this figure shows result of the hologram
reconstruction after both removing zero-order diffraction spot and periodic
interferences.
Figure 8-30. Reconstruction of hologram without pre-processing (a) and with preprocessing
for removing zero order diffraction spot (b) and for removing, in addition, periodic
interferences (c)
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 359
RECONSTRUCTION AND ENHANCEMENT
Fig. 8.31 illustrates elimination of zero order diffraction spot and image
post-processing in digital reconstruction of a hologram recorded in two
exposures by the method ofphase-shifted digital holography.
Normally, phase-shifting method for digital recording optical holograms
assumes taking three or four exposures of the interference pattern of object
and reference beams with phase shifts of the reference beam O,1t /2,1t for
three exposures and O,1t / 2,1t,31t /2 for four exposures ([16]). This gives
the following three or four recorded holograms, correspondingly:
and
With two exposures that produce two holograms, r o and r"/2' the
reconstruction that uses these two holograms as real and imaginary parts of
the object hologram, gives a reconstructed image hidden behind a heavy zero
order diffraction spot due to hologram components IRefl2 and IObjl2 as it is
360 Chapter8
Figure 8-31. Removing zero-order Reeonstrueted image denoising; (a) - image reeonstrueted
direetly from two exposure holograms; (b) - image reeonstrueted from two exposure
holograrns after preproeessing aeeording to Eq. 8.7.8; (e) - reeonstrueted image after filtering
speekle noise using loeal adaptive filtering algorithm (Seet. 8.3.2); (d) - result of subsequent
fihering remaining impulse noise (Seet. 8.5)
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATJON, 361
RECONSTRUCTJON AND ENHANCEMENT
One of the features that determine the capability of vision to detect and
analyze objects in images is contrast of objects in their brightness with
respect to their background. While globally image may utilize the entire
dynamic range of the display device, locally, within small windows image
signal has very frequently much lower dynamic range. Small local contrasts
may make it difficult to detect and analyze objects.
The simplest and the most straightforward way for amplification of local
contrasts is stretching local dynamic range of image signal in a sliding
window to the entire dynamic range of the display device. The two
computationally inexpensive methods for stretching local contrasts are max-
min stretching and local dynamic range normalization by local mean and
variance. They are described correspondingly as
and
"
ak,1 { -
am.. '
= a" , (8.8.2)
amin'
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 363
RECONSTRUCTION AND ENHANCEMENT
where
a - mean(WJ,Wl)
'ii = g
A
ak,l and a k •1 are input and output image sampies in the window position
(k,/), maxk,l' mink,l' are maximal and minimal signal values within the
window, a ma• and amin are boundaries ofthe display device dynamic range,
(8.8.4)
tll:,l
Lh{q)
)11=11....
A
ak,l = qmB>
(
- qmln tI + qmln' (8.8.6)
!h{q)
where qmln and qma. are minimal and maximal quantized values of image
signal (for 256 quantization levels these are 0 and 255).
364 Chapter8
It follows from Eq. 8.8.5 that the slope of the histogram equalization
transfer function ak,1 (a k,/) is proportional to the histogram value:
{).a-kl' och( )
- (8.8.7)
{).a q q=lIk,l'
k,1
(8.8.8)
Figure 8-32. Local histogram modification by means of standardization of local mean and
variance (uper row) and by P-histogram equalization (middle and bottom rows)
366 Chapter8
----r---lf---t---l---j
i l / ! I I
---lf- --1-- !
200
l
! I
100
1 I I I
I j I I!
50 ----r~( r---T--TI
o
50 100 150 200 250
250
i i
200
! I
--r-·-r--r----, ---, 1 ! !
! I
__.---L._I-___ i
L -1.__ I
!
ISO
I
i
Ii I
. I
i
i
!
100 ..-----+--------i--- --4--+---~
! i i I I
50 ---1---- !.---~---t--j
I
OL-~L-~--~--L-~
I
I
i !
i i
SO 100 ISO 200 250
(8.8.9)
~", ,,,,,,·,,,,·; · ······ ·· ,,,, .. 1..·..............;· .... ······,,··!········ ......... , ......... ~.,,;
.............." ............ .
2.2
'. f· . . ······· ; ........, j .. . .
·+. . . ·. . . ·j. . . . . . . . :. . . . . . .
1.8 . ,··j················!···············1· .
........... ;W=3T....·......
",· · , ,·· ·· · · · ~· · · · · ··········!·· .. ············f······· ........ !................!..............
1.4
.. ·······~···············i················r············ · .. ,................ !. . . ......
0.2 0.4 0.6 0.8
Nonnalized frequency
Figure 8-34. I-D sections of unsharp masking frequency responses for window sizes 3 and 15
pixels with amplification coefficient g= I
frequencies is the same for any image. An alternative and adaptive method
for modification of image spectra that results in local contrast enhancement
is P-th law nonlinear modification of absolute values of image spectra
coefficients. It is described by the equation
~ Cl, h
S
Cl" s =~II'""'·s
I,
P
(8.8.9)
JU- ' ,S
where ~ ' .S fand p '.S i are initial and transformed image spectral
coefficients in a selected basis, respectively, and P is a user defined
parameter. As a transform, usually DFI or DeI are used. When 0::; P < 1,
this modification redistributes energy of spectral coefficients in favor of low
energy coefficients. The degree to which individual coefficients are
amplified depends now on the image spectrum.
Such a spectrum modification may be applied globally to the spectrum of
the entire image and locally in a user specified sliding window to spectra of
the window sampIes in each position of the window. In the latter case, the
modification is local adaptive. In Fig. 8-35 one can compare image local
contrast enhancement by means of unsharp masking and global and local
non linear modification of iI!l~ge DeI spectrum.
",.;;in"'.,..",,,....,.,,,.......ItO:"r._-
DCT global speetrum enhaneement, P=O.5 DCT loeal speetrum enhaneement, in the
window 15x 15; P=O.5
Figure 8-35. Unsharp masking and non linear spectrum modifieation for image enhancement
8. SENSOR SIGNAL PERFECTING, IMAGE RESTORATION, 369
RECONSTRUCTION AND ENHANCEMENT
Input image
Artifieial stereoscopie images for left and right eye generated using the image loeal mean as a
depth map.
Figure 8-36. Stereo visualization ofan image and its loeal mean. One ean observe 3_d image
using a stereoseope or by looking at the images with squinted eyes. In the latter ease, one
should tl)' to squint eyes until two pairs of images seen with eaeh eye overlap into 3 images.
Then the eentral ofthose three images with be seen as a 3-D image.
370 Chapter8
References
1. N. Wiener, The interpolation, extrapolation and smoothing of stationary times series,
Wiley, New York, 1949
2. A. N. Kolmogorov, Sur l'interpolation de suits stationaires, C. R. Acad. Sci., 208, 2043-
2045
3. D.L. Donoho and I.M. Johnstone, Ideal spatial adaptation by wavelet shrinkage,
Biometrica, 81(3): 425-455, 1994
4. L. Yaroslavsky and M. Eden, Fundamentals ofDigital Optics, Birkhauser, Boston, 1996
5. L. Yaroslavsky, N. Merzlyakov, Methods of Digital Holography, Consultant Bureau,
N.Y., 1980
6. M.D. Levine, Vision in Man and Mashine, , McGraw-HiII, 1985, pp. 110-130
7. L.P. Yaroslavsky, Local Adaptive Filters for Image Restoration and Enhancement, In:
Int. Conf. on Analysis and Optimization 0/ Systems. Images, Wavelets and PDE's, Paris,
June 26-28, 1996, Eds.: M.-O. Berger, R. Deriche, I. Herlin, J. Jaffre, 1.-M. Morel,
Springer Verlag, Lecture Notes in Control and Information Sciences, 219., 1996, pp. 31-
39.
8. L. Yaroslavsky, "Local Adaptive Image Restoration and Enhancement with the Use of
OFT and DCT in a Running Window", Proceedings. Wavelet Applications in Signal and
Image Processing IV, SPIE Proc. Series, v. 2825, pp. 1-13, 6-9 August 1996, Denver,
Colorado.
9. L. Yaroslavsky, Image Restoration, enhancement and target location with local adaptive
filters, in: International Trends in Optics and Photonics, ICOIV, ed. by T. Asakura,
Springer Verlag, 1999, pp. 111-127
10. B.Z.Shaick, L. Ridel, L. Yaroslavsky, A hybrid transform method for image denoising,
EUSIPC02000, Tampere, Finland, Sept. 5-8, 2000
11. ftp:/ift;p.cis.upenn.edu/pub/eero/matlabPyrTools.tar.gz
12. L. P. Yaroslavsky, H. J. Caulfield, Deconvolution ofmultiple images ofthe same object,
Applied Optics, v. 33, No. 11, pp. 2157-2162,1994
13. L. Yaroslavsky, M. Eden, Correlational Accumulation as a Method for Signal
Restoration, Signal Processing, v. 39, pp. 89-106, 1994
14. N. EI Sherif, G. Turitto, eds., High Resolution Electrocardiography, Futura Mount Cisco,
N.Y., 1992
15. B. L. Trus, M. Unser, T. Pan and A.C. Steven, Digital Image Processing of Electron
Micrographs: the PIC system U", Scanning Microscopy, v. 6,1992, pp. 441-451
16. I. Yamaguchi, 1. Kato, S. Ohta, 1. Mizuno, Image formation in phase shifting digital
holography and application to microscopy, Applied Optics, v. 40, No. 34, pp. 6177-618
17. Photography for the Scientist, Ch. E. Engel, ed., Academic Press, London, 1968, pp. 632
18. V. Kim, L.P. Yaroslavsky, Rank Algorithms for Picture Processing, Computer Vision,
Graphics and Image Processing, v. 35, 1986, p. 234-258.
Chapter 9
..
-~----~----.,.--
I , I
I I I
\~ ... "". .••. I , I
Figure 9-1. The principle ofimage resampling for geometrical transformations: backward co-
ordinate mapping
Consider first a 1-0 model. Let {at} be a set of available signal sampies.
Generate an L -times extended set {at} of sampies by complementing each
available signal sampie with L of zero sampies:
- = ~-
ak L...ak-n hnint , (9.l.2)
n
(9.l.3)
with a sampling kernel {h:mP1 }. Eq. 9.l.3 is a digital counterpart of Eq. 3.3.1,
a) that describes sampling continuous signals.
9. IMAGE RESAMPLING AND GEOMETRICAL 375
TRANSFORMATIONS
_ LN-I
- - ~- h smp'
a k - L... a(k-n)modLN " (9.1.5)
n=O
is the only one that assures preservation of initial signal spectral components
and the absence of aliasing terms in spectra of the interpolated signals. We
will call this interpolation method and its versions the discrete sinc-
interpolation.
2-D signal interpolation can be implemented as either separable or
inseparable digital convolution. When carried out in signal domain,
separable digital convolution is computationally much more efficient then
inseparable one (Sect. 5.1.2). If the convolution is implemented in transform
domain, the use of inseparable convolution kerneIs is possible without any
loss of the computational efficiency. Inseparable interpolation kerneIs are
more natural for image processing. In Sect. 9.3 DFT and DCT based discrete
sinc-interpolation algorithrns will be described that, in principle, enable the
implementation of inseparable convolution kerneIs.
376 Chapter 9
) {I, O:s;
h: . t(O)
n = rect(
niL =
n:S; L-l
0, L:s; n:S; NL-l
. (9.2.1)
l'lr I- ~ _I sin(ltr
I... int(o)1 OCSlDCd(L-, LN-,r",- (
1
IN)). (9.2.2)
sin nrl LN
0.2
Figure 9-2. Frequency response of the nearest neighbor signal interpolation for 16-zooming
signals of 128 sampIes (a) and an example ofimage 8-zooming
9. IMAGE RESAMPLING AND GEOMETRICAL 377
TRANSFORMATIONS
The principle of linear and separable linear interpolation along two co-
ordinates (bilinear interpolation) is illustrated in Fig. 9-3 for interpolation of
signal sampies shifted with respect to available sampies by u -th fraction of
the sampling interval (u -th and v th fractions of the 2-D discretization
interval, for bilinear interpolation).
u i a1,2
::::.:::r.·
........
'. a
············l ...
::::::
a) b)
Figure 9-3. The principle of linear (a) and biIinear (b) interpolation. Available signal sampIes
are shown in bold lines. Interpolated sampIes are shown in doted lines
n / L, 0 :::;: n :::;: L -1
h~nt(l) = tri(n / L) = { 2 - n / L, L :::;: n :::;: 2L -1. (9.2.3)
0, 2L :::;: n :::;: NL-l
I... intl oc
l'lr I·SIßCd(L·, LN.,r11JI -_I sin(1tr
2
( / N)) 1
2
(9.2.4)
sin 1tr / LN
378 Chapter 9
It is shown for 16-zooming signal of 128 sampies in Fig. 9-4 (a) centered
around the symmetry point r=I024. An example of image 8-zooming with
bilinear interpolation is shown in Fig. 9-4, b). As one can see on the figures,
aliasing effects are much less pronounced with linear (bilinear) interpolation
while distortions of original signal frequency components within the base
band is more severe.
0 1 . - - -.........- - -- -
o 500 1000 1500
a) b)
Figure 9-4. Frequency response of linear interpolation for 16-zooming signals of 128 sampies
(a) and an example ofimage 8-zooming with bi linear interpolation
Nearest neighbor and linear interpolations are zero order and first order
special cases of spline (B-spline) interpolation ([1]). In the nearest neighbor
interpolation, one sampie of the original signal is involved in calculation of
each sampie of the zoomed signal. In the linear interpolation, two sampies of
the original signal are involved. In R -th order spline interpolation, R + 1
sampies are involved. Interpolation kerne I in R -th order B-spline is defined
as ([1])
h'pi(R)(X )_~(-l)'(R+l)
- ~ ( "
[O(R+l + x _ t )R]
max, (9.2.5)
t =O R + 1 - t ,.t! 2
O~lxl< 1
l~lxl< 2 (9.2.6)
2~lxl
9. IMAGE RESAMPLING AND GEOMETRICAL 379
TRANSFORMATIONS
0.9
0.8 ...... . , ..........;
0.7
· ··,·~· · ··· · · .. r"'·····r········-r·········
..... .. .... 'f ....... ,. : ....,... ,!, .. ..... .
0.8 ~.,
0.5
······r········r········r········T
·i· ·· · ··· ··t···· ···· ·~····· ·,·+ , · ,··
. . .
Figure 9-5. Cardinal cubic spline interpolation kerne I (a) and frequency response (b) for 16-
zooming signals of 128 sampies
380 Chapter 9
a =IDFT
k LN {[1- reet r - (N + 1)/2]. DFTN {a J } =
LN-N-1
1 N-I
JL ~an sined(N;N;(k 1L-n)) , (9.3.1)
a = IDFT
k LN {[1- reet ~~~ ~]. DFTN{a k }} =
1 N-I
JL ~ a" sined(N -1;N;(k 1L-n)) , (9.3.2)
or
a = IDFT
k LN {[1- reet r-NI2-1] . DFTN {}}
ak =
LN -N-2
1 N-I
JL ~ an sincd(N -1;N;(k 1L-n)). (9.3.3)
9. IMAGE RESAMPLING AND GEOMETRICAL 381
TRANSFORMATIONS
Fig. 9-6 illustrates zero padding for an even N .Upper plot on the figure
shows OFT spectrum of a signal of 8 sampies to be 4-zoomed. The left plot
in the second row is a plot of the zero padded spectrum with discarding
central sampie aN / 2 of the initial spectrum. To the right from it, the
corresponding interpolation kernelsincd(N -1;N; x) is shown. The third
row illustrates zero padding with replication of the initial spectrum central
sample a N/ 2 and the corresponding interpolation kernel
sincd(N + 1;N;x) .
5 __------------------~
~L.dIJ
1 2 3 4
a)
5 6 7 e
...·, .. ......
.. . .
.
4
o
10 20 30 100 200 300 400 500
~ ~
5 ,....----------r----~----_.,
~~ : j ~ ~ ~ ~~I ,~lli=:
10
f)
20 30 100 200
g)
300 400
-I
500
Figure 9-6. Zero padding signal spectrum and corresponding interpolation kerneIs
- 1
a = I r La" sincd[± 1; N; (kIL-n)] ,
N-I
k (9.3.4)
vL "1=0
{exp(in K; 1 (n + p))} =
{IDFrr'l{DFTN{a. ex{-in K; 1k)} {ex{-iln ~ )}}}x
{eXP(in K ;l(n+ p ))} (9.3.6)
~I...F_F_T_~",,-IF_F_T_~
exp(- in K;- 1 k )
(p) _ 1
{exp(- i2npr / N r = O,l,...,(N -1)/2 -1 .
(9.3.8)
11, - '- N(p)_, )' ,
\11 r=(N +) 1 12 + 1,..., N - 1 ,
and, for even N, there are the same three options as for the zero padding
discrete sinc-interpolation method:
option K = N -1 ,
option K = N + 1 ,
9. IMAGE RESAMPLING AND GEOMETRICAL 385
TRANSFORMATIONS
and option K = ±l ,
Flow diagram ofthe algorithm for even N and the most recommended option
K = ±1 is shown in Fig. 9-8.
Iexp(i21tpr 1N) I
r = 0,1,•••, N 12 - 1 a. ,(p)
u,
FFT ~--~-----H
r=N12
Figure 9-8. DFT based discrete sinc-interpolation algorithm for p-shifting signals of even
number N of sampIes (option K =±1 ).
_1_{a DCTn
.J2N 0 '10
+2~aDCTn"cos[1t
f:t' '1,
(k+1I2),]_
N
where
(9.3.13)
and
n =O, .••,[N 12]-1
n= [N /21 ••• ,[N /2]+N-1
n= [N 12]+N,••• ,2N-1
(9.3.15)
with K = N for odd N and K = ±1 for even N .
By virtue ofthe relationships (9.3.9-11), coefficients 1'h, (p) with even
indices can be found direcdy as
Therefore one needs to additionally compute from Eq. 9.3.14 only terms
{r, 2r+l (p)} with odd indices. Flow diagram of this algorithm for generating
p-shifted sinc-interpolated copy of signal is shown in Fig. 9.9
9. IMAGE RESAMPLING AND GEOMETRICAL 387
TRANSFORMATIONS
................................. ~ ..................... .
1.............................
Initial signal la k ) 1. p-+;
{!" (P)}
.........................,...
"'-
Figure 9-9. Flow diagram of a DCT based algorithm for p-shift signal of N sampies
J.
a) b)
Figure 9-10. Signal zooming with DFT based (a) and DCT based (b) discrete sinc-
interpolations
388 Chapter 9
Figure 9-11. Zooming image fragment with DFT based (upper right) and DCT based (bottom
right) discrete sinc-interpolations
from signal sampIes within the window. Interpolation funetion in this ease is
a windowed diserete sine-funetion whose extent is equal to the window size
rather that to the whole signal size required for the perfeet diserete sine-
interpolation. Fig. 9.12 illustrates frequeney responses and PSF of the
eorresponding interpolating low pass filters of two different window sizes
for signal 3-zooming. As one ean see from the figure, interpolation filter
frequeney responses deviate from a reetangle funetion, a frequeney response
ofthe ideallow pass filter.
Windowll
.l
Figure 9-12. Windowed discrete sinc-functions for windows of 11 and 15 sampies and the
corresponding interpolator frequency responses for signal 3-zooming
(9.3.17)
Fig. 9-13 shows flow diagram of such a combined algorithm for signal
restorationlenhancement and fraetional p-shift. It is assumed in the diagram
that signal p-shift is implemented according to the flow diagram of Fig. 9-9.
.
~t.~1
~elrt
f
Input signal Output signal
(9.3.19)
a) b)
Figure 9-14. An example of image denoising and resampling from irregular to regular
sampling grid: a)- noisy image sampled in a irregular sampling grid; b)- Denoised and
resampled (rectified) image
Input
signal Sliding Discrete DeT based Output
window sinc-interpolator signal
DCT
-,
-1 .5 L---50-:'::----:-:I00=------::1~50;::------;:;;200~-----::2!::;;_
MYOCTs,nc-shifted 16.54 pixels
15.---~--~--~---'--~
05
o
~5
-,
-15L---50~---:-:100~---:-:'~50;::------;:;;200
=------::2!:~
AdsIlINlVOCTslnc-sh,ftad 16.54 pixelS wilh lhrashold 0 03
'5
05
o
~5
Figure 9-16. Fractional shift of a signal (upper plot) by non-adaptive (middle plot) and
adaptive (bottom plot) sliding window DCT sinc-interpolations. One can notice disappearance
of oscillations at the edges ofrectangle impulses when interpolation is adaptive
9. IMAGE RESAMPLING AND GEOMETRICAL 393
TRANSFORMATIONS
.--~==-=~~~~~ ..
SL--:;;;;---=-;;=--=.----=. .,. L..-c=---:-::.o:---:"'=---=,.:---:..=---=..=-'
.--~....;.
u-
_---"~-_...- r _.....
' : ·iL',M
:;w. ',i :
-.;. - - ;: -f·'
:
.. - : .
s .., 1IDD 1MJ 3D :.J
.
ur-~;;.:.:.;~==:r--.., ~oc,r.-..: .,..,.,.,.....1Ih.~1'0i . ~I'
Figure 9-17. Signal zooming with non-adaptive and adaptive diserete sine-interpolation. First
row: a model of a eontinuous signal (left plot) and its sampled version (right plot). Seeond
row: initial signal reeonstrueted by nearest neighbor (left plot) and linear (right plot)
interpolation. Third row: signal reeonstruetion using non-adaptive (left plot) and adaptive
(right plot) diserete sine-interpolation. One ean see that smooth sinusoidal signal and
reetangular impulses are reeonstrueted with mueh less artifaets when adaptive interpolation is
used.
a)
b) c)
Figure 9-18. Image (upper) rotation with sliding window non-adaptive (bottom left) and
adaptive DCT sinc-interpolations (bottom right).
9. IMAGE RESAMPLING AND GEOMETRICAL 395
TRANSFORMATIONS
Input Zoomed
signal spectrum
a) b)
Figure 9-19. Schematic diagrams of signal zooming (a) and spectral analysis with sub-sample
resolution (b) using SDFTs for zooming and subsampling factor 3
defined by the zooming faetor. Then one ean re-sample the zoomed image to
the required sampling grid. If the number of pixels of output image is not
less then the number of pixels of the initial image, the simplest nearest
neighbor interpolation ean be used for the resampling. One ean also use the
combination of the diserete sine-interpolated zooming and bilinear
interpolation for the resampling in order to reduce the zoom factor needed
for the given resampling aceuraey. Prineiple of the "zooming-resampling"
method of image geometrieal transformation is illustrated in Fig. 9-20.
For object localization with sub-pixel accuracy, correlation peak (See Ch.
10) should be sub-sampled. In this case, only signal sub-samples in the
vicinity of the highest correlation peak are required. Hence, inverse SDFTs
in the flow diagram of Fig. 9-19, a) can be carried out with pruned
algorithms described in Sect. 6.4.
Schematic diagram of the signal spectral analysis with sub-sample
resolution is show in Fig. 9-18, (b). Fig. 9-21 shows an example of such an
analysis for a sinusoidal signal sin(21tkF / N)( k = O, ...,255;F = 36.7).
Figure 9-21. Image resizing with different resampling filters : (a)- initial image; b) - bilinear
interpolation; c) - bicubic interpolation; d) - nearest neighbor interpolation; e) - sincd-
interpolation in window 11 x 11; t) - sincd-interpolation in window 31 x31
:I ••••• ~ ~ -.+--.--~--
+ ....--····+·····i······{------1··
~
u •.. ..
eo so t~
,-~
b)
.....-~.".--, .-
._+- ~'" ,
..• _............ . -!
. ..
..... --_. .... ':t
a)
.. . .. ... ::: l
'....-,
c)
Figure 9-22. Spectral analysis with sub-sample accuracy. a) - a sinusoidal signal b) - its DFT
spectrum; c) - interpolated spectrum peak
398 Chapter 9
[;]=[~~l (9.4.2)
Eq. 9.4.3 describes the resampling rule that should be applied to input image
to generate its rotated copy. In order to simplify computations, one can
factorize rotation matrix into a product of three matrices each of which
modi fies only one co-ordinate ([5]):
First pass
Figure 9-23. The principle ofthe three pass image rotation algorithm: a) - rotation without
aliasing; b) - aliasing artifacts owing to the cyclicity ofthe convolution
Figure 9-24. Boundary effects and rotation error in image rotation with discrete sinc-
interpolation
400 Chapter 9
00
ä(x)= ~a(x)=
dx
f[(- i21tf):x(f)]exp(- i21tfx)cif
-00
(9.4.6)
ak =
1 N-I (
~ La,exp -i2n-
krJ (9.4.8)
vN ,~O N
•
ak
_
-
I N-I
~ L~, a,]exp( k-
diff
-,2n
• r J, (9.4.9)
vN ,~O N
where {a k } are sampIes of derivative of signal a(x) and f,:iff} are defined
as:
-i2nrl N, r = 0,1,... , N 1 2 -1
TJdiff ={ -n 12 r=N12 (9.4.l0,a)
, i2n (N - ;)/ N, r= N / 2 + 1,... , N - 1
for odd N. Note that the coefficient ll~%') is chosen in Eq. 9.4.10, a)
according to Eqs. 9.3.4 and 9.3.11.
Eq. (9.4.9) describes ~i~ital filtering in the DFT domain with discrete
frequency response f,:iff J and suggests the following algorithm of
computing derivative of signals specified by their sampIes:
(9.4.11)
(9.4.12)
The following simplest differentiating kerneIs of two and four sampIes are
recommended in manuals on numerical methods in mathematics ([6])
hndiff(l} = [-1,1] (94
.. 13 a)
and
h:iff(l} = [-1/12,8/12,-8/12,1/12]. (9.4.14 b)
Both are based on an assumption that, in the vicinity of signal sampIes
signals can be expanded into Taylor series. The first one, obviously, assumes
linear interpolation. One can compare these and above-described ramp-
filtering differentiating method by their frequency responses. Applying N -
point Oiscrete Fourier transform to Eqs. 9.4.13 and 14, obtain:
l.2
: :
....... , ... ,',.,......1"..............................:-... -_ .. -............... .
I
i """" .. ''' ..... ,.......... j.............................. t·········· .L. .. ........... .~
0.8
........ ···j······························i················,·," """""1",,,
0.6
0.4
0.2 ··· · ········1·· · ··· ·· ··· ·· ·· ·· ·· ···· ··· ···· · ~· ···· · · · · .....
Figure 9-25. Frequency responses ofdifferentiating filters described by Eqs. 9.4.13 (curve 1),
Eq. 9.4.14 (curve 2) and Eq. 9.4.10 ("Ramp")
9. IMAGE RESAMPLING AND GEOMETRICAL 403
TRANSFORMATIONS
where
(9.4.16)
where a{x) and a{f) are signal and its Fourier transform spectrum,
respectively, and
(9.4.18)
(9.4.19)
0, r=O
N
r = 1,2000' N /2 - 1 ,
r=N /2
0, r =0
11 (int) ={ N .
r --.-,r=1,2,0.0,(N-l)/2'
,2nr
In the 3/8-Simpson rule, the seeond and the third points need to be
evaluated by the trapezoidal rule and the Simpson rule respeetively. Then
ii(3I8S}
o = 0, ii(3I8S}
k
= ii(3I8S}
k-3
+~
8 (a k-3 + 3ak-l + 3ak-I + a k ). (9.4.23)
(9.4.24)
-es} _ 1 ( 4 )
a-eS}
k - ak-l -"3 a k - 1 + a k _ 1 + a k (9.4.25)
(9.4.26)
0, r=O,
,,!T) = ~
-(Tr) {
= _ cos(nr / N)
r = 1,..•,N -1 '
(9.4.30)
ur 2i sin{nr / N)'
0, r=O
,,!S) = ~
-(S) {
= _ cos(2nr IN)+ 2 r =1,••. ,N -1'
(9.4.31 )
ur 3isin(2nr/N)'
0, r=O
,,!3S) =~
-(3S) {
= _ cos(37tr I N)+ 3cos(nr IN) r = 1,••. ,N -1 .
Ur isin{37trl N) ,
(9.4.32)
In Fig. 9-27, absolute values ofthe frequeney responses ofthe DFT based
method of integration (Eq. 9.4.20) and of the Newton-eotes roles are
represented with a frequency coordinate normalized to its maximal value.
Beeause the absolute values of frequeney responses are symmetrie, only half
of the eurves are shown.
From the figure one can see that the integration methods are low pass
filters, and that they are almost identical in low frequencies while in higher
frequencies 3/8-Simpson, Simpson and trapezoidal integration methods tend
to introduce integration errors. The frequency response of the 3/8-Simpson
rule tends to infinity for the 2/3 of the maximum frequency, and the
frequency response of the Simpson rule has almost the same tendency for the
maximum frequency. This means, in particular, that round off computation
errors and noise that might be present in input data will be amplified by
Simpson and 3.8-Simpson in these frequencies.
,... ...
1-D OFf OCT based Polar-to- Cartesian
spectra zooming coordinate 2-D
zooming along angle r+ conversion of the lDFT
according to coordinate zoomed array of
flow diagram of 1-D OFT spectra of
Fig. 9-24, (b) projection by
T
2-0 Array of
nearest neighbor of
bilinear
interpolation
~
Reconstructed
image projections image
(in coordinates
radius-angle) a)
• Initial polar grid
o 2-Zoomed polar grid
• Polar grid
X Tar~et Cartesian grid
• ' 10 nodes
b) c)
Fig. 9-28, a) by choosing variable zooming factor for zooming the array
along the angle coordinate. As it follows from the geometry of polar
coordinate system (Fig. 9-28, c», zooming along the angular coordinate for
smaller radial coordinates can be made with smaller zooming factor without
compromising interpolation accuracy.
This principle of variable zooming is explained in Fig. 9-29 ([10]). Fig.
9-30 iIIustrates an example of image reconstruction with this method.
Initial array
\-51 step: zooming
along radial coordinate
:.-:~: .
~- - '-
,,,, "",
econd "'''
step:
zooming Zoomed array
along angle
coordinate
Figure 9-30. Shepp-Logan phantom (a) and a result ofits reconstruction from 180 projections
with the DCT based interpolation and variable zooming factor I to 5 (b)
410 Chapter 9
References
1. P. Thevenaz, T. Blu, M. Unser, Image interpolation and resampling, Handbook of
Medieal Imaging, Proeessing and Analysis, I. N. Bankman, Ed., Aeademie Press, San
Diego CA, USA, pp. 393-420, 2000
2. L.P. Yaroslavsky, Effieient algorithm for diserete sine-interpolation, Applied Opties,
Vol. 36, No.2, 10 January, 1997, p. 460-463
3. L. Yaroslavsky, Fast signal sine-interpolation and its applieations in signal and image
proeessing,IS&T/SPIE's 14th Annual Symposium Eleetronie Imaging 2002, Seienee and
Technology, Conferenee 4667 "Image Proeessing: Algorithms and Systems", San Jose,
CA, 21-23 January 2002. Proeeedings ofSPIE vol. 4667
4. L. Yaroslavsky, Boundary Effeet Free and Adaptive Diserete Signal Sine-interpolation
Aigorithms for Signal and Image Resampling, Appl. Opt., v. 42, No. 20,10 July 2003, p.
4166-4175
5. M. Unser, P. Thevenaz, L. Yaroslavsky, Convolution-based Interpolation for Fast, High-
Quality Rotation of Images, IEEE Trans. on Image Proeessing, Oet. 1995, v. 4, No. 10,
p. 1371-1382
6. lH. Mathews, K.D. Fink, Numerieal methods using MATLAB, Prentiee Hall, Ine, 1999
7. S. C. Iriek, R. Krishna Kaza, and W. R. MeKinney, "Obtaining three-dimensional height
profiles from a two-dimensional slope measuring instrument", Rev. Sei. Instrum. 66
(1995),2108-2111
8. W.H. Press, B.P. Flannery, S.A. Teukolsky, W.T. Vetterling. Numerieal recipes. The art
of seientifie eomputing. Cambridge University Press, Cambridge, 1987
9. C. Elster, I. Weingärtner, "High-aecuraey reeonstruetion of a funetion f(x) when only
df(x)/dx or d2f(x)/dx 2 is known at diserete measurements points", Proe. SPIE, v. 4782
10. L. Yaroslavsky, Y. Chemobrodov, DFT and DCT based Diserete Sine-interpolation
Methods for Direet Fourier Tomographie Reeonstruetion, 3-d Int. Symposium, Image
and Signal Proeessing and Analysis, Sept. 18-20,2003, Rome, ltaly
Chapter 10
SIGNAL PARAMETER ESTIMATION AND
MEASUREMENT. OBJECT LOCALIZATION
where AVnO symbolizes the averaging over the sensor's signal ensemble.
Because all values involved in the averaging are non-negative, the
probability of misrecognitions is minimal when, for all b", (x), parameter
estimate p is selected that has maximal aposteriori probability:
(10.1.4)
(10.1.6)
This equation is the basic formula for the design of optimal Bayes parameter
estimating devices. It requires knowledge of the conditional probability
p{ b(x)/ Pq} that depends on the relationship that exists between sensor
signals {b" (x)} and object signals {a{x, pJ} and of apriori probabilities
{p(p q)} of parameter values. The latter may be unknown. Estimates
(10.1.7)
(10.2.1)
(10.2.2)
(10.2.3)
X-I X-I
Finally, because ~:::IbkI2 and Llak(xo,Yof do not depend on the
k=O k=O
coordinates (xo,Yo) we have :
(10.2.6)
(10.2.7)
(10.2.8)
(10.2.10)
where
(10.2.11)
--.... Correlator
(matched filter)
...... Device for localization
of signal maximum
.....
Input
image
i
Target object signal
Figure JO-1. Schematic diagram ofthe optimal device for localization of objects in images
IFFT
Input image Correlation
b(x) image
Figure 10-2. Flow diagram ofthe matched filtering in the Fourier Transform domain
Matched filter
ParaDe)
laser ligh;.;;t_-IIf'IIIt- __ F___ _
beam
a)
,. I ,~...........
,.,- I ,.......... I
,.,.
: " ........... I
,-
I " l .I. . . . . ... ...
,. I I ,
I
,. I
I
I
"
" I
: ' f
I
b)
Figure 10-3. 4-F (a) and a parabolic mirror (b) optical correlators
Correlation
Joint spectrum plane plane
I
,, rVeamera
, I
I
ParaDeI I
beam ......... ~ __ L__ __.f__
oflaser
light . beam ..........•.r..r'-_i'_--
oflaser
light ,,
" ,
lost 2-nd
Spatial
Fourier Fourier
light
lens lens
modulator
In the Joint Transform Correlator, input images and target object image
are placed aside in the input plane. Squared modulus ofFourier Transform of
the sum of these images, or joint spectrum of the images, is converted by TV
camera into an electrical signal that controls the transparence of the spatial
light modulator. The joint spectrum is described by the equation
where b(x,y) and a(x,y) are input and target object images and x is a the
displacement, along coordinate x, of the target object image with respect
the system optical axis. As one can see from this equation, the joint spectrum
consists of four components. First two components are power spectra of the
input and target object images. The third component is spectrum of their
correlation and the fourth component is complex conjugate to the third one.
The second optical part of the Joint Transform Correlator performs
Fourier Transform of the joint spectrum. As it follows from Eq. 10.2.12,
output image of the Joint Transform Correlator also contains four
components. First two components centered at the optical axis are images of
autocorrelation functions of input and target object images. The third
component is the image of their correlation function. It is reconstructed
displaced along x axis by X. The fourth component is the conjugate
correlation image. It is reconstructed in the position - x symmetrical to that
of the correlation image.
Owing to the noise present in the sensor, any measurement device will
estimate the object coordinates with a certain error. The optimal ML
coordinate estimator of Fig. 10-1 that determines the object coordinates by
the position of signal maximum at the output of the correlator (matched
filter) provides the most accurate coordinate estimates possible.
Tbe estimation accuracy may be generally characterized by the
J
probability density p(1) x ,I> of the estimation errors:
(10.2.13)
420 Chapter 10
For the evaluation of p(E x,E J one should analyze the distribution
density of the position of the highest peak of the signal at the output of the
matched filter within the area of all possible positions of the target object.
This signal is a sum of the target object signal auto-correlation function and
of a correlated Gaussian noise resulted from filtering the input white noise
by the matched filter. This mixture is a non-stationary random process. This
fact complicates the analysis very substantially. An approximate analytical
solution of the problem can be obtained using following reasoning.
a)
b) c)
" .. -. - - - • -J,\ -. - - - • - --.
,. ---: -- 'f-.--\\'.---:
o. - - -' . • . . '" . . .'
----:
. ..
-!
~
': r:"~~,:'~ . ~ ~ ~ ..\ ~. :;,.;~
d) e)
Figure JO-5. Nonnal and anomalous localization errors: a) - target object signal; b), c) - target
object signal hindered by additive noise of different intensity; d) ,e) - two corresponding
outputs ofthe filter matched with the target signal that correspond to nonnal and anomalous
localization errors, respectively.
where
x x
Noffa{;, -xl'T), -y,)a{;, -x2,,,, -Y2)~' dT), =
x
No Ra (x2 -XI'Y2 - yJ, (10.3.4)
where AVn U denotes statistical averaging over the noise n(x,y) ensemble.
In the ca se of normal errors, maxima of the signal R(x,y) at the output of
the correlator are located in the close vicinity of the point (xo,Yo) ' the
position of maximum of RJx,y). The following system of equations
determines coordinates the maximum of R(x,y) :
(10.3.4)
X= x o +n x'.
{ (10.3.5)
YA=yo +n·
y'
(10.3.6)
10. Signal parameter estimation. Object localization 423
82 82
+nx - 2 Ra(x-xo'y-yo~x=xo +ny--Ra(x-xo'y-yo~x=xo (10.3.7)
8x Y=Yo 8x8y Y=Yo
=Dxnx + Dxyny,
where
(10.3.8)
(10.3.9)
(10.3.10)
with
(10.3.12)
Denote
(10.3.13)
424 Chapter 10
D D
n = y v - xy v·
x D D _ D 2 x D D _ D 2 Y'
x Y xy x Y xy
(10.3.15)
n = Dx v -
Dxy v.
y D xDY - D xy2 Y D xDY _ D xy2 x
IO.3.16b)
(l0.3.16c)
10. Signal parameter estimation. Object localization 425
y=Yo
00 00 2
(10.3.18)
E a is its energy:
(10.3.19)
and 1: is the inertia moment of its power spectrum along the axis Ix
Similarly,
(10.3.21)
and
426 Chapter 10
(10.3.22)
where
j jl:~(I.,I,rdl"dl,
1 =....:-"""-'-:,""=-""-------
JJ~ (Ix, I, rdlßl,
y
1 (10.3.23)
-00-00
and
r
j j 1,,1, ~ (Ix' I, dlßly
r
1- 1
Jxy
= ....:-""::....-"""'---------- (10.3.24)
J
0000
J~ (I", I y dlßl,
(10.3.25)
-00-00
(10.3.27)
10. Signal parameter estimation. Object localization 427
and
-00-00
(l0.3.29a)
(l0.3.29b)
(l0.3.29c)
(10.3.30)
(10.3.31)
428 Chapter 10
(10.3.32a)
(10.3.32b)
er xy2 = o. (10.3.32c)
The last equation shows that if the signal power spectrum is axes-
symmetrical, normal localization errors along coordinates x and y are
uncorrelated. This situation takes place if the object spec!rum a. (Ix' 1,), or,
which is the same, the object signal a(xJ') is a separable function of the
coordinates. Equations (10.3.32) also coincide with the very well known
relationship for one-dimensional signals ([5,6]).
It is sometimes more convenient to express variances of normal
localization errors in terms of the noise variance er ;"t,n at the output of the
matched filter rather than in terms of the input noise spectral density No.
Because this variance is equal to
(10.3.33)
(l0.3.34a)
(10.3.34b)
10. Signal parameter estimation. Object localization 429
(l0.3.34c)
(l0.3.35a)
(10.3.35b)
(10.3.35b)
(10.3.36a)
(10.3.36b)
(10.3.36c)
(10.3.37a)
430 Chapter 10
(1O.3.37b)
(10.3.38)
(10.3.39)
are discrete inertia moments of the target object spectrum. One can obtain
them by the discretization offormulas (10.3.20), (10.3.22) and (10.3.23) and
using the relationship (10.2.9):
(10.3.40)
(10.3.41)
10. Signal parameter estimation. Object localization 431
(10.3.42)
(10.3.43)
(10.3.45)
(10.3.46)
432 Chapter 10
0.9
0.4
0.3
0.2 ......•......... ; ............... .
0.1
.~ .!-5--::.0~.4--.-::"0.~3"'"""''-:.O~.,--.-::-O.1!-"--~--!:'-0.!-,--::0~.2......;;:;~0.3~--::0-!-.4--::-!0.5
Figure 10-6. Gaussian -shaped impulse and its efficient width shown by solid line. Dashed
line shows limits of interval [- er a,er a]
(10.3.48)
(10.3.49)
10. Signal parameter estimation. Object localization 433
2F = 1 (10.3.50)
a 2,Jna a
(l0.3.51,a)
or
(l0.3.51,b)
(10.3.52, a)
and
a~_2 a~ _ 2 1
a
-2 -
da
I
V1[ a: - I
V1[ PSNR
2' (10.3.52, b)
where
PSNR=~. (10.3.53)
an
is the ratio of signal peak value to standard deviation of noise.
It is instructive to express, using Eqs. 10.3.46 and 10.3.50, the variance of
normal localization errors in terms of the effective impulse width and
efficient bandwidth of its spectrum:
(10.3.54)
434 Chapter 10
and
o .02fi
0 .02
0 _0' S
o 0'
0 .00.5
-'00 ·50 so , 00
Figure 10-7. Experimental distribution density ofthe localization error for Gaussian-shaped
impulse ( 0" a =10 ) and JE IN Q 0 = 1.5 obtained for 100000 realizations.
40 --. -. -.. --··, . - . ---. -.. .. .... --.. --- ..... -. -. --.. ...-. -..
~ ~ ~
· ... . .. ..
35 ··· ..
...
.0 __ • __ ._- ,_ •• _--_ ••••• _---_ •••• " •••• __ • • • • • • ___ ••••
30
· .
- - - - - - - - - - ·:- -- - - - - - - . - .. -- - - - - - -- -.- - -- _. .. . - . -... --.
··· .. ~
.. ....
~ ~.
20 - -- -------. · . . .
.
,----- --
·· .
- - - ,._----_._-
.. .----------".... ... _-_ ... -
.
15 •• -- •• -- •• . . • - --. To'tSI'std' 01 eirOiS' . :. ..... -- .. .
; -- -- • -- -- • : --
·· . .. ..
10 -------- - -.------_._.
·· ,..- _... . .. -... - ... _- -- --., .-.. . .. . .. .
· .. . . . .
oo.~"~'~'~'~"~'~'~'~;'~
0.1
'~'~":':'~'~"0~·~"='='~'~·~~1~~~~:n~'~:~~a~
.2 0.3
l e:rr~om~·tj':'~':"~'~'~'~'0']
0.4
'5
Sq<1{NOlEa)
Figure 10-8. Standard deviation of the localization error as a function of noise-to-signal error
for Gaussian-shaped impulse with (10000 realizations)
the base of data obtained by the same computer simulation as that illustrated
436 Chapter 10
in Fig. 10-7. Lower curve shows standard deviation of errors smaller then
0'Q As one can see, this curve follows theoretical estimations described by
•
Eqs. 10.3.51 and 52. Upper curve shows standard deviation of the
localization error regardless their values. One can see that for low noise-to-
signal ratios localization error standard deviation behaves as it is predicted
by the theory for normal errors. However, for noise-to-signal ratio larger
then about 0.15, error standard deviation starts growing much more rapidly.
This behavior is associated with the appearance of anomalous errors.
Performance of the optimallocalization device in terms of anomalous errors
will be discussed in Sect. 10.3.4
R(x) = Ja (Ix )H(tx )exp{- i21t lrx (x - Xo)]}tlx + Rnh (x), (10.3.61)
~R(X~ =
8x 1x=Xo+nx
00
and
00
Thus, we obtain :
(10.3.65)
47t 2!x2NoIH(lxt. Then the variance (J ~ ofa small random error nx is:
00
N fl:IH(lxtdlx
2 0 _-=-",--00_ _ _ _ _-;:-
(10.3.66)
", ~ 4. ' (If:aVJHVJ1f, )' ,
or,
(10.3.67)
that shows how much the normal error standard deviation for a non-optimal
filter exceeds that for the optimal (matched) filter: according to the Schwarz
inequality LSFR ~ 1. It reaches its minimal value, unity, when the filter is
perfectly matched to the object signal, that is when H(/,,) = a • (J,,).
In order to estimate the order of magnitude of the loss factor we present
results of analytical and numerical evaluation of the loss factor for
localization of Gaussian-shaped, rectangular and triangle impulses with the
help of the filters with, respectively, Gaussian, rectangular and triangle
impulse responses of different width.
For Gaussian-shaped impulses, matched
(10.3.69)
and mismatched
(10.3.70)
2J3/2
LSFR = ( 1+0' mm /0' m
2
(10.3.71)
20' mm /0' m
:~ ~ltJ~~~~
; . ;; i.l ·
10 ··,,···!,···,·,···;·,······1·_······~····· ··;··_·7(. ······ .. ~ ,·····,
o ~ ... ·+·· . ···,, !····· .. ··t. ··· . ·j··· .. ···"i .. ·"···+········ ' li' 'i;~l~
~.6 O~1
" 1 .~ 2 2 S
G." u laflo Im pu l at .... ldl,.. , . [10
0 .8 0.9 t
Sreel/SrO
1.1 1.2 1.3 t .4
a) b)
Figure 10-9. Loss factor for mismatched localization ofGaussian-shaped (a) and rectangular
impulses
(10.3.72)
440 Chapter 10
where
(10.3.74)
(10.3.75)
10. Signal parameter estimation. Object localization 441
lirn
Q~co
[In[<t>( ~{E:
N"; +n)]Q-IJ= Q~co [~ln<t>(E~)/~!]=
lirn
'jdQ dQ Q
ifE 2 /2>t
(10.3.76)
ifE 2 /2st"
Therefore,
.
hrn P
Q~co ae
= {t, if E /2N o slnQ
a
0, if E a / 2N 0 > In Q
• (10.3.77)
This feature implies that if the area of search is large enough with
respect to the size of the object, the probability of anomalous errors may
become enormously high when input signal-to-noise ratio is lower than the
fundamental threshold defined by Eq. 10.3.77. This also means that when
area of search is increased the signal-to-noise ratio must also be increased in
order to keep the probability of anomalous errors low. Moreover, for any
given intensity of noise, there exists a trade off between localization
accuracy defined by the variance of normal errors and localization reliability
described by the probability of anomalous errors. Increasing the accuracy
achieved by widening the object signal spectrum with signal energy fixed
results in increasing the probability of anomalous errors: widening spectrum
width is equivalent to narrowing the signal, and, consequently, to increasing
the ratio Q ofthe area of search to the object signal area.
Results of a numerical verification of the above relationships by
computer simulation of the optimal localization device are shown in Fig.l 0-
10.
In conclusion it should be noted that the formula (10.3.77) has a general
fundamental meaning. Note that In Q may be regarded as the entropy, in
bits, of the results of measurements of object coordinate in which Q different
object positions can be distinguished. Therefore, Eq.(10.3.77) gives the
absolute lower bound for the object signal energy per bit of measurement
information required for reliable parameter estimation in the presence of
white Gaussian noise:
(10.3.78)
442 Chapter 10
0.8
0.6 I
Q=340
0.4
0.2
o
o 0.5 2 2.5
(10.4.1)
-'"
such that noise variance is the same as that of the white noise with spectral
density No'
The problem of target localization in the presence of correlated noise
may be reduced to the white noise case if we pass the observed signal plus
noise mixture through a filter with frequency response
(10.4.2)
(10.4.3)
................................................... ~
Object
Input
coordinates
image Whitening Correlator Device for
filter (matched localization of
filter) signal maximum
Whitening
Optimal filter
::rarget object signal
filter
Figure 10-11. Schematic diagram ofthe optimal device for localizing objects in the presence
of additive correlated Gaussian noise.
From the above reasoning it follows that, for a target signal observed in a
mixture with additive correlated Gaussian noise, the optimal filter generates
a signal which is proportional to aposteriori probability distribution TImction
of target coordinates. Now we will prove that the optimal filter has yet
another important feature. It also provides the highest possible for all linear
filters ratio of its response to the target object signal to standard deviation of
noise at its output (sig~al-to-noise ratio). This feature enables designing
optimal localization devices with linear filtering in case of non-Gaussian
signal observation model. We will use this feature in Ch. 11 for the design of
optimal devices for target location in clutter.
Consider response of an arbitrary filter with a frequency response
H{Ix'/y) to the target signal a(x-xo,y- Yo) and noise n(x) with spectral
density NoIHnVx,/yf. Filter output to the target signal at the point
(xo,Yo) is equal to
Standard deviation ofnoise at the filter output noise may be found as:
(10.4.5)
SNR = _ 0 =
b JJa {tx'/y"p{tx'/y}i/ß/y
-00-00 1/2 • (10.4.6)
er. [N.lllHV•• f,)'IH.V•• f,l'df.df, J
It is follows from Schwarz's inequality ([8]) that
with equality taking place for the optimal filter of Eq.l 0.4.3. Therefore
(10.4.8)
(10.4.9, a)
446 Chapter 10
-2
cr 2 = fx,NW NO. (10.4.9, b)
y,NW - 2 - 2 ( - 2 \2 4 2E '
fy,NW fy,NW - f xy,NW ) 1t a,NW
-2
2 fxy,NW NO (10.4.9, c)
cr xy,NW -,.2 /2 (j2 \2 41t 2E .
J y,NW y,NW - xy,NW) a,NW
where:
(10.4.10)
(10.4.11)
(10.4.12)
(10.4.13)
one-dimensional case. In this case we have from Eq.(IOA.9a) for white noise
that
(10.4.14)
Let us give this formula the form, which explicitly contains the error
variance (J ~ for the case ofwhite noise (Eqs. 10.3.37, a):
-2
2 1 No = er 2 Ealx
er = (10.4.15)
x,NW -2 2 x E 1'2
Ix,NW 411: Ea,NW a,NWJx,NW
The last factor in this formula shows the change of error variance owing
to non-whiteness of noise. Because all functions involved in this factor are
nonnegative, the following inequality takes place:
(10.4.16)
(10.4.17)
(10.4.18)
(10.4.19)
where Sa = J;cr a is the impulse efficient width (Eq. 10.3.46). One can
see from Eq. 10.4.19 that optimal filter is in this case a sort of a band-pass
filter with central frequency 1/ 4nS; Fn • A family of its frequency responses
is illustrated in Fig. 10-12.
Opl lmaf fllI., frequent)' ' •• ponu
0 .9
0 .8
0.7
0 .6
Figure 10-12. Frequency responses of the optimal filter for Gaussian-shaped impulse
2S;)
a(x) = exp(-n x 2 / and non-white noise with exponentially decaying power spectrum
(10.5.1)
Let also assume that components of the additive noise as weIl as sampies of
noise within each component are aIl mutually uncorrelated. This assumption
directly implies, that probability of observing signal {bk •m } provided target
object is located in coordinates (xo,Yo) is
where cr ~.m denotes the variance of the m-th noise component. Therefore,
optimal MAP- and ML-Iocalization devices are defined, respectively, by
equations
and
450 Chapter 10
(10.5.4)
where N O,m is the spectral density of the m-th noise component, and ML-
estimation is
Matched filter
(I-st component)
Input
Matched filter
image (2-nd component)
Device for
components
Adder localization of
Matched filter signal
(3-d component)
maximum
Figure J0- J3. Schematic diagram of the optimal device for localization of a target object in
multi component images
10. Signal parameter estimation. Object localization 451
(10.5.7)
(10.5.8)
where
(10.5.10)
(10.5.11)
M 1
L ~ I I/y1rV",/yf d/"d/y
0000
f-y,ll1 -- ",=1
M
0,'"
1
-00-00
00 00 1
(10.5.12)
~~ I IrV",/y~ d/"d/y
1ft- 1 0,,,, -00-00
M 1 0000 1
L~ I I/"/yrV,,,/y~ d/"d/y
J:-xy,a
1 - ",=1 0,'" -00-00
(10.5.13)
L ~ I Ir v"' /y ~ df"d/y
- M 1 00 00 1
(1.6.14)
(10.5.15)
(10.5.16)
where
(10.5.17)
..
(l-nd component) Device for
localization
Pre- -+ Adder
ofsignal
"whitening" maximum
..... filter
~
Matched filter
(3-d component)
...
po
f
.... +
Object's
coordinates
..... r--+
Matched filter
(M-th component)
......
+-
r----- -
I , f
Pre-whitening filter
f ~
Target image components
,I
Figure 10-14. Schematic diagram ofthe optimal device for localizing objects in multi-
component images with additive correlated noise in the channels.
(10.5.20)
Suppose also that the non-target objects do not overlap one another and the
target object. The most immediate illustration of such a situation would be
the task of locating a specific character in a a printed text.
As non-target objects do not overlap the target object under, it is clear
that the design ofthe filter ensuring the highest localization accuracy, i.e. the
lowest variance of normal errors, is govemed by the same reasonings as
those for above described additive noise model with no non-target objects.
Therefore optimal filter in this case will also be the matched filter and the
variance of normal errors will be defined by the same formulas as those in
the cases of no non-target objects (Eqs. 10.3.34 and 10.3.36). Presence of
non-target objects has an effect only on the probability of anomalous errors.
In order to estimate the probability of anomalous errors in target
localization in the presence of multiple non-target objects let us introduce
some quantitative characteristics of the non-target objects. Let
• p(Qe) be the probability of appearance of exactly Qe non-target objects
in the area of search.
• Non-target objects form some C c1asses, such as, for instance, c1asses
of characters in optical character recognition, according to the maximal
values R!~ of their cross correlation function with the target object
(10.6.2)
c = 1,2,•••,C
10. Signal parameter estimation. Object localization 455
(10.6.3)
where as before
(10.6.4)
(10.6.5)
456 Chapter 10
After substitution of Eq. 10.6.5 into Eq. 10.6.4 and with the use of the
identity
(10.6.6)
we obtain:
(10.6.7)
where
(10.6.9)
10. Signal parameter estimation. Object localization 457
(10.6.11)
E - R(c) ( . )
~ == 1.1-1.6 JlnQo == 1.8-4.4. (10.6.12)
E
a 0
N· .
Pa,e < 10-2 and it increases rapidly with decreasing signal-to-noise ratio
E
Wo .
_R(c)
EaN o
Therefore the probability of anomalous errors in the presence of
multiple false objects may be very high. Note, for example, that the number
of characters on a standard printed page is about 1600. Fig. 10-15 illustrates
the phenomenon of false detection on an example of character recognition.
In order to decrease Pa,e' it is necessary to increase signal- to-noise ratio
by suppressing cross-correlation peaks for non-target objects with respect to
the auto-correlation peak of the target object. This requires an appropriate
modification of the matched filter. Therefore, for localization of a target in
458 Chapter 10
Figure 10-15. Detection, by means of matched filtering, of a character with and without non-
target characters in the area of search. Upper row, left image: noisy image of a printed text
with standard deviation of additive noise 15 (within signal range 0-255). Upper row, right
image: results of detection of character "0" (right); one can see quite a number of false
detections. Bottom row, left: a noisy image with a single character "0" in its center; standard
deviation of the additive noise is 175. Highlighted center of the right image shows that the
character is correctly localized by the matched filter.
References
1. A.Van der Lugt, Signal Detection by Complex Spatial Filtering, IEEE Trans., IT-IO,
1964, No. 2, p. 139
2. L.P. Yaroslavsky, M.G. I1ieva, V.N. Karnaukhov, I.D. Nikolov, G.1. Sokolinov, A
Device for Pattern Recognition, Certificate of Authorship No 1414170, of 24.4.1986,
Reg. 01.04.1988
3. C. S. Weaver and J. W. Goodman, " A technique for optically convoluting two
functions," Applied Optics, 5, pp. 1248-1249 (1966).
4. F. T. S. Yu and X. 1. Lu, "A real-time programmable Joint Transform Correlator", Opt.
Commun., Vol. 52,1984, p. 47.
5. V. Kotelnikov, The Theory of Potential Noise Immunity, Moscow, Energia, 1956 (In
Russian)
6. J. M. Wozencraft, I. M. Jacobs, Principles ofCommunication Engineering, N.Y.,Wiley,
1965
7. L.P. Yaroslavsky, The Theory of Optimal Methods for Localization of Objects in
Pictures, In: Progress in Optics, Ed. E. Wolf, v.XXXII, Elsevier Science Publishers,
Amsterdam, 1993
8. L. Schwartz, "Analyse Mathematique", Hermann, Paris, 1967
Chapter 11
Input Object's
image Device for coordinates
... Linear filter ... locating .....
b(x,y) b'(x,y) signal maximum
(11.1.1)
ffw.(xo,yo}ixodyo = 1; (11.1.2)
s.
s
LW, =1.
s=l
(11.1.3)
Then the quality of estimating coordinates by the device of Fig. 11-1 may
be described by a weighted mean with respect to p(bo)' {w.(xo,Yo)}' and
{W,} ofthe integral ofEq.(11.1.I):
Assume that the target object is defined precisely. In this context, this
means that the response of any filter to this object may be determined
exactly. Therefore, probability density p(b o ) is the delta-function:
(11.2.1)
where bo is the known filter response at the point of the target object
location. Then Eq. (11.1.4) that defines the localization quality becomes:
(11.2.3)
Then obtain:
(11.2.4)
Suppose that weights {Ws} do not depend on sand are equal to IIS. We
refer to this case as to the case of the spatially homogeneous optimality
criterion. Then
-(-) 1 ~ - (- ) (11.2.5)
h \bbg = - L.. h s \bbg
S s=l
is the histogram of the filter output signal amplitudes as measured over the
background part of the whole image and averaged with respect to the
466 Chapter 11
ho = (11.2.7)
-CI)
As for the relationship between Ii(bbg )and H(Jx' I y )' it is, generally
speaking, of a more involved nature. The explicit dependence of Ii(bbg) on
H(/x '/y) may only be written for the second moment m; of the histogram
li(b) using Parseval's relation for the Fourier transform:
Sl Hw(xo,Yo)dxodyo Jbb~(X,y}dxdy=
bg S Sbg
-f-bg Hw(xo,yo)dxodyo
S
11Iß1;·O)(lx,ly fI H(Jx,ly f dlxdly =
y y y , (11.2.8)
where S bg is the area of the background part of the images, or the image
search area minus the area occupied by the target object located at
coordinates (xo,Yo)' ßb~·O)(JX,/J is Fourier spectrum ofthis background
part, and
11. TARGET LOCAT1ng 1N CLUTTER 467
(11.2.9)
Therefore, for the optimization of the localization device filter we will rely
upon the Tchebyshev's inequality
Probability(x ~ x + b o) ~ (J 1 I b; . (11.2.10)
(11.2.11 )
(11.2.12)
It follows from this equation that the mean value of the histogram over the
background part of the picture, b, is determined by the filter frequency
response H(0,0) at the point (Jx = 0, /y = 0). This value defines a constant
bias of the signal at the filter output which is irrelevant for the device that
localizes the signal maximum. Therefore, one can choose H(O,O) = 0 and
disregard b in (11.2.11). Then we can conclude that, in order to minimize
the rate p" of anomalous errors, a filter should be found that maximizes
ratio of its response bo to the target object to standard deviation (m: )1/2 of
its response to the background image component
A V ..
(11.2.13)
468 Chapter 11
(11.2.13)
(11.2.14)
an ensemble of images that may contain the same target. It plays here the
role additive Gaussian noise power spectrum plays in the filter ofEq. 10.4.3.
/a :;0 (Jx' I, r
There is, however, an important difference. When averaging is applied to
the filter will be optimal (in terms of minimization of the
misdetection rate) also on average over the ensemble of possible images.
However, as we mentioned in Sect. 11.1, in applications, one frequently
needs a filter optimal for each particular image rather then on the average.
This means that in this case the averaging over the background image
component AVbg must be dropped out from the optimality equation 11.2.13.
In this way we arrive at the following filter:
11. TARGET LOCAT1ng IN CLUTTER 469
(11.2.15)
This filter will be optimal for the particular observed image. Because its
frequency response depends on the image background component the filter
is adaptive. We will call this filter "optimal adaptive correlator".
To be implemented, optimal adaptive correlator needs knowledge of
power spectrum of the background component of the image. However
coordinates of the target object are not known. Therefore one cannot
separate the target object and the background in the observed image and can
not exact1y detennine the background component power spectrum and
implement the exact optimal adaptive correlator. Yet, one can attempt to
approximate it by means of an appropriate estimation of the background
component power spectrum from the observed image.
(11.2.16)
where the horizontal bar denotes averaging over target coordinates (x o,Yo).
470 Chapter 11
Functions exp[- i21t (fxxo + fyYo )] and exp[i21t (fxxo + fyYo )] are Fourier
transforms ofthe distribution density p(xo,Yo) ofthe target coordinates:
Xy
( 11.2.18)
In the assumption that the target object coordinates are uniformly distributed
within the input image area, nmctions exp[- i21t (Ixx o + lyYo )] and
exp[i21t (Ix X o + I yYo )] are sharp peaks with maximum at Ix = I y =0 and
negligible values for all other frequencies. Therefore, for the additive model
of the target object and image background component, averaged power
spectrum ofthe background component may be estimated as:
(11.2.19)
(11.2.20)
( )= {
wx,y
o within the target object . (11.2.21)
1 elsewhere
In a similar way as it was done for the additive model and in the same
assumption of uniform distribution of the coordinates over the image area,
one can show that in this case power spectrum of the background image
component averaged over all possible target coordinates may be estimated as
where W (Ix' I y) is Fourier trans form of the window function w(x, y). This
method of estimating background component power spectrum resembles the
conventional methods of signal spectral estimation that assurne convolving
11. TARGET LOCATIng IN CLUTTER 471
(11.2.23)
Both models imply that, as a zero order approximation, input image power
spectrum may be used as an estimate of the background component power
spectrum:
(11.2.25)
,,
I
' I I J ~ ['[--J I I
0.6
\)~- 0.8
,/, I i
50 100 150 200 250 20 40 60 80 100 120
a)
b)
Figure 11-3. Localization of correspondent points in stereoscopie images with the matched
filter (a) and the optimal adaptive correlators (b)
11. TARGET LOCAT1ng IN CLUTTER 473
l
Init ial signal Output ofthc optimal COrrclator
220 ...r\ j\.
,I,;'..1
1 !
\; 02
,.
6.5
200ifl \ "
t\ 0.1 I
::,'
6 ii
:l \
\ ! j
\........ /
i ~.~I V 'i'"
I
.. ! 5
( 1
0.8
0.6
0.4
Figure 11-4. Comparison of localization reliability of the matched filter and optimal adaptive
correlators: a) a test X-ray image and outputs ofthe matched filter and ofthe optimal adaptive
correlators; b) - 3-D plot of the target test spot; c) - plots of false detection rates versus test
spot contrast ratio to the image standard deviation for the matched filter correlator (Jeft) and
fOT the optimal adaptive correlator (right).
474 Chapter 11
(11.3.1)
(11.3.2)
(11.3.4)
(11.3.5)
(11.3.6)
Thus, the problem has again been boiled down to that of Sect.11.2, and
one can use Eq. 11.2.14, with corresponding modifications of its
numerator and denominator, to detennine the optimal filter frequency
response. As 'ho is the mean value of the filter output at the target object
location over the distribution p(ho) , a complex conjugate spectrum
u*Vx,tJ of the object in the numerator ofthe fonnula (11.2.14)
should be replaced by a complex conjugate of object spectrum u* Vx' t y)
476 Chapter 11
averaged over variations of the target object. As for the denominator of Eg.
11.2.14, it should be modified by adding to it the variance of the object
spectrum
(11.3.7)
where AVobj (.) denotes the operator of averaging over variations of the
target object uncertain parameters. As a result, we obtain the following
equation for the "optimal-on-average filter" frequency response:
(11.3.8)
(11.3.9)
r:;0 (Ix'
filter, methods of estimating background image component power spectrum
/y )Zdiscussed in the previous section may be used. In a special
case when variations of the target object are caused by additive sensor' s
noise of the imaging system, the variance of the object spectrum is equal to
the spectral density of the noise, and optimal filter becomes:
(11.3.10)
following inequality for "signal-to-clutter ratio" SCR for the filter of Eq.
11.3.9 holds:
(11.3.11)
The right hand part of the inequality is the signal-to-clutter ratio for the
optimal filter, exactly matched to the target object.
478 Chapter 11
Let us return back to the general formula of Eq. 11.1.4 that accounts for
possible inhomogeneity of the localization criterion by means of weight
coefficients {W.}. Depending on the implementation constraint, one of three
ways to attain the minimum of p" may be suggested in this case.
~ ~ ~
This implies that the filter should be re-adjustable for each s-th fragment.
Frequency responses of fragment-wise optimal filters are determined as it
was shown in Sect. 11.2 for the homogeneous criterion on the basis of
measurements of spectra of individual fragments.
According to Eq. 11.1.4, the fragments are assumed not to overlap.
However, from the very sense of Eq. 11.1.4, it is obvious that it gives rise to
a sliding window processing. In the sliding window processing, 10calization
is carried out by applying to image fragments within the sliding window at
each window position the optimal adaptive correlator that is designed upon
estimates of local power spectra within the window. Then the device for
locating signal maximum analyzes obtained output image and generates
coordinates of the highest correlation peak. We will refer to sliding window
optimal adaptive correlators as to loeal adaptive eorrelators.
Experimental verification of the local adaptive correlators confirms that
they substantially outperform global optimal adaptive correlators in terms of
signal-to-cluster ration they provide. Fig. 11-5 illustrates application of the
local adaptive correlator to localization of a small fragment of one of the
stereoscopic images in the second image. Size of the fragment is 8x8 pixels,
image size is 128x256 pixels, size of the sliding window is 32x32 pixels.
Graphs on the figure illustrate difference between filter response to the target
and that to the rest of the image. The same filter applied globally rather than
locally fails to properly localize the fragment (compare with the example
shown in Fig. 11-3, when global optimal adaptive correlator provides
approximately the same signal-to-clatter ration for a target fragment of
16x16 pixels).
11. TARGET LOCATlng IN CLUTTER 479
~ ~--------~----------~
~ r---~----~------------n ~r----,~-----;---;------ ....-i
ror---------~r.-.__1----__H 201---~---t---'--:----'----+-t-i
Figure 11-5. Localization of a small (8x8 pixels, highlighted) fragment of the right of two
stereoscopie images on the left image by the local adaptive correlator.
~ = -1P(b.ldb, t. w, [! J
w, (x" y,){ AV.. h, (b t(x"y, )~b r'dy, =
Jp(ho)dhoAV Jhs(h~b, bg (11.4.2)
bo
Now one can, by analogy with Eqs.l1.3.9, conclude that the optimal filter
may be designed upon averaged power spectra of background components of
image segments as
(11.4.4)
Thus, the optimal filter frequency response in this case depends on the
weights {Ws} and is determined by the averaged power spectrum of all
segments. Of course, such an "averaged" filter may not be as efficient for
individual image fragments as the filter adapted to these fragments but it is
definitely less expensive for computation and hardware implementation.
One can also notice that if a set of images is given for which the filter should
be optimized, the averaging should be extended to this set. In the limit, it
reduces to the estimation of the image statistical power spectrum, and we
arrive at the optimal filter of Eq. 10.4.2 for additive non-white Gaussian
noise.
The above reasoning mayaIso be applied to pattern recognition tasks.
When it is required to recognize an image from a given set of images. It
follows from Eq. (11.4.4) that, prior to recognition, images in the set should
be normalized and the normalization can be implemented in the Fourier
domain by the filter:
(11.4.6)
where k is image index and N is the number of images in the set. The
recognition can then be carried out by filtering the normalized images with
filters
11. TARGET LOCATIng IN CLUTTER 481
(11.4.7)
and by subsequent determining the filter that provides the highest output.
c. Image homogenization. In eases when non re-adjustable loealization
deviee does not provide admissible loealization reliability while one eannot
afford the eomputational expenses for the implementation of the loeal
adaptive eorrelator, an intermediate solution, an image preproeessing that
improves image homogeneity may be used (~IJI~ After the preproeessing,
optimal adaptive eorrelator should be built for the preproeessed image. We
will refer to such a preproeessing as to image homogenization.
The idea of the image homogenization arises from the following
reasoning. In terms of the design of the optimal adaptive eorrelator, image
may be regarded homogeneous if speetra of all image fragments involved in
the averaging proeedure in Eq. (11.4.5) are identieal. While this is, in
general, not the ease, one ean attempt, before applying the optimal adaptive
eorrelator, to reduee the diversity of image loeal speetra by means of an
appropriate image preproeessing.
One of the eomputation-wise simplest preproeessing homogenization
methods is the preproeessing that standardizes image loeal mean and loeal
varianee. In diserete signal representation, it is defined for image fragments
of (2N) + lX2N2 + 1) pixels as
(11.4.8)
where (k,l) are running coordinates of the window eentral pixel, {bk,/} are
image sampIes,
(11.4.9)
(11.4.10)
Output of the rnatched fdter (row) Output of tbe rnatc:bed fdter (column)
o o
-0.5
-0.5 ····· ····i···········i········+·······+········· ~ ........!...
50 100 150 200 250 20 40 60 80 100 120
Figure 11-6. Illustration of improvement of the localization with the use of image
"homogenization"
11. TARGET LOCAT1ng IN CLUTTER 483
(11.5.1)
H~;!){tx,/J=
u*{/x,/J H;:".{tx'/y)x!!·O){tx,/yf =
p.!!.O){tx,/yf/Hlms{tx,/yf p.!!.O){/x,/yf/H;.... {tx,/yf + N{/x'/y)
_ u*{tx,/J [ 1 /Hlms{/x,/yfp.!!·O){/x,/yf]
-i".(0.0)(1" I~l H;•..(/x,/Ji".(o.O)(1" I"~l/H (I" l~l+N (I" 1")'
~bg J x' y~ ~bg Jx'Jy~ Im. Jx' y~ im. Jx'Jy
(11.5.2)
is obviously the optimal adaptive filter designed for the undistorted image.
As for the second factor,
(11.5.4)
(11.5.5)
where SCRld is signal-to-clutter ratio for the nonblurred image. One can
evaluate from Eq. (11.5.5) how much image blur worsen the localization
reliability. In particular, Eq. (11.5.5) shows that so long the noise level in the
imaging system is not very high, image blur may have marginal effect on the
performance of the optimal filter.
11. TARGET LOCATlng IN CLUTTER 485
objects and object features are, on contrary, emphasized. In asense, one can
say that whitening is an operation that automaticaIly enhances dissimilarities
and suppresses similarities of objects in images. This is weIl illustrated by
whitening of a test image of geometrical figures and printed characters
shown in Fig. 11-8 and of a test texture image shown in Fig. 11-9. In Fig.
11-8 one can see that the whitening does enhance edges but the enhancement
is very selective. Vertical and horizontal edges that are common to aIl
figures and characters are enhanced much less then circumferences, slanted
edges and corners in geometrical figures and characters and those are also
enhanced very selectively according to their rate of occurrences in the
image. Whitening of texture image in Fig. 11-9 mayaiso be interpreted as
edge enhancement but the nature of this enhancement is substantiaIly
different from that in Figs. 11-7 and 8. In this example, the whitening
automaticaIly detects boundary between different textures that differ in their
spatial statistical properties rather then in their gray level distribution.
0.8 0.8
~ .
0.6 0.6
0.4 0.4
. - 0.2
0.2
~-~~I · I~~~-~
-0.2 L....._ _ _ __ ~_---' -0.2 L.....--'_--'-~~_ __ ~
Figure 11-10. 1-D sections of impulse responses of whitening filters for two images of
Figs. 11-7 and 8
One can see that those impulse responses are not identical though they
are rather similar. The remarkable fact is that the dissimilarity between the
whitening filter impulse responses is not commensurable with the
dissimilarity between the two images. This may be attributed to the
statistical nature of the image power spectra measured for the entire image
488 Chapter 11
with many objects in it. One can also note that whitening filter impulse
responses resemble very much the impulse response of the filter that
computes image spatial Laplacian defined by the matrix:
(11.6.2)
(11.6.3)
It follows from this formula that the signal-to-clutter ratio at the output ofthe
optimal adaptive correlator is equal to the energy of the whitened spectrum
of the target object. Therefore the higher the energy of the whitened
spectrum, the higher is the signal-to-clutter ratio that one can expect at the
optimal adaptive correlator output for this object. It is this relationship that
quantitatively measures the importance of different signal frequency
components.
This relationship provides also a useful quantitative measure for the
selection of target objects in cases when target objects have to be chosen
from the image. This is the case, for example, in image registration and
matching with a map. Similar problems occur in stereo image analysis, in
robot vision navigation, in multi modal medical imaging. In such
application, the question is how to make the choice of the target objects to
the best advantage. As it follows from Eq. (11.6.2.) those image fragments
will be the best target objects that have maximal energy of their "whitened"
power spectrum. Therefore, the reference objects with intensive high
11. TARGET LOCATlng IN CLUTTER 489
frequency components, that is, image fragments that are visually interpreted
as containing the most intensive edges or texture, will be the best candidates.
This conclusion is illustrated in Fig. II-11 where the "goodness" of the
fragments of 7x7 pixels in terms of the SCR-factor is represented by the
pixel gray levels for test images of Figs. 11-8 and 9. One can see from these
images that edge rich areas are indeed most appropriate potential candidates
for reference object. However, the remarkable observation is that, for
rectangles in the image of geometrical figure and for the image of periodical
textures, corners have much higher "goodness". This result is very natural.
Edges in all rectangles are similar and cannot be regarded as their specific
features. For the texture image, corners of areas of textures with different
orientation are certainly the more specific objects.
Figure 11-11. Potential reference objects for test images of Figs. 11-8 and 9: object
"goodness" is displayed proportional to pixel gray levels
490 Chapter 11
(11.7.1)
11. TARGET LOCATIng IN CLUTTER 491
(11.7.2)
(11.7.3)
POFISNR(/)= { ~(
a' Vx ,/J
)2' or
fi (I
x'J y E
1") F
, (11.7.4)
f'"" /x'/y~
0, otherwise
(11.7.5)
492 Chapter 11
where binary function e(/" , /y) that takes values 0 and 1, and constants
{«PI ,4>2} have to be chosen to optimize the filter figures ofmerits.
"Minimum Average Correlation Energy" filters (MACE- filters, [18]).
They maximize the ratio of the squared peak value of the correlation
function to the average correlation plane energy. The MACE-filters are,
in principle, designed to identify (or detect) multiple targets (in the
presence of virtually different types of uncertainty in their apriori
description), but for precisely known objects they coincide with the
ACMF-filters.
Entropy Optimized Filters (EOF [19]) that provide a strong and narrow
peak for a match between the input and filter function as contrasted with
a uniform signal distribution for any pattern to be rejected. It was shown
by computer simulation, that the discriminatory power of EOF is much
better than that of POF and is comparable with that of MACE-filters. It
was also noted, that a dominant feature of the EOF, observed
experimentally, is substantial enhancement of the high frequency
components of images. Another important feature of the EOF is its
adaptivity, because the filter optimality criterion takes into account the
patterns to be rejected.
Phase-only correlators (POC[20]):
a:(t",fJ
", y-IßV",/y~ ~v",/yr
POC(f /)_ 1 (11.7.6)
where ß(t" '/y) is spectrum of the image in which the target object has
to be detected.
As one can see, the design of only two latter filters takes, in one or
another way, into ac count objects that have to be discriminated from the
target object by the filter. All other filters are optimized only in terms of the
sharpness of the peak of their response to the target object. Obviously, the
latter does not necessarily seeures the high discrimination capability of the
filters with respect to non-target objects. Nevertheless, all these filters may
be regarded as non-adaptive approximates to the optimal adaptive correlator
defined by Eq.(11.2.15) and its implementations described in Sect. 11.2.2.
This explains their improved discrimination capability as compared to the
matched filter.
For instance, frequency response of the optimal adaptive correlator may
be approximated as
(11.7.7)
11. TARGET LOCAT1ng IN CLUTTER 493
Correlator' s
output
Filter formation according
to Eqs. 11.2.15 and 11.2.18
1
Toc-
]k . (11.7.8)
We ill refer to this type of the nonlinearity as to k-th law nonlinearity. The
correlator of Fig. 11-12 with a matched filter a:
(Ix, I y ) installed in its
Fourier plane after the nonlinear medium will implement input image
filtering with frequency response
(11.7.9)
11. TARGET LOCATIng IN CLUTTER 495
Matchcd IiIter
,
,,
,
Parallel ; ,'
laser light ~ __ E __
bea m ............
",
" ,
Input Correlation
I-st Fourier Fourier 2-nd Fourier
image plane
lens plane lens
a)
-=·,~---r~-----
, I
, ' .... I
,,
" ........... I
I "',I .......
I
I " , I ........
I " I
I
I " ,
,
I
I
b)
transformer shown in Fig. 11.-12, b). In this correlator, light passes through
the nonlinear medium twice which means that the required dynamic range of
the nonlinear medium is roughly a square root of that required for the
medium in 4-F correlators.
Computer simulation of nonlinear optical correlator with k-th law
nonlinearity and its comparison with matched filter, phase-only filter and
phase-only correlators demonstrated that ([21])
Optimal value of the nonlinearity index k is c10se to 1.
Input image power spectrum moderate smoothing for the
implementation of the correlator significantly improves the
discrimination capability of the correlator evaluated in terms of
signal-to-c1utter ratio.
k-th law nonlinear correlator substantially outperforms phase-only
correlator, phase-only and filter matched filter in discriminating
target objects against non-target objects. In the experiments, signal-
to-c1utter ratio for these four types of correlators related, by the order
of magnitude, as 7:3:2: 1. At the same time, two to five times gap
was observed between signal-to-c1utter ratio for the optimized k-th
law nonlinear correlator and the ideal optimal adaptive correlator
designed, in the simulation, for the exactly known power spectrum of
the background image component.
A trade-off exists between the optimal nonlinearity index k and the
degree of the dynamic range limitation of the nonlinear medium; a
higher degree of limitation requires a higher value of k.
Combination of input image power spectrum smoothing, high
nonlinearity index k and a high degree of the dynamic range
limitation promises considerable improvement in the nonlinear
correlator's light efficiency while preserving their high
discrimination capability.
Joint Transform correlators offer yet another possibility for implementing
optimal adaptive correlator. In nonlinear JTCs, input image and target object
joint spectrum is non-linearly transformed in an electronic amplifier before
modulating the output spatiallight modulator (Fig. 11-14). Ifthe amplifier is
the logarithmic one, signal recorded on the spatial light modulator is
proportional to
InHIß(Ix,/yf +f(Ix,/yf J+
ß· (Ix'/y )x(/x'/y )exp(i2/xx)+ ß(Ix'/y)x· (Ix'/y )exp(- i2/xx)} =
InlIß(Ix,/yf +f(Ix,/yf J+
11. TARGET LOCATIng IN CLUTTER 497
(11.7.10)
(11.7.11)
is made on the base that the size of the target object is much smaller then the
input image size.
Target
object
Nonlinear
amplifier
Figure 11-14. Schematic diagram ofthe Joint Transform correlator with a nonlinear
transformation of the joint spectrum
As one can see, the last term in Eq. 11. 7.10 approximates the output signal of
the optimal adaptive correlator in which power spectrum of the background
498 Chapter 11
Object's coordinates
Figure 11-15. Schematic diagram of the localization device for multi-component images
~
H(Jx ,Iy ,I.,
r
object. Reformulate Eq.ll.8.1 in terms of power spectrum of the c1utter
~!!,O)(/",/y,fm that may be estimated from the observed image. To this
goal, one can rely on the inequality:
It dV.,f"f.l· HV.,f"f.f
V s
r r
(Eq. 11.8.2)
ML~ c,(J,,'/y,fm ·IH(J,,'/y'/m
I ..
(Eq. 11.8.3)
Right part of Eq. 11.8.3 defines the lower bound of SCR for any
spectrum of c1utter and any filter frequency response. It follows now from
Cauchy-Schwarz-Buniakowsky inequality that choosing
(Eq. 11.8.4)
(Eq. 11.8.5)
11. TARGET LOCATIng IN CLUTTER 501
(11.8.6)
H (J f f. )= CI. *Vx,I,.,/J
Opl x' ,.' '" Iß(Ix ,1,.,1... )2~ • w (Ix,I,.,I. . ) , (11.8.7)
(11.8.9)
502 Chapter 11
loSt ebaDuel
eorrelator
,,
Primary image
components
Figure 11-16. A separable implementation ofthe multi-component localization device
(11.8.10)
and
(11.8.11)
r
spectrum, the following empirical estimations from input image power
spectrum Iß(Jx' I y' Im may be suggested:
(11.8.12)
11. TARGET LOCATIng IN CLUTTER 503
and
r;:){tx,/yf == IIß{tx,fy,f",f .
I ..
(11.8.13)
1--" - - -
Chanoel correl.tor
I-D
component-wise·
inverse DU
1-0
2-wbitening
component -wise
filter
r
2-whitenlng ruter
lIri;)(f~
---
component- wise i.."..,', ................. ____ .. __ ................. ,.. ," ....................... j inverse
DFl' DFr
- - ----
Channel correl.tor
Localization
of signal maximum
Input image
componenls
Target co-ordinates
Figure 11-17. Flow diagram of computer implementation of the separable optimal adaptive
multi-component correlator
m = 0,1,2; (11.8.14)
_ b(m+l)mod3 + b(",+2)mod3
b (",) = b - kJ kJ • m = 0"1 2 , (11.8.15)
kJ", 2 '
Figure 11-18. Color image RGB components (upper row) and the result of component wise
whitening (bottom row)
-0. 5 2 3
::1
0 .'
oe
0'
o~
Figure 13-20. Localization of a target in a color image: separable optimal correlator (images
are printed in gray scale)
506 Chapter 11
..
Ct'Ioft'''&A.t&ir". 0UI1'loIA
~.
.t .,.:).nl!!l
..
C ~lW"".
. .
&I."I'U! 6' y.~1"
.
00 0 .0
I
0."
0 .4 0."
I 1III \, r l
I.'
J I
_0" "
I
I
~ (I " I
I.
d "
r 'r ' '.
." I .1
J
Figure 11-21. Localization of a target in the single component image (Green component
ofthe image in Fig. 11-20)
References
I. L. Schwartz, "Analyse Mathematique", Hermann, Paris, 1967
2. L. Yaroslavsky, "Adaptive Nonlinear Correlatorsfor Object location and Recognition",
In:Advanced Optical Correlators for Pattern Recognition and Association, Ed.
Kazuyoshi Itoh,Research Signpost, Trivandrurn, Kerala, State, India, 1997, pp. 47-61.
3. L. Yaroslavsky, "Nonlinear Optical Co"elators with improved discrimination
Capability for Object Location and Pattern Recognition", In : Optical Patern
Recognition, Ed. F.S.T. Yu and S. Jutamulia, Cambridge University Press, 1998, 141-
170
4. R.B. B1ackman and 1.W. Tukey, The Measurement ofPower Spectra, New York: Dover,
1958
5. L. Yaroslavsky, O. Lopez-Coronado, Juan Campos, Input image "homogenization"
method for improvement the correlator's discrimination capability, Optics Letters, 15
July, 1998
6. D. Marr, Vision. A Computational Investigation into the Human Representation and
Processing ofVisual Information, W.H. Freeman and Co., N.Y. 1982
7. A. Vander Lugt, Signal Detection by Complex Spatial Filtering, IEEE Trans., IT-IO,
1964, No. 2, p. 139
8. S. Lowenthal, Y. Belvaux, Reconnaissance des Formes par Filtrage des Frequences
Spaciale~ Optica Acta, v. 14, p. 245-258, 1967
9. A. Vander Lugt, A Review ofOptical Data-Processing Techniques, Optica Acta, v. 15,
1-33, 1968
10. R. A. Binns, A. Dickinson, B. M. Watrasiewicz, Method ofIncreasing Discrimination of
Optical Filtering, Appl. Optics, v. 7,1047-1051, 1968
11. 1. L. Horner, P.D. Gianino, Phase-only Matched Filtering, Appl. Opt., v. 23, p. 812-816,
1984
12. J.L Horner, J. R. Leger, Pattern Recognition with Binary Phase-only Filters, Appl. Opt.,
v. 24, 609-611, 1985
13. G.-G. Mu, X.-M. Wang, Z.-O. Wang, Appl. Opt., v. 27, 3461,1988
14. A.A. S. Awwal, M. A. Karim, S. R. Jahan, Improved Correlation Discrimination Using
an Amplitude-modulated Phase-only Filter, Appl. Opt., v. 29, 233-236,1990
15. B. V. K. Vijaya Kumar, Z. Bahri, Phase-only Filters with Improved Signal to Noise
Ratio, Appl. Opt.,18, pp. 250-257, 1989
16. F. M. Dickey, B. V. K. Vijaya Kumar, L. A. Romero,1. M. Connelly, Complex Ternary
Matched Filters Yielding High Signal to Noise Ration, Opt. Eng. , v. 29, p. 994-1001,
1990
17. W. W. Farn, 1. W. Goodman, Optimal Binary Phase-only Matched Filters, Appl. Opt., v.
27,pp.4431-4437,1988
18. A. Mahalanobis, B. V. K. Vijaya Kumar, D. Casasent, Minimum Average Correlation
Energy Filters, Appl. Opt., v. 26, pp. 3633-3640, 1987
19. M. Fleisher, U. Mahlab, 1. Shamir, Entropy Optimized Filter for Pattern Recognition,
Appl. Opt. V. 29, pp. 2091-2098,1990
20. K. Chalasinska-Macukow, "Generalized Matched Spatial Filters with Optimum Light
Efficiency", in Optical Processing and Computing. H. H. Arsenault, T. Szoplik and B.
Macukow, Ed. (Academic Press, Boston, 1989), p. 31-45.
21. L.P. Yaroslavsky, "Optical Correlators with (-k)th Law Nonlinearity: Optimal and
Suboptimal Solutions", Applied Optics, 34, pp. 3924-3932 (1995)
22. L. Yaroslavsky, E. Marom, Nonlinearity Optimization in Nonlinear Joint Transform
Correlators, Applied Optics, vol. 36, No. 20,10 July, 1997, pp. 4816-4822
23. L. Yaroslavsky, Optimal Target Location in Color and Multi-component Images, Asian
Journal ofPhysics, Vo18, No 3,1999, pp. 100-113
508 Chapter 11
24. K. Yamaba and Y.Miyake, "Color character recognition method based on human
perception," Opt. Eng. 32, 33-40 (1993).
25. S.L.Guth, "Model for color vision and light adaptation," J. Opt. Soc. Am. 8, 976-993
(1991).
26. E.Badique, Y.Komiya, N.Ohyama, J.Tsujiuchi and T.Honda, "Color image correlation,"
Optics Comm. 62, 181-186, (1987).
27. E.Badique, N.Ohyama, T.Honda and J.Tsujiuchi, "Color image correlation for
spatial/spectral recognition and increased selectivity," Optics Comm. 68, 91-96, (1988).
28. V. Kober, V. Lashin, L. Moreno, J. Campos, L. Yaroslavsky, and M.J. Yzuel, Color
component transformations for optical pattern recognition, JOSA, A/vol. 14, No.IO,
October 1997, pp.2656-2669
Chapter 12
We will assume that images are single component signals with scalar and
quantized values.
Main common properties of the filters that constitute the base of thei
classification are:
Filtering is performed within a filter window. For processing signals
and images filter window scans input image sampie by sampie.
In each position k of the window, with k being a coordinate in the
signal domain, filters generate, from input signal sampies {b~)}
within the window, an output value a for this position by means of
k
(12.1)
T.ablel 22
- . EstlmatlOn ooeratIons
Ooeration Denotation I Definition
SMTH (NBH): Data smoothing operations
Arithmetie and MEAN(NBH) Arithmetic mean ofsamples ofthe neighborhood
geometrie PROD(NBH) Product of sampies of the neighborhood
mean
Order statisties K_ROS(NBH) Value that occupies K-th place (has rank K) in the
variational row over the neighborhood. Special
cases:
MIN(NBH) Minimum over the neighborhood (the first term of
the variational row)
MEDN(NBH) Central element (median) ofthe variational row
ROS operations may be optimal MAP estimations for other then additive
Gaussian noise models. For instance, if neighborhood elements are
observations of a constant distorted by addition to it independent random
514 Chapter 12
a)
d) e)
Figure 12-2. Illustration ofRAND(NBH) as a data smoothing operation: a) - initial test image
with 256 quantization levels; b) - result of applying operation RAND(Wnbh) in the window
of 7 x 7 pixels; c) - result of applying operation RAND(EV-nbh) operation in the filter
window of 7x 7 pixels and E: =E; = 10 (for the definition of EV-neighborhood see
Sect. 12.1.4); d) - difference between images (a) and (d); standard deviation ofthe difference
is 26; t) - difference between images and (e); standard deviation ofthe difference is 5.6.
12. NONL1NEAR FILTERS IN IMAGE PROCESSING 515
FLAT(NBH)-
"Flat"-neighborhood
I
G-neighborhoods: Geometrical attributes
Neighborhood elements with values of Laplacian (or
module of gradient) lower than a certain threshold
Linear combination of elements of nei2hborhood
T(NBH) I Orthogonal transform T of neighborhood elements
DEV(NBH,a) I Differences between elements of the neighborhood and certain value a
between window sampies and certain value, such as, for instance, that of the
window central pixel.
J) •.•••• ~. • •••.•••••••••••••••••••.•
;n • r · ••
15 ••• • •••••
- _._ .. _· · ·······r······
r
a special case of transform domain filters in which signal squared transform
coefficients IT(Wnbh (signal spectral estimations) are replaced by their
mean values (STD(Wnbh)Y .
Stack filters belong to yet another large family of filters. They originate
from the idea of threshold decomposition of multilevel signals to binary
signals to which Boolean functions are then applied ([13]).
Four stage neighborhood filters are exemplified in Table 12-7 by a family
ofpolynomial filters and by Weighted Majority of m Values with Minimum
520 Chapter 12
"Ranked order"
("percentile") filters
a = K _ ROS(Wnbh)
k
R-nei2hborhoods
Alpha-trimmed mean,
median
ak = MEAN(Qnbh(Wnbh,R/eft,Rr;ghl ));
a = MEDN(Qnbh(Wnbh,R,eft,Rrighl))
k
Impulse
noise
General a = MEMB(Qnbh(wnbh, Rieft' R"ght1a a
k k )· k +
filtering [1-MEMB(Qnbh(wnbh,R,eft,Rr;ghI1ak)]x
filters
SMTH(Qnbh(Wnbh, Rieft' R nghl ))
Rank a = MEMB(Qnbh(Wnbh,Rleft,R,;ghJa
k k )· ak +
II - MEMB(Qnbh(Wnbh, Rieft' Rrighl 1a
Conditioned
Median filter k) j X
{o.x,;; °
where 0 is a filter parameter,
STEP(x)=
1,x> 0
LLMMSE-
filter a k = 1- (
A (
(0
STD Wnbh
))2 ak + ( (0
2
STD Wnbh
J 2
)y MEAN(Wnbh) ,
where cr 2 is a filter parameter
Double Window
Modified Trimmed
a = MEAN(EVnbh(Wnbh;MEDN(SHnbh~E;;E; )).
k
Mean filter
...
Stack filters
ak =
MAX(MIN(SubWnbhl~MIN(SubWnbh2~ •••,MIN(SubWnbhn)).
WMVMS ak = MEAN(MULT_R(MIN_RNG({SubRnbh}m)}))),
- filters
where {SubRnbhf m )} are rank based sub-neighborhoods of m
elements.
522 Chapter 12
(12.2.1)
(12.2.2)
Size controlled
a. =sw . MEAN(EVnbh(a. ))+ (1- sw )MEDN(SHnbh)
(SC-) "Sigma"- sw =RECT(SIZE(EVnbh(Wnbh;a. ;E;;E~ ))- Thr)
filter
Size_EV -filter ak = SIZE(EVnbh(Wnbh;E Vpl;E Vmmn ))
p- histogram equalization
a = f(h(q)Y '!(h(q)Y
k
q=O q=O
"Cardnl-filter" ak = HIST(Wnbh,ak)
NBH1-histo2ram equalization
EVnbh-histogram ak =RANK(EVnbh(Wnbh;ak;E Vpl;E Vmn))
equalization
KNVnbh-histogram ak = RANK(KNV(Wnbh;ak;K»
equalization
"SHnbh-histogram ak =RANK(SHnbh)
equalization
In the design of filters for image denoising one should distinguish cases
of additive noise, impulse noise and of mixture of both types of noise. The
following equation describes the principle of nonlinear filtering for cleaning
additive noise:
alt)
kJ
= SMTH(NBH{a(t-l»)~
~ kJ ~, (12.3.1)
(12.3.2)
10 1\ I
8 \
6 \ --
4
- f\
2 I \ l
\1
2 3 4
Iteration
b)
a)
Figure 12-4. Iterative EV-neighborhood filtering additive noise: a) - test piece-wise constant
image (I), image distorted by additive noise with a limited dynamic range (2), filtered image
after 4 iterations (3), difference between noisy (2) and filtered (3) images amplified for better
visibility; b) - standard deviation of difference image as a function of the iteration index
528 Chapter 12
(12.3.3)
d) d)
Figure 12-5. Filtering additive noise with nonnal distribution a) - initial test piece-wise image
distorted by additive Gaussian noise with standard deviation 20 within image dynamic range
0-255; b) - result of 5 iterations of "Sigma" filter, with Wnbh=5x5 , c; c:
= = 20; c) -
result of 5 iterations of size-controlled "Sigma" filter (Eq. 12.3.4) with the same parameters
=
and threshold size Thr 5; d) filtering same image with median filter in the window of 5x5
pixels.
12. NONLINEAR FILTERS IN IMAGE PROCESSING 529
a = MEDN(Wnbh)
k (12.3 .5)
Figure 12-6. Iterative SC-"Sigma" filtering mixture of additive and impulse noise: a) a piece-
wise test image distorted by additive Gaussian noise with standard deviation 20 (in image
range 0-255) and by impulse noise with probability 0.15; b) SC "Sigma"-filtering result with
the same filter parameters as in Fig. 12-6.
530 Chapter 12
Figure 12-7. Filtering impulse noise: a) initial image; b) - image distorted by impulse noise
with probabilityO.3; c) -filtering result; d)
a(tl
k,l
= sw·a(t-I)
k,l
+ (l-sw) .SMTH es. (NBH';i-I))~
~ k,/ ~, (12.3.6)
(12.3.8)
(12.3.9)
a) c)
Figure 12-8. Image unsharp masking using MEAN (Wnbh) and MEDN(Wnbh) in the
window of 15x 15 pixels as smoothing operations
Qk = RANK(NBH,a k )· (12.3.10)
Figure 12-9. Loeal histogram equalization and P-histogram equalization: a) - initial image
(256x256 pixels) of eye blood vessels; b) - loeal histogram equalization over spatial window
of 35x35 pixels; e) - P-histogram equalization with P=O.5.
534 Chapter 12
a) b) e) d)
e) d)
Figure 12-11. Comparison ofloeal histogram equalization over spatial neighborhood and over
EV - and KNV -neighborhoods applied to an angiogram. A) - initial image (256 quantization
levels, 256x256 pixels); b) loeal histogram equalization over spatial window of 15xl5 pixels;
e) - loeal histogram equalization over KNV-neighborhood in the same spatial window with
K= 113 (half of the window size); d) loeal histogram equalization over EV -neighborhood in
=
the same spatial window with 8; 8; 10. =
12. NONLINEAR FILTERS IN IMAGE PROCESSING 535
Other pixel attributes that may be useful for visual image analysis are size
ofEV-neighborhood and pixels cardinality. Image enhancement by means of
measuring and displaying these attributes is illustrated in Fig. 12-12 on an
example of a microscopic image of crystals.
Figure J2-12. Local contrast enhancement using Size-EV and cardinality as pixel attributes:
a) - original image (256 gray levels); b) -c): Size-EV (C; =C; =3) and pixel cardinality
displayed as images.
(12.3.11)
15 - - I-
_L __ L __ ""' __ -I _ _
I I I I I
I I
1 - _!..
I
I
05 - - I"
I
c)
e) f)
a) b)
c) d)
Figure 12-14. Edge detection using different image attributes: a) -Iocal standard deviation in
the window of 7x7 pixels; b) - local standard deviation in the window of 3x3 pixels; c) -
SIZE(EV(Wnbh) in the window of 7x7 pixels, 8; = 8; = 3 (256 level of quantization); d)
- differenee between loeal maxima and minima over spatial window of 3x3 pixels. Image c) is
identieal to the image ofFig. 12-12 b) shown in negative.
Figure 12-15. The importance ofimage smoothing before edge detection: a) initial image ofa
printed circuit board; b) - edge image generated by the algorithm MAX(Wnbh3x3)-
MIN(Wnbh3x3; c) - edge image generated by the same algorithm after image smoothing by
Sigma-filter (MEAN(EV(Wnbh5x5,3,3))); d), e) - corresponding magnified fragments of the
edge images.
References
1. J. W. Tukey, Exploratory Data Analysis, Addison Wesley, 1971
2. J. Serra, Image Analysis and Mathematical Morphology, Academic Press, 1983, 1988
3. V. Kim, L. Yaroslavsky, "Rank algorithms for picture processing", Computer Vision,
Graphics and Image Processing, v. 35, 1986, p. 234-258
4. I. Pitas, A. N. Venetsanopoulos, Nonlinear Digital Filters. Principles and Applications.
Kluwer, 1990
5. E. R. Dougherty, An Introduction to Morphological Image Processing, SPIE Press, 1992
6. H. Heijmans, Morphological Operators, Academic Press, 1994
7. E. R. Dougherty, J. Astola, An Introduction to Nonlinear Image Processing, SPIE
Optical Engineering Press, 1994
8. L. Yaroslavsky, M. Eden, Fundamentals 0/Digital Optics, Birkhauser, Boston, 1996
9. J. Astola, P. Kuosmanen, Fundamentals o/Nonlinear Digital Filtering, CRC Press, Roca
Baton, New York, 1997
10. E. Dougherty, J. Astola, Nonlinear Filters/or Image Processing, Eds., IEEE publ., 1999
1l. L. P. Yaroslavsky, K. o. Egiazarian, J. T. Astola, "Transform domain image restoration
methods: review, comparison and interpretation", Photonics West, Con/erence 4304,
Nonlinear Processing and Pattern Analysis, 22-23 January, 2001, Proceedings 0/ SPIE,
v.4304.
12. D. L. Donoho, I.. M. Johnstone, "Ideal Spatial Adaptation by Wavelet Shrinkage",
Biometrica, 81(3), pp. 425-455,1994
13. P. D. Wendt, E. J. Coyle, and N. C. Gallagher, Jr., "Stack Filters ", IEEE Trans. On
Acoust., Speech and Signal Processing, vol. ASSP-34, pp. 898-911, Aug. 1986.
14. T. Szoplik, Selected Papers On Morphological Image Processing: Principles and
Optoelectronic Implementations, SPIE, Bellingham, Wa., 1996 (MS 127).
Chapter 13
COMPUTER GENERATED HOLOGRAMS
which are functions of the object's surface coordinates (x, y, z) and are
connected with the relationship: O(x,y,z)= lo(x,y,zf .
C 1. ~ __
_---,- ----------
..
lIIum'",,"on
---- --------- ;7 \
/-~7_--.
\
. ,-- ---- ---- \
,,,
\
--.',
Light
scattered by
angle the object surface
whose kernel T(x, y, z;;, 11 ,l;) depends on the spatial disposition of the
object and the observation surface. Reconstruction of the object can be
described by a back propagation integral over the observation surface S s :
In view of these limitations, one may break down the observation surface
into areas approximated by planes, and, using the laws of geometrical
optics, replace the wave amplitude and phase distributions over the object
surface by those of the wave in a plane tangent to the object ( or sufficiently
elose to it, so as to make diffraction effects negligible) and parallel to the
given plane area of the observation surface. In this way the problem of
hologram synthesis over the whole observation surface boils down to the
synthesis of fragmentary holograms for the plane areas of this surface, with
the complete hologram being composed as a mosaic of fragmentary ones.
Instead ofEq.(13.1.4) one then obtains for fragmentary holograms:
(13.1.8)
-00-00
(13.1.9)
546 Chapter 13
Point source of
Reconstructing reconstructing spherical
parallel light beam lightbeam
- -----------~
-------------~
--------------~
Hologram Hologram
Figure J3-3. Scheme of visual observation of virtual objects using computer generated
Fresnel (a) and Fourier (b) holograms
Having passed from the spatial 3D-problem to the problem of treating its
20 approximation, we, strictly speaking, lose the possibility of taking into
account the precise effect of object depth relief on the wave front. Even
Fresnel holograms use only a single distance value from the object to the
observation plane rather than object relief depth. Nevertheless, it is still
possible to synthesize holograms which are capable of reconstructing 3-D
images for visual observation and retain the most important property of
holographic visualization - naturalness of object visual observation. At
least two options exist for this: composite computer generated stereo- and
macro-holograms and programmed diffuser holograms,
Stereo-holograms are pairs of holograms synthesized to be viewed by left
and right eye and to reconstruct the corresponding stereo views of the object
([I]). Composite stereo holograms are composed of multiple holograms that
reproduce different aspects of the object as observed from different
positions. Macro-holograms are large composite stereo-holograms that
cover wide observation angle and can be used as a wide observation window.
An example of a composite stereo-hologram that cover full 3600 of
observation angle with horizontal parallax is shown in Fig. 13-4.
Programmed diffuser holograms are computer generated Fourier
holograms that imitate properties of diffuse surfaces to scatter irradiation
non-uniformly in the space and in this way provide visual clue about the
object surface shape (See Sect. 7.2.4).
Figure 13-4. Composite stereo maero hologram, a holographie "movie", eomposed of 800
e1ementary holograms (adopted from [I)). I - holograms reeorded on a photographie film, 2 -
annular holder, 3 - eonie refleetor. Holograms are iIIuminated from above through the eonie
refleetor and are viewed as through a window.
548 Chapter 13
sampling step defines the angular dimensions ofthe reconstructed image. For
instance, in order to make the image's angular dimensions about 100 or
more with a reconstruction light wavelength of about 0.5 ~, A~ and AT)
should be 3 ~ at most. The existing recording devices have a sampling step
of about 1 through 10 ~, and the total number of resolution cells, by the
order of magnitude, from 10 3 x 10 3 to 10 4 X 10 4 • At the early stages of
digital holography, holograms were recorded by means of standard computer
plotters and printers. The obtained hologram hard copies were then
photographically reduced in order to achieve acceptable values of the
sampling step ([1, 5,6]).
The existing methods for recording synthesized holograms on amplitude-
only, phase-only, binary, and combined amplitude/phase media may be
classified with respect to different features. Methods of representing complex
numbers describing sampIes of the mathematical hologram will be chosen
here as the classification key (Fig. 13-5). For other possible classifications
the reader is referred to the reviews [5-7].
Complex numbers may be represented in two ways: in the exponential
form A expVcI» where A and cl> are, respectively, the amplitude and phase
of the numbers or additively as a sum of two or more basis complex
numbers. Obviously, for hologram recording on the combined
amplitude/phase media, the exponential form is the most suitable. It does not
require any special encoding. Computer generated holograms recorded on
combined media are capable of wave front reconstruction on optical axis of
the reconstruction set-up at the zero-order diffraction order and are called
on-axis holograms.
For recording on phase-only or amplitude-only media, special hologram
encoding methods are required. One of the most straightforward encoding
methods oriented on using phase-only media is that of kin%rm ([8,9]).
Kinoform is a computer generated hologram in which amplitude data of the
mathematical hologram intended for recording are disregarded and only the
hologram sampIe phases are recorded on a phase-only medium. Although
disregarding hologram sampIe amplitudes results in substantial distortions
such as appearance of speckle noise in the reconstructed wave front (see
Sect. 7.3), kinoforms are advantageous in terms of the energy of
reconstruction light because the total energy of the reconstruction light is
transformed into the energy of the reconstructed wave field without being
absorbed in the hologram. Moreover, the distortions may to some degree be
reduced by an appropriate choice of the diffusion component of the wave
field phase on the object using an iterative optimization procedure similar to
that described in Sect. 7.2.4 ([6]).
550 Chapter 13
I
tbe wave field
I
Phase-only
media I I " Kinororm"
Encoding by
Amplitude- I .,' "symmetrization"
, ' ,
onlymedia ~
Binary media 1' ", "Detour phase"
.- I ' method
,* " , ,
Addjtive Double-phase
representation Phase-only method
orcomplex media
amplitude of Multiple phase
the wave Held method
Binary
media Representation by
orthogonal and bi-
orthogonal
components
Ampljtude-
only
media
~ - 2-D symplex
representation
~ Explicit spatial
carrier methods
symmetrization rule for the object wave front specified by the array of its
sampIes {Au}' k = O,I,•••,N) -I, 1 = O,I,•••,N1 -I becomes
(13.2.1)
Encoded computer
~enerated holo~ram
Nodes ofthe
\ '. discretization raster
\ '.
\
"'\ \ \ '.'.\
\
\ \
\ \
,
\ \
\ \
~\. \,
"...
\ \
~ ~ ~ ~
Direction to the Direction of image
observation plane reconstruction
Figure 13-6. Detour phase method for coding phase shift by spatial shift ofthe transparent
aperture
As was already noted, only spatial degrees of freedom are used for
recording in binary media; therefore, the number of binary medium cells
should exceed the number of hologram sampies by a factor equal to the
product of the amplitude and phase quantization levels. This product may run
into several tens or even hundreds. The low effectiveness of using the
degrees of freedom of the hologram carrier is the major drawback of the
binary hologram method. However many modern technologies such as
lithography and laser and electron beam pattern generators ([ 12]) provide
much higher number of spatial degrees of freedom that the required number
of hologram sampies. Therefore this drawback is usually disregarded and
such merits ofbinary recording as simpler recording and copying technology
of synthesized holograms and the possibility ofusing commercially available
devices made it the most widespread. Numerous modifications of the method
are known, oriented to different types of recording devices.
With an additive representation, the complex number (regarded as a
vector in the complex plane) is a sum of several components. For recording
on amplitude-only media these components should be assigned standard
13. COMPUTER GENERA TED HOLOGRAMS 553
+,
Vif
- ~ r - ~ r eo
e;m~ e~~I
____ I
I
_L.
_
I I
. . . __ 1 _J_.
I
I
I
I
k'-:e
- I _ /" , I
•
I e re e-re I e re /" 'I
I I _'~ ----
I 240 e120
I
a) b) e) d)
Figure 13-7. Additive representation of eornplex nurnbers over orthogonal basis (a), over
biorthogonal basis (b), over 2-D syrnplex (e) and as a surn oftwo equallength veetors (d).
(13.2.2)
where e re and e im are orthogonal unit vectors. When recording a hologram,
the phase angle between the orthogonal components may be encoded by the
detour phase method, rre and r;m being recorded into neighboring
hologram resolution cells neighboring in araster rows as shown in Fig.13-7,
a. The reconstructed image will be then observed at an angle e~ to the axis
defined as folIows:
(13.2.3)
For recording negative values of rre and r;m a constant positive bias may
be added to recorded values. We will refer to this method as to orthogonal
encoding method.
Since two resolution elements are used here for recording one hologram
sampIe, such a hologram has double redundancy, as in the symmetrization
with duplication.
With such encoding and recording of a hologram, one must take into
account that the optical path differences A. / 2 and 3A. /4 will correspond to
the next pair of hologram resolution cells (see Fig.13-8 a); that is, values of
r,e and r;m for each odd sampIe of the mathematical hologram should be
recorded with opposite signs.
554 Chapter 13
s r r+l s r
mo=O mo=1
a) b)
r r+l
ere e/m e-
re
e.,--
,m no=O
0 0 90 0 180 0 2700
s
e-n e.,--
,m ere elm
00 900
no=1
180 0 270 0
mo=O mo=l
c)
Figure 13-8. Hologram encoding by decomposition of complex numbers in orthogonal (a) and
biorthogonal (b, c) bases. Hologram encoding cells corresponding to one hologram sampIe is
outlined with a double line.
(13.2.5)
One can avoid the constant biasing for recording r,e and r im byallocating,
for recording one hologram sampie, four neighboring-in-raster medium
resolution cells rather then two ([5]). This arrangement is shown in Fig.13-8,
b. With a reconstruction angle as defined by Eq.(13.2.3), phase detours 0,
1t / 2, 1t and 3n: / 2 correspond to them. Therefore, cells in each group of
four cells allocated for recording of one sampie of the mathematical
hologram should be written in the following
1- 1-
order: (r,e + Ir,el)/2; (rim + Irim j)/2; ~r,e r,e )/2; ~rim r im )/2. One can
see that in this case all the recorded values are nonnegative. The method
represents a vector in the complex plane in abiorthogonal basis
(e,e ,e-Fe ,e im ,e:-):
Im
(13.2.6)
rm,n
=!4 {;mo r[(_l)mor +r' ]+ (_l)noi moo 1(_l)mor +r'
',S 1,S ',S
1 (13.2.9)
1,S'
556 Chapter 13
(13.2.12)
where
A = (r + r* )/2;
00 1200 2400
n=2
-
00 1200 2400
J
a)
"
- b)
c)
00 1200 2400 00 1200
Figure J3-9. Three methods of arrangement of recording medium resolution cells for
hologram encoding by decomposition of complex numbers with respect to a 2-D simplex.
Tripies of medium resolution cells used for recording one hologram sampie are outIined with
abold line.
{Re{r". exp[i2rc (mo + 1)/ 3]}+ /Re{r". exp[i2rc (mo + 1)/ 3m}l.
(13.2.19)
where horizontal and vertical indices {m,n} (Fig. 13-8, a» are counted as:
m = 3r + m o, m o = 0,1,2, n = sand Re(X) is real part of X .
The above described encoding methods based on additive representation
of complex vectors are applicable for recording on amplitude-only media,
both continuos tone and binary. In the latter case, projections of the complex
number on basic vectors are represented by varying the size of the
transparent aperture in each of the appropriate resolution cells.
For recording on phase media, additive representation of complex vectors
assumes complex vector representation as a sum of complex vectors of a
standard length. Two versions of such encoding are known: double phase
method and multiple phase method.
In the double-phase method ([ 10,15]), hologram sampIes are encoded as
where (<p I ,<p 1 ) are component phase angles defined by the following
equations:
(13.2.21)
that can be easily derived from the geometry of the method illustrated in Fig.
13-6, d.
Two neighboring medium resolution cells should be allocated for
representing two vector components. Formally, the encoded in this way
hologram may be written as
where {r ',' =Ir ".1 exp(i<p ',S )} are sampIes of the mathematical hologram,
indices of the recording medium resolution cells (m,n) are counted as
m = 2r + mo' m o = 0,1, n = s and Ao is found as half of maximal value of
{I r".I}·
560 Chapter 13
I...
1
~~~ 1 l.sl~):
27t 1 1 27t '"
1
1
~~
a) b)
Figure 13-10. Double-phase encoding method for hologram recording on a binaty phase
medium: two separate cells (a); decomposition of cells into alternating suh-cells
13. COMPUTER GENERA TED HOLOGRAMS 561
r =L Ao exp(i<p q).
Q
(13.2.23)
q=1
In this case, the following equations can be derived for the increment Ll<p :
(13.2.27)
X, X 2 0
retf(X)= { , (13.2.29)
0, X <0
13. COMPUTER GENERA TED HOLOGRAMS 563
I, X ~O
hlim(X) = { . (13.2.30)
0, X <0
As one may see from these expressions, the spatial carrier has aperiod no
greater than one half that of hologram sampling. At least two sampies of the
spatial carrier have to correspond to one hologram sampie in order to enable
reconstruction of amplitude and phase of each hologram sampie through the
modulated signal of the spatial carrier. This redundancy implies that, in order
to modulate the spatial carrier by a hologram, one or more intermediate
sampies are required between the basic ones. They may be determined by
any of variety of interpolation methods.
The simplest interpolation method, zero order interpolation, simply.
repeats sampies. It is this interpolation that is implied in the recording
methods described above. For instance, according to Eq.(13.2.5') each
hologram sampie is repeated twice for two sampies of the spatial carrier, in
Eq.(13.2.8') it is repeated four times for four sampies, etc. Of course such an
interpolation found in all modifications of the detour phase method yields a
very rough approximation of hologram intermediate sampies. In Sect.l3.3 it
will be shown that the distortions of the reconstructed image caused by
improper interpolation of hologram sampIes manifest themselves in aliasing
effects.
As we have shown in eh. 9 aliasing free interpolation can be achieved
with discrete sinc-interpolation. Discrete sinc-interpolated values of the
desired intermediate sampies can be obtained by using, at the hologram
synthesis, Shifted DFT with appropriately found shift parameters that
correspond to the position of required intermediate sampies of the hologram.
It should be also noted that the symmetrization method may be regarded as
an analog of the method of Eq.(l3.2.5) with discrete sinc-interpolation of
intermediate sampies. In the symmetrization method, such an interpolation is
secured automatically and, as will be seen in Sect.13.5, the restored images
are free of aliasing images.
Along with methods that introduce the spatial carriers implicitly, explicit
introduction of spatial carriers is also possible that directly simulates optical
recording ofholograms and interferograms.
Of the methods oriented to amplitude-only media, one can mention those
of Burch ([ 17]):
(.n = 1+ Ir Icos(2njm +
m,n <j> m,n ) (13.2.31)
(13.2.34)
, .
where b is a constant bias required for eliminating negative values in
recording {r".} and indices (r, s) run over all available hologram sampIes.
Eq. (13.3.1) may be rewritten in a form ofthe convolution:
where ® stands for the convolution and summation limits are changed to
[- 00,00] having in mind that hologram window function w(~ ,T)) selects
available hologram sampIes from virtual infinite number of sampIes.
Let now (x,y) be coordinates in reconstructed image plane situated at a
distance D from the hologram plane and A. be wavelength of the
reconstructing light beam. According to the convolution theorem of the
integral Fourier transform, the result of the optical Fourier transform
performed at the reconstruction of the hologram r(~ ,T)) can then be
represented as:
0000
W(x,y)®{ H,ee(x,y )_!Lt.t (r". +b~ (~ +~o - rL1~)3 (T) +T)o - SL1T))x
where
f fW(~,ll)exp[i21t(X~ + Y1l)/A.D~d1l
00 00
W(x,y)= (13.3.4)
and
f fh,j~,ll)exp[i21t(X~ + Y1l)/W~dll
00 00
H,ec(x,y)= (13.3.5)
are Fourier transforms of the hologram window function and of the aperture
function of the hologram recording device.
Introduce now an array A}~), k = O,...,2NJ -1, 1== O,I,••. ,N2 -1, of
sampies of the object wave front. In the symmetrization method, the array is
symmetrical as defined by (Eq. 13.2.1). For Fourier holograms, mathematical
hologram r"s is computed as Shifted OFT ofthis array:
r rs oc 2~J~JA(o)
L L k,l exp {"21t [(k+u)(r+ p)+ (I+V)(S+q)]} ,
I (13.3.6)
, k~O I~O 2N J N2
where (u, v) and (p,q) are shift parameters. Substitute Eq. 13.3.6 ioto Eq.
13.3.3 aod obtaio:
(13.3.8)
H rec ( x,y'
) ~
{ m~-oo ~2~1~IA-(O)W[k+U
L. L. L L k./
ß~ l+v ß1lY
-----+m,-----+n
)
n~-oo k~O /~O 2N f,J)
I N 'AD 2
+b f f w[k2N+
m~-oon~-oo
U _
I
ß~ x+ m, I + v_ßll Y + n)}.
f,J) N2 f,J)
(13.3.9)
(13.3.10)
., IU . . . . . . . . . . .
.
1
b)
a)
A~(X'Y)OCH~(X,+oH~+ ";;)] x
f fffA(O)W(k+U-Ll~X+m+!
m:-oo n:-ook:O /:0 k ,/ NI ').J)
l+v _Ll11Y +n)+
2' N 2 ').J)
~ .t.....t.....t....
.t.... ~ ~ ~A(O) N,-kN 2 -/
'w(k- -+- - -Ll~+x m + - ,I-+-v- - - +Y n) +
U 1 Llll
m:-oo n:-ook:O / : 0 ' NI AD 2 N2 ').J)
(13.3.11)
where, as above, (u, v) and (p,q) are shift parameters of SOFT used in
computing the mathematical hologram, NI and N 2 are dimensions of the
array {A1~)}. As in the above case of hologram encoding by symmetrization,
one should take p = q = 0 and ~o =110 = 0 at hologram synthesis and
recording.
One can see from Eq. (13.3.11) that, similarly to the above case, the
reconstructed image contains a number of diffraction orders, masked by the
function H,Jx,y), spatial frequency response of the hologram recording
device, and a central spot in several diffraction orders due to the constant
bias in the hologram. But here, in contrast to the above ca se, each
diffraction order contains two superimposed images of the object, - a direct
image and its conjugate rotated by 1800 with respect the former. Each of
them is additionally masked by functions
The pattern of diffraction orders of the direct and conjugate images and
their corresponding masking functions are shown in Fig.ll-13 for
u = -NI /2 and v = - N 2 /2 . One may easily see the direct image and its
conjugate, an aliasing image whose contrast increases along the horizontal
axis from the center to the periphery.
COS[J!(.!.4 + 6,;x
AD
Figure 13-13. Image reconstruction for the orthogonal encoding method. At the bottom,
spatial weighting functions ofthe direct and conjugate images are depicted.
(13.3.12)
Aree (x, y) =
W(x,y)® {2H ree (X,y )exp(- i21t {[(~o - Ll~ /2)x +TloY ]}/Al» x
sm
o ( Ll~ XJ L..~ L..~
1t - -
1- Irr,s 1 exp(0lCJ> r s )exp[- I021t
--2
2
(2rLl~x + SLlTly)]l .
Al> r~-oos~-oo 4A o ' "AD
(13.3.13)
Irr,S IexpvCJ>
(o
r,S
)= ~1~IA(o)
L.. L.. exp{02
I 1t
[(k+uXr+p) + (I + vXs +q)]}
k,1
k~O I~O NI N2
(13.3.14)
(13.3.15)
{ cos ( 7t
Ll~X)~L.J ~~~A(O)W(k+U
-- L.JLL lLl~X
----- +m,- - - -LlllY
I+v kJ- + n) x
')J) m~-oo ,,~-ook~O '~O NI ')J) N1 ')J)
Ll~X)~ ~~~A-(O)W(k+U
• ( 7t - - L.J L.JLL
sm lLl~X
----- +m,- - - -LlllY
I+v - + n )}
kJ
')J) m~-oo "~-,,,k~O '~O NI ')J) N1 ')J)
(13.3.16)
sin( 7t Ll~X ) , respectively. In the center of the proper image aliasing image
is fully attenuated but over the peripheral area it may be of the same intensity
as the proper image. In contrast to the orthogonal coding method, the aliasing
image here is not conjugate to the original one, but is similar to it, in asense,
because, according to Eq. 13.3.15, it has the same phase spectrum and a
distorted amplitude spectrum. Unlike the holograms coded via the
symmetrization or orthogonal methods, double- phase encoding does not
produce a central spot in the diffraction orders of the reconstructed image
because the hologram is recorded in the phase medium without an amplitude
574 Chapter 13
Figure J3- J4. Hologram reconstruction for double phase recording method: a) -arrangement
of diffraction orders (Solid and dashed arrows indicate the basic and aliasing images with
diffraction order shown in boxes; spatial masking functions for both images are depicted at
bottom correspondingly by solid and dashed Iines); b) - reconstructed image for the test object
ofFig. 13-12, c).
13. COMPUTER GENERA TED HOLOGRAMS 575
References