A Tutorial of The Wavelet Transform: Chun-Lin, Liu February 23, 2010
A Tutorial of The Wavelet Transform: Chun-Lin, Liu February 23, 2010
Chun-Lin, Liu
Overview
1.1 Introduction
The Fourier transform is an useful tool to analyze the frequency components
of the signal. However, if we take the Fourier transform over the whole time
axis, we cannot tell at what instant a particular frequency rises. Short-time
Fourier transform (STFT) uses a sliding window to find spectrogram, which
gives the information of both time and frequency. But still another problem
exists: The length of window limits the resolution in frequency. Wavelet
transform seems to be a solution to the problem above. Wavelet transforms
are based on small wavelets with limited duration. The translated-version
wavelets locate where we concern. Whereas the scaled-version wavelets allow
us to analyze the signal in different scale.
1.2 History
The first literature that relates to the wavelet transform is Haar wavelet. It
was proposed by the mathematician Alfrd Haar in 1909. However, the con-
cept of the wavelet did not exist at that time. Until 1981, the concept was
proposed by the geophysicist Jean Morlet. Afterward, Morlet and the physi-
cist Alex Grossman invented the term wavelet in 1984. Before 1985, Haar
wavelet was the only orthogonal wavelet people know. A lot of researchers
even thought that there was no orthogonal wavelet except Haar wavelet. For-
tunately, the mathematician Yves Meyer constructed the second orthogonal
wavelet called Meyer wavelet in 1985. As more and more scholars joined in
this field, the 1st international conference was held in France in 1987.
1
In 1988, Stephane Mallat and Meyer proposed the concept of multireso-
lution. In the same year, Ingrid Daubechies found a systematical method to
construct the compact support orthogonal wavelet. In 1989, Mallat proposed
the fast wavelet transform. With the appearance of this fast algorithm, the
wavelet transform had numerous applications in the signal processing field.
Summarize the history. We have the following table:
2
Chapter 2
2|t|
x(t) = 1 − , |t| < T. (2.1)
T
Find its Fourier series coefficient.
3
Figure 2.1: An illustration of the approximation example. The red line is
the original signal. The blue dash line is the approximated signal with (a)
K = 0 (b) K = 1 (c) K = 2 (d) K = 3 (e) K = 4 (f ) K = 5
4
the variable k. The higher k is, the higher frequency refers to. In general, the
power of a signal concentrates more likely in the low frequency components.
If we truncate the Fourier series coefficient in the range of [−K, K] and set
ak = 0 outside this range, we expect the synthesis version of the truncated
coefficients be
K
X 2πkt
x̃K (t) = ak exp(j ). (2.5)
k=−K
T
Intuitively, as K approaches infinity, the reconstructed signal is close to the
original one. Refer to Fig. 2.1, we can see the approximation signal is very
close to the original one as k grows higher.
5
2.2 Abstract Idea in the Approximation Ex-
ample
From the point of linear algebra, we can decompose the signal into linear
combination of the basis if the signal is in the the space spanned by the
basis. In pp. 364-365 [1], it is,
X
f (t) = ak φk (t). (2.6)
k
where k is an integer index of the finite or infinite sum, the ak are expansion
coefficients, and the φk (t) are expansion functions, or the basis. Compare
to 2.2, we know that Fourier series expansion is a special case when φk (t) =
exp(j2πkt/T ).In general, if we choose the basis appropriately, there exists
another set of basis {φ̃k (t)} such that {φk (t)} and {φ̃k (t)} are orthonormal.
The inner product 1 is
Z
< φi (t), φ̃j (t) >= φi (t)φ̃∗j (t)dt = δij . (2.7)
where {φ̃k (t)} is called the dual function of {φk (t)}. With this orthonormal
property, we can find the coefficients by
Z
< f (t), φ̃k (t) > = f (t)φ̃∗k (t)dt
Z X
= ( ak0 φk0 (t))φ̃∗k (t)dt
k0
X Z
= ak0 ( φk0 (t)φ̃∗k (t)dt)
k0
X
= ak 0 δk 0 k
k0
= ak .
Rewrite as follows:
Z
ak =< f (t), φ̃k (t) >= f (t)φ̃∗k (t)dt. (2.8)
The satisfying result comes because of the orthonormal property of the basis.
Review 2.3, the Fourier series analysis equation is a special case when φ̃k (t) =
1 ∗
R
If x(t) and y(t) are both complex signals, < x(t), y(t)
R >= x(t)y (t)dt. If they are
both real-valued, the inner product can be reduced to (x(t)y(t)dt. The latter is more
well-known.
6
(1/T ) exp(j2πkt/T ). Therefore, it is important to choose an appropriate set
of basis and its dual. For the signal we want to deal with, apply a particular
basis satisfying the orthonormal property on that signal. It is easy to find
the expansion coefficients ak . Fortunately, the coefficients concentrate on
some critical values, while others are close to zero. We can drop the small
coefficients and use more bits to record the important values. This process
can be applied to data compression while preserving the resemblance to the
original signal at the same time. In the example in the previous section,
we can approximate the original by only some critical coefficients. Data
compression can be achieved.
7
2.3 Two Examples about Multiresolutions
2.3.1 Example 1: Approximate discrete-time signals
using delta function
Consider a discrete-time signal x[n] defined as
1
x[n] = ( )|n| . (2.9)
2
Now we choose a basis {φk [n]} = {δ[n − k]} and its dual {φ̃k [n]} = {φk [n]} =
{δ[n−k]} to find the expansion coefficients of x[n]. Note that for the discrete
case, the variable is changed to n and use brackets instead of parentheses,
and the inner product is revised as summations instead of integrations. But
the concept of function expansion and inner product is the same. First, we
check whether the basis satisfies the orthonormal property:
∞
X
< φi [n], φ̃j [n] >=< δ[n − i], δ[n − j] >= δ[n − i]δ[n − j] = δij . (2.10)
n=−∞
8
Figure 2.2: The result of approximation using delta functions. (a) Original
signal
P0 x[n] (b) Obtained coefficients ak (c) ReconstructedP2 signal x1 [n] =
k=−2 ak δ[n − k] (d) Reconstructed signal x2 [n] = k=0 ak δ[n − k]
9
Figure 2.3: Plot of the scaled version of the basis.(a) φn (t) = φ0 (t) (b)
φn (t) = φ1 (t) (c) φn (t) = φ−1 (t) (d) φn (t) = φ2 (t).
10
Here comes another question: is the family {φn (t)} a set of orthogonal
family? Check the first two functions, φ0 (t) and φ1 (t)
1
Z ∞ Z
2 1
< φ0 (t), φ1 (t) >= φ0 (t)φ1 (t)dt = dt = .
−∞ 0 2
Unfortunately, the two functions are not orthogonal to each other, so as the
others. But don’t be frustrated here, we can apply Gram-Schmidt process to
obtain a set of orthonormal basis from existing {φn (t)}. The procedure is
= ψ(t)/2
We can continuously apply this process to extend the basis. But we look
at the first two basis and its Fourier transform. By inspection, φ(t) has
a nonzero mean while ψ(t) is zero in average. The φ(t) has two jumps at
x = 0, 1 while ψ(t) jumps at x = 0, 1/2, 1. Therefore, the power of φ(t)
is more compact at low frequencies while the power of ψ(t) concentrates at
relatively high frequencies. In formal φ(t) is called scaling function to do
approximation and ψ(t) is called wavelet function to find the details.
11
2.4 First Look at Multiresoltuion Analysis
In the translating example, we found a function which is orthonormal to
its translating version. Different scaled version can see different frequency
resolutions. Combined with the two properties, we can construct a basis from
the scaling function and wavelet function with two parameters: scaling and
translating, formally defined in pp. 365-372 [1] as
2. The subspaces spanned by the scaling function at low scales are nested
within those spanned at higher scales: From Fig.2.3, it is clear to show
that φ−1 (t) = φ0 (t) + φ0 (t − 1). In the notations defined above, it’s
Vj ⊂ Vj+1 .
Requirement 2 is worthy to note. For Haar scaling function (in the ex-
ample), it is
1 1
φ(t) = φ0,0 (t) = √ φ1,0 (t) + √ φ1,1 (t).
2 2
12
Figure 2.4: The relationship between scaling and wavelet function spaces.
1 √ 1 √
φ(t) = √ ( 2φ(2t)) + √ ( 2φ(2t − 1)). (2.19)
2 2
In general, this equation is called refinement equation, multiresolution anal-
ysis equation, or dilation equation
X √
φ(t) = hφ [n] 2φ(2t − n). (2.20)
n
√ √
It is evident to show that hφ [n] = {1/ 2, 1/ 2} for Haar scaling functions.
A physical meaning of 2.20 is that φ(2t) is a function with higher frequency
components than φ(t). It is designed properly such that we apply a discrete
low pass filter hφ [n] to have φ(t).
Similar relationship exists for the wavelet functions. It is
X √
ψ(t) = hψ [n] 2φ(2t − n). (2.21)
n
√ √
Again, for Haar wavelets, hψ [n] = {1/ 2, −1/ 2}. These two filters are
related by
hψ [n] = (−1)n hφ [1 − n]. (2.22)
Finally, a diagram shows the relationship between the sets and the func-
tions. V0 is the approximation in scale 0. We can union the set spanned by
different levels of wavelet functions to have higher order of approximation.
13
Figure 2.5: The relationship between scaling and wavelet functions.
Once we union infinite wavelet sets, the sets are equal to the L2 (R) set2 .
This is requirement 4. In mathematical words, it is
2
L2 (R) = {f (x)| |f (x)|2 dx < ∞}
R
14
Chapter 3
3.1 Definition
In the previous chapter, we discuss the scaling function, wavelet function,
and their properties. Suppose the scaling function and wavelet function are
given as Haar, Daubechies,... Say, the basis are known. We can approximate
a discrete signal in l2 (Z)1 by
∞
1 X 1 XX
f [n] = √ Wφ [j0 , k]φj0 ,k [n] + √ Wψ [j, k]ψj,k [n]. (3.1)
M k M j=j0 k
Here f [n], φj0 ,k [n]andψj,k [n] are discrete functions defined in [0, M − 1], to-
tally M points. Because the sets {φj0 ,k [n]}k∈Z and {ψj,k [n]}(j,k)∈Z2 ,j≥j0 are
orthogonal to each other. We can simply take the inner product to obtain
the wavelet coefficients
1 X
Wφ [j0 , k] = √ f [n]φj0 ,k [n]. (3.2)
M n
1 X
Wψ [j, k] = √ f [n]ψj,k [n] j ≥ j0 . (3.3)
M n
3.2 are called approximation coefficients while 3.3 are called detailed co-
efficients.
1 2
P∞
l (Z) = {f [n]| n=−∞ |f [n]|2 < ∞}
15
3.2 The Fast Wavelet Transform
Start from the definition, if the form of scaling and wavelet function is known,
its coefficients is defined in 3.2 and 3.3. If we can find another way to find
the coefficients without knowing the scaling and dilation version of scaling
and wavelet function. The computation time can be reduced. From 2.20, we
have
For the commonly used discrete signal, say, a digital image, the original
data can be viewed as approximation coefficients with order J. That is,
f [n] = Wφ [J, n] By 3.7 and 3.8, next level of approximation and detail can
be obtained. This algorithm is ”fast” because one can find the coefficients
level by level rather than directly using 3.2 and 3.3 to find the coefficients.
This algorithm was first proposed in [2].
16
Figure 3.1: The schematic diagram to realize discrete wavelet transform.
Here the filter names are changed.
X(2) (z) = ...+x(2) [−2]z 2 +x(2) [−1]z 1 +x(2) [0]+x(2) [1]z −1 +x(2) [2]z −2 +....
X(2) (−z) = ...+x(2) [−2]z 2 −x(2) [−1]z 1 +x(2) [0]−x(2) [1]z −1 +x(2) [2]z −2 +....
17
Add the two equations above
X(2) (z)+X(2) (−z) = 2(...+x(2) [−4]z 4 +x(2) [−2]z 2 +x(2) [0]+x(2) [2]z −2 x(2) [4]z −4 +...).
1 1
X(2) (z 2 )+X(2) (−z 2 ) = 2(...+x(2) [−4]z 2 +x(2) [−2]z 1 +x(2) [0]+x(2) [2]z −1 x(2) [4]z −2 +...).
so 1 1
X(2) (z 2 ) + X(2) (−z 2 )
X(4) (z) = .
2
5. Time domain: x(5) [n] = x(3) [2n]
z-domain: 1 1
X(3) (z 2 ) + X(3) (−z 2 )
X(5) (z) = .
2
6. Time domain: x(6) [n] = x(4) [n/2] for n=multiple of 2, zero for others.
z-domain:
∞ ∞ ∞
X X m=n/2 X
X(6) (z) = x(6) [n]z −n = x(4) [n/2]z −n = x(4) [m]z −2m = X(4) (z 2 ).
n=−∞ n=−∞ m=−∞
8. Convolve x(6) [n] with a LTI filter with impulse response h1 [n]. The
z-domain representation is
10. The reconstructed signal x0 [n] is the sum of node 8 and 9. Finally, the
z-domain representation is
18
3.9 gives us some constraint to make the reconstructed signal lossless.
For any signal, we let X(z) = X0 (z). The following two conditions can be
derived under the perfectly reconstruction assumption,
H(z)H1 (z) + G(z)G1 (z) = 2 (3.10)
H(−z)H1 (z) + G(−z)G1 (z) = 0. (3.11)
Rewrite 3.10 and 3.11 into matrix form, we obtain
" #" # " #
H(z) G(z) H1 (z) 2
= . (3.12)
H(−z) G(−z) G1 (z) 0
Solve the linear system, the relationship between the decomposition and
reconstruction filters is
" # " #
H1 (z) 2 G(−z)
= (3.13)
G1 (z) det(Hm ) −H(−z)
where " #
H(z) G(z)
Hm = . (3.14)
H(−z) G(−z)
3.13 shows once the decomposition filters is determined the correspond-
ing reconstruction filters are unique. For most case, Hm is invertible so
det(Hm ) 6= 0. Another sufficient and necessary condition of 3.13 is the
biorthogonal relationship. It consists of four equations:
H(z)H1 (z) + H(−z)H1 (−z) = 2 (3.15)
G(z)G1 (z) + G(−z)G1 (−z) = 2 (3.16)
H(z)G1 (z) + H(−z)G1 (−z) = 0 (3.17)
G(z)H1 (z) + G(−z)H1 (−z) = 0. (3.18)
Rewrite in the time domain, we will clearly see the biorthogonality:
< h[k], h1 [2n − k] > = δ[n] (3.19)
< g[k], g1 [2n − k] > = δ[n] (3.20)
< g1 [k], h[2n − k] > = 0 (3.21)
< h1 [k], g[2n − k] > = 0. (3.22)
The biorthogonality is an important guideline to design the subband coding
algorithm. A regular requirement is to make the four filter FIR filter for
simplicity. From 3.13, if we let det(Hm ) = constant or αz −(2k+1) , H1 (z) and
G1 (z) are in FIR form. The reconstruction filter can be discussed in three
cases:
19
1. det(Hm ) = 2:
2. det(Hm ) = −2:
3. det(Hm ) = 2z −(2k+1) :
A wavelet interpretation of this result is that the set of filters, {h[n], g[n]},
corresponds to an arbitary set of scaling and wavelet functions, {φ(t), ψ(t)},
while another set of filters, {h1 [n], g1 [n]}, maps another set of wavelet func-
tions, {φ̃(t), ψ̃(t)}. These filters are tied with biorthogonal relations. This
result gives us more flexibility to design the filters.
20
3.4 2D Wavelet Transform
Recall the 2D case in the Fourier transform, the basis are modified into
exp(j(ω1 t1 + ω2 t2 )) (3.29)
There are three different wavelet functions, ψ H (x, y), ψ V (x, y) and ψ D (x, y).
Conceptually, the scaling function is the low frequency component of the
previous scaling function in 2 dimensions. Therefore, there is one 2D scaling
function. However, the wavelet function is related to the order to apply the
filters. If the wavelet function is separable, i.e. f (x, y) = f1 (x)f2 (y). These
functions can be easily rewritten as
21
Figure 3.2: Schematic diagram of 2D wavelet transform
22
Figure 3.4: Lena image after wavelet decomposition
In addition to decompose the image in two axis, Yu-Si Zhang [6] intro-
duced a method to decompose the image along the natural edge of the image.
But in general, the structure of Fig.3.4 is used due to the implementation
complexity.
23
Figure 3.5: Reconstructed Lena image. The border is black due to the con-
volution.
24
Chapter 4
4.1 Introduction
In this chapter, we talk about continuous wavelet transform. This transform
works when we use a continuous wavelet function to find the detailed coef-
ficients of a continuous signal. Like the concept in chapter 2, we have to
establish a basis to do such analysis. First, we give the definition of continu-
ous wavelet transform and do some comparison between that and the Fourier
transform.
25
4.2 Definition
We define a mother wavelet function ψ(t) ∈ L2 (R), which is limited in time
domain. That is, ψ(t) has values in a certain range and zeros elsewhere.
Another property of mother wavelet is zero-mean. The other property is
that the mother wavelet is normalized. Mathematically, they are
Z ∞
ψ(t)dt = 0 (4.1)
−∞
Z ∞
2
kψ(t)k = ψ(t)ψ ∗ (t)dt = 1. (4.2)
−∞
As the dilation and translation property states, the mother wavelet can
form a basis set denoted by
n 1 t − u o
ψs,u (t) = √ ψ( ) . (4.3)
s s u∈R,s∈R+
26
4.3 An example of wavelet transform
In the previous section, we assume the mother wavelet is given with some
satisfactory properties. Here we illustrate a famous mother wavelet function,
called Mexican hat wavelet with
2 t2 t2
ψ(t) = √ ( − 1) exp(− ). (4.7)
π 1/4 3σ σ 2 σ2
27
Figure 4.1: Mexican-hat wavelet for σ = 1 and its Fourier transform.
28
4.4 Comparison Among the Fourier Trans-
form, Short-time Fourier Transform(STFT)
and Wavelet Transform
In this section, we talk about features among the three transforms.
STFT
Z ∞
Sf (u, ξ) = f (t)w(t − u) exp(−jξt)dt. (4.10)
−∞
Wavelet transform
∞
t−u
Z
1
W f (s, u) = f (t) √ ψ ∗ ( )dt. (4.11)
−∞ s s
Wavelet transform overcomes the previous problem. The wavelet function
is designed to strike a balance between time domain (finite length) and fre-
quency domain (finite bandwidth). As we dilate and translate the mother
wavelet, we can see very low frequency components at large s while very high
frequency component can be located precisely at small s.
29
4.4.2 Inverse transform
Fourier transform
Z ∞
1
f (t) = F (ω) exp(jωt)dt. (4.12)
2π −∞
STFT
Z ∞ Z ∞
1
f (t) = Sf (u, ξ)w(t − u) exp(jξt)dξdu. (4.13)
2π −∞ −∞
Wavelet transform
∞ ∞
t−u
Z Z
1 1 ds
f (t) = W f (s, u) √ ψ( )du 2 , (4.14)
Cψ 0 −∞ s s s
Z ∞
|Ψ(ω)|2
Cψ = dω < ∞. (4.15)
0 ω
4.4.3 Basis
Here we discuss the different basis for each transforms. Fig. 4.4.2 shows the
results.
Fourier transform
Complex exponential function with different frequencies:
exp(jωt). (4.16)
STFT
Truncated or windowed complex exponential function:
Wavelet transform
Scaled and translated version of mother wavelets:
1 t−u
√ ψ( ). (4.18)
s s
30
Figure 4.2: Different basis for the transforms. (a) Real part of the basis for
Fourier transform, exp(jπt). (b) Basis for different frequency, exp(j4πt). (c)
Basis for STFT, using Gaussian window of σ = 1. It is exp(−t2 /2) exp(jπt).
(d) Basis for different frequency, exp(−t2 /2) exp(j4πt). (e) Mexican-hat
mother wavelet function and (f ) s = 4.
31
Figure 4.3: Different time-frequency tile allocation of the three transforms:
(a) Fourier transform, (a) STFT and (a) wavelet transform.
32
4.4.4 Time-frequency tiling
In quantum physics, the Heisenberg uncertainty principle states that cer-
tain pairs of physical properties, like position and momentum, cannot both
be known to arbitrary precision. The same principle holds in signal pro-
cessing. We cannot locate both time and frequency very precisely. The
product of variation in time and variation in frequency is greater than 1/2,
i.e. σt σω ≥ 1/2. It can be viewed as a rectangle with constant area and
different transform adjusts the width and height of the rectangle. The three
transforms are illustrated in Fig. 4.4.4.
Fourier transform
The time information is completely lost. Frequency axis is divided uniformly.
Frequency resolution can be very precise if we integrate along the whole time
axis.
STFT
Add a window to take the time domain information into consideration. The
frequency resolution depends on the time resolution, or the size of the win-
dow. We cannot zoom in a particular frequency range because the box is
uniformly placed.
Wavelet transform
The s parameter is inversely proportional to the frequency. As we see, if we
want to focus on low frequencies, larger s is used while higher frequencies
uses small s. This flexibility increases the time-frequency analysis.
33
Chapter 5
34
form of the discrete filters. If we substitute the recursive equation continu-
ously, we obtain
∞
Y Hφ (ω)
Φ(ω) = √ Φ(0). (5.5)
p=1
2
By (5.5), once we have the lowpass filter, we can calculate the correspond-
ing scaling function by its Fourier transform. If we know the scaling function
in advance, 5.3 helps us to find the lowpass filter.
and
Hψ (ω)Hφ∗ (ω) + Hψ (ω + π)Hφ∗ (ω + π) = 0. (5.9)
These equations might not strange to you because similar equations appear
in (3.15), using subband coding approach. We usually set
35
5.1.4 Design lowpass filter
Here we have all the design equations and we can focus on the design of the
lowpass filter. Note that we only have to design the frequency band ω ∈
[0, π/2]. the value in [π/2, 3π/2] can be obtained using (5.6) and [3π/2, 2π]
band is the reflection version due to the real filter property.
36
5.2 Some Design Issues
Besides the above design equations, we want other properties which are useful
in the general case. Here, we mention the vanishing moments, the size of
support, support versus moments, and regularity. These properties are more
strongly related to the practical use of these functions.
The definition seems not useful in design of the wavelet functions because
it involves a continuous integration. A theorem (pp. 284 in [8]) shows the
four statements are equivalent if |ψ| = O((1 + t2 )−p/2−1 ) and |φ(t)| = O((1 +
t2 )−p/2−1 ),
1. The wavelet ψ has p vanishing moments.
2. The Fourier transform of φ(t), ψ(t) and its first p − 1 derivatives are
zero at ω = 0.
3. The Fourier transform of hφ [n], Hφ (ejω ) and its first p − 1 derivatives
are zero at ω = π.
4. For any 0 ≤ k < p,
∞
X
qk (t) = nk φ(t − n) (5.16)
n=−∞
is a polynomial of degree k.
37
If we combine statement 3 with the conjugate mirror filter condition (5.6),
we can write the lowpass filter as
jω
√ 1 + ejω p
Hφ (e ) = 2( ) L(ejω ) (5.17)
2
Where L(x) is a polynomial. This result simplifies the process of wavelet
design.
5.2.4 Regularity
The regularity of ψ has significant influence on the error introduced by thresh-
olding or quantizing the wavelet coefficients. When reconstructing a signal
from wavelet coefficients
∞
X ∞
X
f= < f, ψj,n > ψj,n , (5.18)
j=−∞ n=−∞
38
5.3 Daubechies wavelets
Based on these equation, Daubechies [9], designed a type of wavelet for a
given vanishing moment p and find the minimum size discrete filter. The
conclusion is that if we want the wavelet function with p vanishing moments,
the minimum filter size is 2p. The derivation first starts from 5.17, rewrite
as
√ 1 + e−jω p
Hφ (ejω ) = 2( ) R(ejω ). (5.19)
2
The absolute-square of this function is
|Hφ (ejω )|2 = Hφ (ejω )Hφ∗ (ejω )
1 + e−jω 1 + ejω p
= 2( ) R(ejω )R∗ (ejω )
2 2
2 + ejω + e−jω 2p
= 2( ) |R(ejω )|2
4
ω 2p ω
= 2(cos ) P (sin2 ). (5.20)
2 2
The last step makes P (sin2 ω2 ) = |R(ejω )|2 . Recall (5.6), we can determine
the form of P (x). Let y = sin2 ω2 . We have
(1 − y)p P (y) + y p P (1 − y) = 1. (5.21)
A theorem in algebra, called Bezout theorem, can solve this equation. The
unique solution is
p−1
!
X p−1+k
P (y) = yk . (5.22)
k=0
k
The polynomial P (y) is the minimum degree polynomial satisfying (5.21).
Once we have P (y), the polynomial R(ejω ) can be derived. First we decom-
pose R(ejω ) according to its roots
m
X m
Y
−jkω
jω
R(e ) = rk e = r0 (1 − ak e−jω ). (5.23)
k=0 k=0
39
Taking p = 2 for a example. The obtained polynomial P (y) is
1
!
X 1+k
P (y) = y k = 1 + 2y. (5.25)
k=0
k
2 − z − z −1 1 1
P( ) = 2 − z − z −1 . (5.26)
4 2 2
√ √
The roots are 2 + 3 and 2 − 3 . After factorization, we have the lowpass
filter to be
√ √ √ √ √ √ √ √
jω 2 + 6 3 2 + 6 −jω 3 2 − 6 −j2ω 2 − 6 −j3ω
Hφ (e ) = + e + e + e .
8 8 8 8
(5.27)
The discrete-time domain representation is
√ √ √ √ √ √ √ √
2+ 6 3 2+ 6 3 2− 6 2− 6
hφ [n] = δ[n]+ δ[n−1]+ δ[n−2]+ δ[n−3].
8 8 8 8
(5.28)
The result is the minimum size filter with 2 vanishing moments and the
corresponding filter size is 4. Recall the conclusion mentioned above, the filter
size is two times the vanishing moment. Higher order Daubechies wavelets
are derived at similar way. The coefficient and the plot are shown in the
appendix A.
5.4 Symlets
Take a look at the discrete filters and the scaling/wavelet functions of Daubeches
wavelets. These functions are far from symmetry. That’s because Daubechies
wavelets select the minimum phase square root such that the energy concen-
trates near the starting point of their support. Symmlets select other set of
roots to have closer symmetry but with linear complex phase. The coeffi-
cients of these filters are list in the Appendix B.
5.5 Coiflets
For an application in numerical analysis, Coifman asked Daubechies [9] to
construct a family of wavelets ψ that have p vanishing moments, minimum-
size support and Z ∞
φ(t)dt = 1, (5.29)
−∞
40
Z ∞
tk φ(t)dt = 0 for 1 ≤ k < p. (5.30)
−∞
The equation above can be taken as some requirement about vanishing mo-
ments of the scaling function. The resulting coiflets has a support of size
3p − 1. These coefficient of the filters are shown in Appendix C.
41
Chapter 6
Applications of Wavelet
Transform
6.1 Introduction
In the previous sections, we mention the fundamental concepts of the wavelet
transform. More than a theory, the wavelets are widely used in many appli-
cations. Its ability of multiresolution outperforms Fourier-based transform
such as discrete cosine transform in terms of compression and quality. Here,
we talk more about JPEG200, which is wavelet-based, and some applications
of the wavelet transform.
42
Figure 6.1: The simple system diagram of JPEG
43
Figure 6.2: Comparison between JPEG and JPEG 2000. CR stands for
compression ration and RMSE means root mean square error.
convolution.
An overview of the JPEG 2000 system can be found in [12], [13].
44
6.3.2 Pattern recognition
Pattern recognition depends on computing the features of the data. If we
first condense the signal by the wavelet transform, the computation effort
can be reduced. Once the pattern is probably recognized, we can use wavelet
transform to do finer analysis.
45
Chapter 7
Conclusions
In this tutorial, we explore the world of wavelets. From the abstract idea
in approximation, multiresolition theory, we generalize the Fourier transform
and start the journey of wavelets. The discrete wavelet is more useful in real-
ization. We often use 2D wavelets to do image compression. The continuous
wavelet transform analyze the continuous-time signal in a different perspec-
tive. By the advantage of multiresolution, we can locate time and frequency
more accurately. The wavelet design is more complicated in mathematics but
the design procedure completes the existence of the wavelets. The application
chapter mentions the nowadays JPEG and JPEG2000 standard.
[8] and [9] give a thorough approach to the wavelets but in purely math-
ematically words. [1] and [15] illustrate some examples on the wavelet and
then the abstract concepts. Other references mainly focus on some parts of
the materials.
The wavelets bring us to a new vision of signal processing. It tactically
avoids the problem that Fourier analysis encounters. Its implementation
is simple. We need some designed filters to do the task. Although the
implementation is quite easy, the filter design includes lots of mathematical
originality and this is the research topic. Once we establish wavelets with
more ideal properties, lots of difficult problems might be solved.
46
Appendix A
Daubechies wavelets
A.1 Coefficients
A.1.1 p=2
1
A.1.2 p=3
47
A.1.3 p=4
A.1.4 p=5
48
A.1.5 p=6
A.1.6 p=7
49
A.1.7 p=8
50
A.1.8 p=9
51
A.1.9 p=10
52
A.2.1 p=2
A.2.2 p=3
53
A.2.3 p=4
A.2.4 p=5
54
A.2.5 p=6
A.2.6 p=7
55
A.2.7 p=8
A.2.8 p=9
56
A.2.9 p=10
57
Appendix B
Symlets
B.1 Coefficients
B.1.1 p=2
B.1.2 p=3
58
B.1.3 p=4
B.1.4 p=5
59
B.2 Function Plot
B.2.1 p=2
B.2.2 p=3
60
B.2.3 p=4
B.2.4 p=5
61
Appendix C
Coiflets
C.1 Coefficients
C.1.1 p=1
62
C.1.2 p=2
63
C.1.3 p=3
64
C.1.4 p=4
65
C.1.5 p=5
66
C.2 Function Plot
C.2.1 p=1
C.2.2 p=2
67
C.2.3 p=3
C.2.4 p=4
68
C.2.5 p=5
69
References
[4] A. Said and W. A. Pearlman, “A new, fast, and efficient image codec
based on set partitioning in hierarchical trees,” vol. 6, pp. 243–250, June
1996.
[8] S. Mallat, A Wavelet Tour of Signal Processing, 3rd ed., Third Edition:
The Sparse Way. Academic Press, 3 ed., December 2008.
70
[11] N.-C. Shen, “Sectioned convolution for discrete wavelet transform,”
Master’s thesis, 2008.
[12] M. Rabbani and R. Joshi, “An overview of the jpeg 2000 still image
compression standard,” pp. 3–48, Elsevier, January 2002.
71