Entropy
Entropy
Compression
Myung-Sin Song1
1 Introduction
In the process of wavelet image compression, there are three major steps
that makes the compression possible, namely, decomposition, quanti-
zation and entropy encoding steps. While quantization may be a lossy
step where some quantity of data may be lost and may not be re-
covered, entropy encoding enables a lossless compression that further
compresses the data. [13], [18], [5]
In this paper we discuss various entropy encoding schemes that are
used by engineers (in various applications).
section 3.3), Huffman coding (see section 3.4), Kolmogorov entropy (see
section 3.2) and arithmetic coding (see section 3.5) are ones that are
used by engineers. Here, optimal means it uses least memory space to
represent the data. i.e., instead of using 16 bits, use 11 bits. Thus, the
best basis found would make it possible to represent the digital image
with less storage memory. In addition, the choices made for entropy
encoding varies; one might take into account the effectiveness of the
coding and the degree of difficulty of implementation step into pro-
gramming codes. We will also discuss how those preferences are made
in section 3.
zG
{
y GGGGp
GGGGGp
Gi GGl j
G{ x¡ Gk GGGGp
· · · ⊂ V−1 ⊂ V0 ⊂ V1 ⊂ V2 ⊂ · · · ⊂ Vn ⊂ Vn+1 ⊂ · · ·
^ _
Vn = Hu , and Vn = H.
n n
Scaling Operator
ϕ ψ
W0
... 2
S0S1 S0S1 S1
... S0 S0
Fig. 3. Multiresolution. l2 (Z)-version (discrete); ϕ ∈ V0 , ψ ∈ W0 .
· · · ⊂ V−1 ⊂ V0 ⊂ V1 ⊂ · · · , V0 + W0 = V1
The word “multiresolution” suggests a connection to optics from
physics. So that should have been a hint to mathematicians to take a
closer look at trends in signal and image processing! Moreover, even
staying within mathematics, it turns out that as a general notion this
same idea of a “multiresolution” has long roots in mathematics, even
in such modern and pure areas as operator theory and Hilbert-space
geometry. Looking even closer at these interconnections, we can now
recognize scales of subspaces (so-called multiresolutions) in classical
algorithmic construction of orthogonal bases in inner-product spaces,
now taught in lots of mathematics courses under the name of the Gram–
Schmidt algorithm. Indeed, a closer look at good old Gram–Schmidt
reveals that it is a matrix algorithm, Hence new mathematical tools
involving non-commutativity!
Entropy Encoding in Wavelet Image Compression 5
Hilbert space which form frames (see [10]). In some cases parts of the
frame vectors fuse (fusion-frames) onto closed subspaces, and we will
be working with the corresponding family of (orthogonal) projections.
Either way, we arrive at a family of selfadjoint positive semidefinite
operators in Hilbert space. The particular Hilbert space depends on
the application at hand. While the Spectral Theorem does allow us to
diagonalize these operators, the direct application the Spectral Theo-
rem may lead to continuous spectrum which is not directly useful in
computations, or it may not be computable by recursive algorithms.
The questions we address are optimality of approximation in a va-
riety of ONBs, and the choice of the “best” ONB. Here “best” is given
two precise meanings: (1) In the computation of a sequence of approxi-
mations to the frame vectors, the error terms must be smallest possible;
and similarly (2) we wish to minimize the corresponding sequence of en-
tropy numbers (referring to von Neumann’s entropy). In two theorems
we make precise an operator theoretic Karhunen-Loève basis, which
we show is optimal both in regards to criteria (1) and (2). But before
we prove our theorems, we give the two problems an operator theo-
retic formulation; and in fact our theorems are stated in this operator
theoretic context.
2 How it works
In most images their neighboring pixels are correlated and thus contain
redundant information. Our task is to to find less correlated representa-
tion of the image, then perform redundancy reduction and irrelavancy
reduction. Redundancy reduction removes duplication from the signal
source (for instance a digital image). Irrelavancy reduction omits parts
of the signal that will not be noticed by the Human Visual System
(HVS).
Entropy encoding further compresses the quantized values in lossless
manner which gives better compression in overall. It uses a model to
accurately determine the probabilities for each quantized value and
produces an appropriate code based on these probabilities so that the
resultant output code stream will be smaller than the input stream.
Some Terminology
S0SH
S H
S0SV S0SD
S V S D
(
0 if |x| ≤ λ
Thard (x) = (6)
x if |x| > λ
One might think that the quantization step suffices for compression. It
is true that the quantization does compress the data tremendously. Af-
ter the quantization step many of the pixel values are either eliminated
or replaced with other suitable values. However, those pixel values are
still represented with either 8 or 16 bits. See 1.1. So we aim to minimize
the number of bits used by means of entropy encoding. Karhunen-Loève
transform or PCAs makes it possible to represent each pixel on the digi-
tal image with the least bit representation accoding to their probability
thus yields the lossless optimized representation using least amount of
memory.
The Algorithm
with Z b
xαi = ψi∗ (ξ)f α (ξ)dξ. (12)
a
14 Myung-Sin Song
in definition of ONB.
Let H = L2 (a, b). ψi : H → l 2 (Z) and U : l 2 (Z) → l 2 (Z) where U is
a unitary operator
Note that the distance is invariant under a unitary transformation.
Thus, using another coordinate system {φj } in place of {ψi }, would
not change the distance.
Let {φj }, j = 1, 2, ..., be another set of ONB functions instead of
{ψi (ξ)}, i = 1, 2, ...,. Let yjα be the component of f α in {φj } where it can
be expressed in terms of xαi by a linear relation yjα = ∞ α
P
P∞ i=1 hφj , ψi ixi =
α
i=1 Ui,j xi where U : l 2 (Z) → l 2 (Z), U is a unitary operator matrix
Rb ∗
Ui,j = hφj , ψi i = a φj (ξ)ψi (ξ)dξ Also, xαi can be written in terms of
yjα under the following relation xαi = ∞
P α
P∞ −1 α
j=1 hψi , φj iyj = j=1 Ui,j yj
−1 ∗
where Ui,j = Ui,j and Ui,j = Uj,i
∞
X X
f α (ξ) = xαi (ξ)ψi (ξ) = yiα (ξ)φi (ξ)
i=1
P∞ α
P∞ α
So U (xi ) = (yi ) and i=1 xi ψi (ξ) = j=1 yj φj (ξ)
Z b
xαi =< ψi , f α >= ψi (ξ)f (α) (ξ)dξ
a
(α)
The squared magnitude |xi |2
of the coefficient for ψi in the expan-
sion of f (α) can be considered as a good measure of the average in the
ensemble n
(α)
X
Qi = w(α) |xi |2
α=1
can be considered as the measure of importance of {ψi }.
X
Qi ≥ 0, Qi = 1
i
Entropy Encoding in Wavelet Image Compression 15
0 · · · λk
16 Myung-Sin Song
Implementation
Acknowledgement
The author would like to thank Professor Palle Jorgensen, the members
of WashU Wavelet Seminar, Professors David Larson, Gestur Olafsson,
18 Myung-Sin Song
References
1. Ash RB (1990) Information theory. Corrected reprint of the 1965 original.
Dover Publications, Inc., New York
2. Bell TC, Cleary JG, Witten IH (1990) Text Compression. Prentice Hall,
Englewood Cliffs
3. Cohen A, Dahmen W, Daubechies I, DeVore R (2001) Tree Approxima-
tion and Optimal Encoding. Applied Computational Harmonic Analysis
11:192–226
4. Daubechies I (1992) Ten Lectures on Wavelets. SIAM
5. Donoho DL, Vetterli M, DeVore RA, Daubechies I (Oct. 1998) Data Com-
pression and Harmonic Analysis. IEEE Trans. Inf. Theory, 44 (6):2435–
2476
6. Dirac PAM (1947) The Principles of Quantum Mechanics. 3d ed Oxford,
at the Clarendon Press
7. Effros M, Feng H, Zeger K (Aug. 2004) Suboptimality of the Karhunen-
Loève Transform for Transform Coding. IEEE Trans. Inf. Theory, 50
(8):1605–1619
8. Field DJ (1999) Wavelets, vision and the statistics of natural scenes. Phil.
Trans. R. Soc. Lond. A 357:2527–2542
9. Gonzalez RC, Woods RE, Eddins SL (2004) Digital Image Processing
Using MATLAB. Prentice Hall
10. Jorgensen PET (2006) Analysis and Probability Wavelets, Signals, Frac-
tals. Springer, Berlin Heidelberg New York
11. Jorgensen PET, Song M-S (2007) Entropy Encoding using Hilbert Space
and Karhunen-Loève Transforms. preprint
12. Pierce JR (1980) An Introduction to Information Theory Symbols, Sig-
nals and Noise. 2nd Edition Dover Publications, Inc., New York
13. Schwab C, Todor RA (2006) Karhunen-Loève approximation of random
fields by generalized fast multipole methods. Journal of Computational
Physics 217, Elsevier
14. Skodras A, Christopoulos C, Ebrahimi T (Sept. 2001) JPEG 2000 Still
Image Compression Standard. IEEE Signal processing Magazine 18:36–
58
15. Shannon CE, Weaver W (1998) The Mathematical Theory of Communi-
cation. University of Illinois Press, Urbana and Chicago
16. Smith LI (2002) A Tutorial on Principal Components Analysis.
http://csnet.otago.ac.nz/cosc453/student tutorials/principal components.pdf
17. Song M-S (2005) Wavelet image compression. Ph.D. thesis The University
of Iowa, Iowa
Entropy Encoding in Wavelet Image Compression 19
18. Song M-S (2006) Wavelet image compression. Operator theory, operator
algebras, and applications Contemp. Math. 414:41–73, Amer. Math. Soc.,
Providence, RI
19. Usevitch BE (Sept. 2001) A Tutorial on Modern Lossy Wavelet Image
Compression: Foundations of JPEG 2000. IEEE Signal processing Mag-
azine 18:22–35
20. Walker JS (1999) A Primer on Wavelets and their Scientific Applications.
Chapman & Hall, CRC
21. Walnut DF (2002) An Introduction to Wavelet Analysis. Birkhäuser
22. Watanabe S (1965) Karhunen-Loève Expansion and Factor Analysis The-
oretical Remarks and Applications Transactions of the Fourth Prague
Conference on Information Theory Statistical Decision Functions Ran-
dom Process. Adademia Press
23. Witten IH, Neal RM, Cleary JG, (June 1987) Arithmetic Coding for Data
Compression. Communications of the ACM 30 (6):520–540