Chapter 8 of the Multimedia Systems course discusses loss compression algorithms, focusing on distortion measures, rate distortion theory, quantization, and transform coding. It highlights the importance of perceptual distortion in lossy compression and introduces techniques such as uniform and nonuniform scalar quantization, vector quantization, and discrete cosine transform (DCT). The chapter concludes with the Karhunen-Loeve Transform (KLT) as a method for optimal decorrelation of input data.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0 ratings0% found this document useful (0 votes)
9 views13 pages
Chapter 8
Chapter 8 of the Multimedia Systems course discusses loss compression algorithms, focusing on distortion measures, rate distortion theory, quantization, and transform coding. It highlights the importance of perceptual distortion in lossy compression and introduces techniques such as uniform and nonuniform scalar quantization, vector quantization, and discrete cosine transform (DCT). The chapter concludes with the Karhunen-Loeve Transform (KLT) as a method for optimal decorrelation of input data.
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 13
Course Title
Multimedia Systems CoSc4072
Chapter 8 Loss Compression Algorithms
By: Ins. Solomon S.
Outlines Distortion Measures The Rate Distortion Theory Quantization Transform Coding Introduction The compression ratio for image data using lossless compression techniques (e.g., Huffman Coding, Arithmetic Coding, LZW) is low when the image histogram is relatively flat. For image compression in multimedia applications, where a higher compression ratio is required, lossy methods are usually adopted. In lossy compression, the compressed image is usually not the same as the original image but is meant to form a close approximation to the original image perceptually. To quantitatively describe how close the approximation is to the original data, some form of distortion measure is required. Distortion Measures A distortion measure is a mathematical quantity that specifies how close an approximation is to its original, using some distortion criteria. When looking at compressed data, it is natural to think of the distortion in terms of the numerical difference between the original data and the reconstructed data. However, when the data to be compressed is an image, such a measure may not yield the intended result. For example, if the reconstructed image is the same as original image except that it is shifted to the right by one vertical scan line, an average human observer would have a hard time distinguishing it from the original and would therefore conclude that the distortion is small. However, when the calculation is carried out numerically, we find a large distortion, because of the large changes in individual pixels of the reconstructed image. The problem is that we need a measure of perceptual distortion. The rate distortion theory Rate–distortion theory is a major branch of information theory which provides the theoretical foundations for lossy data compression; It addresses the problem of determining the minimal number of bits per symbol, as measured by the rate R, that should be communicated over a channel, so that the source (input signal) can be approximately reconstructed at the receiver (output signal) without exceeding an expected distortion D. The tradeoff between rate and distortion is represented in the form of a rate-distortion function R(D).
It is easy to see that when D = 0, we have a lossless compression of the source.
Quantization Quantization in some form is the heart of any lossy scheme. Without quantization, we would indeed be losing little information. The source we are interested in compressing may contain a large number of distinct output values (or even infinite, if analog). To efficiently represent the source output, we have to reduce the number of distinct values to a much smaller set, via quantization. Uniform Scalar Quantization
A uniform scalar quantizer partitions the domain of input values into
equally spaced intervals, except possibly at the two outer intervals. The endpoints of partition intervals are called the quantizer's decision boundaries. Nonuniform Scalar Quantization If the input source is not uniformly distributed, a uniform quantizer may be inefficient. Increasing the number of decision levels within the region where the source is densely distributed can effectively lower granular distortion. In addition, without having to increase the total number of decision levels, we can enlarge the region in which the source is sparsely distributed. Such nonuniform quantizers thus have nonuniformly defined decision boundaries. Vector Quantization One of the fundamental ideas in Sharmon's original work on inflation theory is that any compression system performs better if it operates on vectors or groups of samples rather than on individual symbols or samples. We can form vectors of input samples by concatenating a number of consecutive samples into a single vector. For example, an input vector might be a segment of a speech sample, a group of consecutive pixels in an image, or a chunk of data in any other format. Transform Coding
From basic principles of information theory, we know that coding vectors
is more efficient than coding scalars. To implement such intention, we need to group blocks of consecutive samples from the source input into vectors. Discrete Cosine Transform (DCT)
The Discrete Cosine Transform (DCT), a widely used transform coding
technique, is able to perform decorrelation of the input signal in a data- independent manner. Because of this, it has gained tremendous popularity. Karhunen-Loeve Transform
The Karhunen-Loeve Transfonn (KLT) is a reversible linear transform that
exploits the statistical properties of the vector representation. Its primary property is that it optimally decorrelates the input. To do so, it fits an n-dimensional ellipsoid around the (mean-subtracted) data. The main ellipsoid axis is the major direction of change in the data End of Chapter 8