Chapter 1 (A)
Chapter 1 (A)
1
Chapter 1
Information Sources
&
Source Coding
2
Reading Material
3
Part a
• Introduction
• Information Sources
• Uncertainty and Information
• Amount of Information
• Average Information and Entropy
4
Introduction
The information theory came into existence in 1948,
with the publication of Claude Shannon’s paper
"A mathematical theory of communication”
6
Communication System Model
Coding
Information Source Channel
source coding coding
Information
channel
Source Channel
Destination
decoding decoding
Decoding
7
Information Theory ??
8
Introduction
The purpose of a communication system is to transmit information
from one point to another with high efficiency and reliability.
Information theory provides a quantitative measure of the
information contained in message signals and allows us to
determine the capacity of a communication system to transfer this
information from source to destination.
Through the use of coding, redundancy can be reduced from
message signals so that channels can be used with improved
efficiency. This is called Source Coding.
In addition, systematic redundancy can be introduced into the
transmitted signal so that channels can be used with improved
reliability. This is called Channel Coding or Error Control
Coding. 9
Introduction
• Information theory applies the laws of
probability theory, and mathematics in general,
to study the collection and processing of
information.
10
Introduction
In particular, Information Theory provides answers to
the following two fundamental questions:
What is the minimum number of bits per entropy of a source
symbol required to fully represent the
source? capacity of a channel
What is the ultimate transmission rate for
11
Information Sources
An information source is an object that produces
an event, the outcome of which is random and in
accordance with some probability distribution.
A practical information source in a communication system is a
device that produces messages. It can be either analogue or digital.
Here, we shall deal mainly with the discrete sources, since the
analogue sources can be transformed to discrete sources through
the use of sampling and quantisation techniques.
14
Uncertainty and Information
• Suppose that a probabilistic experiment involves the
observation of the output emitted by a discrete memory-less
source during every unit of time (signalling interval).
• The source output is modelled as a discrete random variable
S which takes on K-symbols from a fixed finite alphabet
I(sk) = log(1/pk)
It is the standard practice to use a logarithm to base 2. The resulting unit of
information is called the bit.
I(sk) = log2(1/pk)
= -log2(pk) for k = 0,1, 2, …., K-1
• When pk=1/2, we have I(sk)= 1 bit. Hence, one bit is the amount of
information that we gain when one of two possible and equally likely
events occurs.
19
Amount of Information
• Note that the information-theoretic concept of “bit” is quite
unrelated to the computer science usage of the term “bit” .
Solution
Let the event S=sk denote the emission of symbol sk by the source.
Hence, I(sk) = log2(1/pk) bits
I(s0) = log2(1/0.4) = 1.322 bits
I(s1) = log2(1/0.3) = 1.737 bits
I(s2) = log2(1/0.2) = 2.322 bits
I(s3) = log2(1/0.1) = 3.322 bits
22
Average Information and Entropy
• Messages produced by information sources consist of sequences of symbols. While
the receiver of a message may interpret the entire message as a single unit,
communication systems often have to deal with individual symbols.
For example, if we are sending messages in English language, the user at the receiving end is
interested mainly in words, phrases and sentences, whereas the communication system has to
deal with individual letters or symbols.
• Hence it is desired to know the average information content per source symbol, known
also as entropy, (H).
H = E[I(sk)]
= pkI(sk)
23
= pklog2(1/pk)
Average Information and Entropy
The quantity H is called the entropy of a discrete memory-less
source. It is a measure of the average information content per
source symbol. It may be noted that the entropy H depends on the
probabilities of the symbols in the alphabet of the source.
i.2 2 bits
i
i 1
25
Properties of Entropy
For a discrete memory-less source with a fixed alphabet:
• H=0, if and only if the probability pk=1 for some k,
and the remaining probabilities in the set are all zero.
This lower bound on the entropy corresponds to ‘no
uncertainty’.
• H=log2(K), if and only if pk=1/K for all k (i.e. all the
symbols in the alphabet are equiprobable). This upper
bound on the entropy corresponds to ‘maximum
uncertainty’.
Hence,
0 H log2(K)
K is the radix (number of symbols) of the alphabet S of the source.
26
Properties of Entropy
H log2(K)
29
Properties of Entropy
• In Case I, it is very easy to guess whether the message s 0 with a
probability =0.01 will occur or the message s1 with probability =0.99
will occur.(Most of the time message s1 will occur). Thus in this case,
the uncertainty is less.
• In Case II, it is somewhat difficult to guess whether s 0 will occur or s1
will occur as their probabilities are nearly equal. Thus in this case,
the uncertainty is more.
• In Case III, it is extremely difficult to guess whether s 0 or s1 will
occur, as their probabilities are equal. Thus in this case, the
uncertainty is maximum.