ITC Module - I
ITC Module - I
Units
WHY LOG OF PROBABILITY?
Monotonically decreasing function of
probability
If a particular event has a 100% probability of
occurring,- then its self-information is {log(1)=0}:
its occurrence is "perfectly non-surprising" and
yields no information.
If a particular event has a 0% probability of
occurring, then its self-information is {log(0)=α }: -
its occurrence is "infinitely surprising."
TO CALCULATE LOG2 FROM ANOTHER LOG
BASE
ENTROPY- AVERAGE INFORMATION
Find entropy of X.
PROBLEM 2
Suppose we have a horse race with eight horses
taking part. Assume that the probabilities of
winning for the eight horses are( ½ ¼,1/8,
1/16,1/64, 1/64,1/64,1/64). Calculate the entropy
of horse race .
Ans = 2 bits
PROPERTIES OF ENTROPY
H(X) ≥0
The entropy is a continuous function with respect to each
variable pi because it is a sum of continuous functions.
Additivity: self information is additive and so it is H(X),
which represents the mean value of self information.
Entropy is maximum when the symbols have equal
probabilities; this maximum value is also called
decision quantity, D(X): H max ( X) =: D( X) = log 2 m
Symmetry: H(X) is unchanged if the events xi are
reordered.
The maximum value of the entropy s obtained for
equiprobabile messages.
Hb(X) = logb(a) Ha(X)
MINIMUM AVERAGE NUMBER OF BITS IS
MESSAGES WITH EQUAL PROBABILITY
Consider a discrete binary source – emits 0 or 1-
a sequence -statistically independent.
The output is “0” with probability “ p” or 1 with a
probability 1 – p.
The entropy of this binary source is
H ( X )
i
p i log p i
= – p log2(p) – (1 – p) log2(1 – p)
=H(p) ( def)
MESSAGE WITH EQUAL PROBABILITY FOR
BINARY SOURCE.
Ans=1.875 bits ms
PROBLEM 4
An analog signal band limited to 10 kHz is
quantized in 8 levels of a PCM system with
probabilities of ¼, 1/5, 1/5, 1/10, 1/10, 1/20, 1/20,
1/20 respectively. Find the entropy and the rate
of information.
REMEMBER
Information increases Entropy
SOURCE ENTROPY
The minimum average number of binary digits
needed to specify a source output (message) uniquely
is called “SOURCE ENTROPY”
JOINT ENTROPY
measures how much entropy is contained in a
joint system of two random variables.
If the random variables are X and Y, the joint
entropy - H(X,Y).
measured in bits, nats, or Hartleys depending on
the base of the logarithm.
JOINT ENTROPY
The Joint Entropy of a pair of discrete random
variables (X, Y) with a joint distribution
p(x, y ) is given by
PROPERTIES OF JOINT ENTROPY
Nonnegative, i.e., H(X,Y)≥0
greater than individual entropies
I(X;Y) ≥0
Mutual information can be expressed in terms of
entropy of the channel output.
Mutual information is additive for independent
variables
I(X,W;Y,Z) =I(X;Y) +I(W;Z)
If X and Y are independent I(X;Y) =0
DISCRETE MEMORY LESS SOURCE
A source from which the data is being emitted at
successive intervals,
w independent of previous values
Efficiency of a code :
H (s)
L
Usually expressed in % (multiplied by 100)
Redundancy R= 1-η
PROBLEMS
1. A scanner converts a black and white document,
line-by-line into binary data for transmission.
The scanner produces source data comprising
symbols representing runs of up to six similar
image pixel elements with the probabilities as
shown below:
No. of 1 2 3 4 5 6
consecutive
pixels
Answers
H(s) = 2.29 bits
0111110010110101001111100100
Now
|0|111|110|0|1011|0|1010|0|111
|110|0|100
|0|111|110|0|1011|0|1010|0|111|110|0|100
a b r a c a d a b r a !
PREFIX CODE- PROPERTY
The encoding scheme to reduce the number of
bits in a message, always use short encodings for
frequently used characters, and long encodings
for infrequent ones.
A second fundamental property of prefix codes is
that messages can be formed by simply stringing
together the code bits from left to right
VENN DIAGRAM REPRESENTATION
HUFFMAN CODING
Invented by David A Huffman as a part of his
class assignment.
Idea : to assign variable-length codes to input
characters,-lengths of the assigned codes are
based on the frequencies of corresponding
characters
Prefix code
H(s) L
By Kraft in 1949
ni
D 1