Ece458 L6
Ece458 L6
[email protected]
2
TODAY’S TOPICS
Joint entropy and Conditional Entropy
Mutual Information
[email protected]
Channel Capacity
3
JOINT ENTROPY AND CONDITIONAL ENTROPY
[email protected]
Definition: Conditional entropy of Y given a random variable X
(average over X) is:
4
Theorem: Chain rule
[email protected]
Corollary
5
MUTUAL INFORMATION
[email protected]
H(x|y) I(x,y)
H(y|x)
H(y)
H(x)
6
MUTUAL INFORMATION
[email protected]
Receiver entropy : H ( y ) p ( y ) log p( y )
Conditional Entropy : H ( x | y ) p ( x, y ) log p ( x | y )
H ( y | x) p( x, y ) log p( y | x)
p( x | y )
Information transfer : I ( x, y ) p ( x, y ) log [ ]
p( x) p( y )
H ( x) H ( x | y )
H ( y ) H ( y | x)
Note:
7
[email protected]
8
H(Y|X)
SUMMARY
Quantity Definition
Source information I ( X i ) log 2 P( X i )
Received I ( Y j ) log 2 P( Y j )
information
[email protected]
P( X i | Y j )
Mutual information I ( X i ,Y j ) log 2
P( X i )
[email protected]
C=max I(X,Y)
that is maximum information transfer
10
BINARY SYMMETRIC CHANNEL CAPACITY
P(Y 0 | X 0)
I ( X ; Y ) P( X 0) P(Y 0 | X 0) log 0 1-p 0
P(Y 0)
P(Y 1 | X 0)
P( X 0) P(Y 1 | X 0) log p
[email protected]
P(Y 1)
Input Output
P(Y 0 | X 1)
P( X 1) P(Y 0 | X 1) log p
P(Y 0)
P(Y 1 | X 1) 1 1
P( X 1) P(Y 1 | X 1) log
P(Y 1)
1-p
1 p
P( X 0)(1 p ) log 1 p p
(1 p ) P( X 0) pP( X 1) P
P( X 0) p log
p
p 1 p
pP( X 0) (1 p ) P( X 1)
p
P( X 1) p log
(1 p ) P( X 0) pP( X 1)
1 p
P( X 1)(1 p ) log 11
pP( X 0) (1 p ) P( X 1)
BINARY SYMMETRIC CHANNEL CAPACITY
Channel Capacity is Maximum Information
1
max( I ( X ; Y )) P( X 1) P( X 0)
2
[email protected]
1 p
C max( I ( X ; Y )) P( X 0)(1 p ) log
(1 p ) P ( X 0) pP ( X 1)
p
P( X 0) p log
pP( X 0) (1 p ) P( X 1)
p
P( X 1) p log
(1 p ) P( X 0) pP( X 1)
1 p
P( X 1)(1 p ) log
pP( X 0) (1 p ) P( X 1) P ( X 1) P ( X 0 )
1
2
[email protected]
four symbols. The joint distribution of these two random
variables is as follows:
14
EXAMPLE
(a) Write down the marginal (individual) distribution for X
and compute the marginal entropy H(X) in bits.
[email protected]
(b) Write down the marginal distribution for Y and compute
the marginal entropy H(Y ) in bits.
[email protected]
𝑝 𝑥𝑖 , 𝑦𝑗 = 1, ∀ 𝑖, 𝑗 ∈ *𝑎, 𝑏, 𝑐, 𝑑+
𝑖 𝑗
𝑝 𝑥, 𝑦 = 𝑝 𝑦 𝑥 . 𝑝 𝑥 = 𝑝 𝑥 𝑦 𝑝 𝑦 (1)
From the total probability theory:
𝑝 𝑥 = 𝑗𝑝 𝑥 𝑦𝑖 𝑝 𝑦𝑖 = 𝑗𝑝 𝑥, 𝑦𝑗 From Eq(1)
i.e., p(x_a) = p(y_a)+ p(y_b)+ p(y_c)+ p(y_d) =1/8+1/16+1/32+1/32 =
¼ and so on.
p(x_b) = ¼ also p(x_c) = ¼ also p(x_d) = ¼ also
[email protected]
(c) joint entropy H(X, Y):
𝐻 𝑋, 𝑌 = 𝑖 𝑗𝑝 𝑥𝑖 , 𝑦𝑗 log 2 𝑝(𝑥𝑖 , 𝑦𝑗 )
i.e., over all 16 probabilities in the joint distribution(of which only 4
different non-zero values appear, with the following frequencies):
(1)(2/4) + (2)(3/8) + (6)(4/16) + (4)(5/32) = 1/2 + 3/4 + 3/2 + 5/8 = 27/8 bits.
[email protected]
Or,
I(X; Y ) = H(X) + H(Y ) − H(X, Y ) = 2 + 7/4 - 27/8
= (16+14-27)/8 = 3/8bits.