Information Theory: Mohamed Hamada
Information Theory: Mohamed Hamada
Information Theory: Mohamed Hamada
Mohamed Hamada
Software Engineering Lab
The University of Aizu
Email: [email protected]
URL: http://www.u-aizu.ac.jp/~hamada
Today’s Topics
• Communication Channel
• Noiseless binary channel
• Binary Symmetric Channel (BSC)
• Symmetric Channel
• Mutual Information
• Channel Capacity
1
Digital Communication Systems
1. Huffman Code.
3. Lemple-Ziv Code.
Source Source
Encoder Decoder
Channel Channel
Encoder Decoder
Modulator De-Modulator
Channel
2
Digital Communication Systems
1. Memoryless
2. Stochastic
3. Markov
Information 4. Ergodic User of
Source Information
Source Source
Encoder Decoder
Channel Channel
Encoder Decoder
Modulator De-Modulator
Channel
3
INFORMATION TRANSFER ACROSS CHANNELS
Sent Received
messages messages
symbols
Source Channel Channel Source
source Channel decoding decoding receiver
coding coding
4
Communication Channel
Examples of channels:
5
Communication Channel
Channel
Transition probabilities
Memoryless:
- output only on input
- input and output alphabet finite
6
Noiseless binary channel
Noiseless binary channel
Channel
0 0
1 1
Transition Matrix
0 1
p(y | x) = 0 1 0
1 0 1
7
Binary Symmetric Channel (BSC)
(Noisy channel)
1-p
Error Source
0 0
e
p
xi y i = xi e
+
Output 1 1
Input
1-p
8
Binary Symmetric Channel (BSC)
(Noisy channel)
BSC Channel
0 01
p 1-p
BSC Channel
1 10
p 1-p
p
0 0
1-p
1 1
p
BSC Channel
9
Symmetric Channel
(Noisy channel)
Channel
X Y
y1 y2 y3
x1 0.3 0.2 0.5
p(y | x) = x2 0.5 0.3 0.2
x3 0.2 0.5 0.3
10
Mutual Information
11
Mutual Information (MI)
Channel
X p(y|x) Y
I(X,Y)=H(Y)-H(Y|X)
Proof:
p( y | x)
I ( X , Y ) p ( x, y ) log 2 ( )
x y p( y)
I ( X , Y ) p ( x, y ) log 2 ( p ( y )) p ( x, y ) log 2 ( p ( y | x ))
x y x y
• I(X,Y)=H(Y)-H(Y|X)
• I(X,Y)=H(X)-H(X|Y)
• I(X,Y)=H(X)+H(Y)-H(X,Y)
• I(X,Y)=I(Y,X)
• I(X,X)=H(X)
H(X) H(Y) 14
Mutual Information with 2 channels
Channel 1 Channel 2
Y Z
X p(y|x) p(z|y)
15
Mutual Information with 2 channels
Theory: I(X,Y) I(X, Z)
16
Channel Capacity
17
Transmission efficiency
I need on the average H(X) bits/source output to describe the
source symbols X After observing Y, I need H(X|Y) bits/source
output.
Channel
H(X) X Y H(X|Y)
Note that :
• During data compression, we remove all redundancy in the
data to form the most compressed version possible.
Channel
Probability
Transition Matrix
1/2 0 0
0 1
p(y|x) = 0 1 0
1 0 1
1/2 1 1
H(X) = H(p, 1-p)= H(1/2,1/2) = -1/2 log 1/2 - 1/2 log 1/2 = 1 bit
p
0 0
1-p Transition Matrix
1 1
p 0 1
BSC Channel p(y|x)= 0 p 1-p
Here the mutual information 1 1-p p
I(X,Y) = H(Y) - H(Y | X) = H(Y) - ∑x p(x) H(Y | X=x)
= H(Y) - ∑x p(x) H(P) = H(Y) - H(P) ∑x p(x) (Note that ∑p(x) = 1)
x
For example: Consider a symmetric channel with the following transmission matrix
It is clear that such channel is symmetric y1 y2 y3
since rows (and columns) are permutations. x1 0.3 0.2 0.5
Hence the capacity C of this channel is: P (y|x) = x2 0.5 0.3 0.2
x3 0.2 0.5 0.3
C = log 3 - H(0.2, 0.3, 0.5) = 0.1 bits (approx.)
Transition Matrix
22
Channel Capacity
(Noisy) Symmetric channel
Channel
X Y
C = log |Y | - H (r)
23