Introduction To Information Theory Channel Capacity and Models
Introduction To Information Theory Channel Capacity and Models
theory
channel capacity and models
A.J. Han Vinck
University of Duisburg-Essen
May 2012
This lecture
Some models
Channel capacity
P(y|x)
output Y
transition probabilities
memoryless:
- output at time i depends only on input at time i
- input and output alphabet finite
1-p
Error Source
E
X
Input
Y X E
Output
0
p
1
1-p
from AWGN
to BSC
Other models
0
0 (light on)
X
p
1-p
1-e
e
Y
1 (light off)
e
P(X=0) = P0
1
P(X=0) = P0
Z-channel (optical)
(MAC)
Erasure channel
p
p
1-p-e
e
e
1-p-e
Error Source
Error Source
good
bad
Pbg
transition probability
Pbb
channel capacity:
I(X;Y) = H(X) - H(X|Y) = H(Y) H(Y|X) (Shannon 1948)
H(X)
H(X|Y)
channel
notes:
max I(X; Y) capacity
capacity depends onP (input
x ) probabilities
because the transition probabilites are fixed
message
2k
receive
estimate
channel
with errors
decoder
Code book
n
There are 2k code words of length n
k is the number of information bits transmitted in n channel uses
Channel capacity
Definition:
The rate R of a code is the ratio k/n, where
k is the number of information bits transmitted in n channel uses
P( 1) (2 k 1)
2 nh ( p)
2n
2 n (1 h ( p) R ) 2 n (CBSC R ) 0
for
and
k
R 1 h ( p)
n
n
1-p
0
X
since Y is binary
H(Y|X) = h(p)
1-p
= P(X=0)h(p) + P(X=1)h(p)
1.0
0.5
Bit error p
1.0
0 (light on)
X
p
1-p
P(X=0) = P0
Y
1 (light off)
1-e
e
Y
e
H(X) = h(P0 )
H(X|Y) = e h(P0)
1
P(X=0) = P0
Thus Cerasure = 1 e
(check!, draw and compare with BSC and Z)
P( 1) (2 k 1)2 n (1e )
2 n (1e R ) 2 n ( Cerasure R ) 0
k
for R 1 e, n
n
p
p
1-p-e
e
e
1-p-e
example
0
1/3
1
1/3
x1
x2
y1
P2|1
P1|2
:
:
:
xn
y2
P2|2
:
:
:
In general:
Pm|n
ym
* clue:
I(X;Y)
is convex in the input probabilities
i.e. finding a maximum is simple
Pe
k/n
C
Converse:
Xi
channel
Yi
i 1
i 1
i 1
i 1
source
encoder
Xn
channel
Yn
decoder
converse
k = H(M) = I(M;Yn)+H(M|Yn)
I(Xn;Yn) +H(M|Yn M)
Xn is a function of M
I(Xn;Yn) +H(M|M)
M is a function of Yn
Fano inequality
nC + 1 + k Pe
Pe 1 C/R - 1/nR
Hence:
Appendix:
Assume:
binary sequence P(0) = 1 P(1) = 1-p
t is the # of 1s in the sequence
Then n , > 0
Weak law of large numbers
Probability ( |t/n p| > ) 0
i.e. we expect with high probability pn 1s
Appendix:
Consequence:
1.
2.
3.
n ( p )
n
n
2n
2n 2 nh ( p)
t
pn
1
lim n log 2 2n
n
n
h (p)
pn
N ! 2 N N e
N N
Binary Entropy:
Note:
0.9
h(p) = h(1-p)
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
0.1
0.2
0.3
0.4
0.5
0.6
0.7
0.8
0.9
Output Y
x 2 S / 2 W
Noise
X
2noise 2x
Cap W log 2 (
p(z)
1
22z
z2 / 2 2z
2
noise
) bits / trans.
2noise S / 2W
2noise
) bits / sec .
Transition
probability P(0)
channel 1
channel 2
Select channel k
with probability
Q(k)
channel k has
transition
probability p(k)
Fritzman model:
multiple states G and only one state B
Closer to an actual real-world channel
G1
1-p
Gn
Error probability 0
B
Error probability h
interleaver
channel
interleaver
encoder
-1
message
decoder
random error
Note: interleaving brings encoding and decoding delay
Homework: compare the block and convolutional interleaving w.r.t. delay
Interleaving: block
Channel models are difficult to derive:
- burst definition ?
- random and burst errors ?
for practical reasons: convert burst into random error
read in row wise
transmit
column wise
De-Interleaving: block
read in column
wise
this row contains 1 error
read out
row wise
Interleaving: convolutional
input sequence 0
input sequence 1
delay of b elements