ECEVSP L02 Info Theor
ECEVSP L02 Info Theor
Sink
📢$ 🔋
🔋
Source
A •– J •–––
✍
S ••• A
Sink
•– J
$
•––– S •••
B –••• K –•– T – B –••• K –•– T –
E • N –• W •–– E • N –• W •––
F ••–• O ––– X –••– F ••–• O ––– X –••–
I •• R •–• I •• R •–•
📢
🔋 🔋
Source Sink
Communication system (link) model
Error
Compression
Source Correction Modulator
Encoder
Encoder
Medium
Analog or
Digital Digital Analog
Error
Compression
Sink Correction Demodulator
Decoder
Decoder
Communication channel
Error
Compression
Source Correction Modulator
Encoder
Encoder
Medium
Error
Compression
Sink Correction Demodulator
Decoder
Decoder
Communication channel
For example:
optical fibre, cable, wires
Medium
air + (moving) objects
paper
solid state memory
paintings
…
Communication channel
Medium
antennas, couplers (e.g.
handwriting style), amplifiers,
digital-analog converters, Demodulator
filters (e.g. circuits or digital
processing).
Communication channel
Error
Compression
Source Correction Modulator
Encoder
Encoder
Medium
Error
Compression
Sink Correction Demodulator
Decoder
Decoder
Error
Compression
Source Correction Modulator
Encoder
Encoder
Medium
Error
Compression
Sink Correction Demodulator
Decoder
Decoder
Medium
Error
Compression
Sink Correction Demodulator
Decoder
Decoder
Clock strikes 12
A
at noon!
Flash flood in
B
the Sahara
desert.
UBC B-Line
C
filled to capacity.
Your inner information detector
Is ‘information’ somehow
Clock strikes 12
A related to the probability of
at noon!
an event?
Flash flood in
B
the Sahara
desert.
UBC B-Line Least likely of the
C three events.
filled to capacity.
Heuristic measure of information
(April 30, 1916 –February 24, 2001) was an American mathematician, electronic engineer,
and cryptographer known as "the father of information theory". He was also a juggler.
https://en.wikipedia.org/wiki/Claude_Shannon
See http://eecs.umich.edu/eecs/about/articles/2016/shannonposters/poster10.pdf
Where does randomness exist in
a telecommunication system?
Error
Compression
Source Correction Modulator
Encoder
Encoder
Medium
Error
Compression
Sink Correction Demodulator
Decoder
Decoder
Probabilities and Ensembles
ensemble
X : (x, AX , PX )
x outcome, value of random variable
AX = {a1 , a2 , . . . , aI }
set of possible alphabet
values, alphabet size
PX = {p1 , p2 , . . . , pI }
Probability: P (x = ai ) = pi (= p(ai ))
Facts: pi 0 8i
P
I
pi = 1
i=1
Joint Ensembles
XY : outcome is ordered pair (x, y) with
x 2 AX = {a1 , . . . , aI }, y 2 AY = {b1 , . . . , bJ }
Figure 2.2
Figure 2.3
Joint Ensembles
P
I
1) P (x = ai |y = bj ) =
i=1
?
2) P (x = ai , y = bj ) = P (x = ai )P (y = bj )
2) Monty-Hall Problem:
Suppose you are on a game show, and you are given the choice of
three doors: Behind one door is a car; behind the others, goats.
You pick a door, say No. 1, and the host, who knows what is behind
the doors, opens another door, say No. 3, which has a goat.
He then says to you, “Do you want to pick door No. 2?”
Is it to your advantage to switch your choice?
https://en.wikipedia.org/wiki/Monty_Hall_problem
Measure of information
- what is 1 bit?
Claude E. Shannon, “A mathematical theory of communication,” Bell System Technical Journal, vol.
27, pp. 379–423 and 623–656, July and October 1948.
Available: http://moser.cm.nctu.edu.tw/nctu/doc/shannon1948.pdf
http://www.inference.org.uk/itprnn/book.pdf
Information from a message
Message is the outcome of rolling dice
1) roll with one die: 6 possible outcomes 🎲
2) roll with two dice: 36 possible outcomes 🎲
🎲
Which observation resolves more uncertainty about the outcome?
1
Define: IS (x) = log2 p(x)
Lottery 6/49:
x=1: you win the jackpot, P(x=1)=1/13,983,816 ≣ 23.7 bit
x=0: you did not win it, P(x=0)=1-P(x=1) ≣ 10-7 bit
Entropy of a Source
Average information (reduction of uncertainty when observing the
outcome) of a message (random experiment)
P ⇣ ⌘
Define: H(X) = 1
p(x) log2 p(x)
x2AX
0.9
Coin toss:
0.8
0.7
H(X) 0.6
0.5
p1 log2 (p1 ) p0 log2 (p0 )
0.4
0.3
0.2
0.1
Radnat at freedigitalphotos.net
0
0 0.1 0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1
P(x=Head)
Entropy of a message
joint entropy
P ⇣ ⌘
1
H(X, Y ) = p(x, y) log2 p(x,y)
(x,y)2AX ⇥AY
conditional entropy
P ⇣ ⌘
1
H(X|Y ) = p(x, y) log2 p(x|y)
(x,y)2AX ⇥AY
H(X|Y ) H(X)
with equality if and only if X and Y are
statistically independent
Gedankenexperiment: Weighing
babies problem
# Weighings
A 3-4
B 5-6
C 7-8
D 9-10
E 11-12
What is your ensemble for
baby weighing?
Set of outcomes Random Variable is Probabilities
mapping to numbers associated with an
. outcome……..
1
How should you
distribute the
0 babies to get the
most information in
the 1st weighing
-1 (trial)?
What is your first weighing distribution?
# on Left
vs.Right
A 6v6
B 5v5
C 4v4
D 3v3
E 2v2
Probabilities associated with weighing distributions
C={# on
P ( x = '\ ' = 1) P ( x = '- ' = 0) P ( x = '/ ' = -1) H (c) bits
Left vs.
Right}
6v6 1/2 0 1/2 1.00
5v5 5/12 1/6 5/12 1.48
4v4 1/3 1/3 1/3 1.58
2v2 1/6 2/3 1/6 1.25
1v1 1/12 5/6 1/12 0.82
Which weighing distribution do you think is the best and why?
I secretly pick one number from the set and you have to guess it.
You may ask me any yes/no question you like.
1. What strategy would you use in asking questions?
2. What is the minimum number of questions you need to ask to
guarantee guessing the number?
What is the minimum number of questions you
need to ask to guarantee finding the number?
# Questions
A <=4
B 5
C 6
D 7
E >=8
What have we learned from the
Gedankenexperiments …