Module-1
Module-1
Claude Shannon
• Shannon proposed that certain types of random processes, such as music and speech,
have a fundamental or irreducible complexity. This means there is a lower bound to
how much these signals can be compressed without losing essential information. This
is a cornerstone of source coding theory, which addresses limits to data compression.
DATA and Information
What is Information
Let us consider the following statements:
Tomorrow, the sun will rise from the East.
From above statements, the amount of information is nothing to do with the length of the
sentences.
Also, it is interesting to note that the amount of information carrying is inversely proportional to
the probability of occurrences of the events.
Shannon gave the mathematical definition to quantify the amount of information through his
published work titled “A Mathematical Theory of Communication”.
Uncertainty and Information
Suppose, an experiment involves the observation of the output. (Ex- Tossing a fair
dice).
Every time we throw a dice, the outcome belongs to be any one of the symbol 1,
2, 3, 4, 5, 6}.
After the occurrence of event there is a gain in the amount of information (which
may be viewed as resolution of uncertainty).
Information Sources
• In communication system, it can be a device sending message, it can be
continuous or discrete.
• Continuous sources can be converted to discrete sources by sampling
and quantization.
• Therefore, discrete information source is a source which has only a finite
set of symbols as possible output.
A source for which each symbol produced is independent of the previous symbols is
called DMS.
Ex:
When the symbols are binary 1’s and 0’s then it is called binary memoryless source
(BMS).
Measure of Information (Very
Important)
• For a source alphabet where is the number of symbols each with
probability distribution . Then the information gained for the outcome
is
1. =0 for =1
𝐼 ( 𝒮= 𝑠 𝑖 ) =− log 𝑏 𝑝𝑖
• nat (natural unit) if b = e.
Example
You are watching the weather forecast, and the announcer says that it will rain
tomorrow. The probability of rain in your area is known to be P(rain)=0.2, and the
probability of no rain is P(no rain)=0.8.
Calculate the amount of information content (measured in bits) in the weather forecast.
Imagine you are a detective solving a mystery, and your job is to find out which
suspect committed the crime. There are 8 suspects.
Imagine you're trying to guess someone's 4-digit PIN code, and you know the PIN is
randomly chosen. Let’s calculate the information content of finding out the PIN.
Example on Information content
Question: A source produces one of four possible symbols, having probability P(x1)= ½,
P(x2)= ¼, P(x3)= P(x4)=1/8. Obtain the information content of each of these symbol.
Answer:
Example on Information content
Question: In a binary channel, of ‘0’ occurs with probability ¼ and ‘1’ occurs with
probability of ¾. Then, calculate the amount of information carried by each binit.
Entropy (Average Information
per symbol)
• In practical communication, we transmit long sequences of symbols.
vs
Example
• You have a biased coin where the probability of getting Heads is ppp, and the probability of getting Tails is
1−p1-p1−p.
1. Case 1: The coin is fair (p=0.5).
2. Case 2: The coin is biased, with p=0.9 for Heads and 1−p=0.1 for Tails.
3. Case 3: The coin is so biased that it always lands on Heads (p=1).
Example
• A weather forecast predicts the chance of rain or no rain for a given day. The probabilities for different
scenarios are as follows:
1. Case 1: The weather forecast says:
1. Rain: p=0.5,
2. No Rain: 1−p=0.5
2. Case 2: The forecast predicts:
1. Rain: p=0.9
2. No Rain: 1−p=0.1
3. Case 3: The forecast is certain:
1. Rain: p=1
2. No Rain: 1−p=0
Example
• A light bulb can either be ON or OFF. Consider the following three cases:
1. Case 1: The light bulb is almost always ON with probability p=0.9999, and OFF with probability 0.0001.
2. Case 2: The light bulb is equally likely to be ON or OFF, with p=0.5
3. Case 3: The light bulb is always ON with probability p=1
• (a) Information Content:
• For each case, calculate the information (III) for the ON and OFF events.
• (b) Entropy:
• For each case, calculate the entropy (HHH) of the system. Explain why the entropy is low or high in each
scenario.
Example
•A bag contains 16 balls:
•12 balls are red,
•4 balls are white.
•One ball is picked randomly.
Calculate the entropy HHH of the system using the formula:
Information Rate
• If the time rate at which emits symbols is (symbols/sec), then the
information rate of the source is given by
bits/second
The capacity of a channel (C) is proportional to the bandwidth (BW) and S/N.
So, in summary, entropy decides the rate at which you generate messages