0% found this document useful (0 votes)
203 views7 pages

Unit I Digital Communcation R2017 MCQ

The document contains questions and answers related to information theory and coding. It covers key topics like Shannon's channel capacity theorem, entropy, mutual information, error correction coding, Huffman coding. Some key points covered are: Shannon's theorem states the maximum error-free communication is defined by the channel capacity C. Entropy H is the average information per message. Mutual information I(X;Y) is always non-negative and symmetric. Error probability reduces to arbitrarily small value if transmission rate R is less than channel capacity C.

Uploaded by

Muthumari
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
203 views7 pages

Unit I Digital Communcation R2017 MCQ

The document contains questions and answers related to information theory and coding. It covers key topics like Shannon's channel capacity theorem, entropy, mutual information, error correction coding, Huffman coding. Some key points covered are: Shannon's theorem states the maximum error-free communication is defined by the channel capacity C. Entropy H is the average information per message. Mutual information I(X;Y) is always non-negative and symmetric. Error probability reduces to arbitrarily small value if transmission rate R is less than channel capacity C.

Uploaded by

Muthumari
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

UNIT I - INFORMATION THEORY AND CODING

1.What conditions must be fulfilled in a good digital communication system?


a) High data rate
b) High fidelity
c) Low transmit power
d) All of the mentioned

ANSWER: d

2. The channel capacity according to Shannon’s equation is


a. Maximum error free communication
b. Defined for optimum system
c. Information transmitted
d. All of the above

ANSWER: d

3. For a binary symmetric channel, the random bits are given as


a. Logic 1 given by probability P and logic 0 by (1-P)
b. Logic 1 given by probability 1-P and logic 0 by P
c. Logic 1 given by probability P2 and logic 0 by 1-P
d. Logic 1 given by probability P and logic 0 by (1-P)2

ANSWER: a

4. The technique that may be used to increase average information per bit is
a. Shannon-Fano algorithm
b. ASK
c. FSK
d. Digital modulation techniques

ANSWER: a

5. Code rate r, k information bits and n as total bits, is defined as


a. r = k/n
b. k = n/r
c. r = k * n
d. n = r * k

ANSWER: a

6. The information rate R for given average information H= 2.0 for analog signal band limited to B Hz is
a. 8 B bits/sec
b. 4 B bits/sec
c. 2 B bits/sec
d. 16 B bits/sec

ANSWER:(b)
7. Information rate is defined as
a. Information per unit time
b. Average number of bits of information per second
c. rH
d. All of the above

ANSWER: (d)

8. The mutual information


a. Is symmetric
b. Always non negative
c. Both a and b are correct
d. None of the above

ANSWER: (c)

9 . The relation between entropy and mutual information is


a. I(X;Y) = H(X) – H(X/Y)
b. I(X;Y) = H(X/Y) – H(Y/X)
c. I(X;Y) = H(X) – H(Y)
d. I(X;Y) = H(Y) – H(X)

ANSWER:(a)

10. Entropy is
a. Average information per message
b. Information in a signal
c. Amplitude of signal
d. All of the above

ANSWER: (a)

11. The memory less source refers to


a. No previous information
b. No message storage
c. Emitted message is independent of previous message
d. None of the above

ANSWER: (c)

12. The information I contained in a message with probability of occurrence is given by (k is constant)

a. I = k log2(1/P)
b. I = k log2(2P)
c. I = k log2(1/2P)
d. I = k log2(1/P2)

ANSWER:(a)
13. The expected information contained in a message is called
a. Entropy
b. Efficiency
c. Coded signal
d. None of the above

ANSWER: (a)

14. Probability density function defines


a. Amplitudes of random noise
b. Density of signal
c. Probability of error
d. All of the above

ANSWER: (a)

15.The capacity of Gaussian channel is


a. C = 2B(1+S/N) bits/s
b. C = B2(1+S/N) bits/s
c. C = B(1+S/N) bits/s
d. C = B(1+S/N)2 bits/s

ANSWER: (c)

16. For M equally likely messages, the average amount of information H is


a. H = log10M
b. H = log2M
c. H = log10M2
d. H = 2log10M

ANSWER:(b)

17. The channel capacity is


a. The maximum information transmitted by one symbol over the channel
b. Information contained in a signal
c. The amplitude of the modulated signal
d. All of the above

ANSWER: (a)

18 . The capacity of a binary symmetric channel, given H(P) is binary entropy function is
a. 1 – H(P)
b. H(P) – 1
c. 1 – H(P)2
d. H(P)2 – 1

ANSWER:(a)
19. Entropy of a random variable is
a) 0
b) 1
c) Infinite
d) Cannot be determined

Answer: c

20. Which is more efficient method?


a) Encoding each symbol of a block
b) Encoding block of symbols
c) Encoding each symbol of a block & Encoding block of symbols
d) None of the mentioned

Answer: b

21. According to Shannon Hartley theorem,


a. The channel capacity becomes infinite with infinite bandwidth
b. The channel capacity does not become infinite with infinite bandwidth
c. Has a tradeoff between bandwidth and Signal to noise ratio
d. Both b and c are correct

ANSWER: (d)

22. The negative statement for Shannon’s theorem states that


a. If R > C, the error probability increases towards Unity
b. If R < C, the error probability is very small
c. Both a & b
d. None of the above

ANSWER: (a)

23.For M equally likely messages, M>>1, if the rate of information R ≤ C, the probability of error is
a. Arbitrarily small
b. Close to unity
c. Not predictable
d. Unknown

ANSWER: (a)

24. For M equally likely messages, M>>1, if the rate of information R > C, the probability of error is
a. Arbitrarily small
b. Close to unity
c. Not predictable
d. Unknown

ANSWER: (b)
25. The value of the probability density function of random variable is
a) Positive function
b) Negative function
c) Zero
d) One

Answer: a

26. Which gives the measure of randomness of the random variable?


a) Mean
b) Variance
c) Standard variance
d) Pdf

Answer: b

27. The process of data conversion along with formatting the data is called as ______
a) Formatting
b) Modulation
c) Source coding
d) Amplifying

Answer: c

28. The process that transforms text into binary digits is called as _______
a) Binary coding
b) Data coding
c) Character coding
d) Sampling

Answer: c

29. For the number of bits, k=1 and number of symbols, M=2 the system is called as
a) Unary
b) Binary
c) Quarternary
d) None of the mentioned

Answer: b

30. Entropy is the measure of


a) Randomness
b) Information
c) Randomness & Information
d) None of the mentioned

Answer: c
31. Entropy calculation returns
a) Random variable
b) Deterministic number
c) Random number
d) None of the mentioned

Answer: b

32. Entropy of N random variables is the _____ of the entropy of individual random variable.
a) Sum
b) Product
c) Sum of squares
d) Average

Answer: a

33.The information source of a digital communication system can be


a) Packetized
b) Continuous
c) Packetized & Continuous
d) None of the mentioned

Answer: c

34. The unit of average mutual information is


a) Bits
b) Bytes
c) Bits per symbol
d) Bytes per symbol

Answer: a

35. When probability of error during transmission is 0.5, it indicates that


a) Channel is very noisy
b) No information is received
c) Channel is very noisy & No information is received
d) None of the mentioned

Answer: c

36. Binary Huffman coding is a


a) Prefix condition code
b) Suffix condition code
c) Prefix & Suffix condition code
d) None of the mentioned

Answer: a
37. The event with minimum probability has least number of bits.
a) True
b) False

Answer: b

38. When the base of the logarithm is 2, then the unit of measure of information is
a) Bits
b) Bytes
c) Nats
d) None of the mentioned

Answer: a

39. When X and Y are statistically independent, then I (x,y) is


a) 1
b) 0
c) Ln 2
d) Cannot be determined

Answer: b

40. The self information of random variable is


a) 0
b) 1
c) Infinite
d) Cannot be determined

Answer: c

41. Coded system are inherently capable of better transmission efficiency than the uncoded system.
a) True
b) False

Answer: a

You might also like