MMC Module 2 Notes
MMC Module 2 Notes
MULTIMEDIA COMMUNICATION
BEC613A
VI SEMESTER
MODULE 2 NOTES
Prepared by,
Mrs. MEGHANA M N
Assistant Professor
Dept. of ECE
MIT Thandavapura
MODULE 2 CONTENTS
INFORMATION REPRESENTATION
• Introduction
• Digitization principles
• Text
• Images
• Audio and Video (Chapter 2 of Text1)
Text Book 1: Multimedia Communications –Fred Halsall, Pearson Education,2001,ISBN-978813170994
MULTIMEDIA COMMUNICATION (BEC613A) MODULE 2
MODULE 2
MULTIMEDIA INFORMATION REPRESENTATION
1. INTRODUCTION
• All types of multimedia information are process and store within the computer in a digital form.
• Textual information: contains strings of characters entered through keyboard.
• Code word: each character represented by a unique combination of fixed number of bits.
• Analog Signal: signal whose amplitude varies continuously with time. In order to store and
process analog signal type of media in a computer we should convert any time varying analog
signals into a digital form is necessary.
• Signal encoder: Conversion of analog signal into digital signal is carried out using an electrical
circuit known as Signal Encoder, it includes following steps:
Ø Sampler: It samples the amplitude of analog signals at repetitive time intervals.
Ø Quantization: converting amplitude of each sample into a corresponding digital value.
• Signal decoder: Conversion of stored digital sample relating to a particular media type into their
corresponding time-varying analog form is performed by a electrical circuit is known as a signal
decoder.
Figure 1: Signal properties (a) time-varying analog signal; (b) sinusoidal frequency components; (c)
signal bandwidth example; (d) effect of a limited bandwidth transmission channel
• In figure 1(a), the amplitude of signals varies continuously with time. A mathematical technique
known as Fourier analysis can be used to show that any time varying analog signal is made up of
infinite number of single frequency sinusoidal signals whose amplitude and phase vary
continuously with time.
• The highest and lowest frequency components of the signal are shown in the figure 1(a) and (b).
• The range of frequencies of the sinusoidal components that make up a signal is called the signal
bandwidth which is shown in the figure 1(c).
• In terms of speech, human produce sounds which are converted into electrical signals by a
microphone that are made up of a range of sinusoidal signals varying frequency between 50Hz and
10kHz.
• In case of a music signal, the range of signals is wider and varies between 15Hz and 20kHz which
is sensitive to ear.
• When an analog signal is being transmitted through a network, the bandwidth of the transmission
channel, the range of frequencies of the channel should be equal or greater than the bandwidth of
the signal.
• If the bandwidth of the channel is less than this, then some of the low and high frequency
components will be lost which degrades the quality of the received signal. This type of
transmission channel is called bandlimiting channel and its effect is shown in figure 1(d).
• The conversion of a time varying analog signal (audio signal) into a digital form is carried out
using an electronic circuit is known as signal encoder.
• The principles of an encoder are shown in figure 2 (a) which consists of 2 main circuits:
Ø Bandlimiting filter
Ø ADC (Analog to Digital Converter) which has 2 components:
v Sample-and-hold circuit
v Quantizer
• A typical waveform for a signal encoder is shown in figure 2
• Bandlimiting filter: remove selected higher-frequency components from the source signal (A).
• Sample-And-Hold: got output of bandlimiting filter, (B) signal used to sample amplitude of the
filtered signal at regular time intervals (C) and to hold the sample amplitude constant between
samples (D) signal Quantizer circuit got signal (D) which converts each sample amplitude into a
binary value known as a codeword like (E) signal.
• To represent the amplitude of a time-varying analog signal precisely require 2 things:
1. Sampling Rate:
Ø Nyquist Sampling Theorem: states that a time- varying analog signal it's amplitude
must be sampled at a minimum rate that is equal to or greater than twice the highest
sinusoidal frequency component that is present in the signal and is known as Nyquist
rate, represented as either Hz or samples per second (sps).
Ø Sampling a signal at a rate which is lower than the Nyquist rate results in additional
frequency components being generated that are not present in the original signal which
cause the original signal to become distorted and such signals are called alias signals.
Ø Band limiting filter/Antialiasing filter: source signal is passed into the band limiting
filter to pass only those frequency components up to that determined by Nyquist rate any
higher-frequency components in the signal which are higher than this are removed
before the signal is sampled.
2. Quantization intervals:
Ø If Vmax, is the maximum positive and negative signal amplitude, n is the number of
binary bits used and q is the magnitude of each quantization interval is given by:
Ø Signal - anywhere within a Quantization intervals will be represented by the same binary
codeword.
Ø The difference between the actual signal amplitude and the corresponding amplitude is
Ø called quantization error.
Ø The error values will vary randomly from samples to sample and hence quantization
error is also known as quantization noise.
Ø The ratio of peak amplitude of a signal to its minimum amplitude is known as Dynamic
range of the signal D.
𝑉#$%
𝐷 = 20 𝑙𝑜𝑔!" ) + 𝑑𝐵
𝑉#&'
• Analog signal is processed and transmitted in a digital form. Prior to their output they must be
converted back again into their analog form. The electronic circuit that performs this conversion is
known as signal decoder.
• The circuit components of decoder and signal waveform is shown in the figure 3.
Figure 3: Signal decoder design: (a) circuit components, (b) signal waveform
• Digital To Analog Converter (DAC) is a circuit which converts each digital codeword (A) into an
equivalent analog sample (B), amplitude of each level being determined by, corresponding
codeword.
• Fourier analysis: used to show that output of DAC comprises sinusoidal frequency components
make up the original (filtered) analog signal + an infinite number of additional higher-frequency
components.
• For original signal to reproduce DAC output is passed through a LPF (Low Pass Filter) which only
passes those frequency components that made up the original filtered signal (C).
• In order to produce the original signal, the output of the DAC is passed through a LPF which only
passes low frequency components of original filtered signal. Therefore, LPF is also known as
recovery or reconstruction filter.
• Since most of the multimedia applications involve audio and video, the communication channel is
two way simultaneous in order the terminal equipment must support both input and output
simultaneously. The audio/video signal encoders and decoders in each terminal equipment are
combined into single unit called audio/video encoder decoder or audio/video codec.
3. TEXT
• Definition: Text is a human-readable sequence of characters and the words they form that can be
encoded into computer- readable formats such as ASCII.
• There are 3 Types of text:
1) Unformatted Text: alternative name plaintext and it enables pages to be created comprises of
strings of fixed-sized characters from a limited character set.
2) Formatted Text: alternative name rich text and it enables pages and complete documents to
be created which, comprise of strings of characters of different styles, size and shape with
tables, graphics, and images inserted at appropriate points.
3) Hypertext: It enables an integrated set of documents (each comprising formatted text) to be
created which have defined linkages between them.
• Two examples of character sets widely used to create pages consisting of unformatted text strings
are:
1) ASCII character set
2) Mosaic character set
• Table 1(a) shows set of characters, available in the ASCII (American Standard Code for
Information Interchange) character set is one of the most widely used character sets.
• Table shows binary codewords represented by a unique 7-bit binary codeword means there are 128
alternate characters and codeword used to identify each character and is obtained by combining the
corresponding column (bits 7-5) and row (bits 4-1) bits together.
• Bit 7 is MSB and Bit 0 is LSB thus codeword for uppercase M is 1001101.
• Printable Characters: It is a collection of normal alphabetic, numeric and punctuation characters
but ASCII total characters also include a number of control characters including:
1) Format control characters: BS (backspace), LF (Linefeed), CR (Carriage Return), SP (Space),
DEL (Delete), ESC (Escape), and FF (Formfeed).
2) Information separators: FS (File Separator), RS (Record Separator).
3) Transmission control characters: SOH (Start-Of-Heading), STX (Start-Of- Text), ETX (End-Of-
Text), ACK (Acknowledge), NAK (Negative Acknowledge), SYN(Synchronous Idle).
• Table 1(b) shows set of characters, available in the mosaic character set for table 1(a).
• Characters in columns 010/011 and 110/111 are replaced by set of mosaic characters.
• Mosaic characters used with uppercase characters to create relatively simple graphical images.
• An example application which uses mosaic character set are Videotext and Teletex which are
mosaic general broadcast information services available, through a standard television set.
3.3 HYPERTEXT
• This enables an integrated set of documents referred to as pages to be created which have defined
linkage points referred to as hyperlinks between them. Documents consisting of only text are
created using hypertext.
• Each document has a unique address known as Uniform Resource Locator (URL). The first page of
the document is known as home page.
• The linked set of pages that are stored in a particular server are accessed and viewed using a client
program known as a browser.
• Hypertext is used to create an electronic version of such documents with the index, description of
the department, courses offered, library and other facilities all written in hypertext as pages with
4. IMAGES
• Images include computer generated images referred to as computer graphics or simply graphics
and digitized images of both documents and pictures.
• These images are displayed in the form of a two-dimensional matrix of individual picture elements
known as Pixels or Pels.
• There are 3 types of images:
1. Graphics
2. Digitized documents
3. Digitized pictures
4.1 GRAPHICS
• Software packages provide easy to use tools to create graphics that are composed of all kinds of
visual objects including lines, arcs, squares, rectangles, circles, ovals, diamonds, stars and so on, as
well as any form of hand drawn (freeform) objects.
• These are produced by drawing the desired shape on the screen by means of a combination of a
cursor symbol on the screen.
• Textual information can also be included in a graphics, together with prescribed tables and graphs
and digitized pictures and photographs.
• A computer’s display screen can be considered as being made up of a two-dimensional matrix of
individual picture elements each of which can have a range of colors associated with it.
• Ex.: VGA (Video Graphics Array) common type of display.
• Figure 6(a) Shows a matrix of 640 horizontal pixels by 480 vertical pixels. 8 bits/pixel which
allows each pixel to have one of 256 different colors.
• All objects including the free-form objects made up of a series of lines connected to each other as
shown in figure 6(b).
• Attributes: each object has a number of attributes associated with it they include:
Ø Its shape - a line, a circle, a square, and so on.
Ø Its size - in terms of pixel positions of its border coordinates.
Ø Color of border.
Ø Its shadow, and so on
• Editing of an object involves simply, changing selected attributes associated with the object.
• Object shape can be either open or closed.
Ø Open object: start of the first line and end of the last line that make up the object's border
are not connected i.e. they do not start and end on the same pixel.
Ø Closed object: start of the first line and end of the last line that make up the object's border
are connected i.e. they start and end on the same pixel.
• For closed objects color-fill can be done as in Figure 6(c). coloring the solid block with the same
color is known as rendering.
Figure 6: Graphic principles (a)Example of screen format (b)Simple object example (c)Change of object, solid
object
• 2 forms of representation of computer graphic:
1. High-level version:
v Actual pixel-image of the graphic
v They require less memory to store the image and less bandwidth for its transmission.
v Ex: png, jpeg
Figure 9: Color principles: (a) additive color mixing (b) subtractive color mixing
• The mixing technique used in figure (a) is known as additive color mixing. Since black is
produced when all the three primary colors are zero, which is particularly useful for producing a
color image on a black surface as in the display applications.
• It is also possible to perform the complementary subtractive color mixing operation to produce a
similar range of colors which is shown in the figure (b).
• In subtractive color mixing, a white is produced when the three chosen primary colors cyan (C),
magneta (M) and yellow (Y) are all zero.
Page No - 13
MULTIMEDIA COMMUNICATION (BEC613A) MODULE 2
• Hence this choice of colors is useful for producing a color image on a white surface as in the case
of printed applications.
• The same principle is used in the picture tubes associated with the color television sets with the
three primary colors R, G and B. Also, in most computer monitors uses the same picture tubes as
are used in television sets.
• Inside the display screen of the picture tube is coated with a light sensitive phosphor which emits
light when energized by the electron beam.
• Brightness: It is the amount of light emitted which is determined by the power in the electron beam
at that instant.
• In case of black and white picture tubes just a single electron beam is used with a white sensitive
phosphor. Color tubes use three separates closely located beam and a 2-D matrix of pixels.
• Each pixel consists of set of three-color sensitive phosphors associated with each pixel is called
phosphor triad which is shown in figure 12.
• Video controller is a hardware subsystem that read the pixel values stored in the video RAM in
time with the scanning process converts each set of pixel values into equivalent set of R, G and B
analog signals for output to display.
1. US standard: US: NTSC (National Television Standards Committee) uses 525 scan lines per
frame.
2. Europe standard: Europe three color standards namely: PAL (UK), CCIR (Germany) and
SECAM (France) uses 625 lines per frame.
Figure 14: screen resolution for 525 and 625 lines (a) visible lines per frame and (b) spatial resolution
• Figure 14 Shows the screen resolution for 525 and 625 lines visible lines per frame and spatial
resolution.
• To produces a square picture avoiding distortion on the screen with 4/3 aspect ratio it is necessary
for displaying a square of (N X N) pixels to have:
1. 640 pixels (480 * 4/3) per line, with an NTSC monitor
2. 768 pixels (576 * 4/3) per line, with a European monitor
• Memory requirements to store a single digital image can be high, vary between 307.2 Kbytes for
an image displayed on a VGA screen through to approximately 2.36Mbytes for a SVGA (Super
VGA) screen.
• In the figure it is assumed that, the captured image is transferred to the computer directly as it is
produced.
• In the case of digital camera, a set of digitized images can be stored within the camera itself and
then downloaded into the computer later.
• Image sensor is a device that captures the image within the camera. This is a silicon chip in digital
cameras consisting of a 2-D grid of light sensitive cells called photosites.
• When the camera shutter is activated, each photosites stores the level of intensity of the light that
falled on it.
• A widely used image sensor is a charge coupled device (CCD). This consists of an array of
photosites on its surface and operates by converting the level of intensity that falls on each
photosites into an equivalent electrical charge.
• The level of charge and light intensity stored at each photosites position is then read out and
converted into a digital value using an ADC.
• The three methods to obtain color associated with each photosites and pixel position is as shown in
the figure 16 and the methods are as follows:
• Method 1: Surface of each photosite coated with either R, G, or B filter so, that its charge is
determined only by the level of R, G, and B lights that falls on it. Coatings are arranged on 3 X 3
grid structure as in Figure above then, color associated with each photosite/pixel determined by the
output of photosite R, G, and B together with each of its 8 immediate neighbors. The levels of two
other colors in each pixel are estimated by interpolation procedure involving all 9 values.
Application: Consumer-friendly cameras.
• Method 2: It involves use of 3 separate exposures of a single image sensor 1st through a Red filter,
2nd through a Green filter, and 3rd through a Blue filter. Color associated with each pixel position
is then, determined by the charge obtained with each of 3 filters - R, G, and B.
Application: High-resolution still-image cameras in locations such as photographic studios where,
cameras can be attached to a tripod.
• Method 3: uses 3 separate image sensors one with all photosites coated with a red filter, 2nd with
a green filter, 3rd with a blue filter. A single exposure is used with the incoming light split into 3
beams each of which exposes a separate image sensor.
Application: Professional-quality-high-resolution still and moving image camera.
5. AUDIO
• There are two types of audio signals:
1. Speech audio signal: Used in interpersonal applications including telephony and video
telephony.
2. Music audio signal: Used in applications such as CD on demand and broadcast television.
• Audio can be produced by means of a microphone or electronically using some form of
synthesizer.
• In case of synthesizer, the audio is created in a digital form and stored within the computer
memory.
• Audio can be produced in 3 ways:
1. Microphone: generates a time-varying analog signal audio signal encoder take analog signals,
convert them into digital form.
2. Synthesizer: Create audio in the digital form can be readily stored in the computer memory
output.
3. Loudspeaker: generates digital form signal to a time-varying analog audio signal.
• Bandwidth of the audible signal ranges from 20Hz to 20KHz and BW of speech and music is:
1. BW of a typical speech signal is from 50Hz to 10kHz thus sampling rate of 20ksps (2 X 10
kHz) is used for speech.
2. BW of a typical music signal is from 15Hz to 20kHz thus sampling rate of 40ksps (2 X 20
kHz) is used for speech.
Figure 17: PCM principles (a) Signal encoding and decoding schematic (b) Compressor characteristics
(c) Expander characteristics
• This can be achieved by compressor circuit and the reverse operation is performed by expander
circuit. The overall operation is known as companding.
• The input and output relationship of both circuits is shown in figure (b) and (c). Figure (b) shows
compression characteristic and figure (c) shows the expansion characteristic.
• Prior to the input signal being sampled and converted into a digital form by the ADC, it is passed
through the compressor circuit by passing the amplitude of the input signal.
• The level of compression and quantization intervals increases as the amplitude of the input signal
increases.
• 2 different compression-expansion characteristics are in use:
Ø µ-law: used in North America and Japan.
Ø A-law: used in Europe and some other countries.
6. VIDEO
• Video features in a range of multimedia applications such as:
1. Entertainment: Broadcast television and VCR/DVD recordings.
2. Interpersonal: Video telephony and Videoconferencing
3. Interactive: Short video clips
• The total screen contents are then refreshed at a rate of either 60 or 50 frames per second.
• The computer monitors used with most personal computers use the same picture tubes as those in
broadcast television receivers and hence operate in similar way.
• The three digitized color signals that make up a stored picture/image are read from the computer
memory in time synchronous with the scanning operation of the display tube.
• After each complete scan of the display, the procedure repeats producing a flicker free color image
on the screen.
• Practically Broadcast television operates slightly different than the above logic in terms of:
Ø Scanning sequence
Ø Choice of color signal
• In order to fit the Y, Cb and Cr signals in the same bandwidth, the 3 signals must be combined
together for transmission. The resulting signal is then known as the composite video signal.
• If the two-color difference signals are transmitted at their original magnitudes, the amplitude of
the luminance signal can become greater than that of the equivalent monochrome signal. This
leads to a degradation in the quality of the monochrome picture and hence is unacceptable.
• To overcome this effect, the magnitude of the two color difference sign are both scaled down.
• The two color difference signals are referred to by different symbols in each system:
v PAL system: Cb, and Cr are referred to as U and V respectively, scaling factors used for
the three signals are:
Y = 0.299R + 0.587 G + 0.114 B
U = 0.493 (B - Y)
V = 0.877 (R – Y)
v NTSC system: two color difference signals are combined to form two different signals
referred to as I and Q, scaling factor used are:
Y = 0.299R + 0.587 G + 0.114 B
I = 0.74 (R – Y) – 0.27 (B – Y)
Q = 0.48 (R – Y) + 0.41 (B – Y)
• Digitization of video signals has been carried out in television studios for many years in order
to perform conversion from one video format into another.
• In order to standardize this process and to make the exchange of television programmes
internationally easier, the international body for television standards, the International
Telecommunications Union-Radiocommunication Branch (ITU-R) formerly known as the
Consultative Committee for International Radiocommunications (CCIR) define a standard for
the digitization of video pictures known as Recommendation CCIR-601.
• The number of bits per sample is 8bits for all three signals which correspond to 256 quantization
intervals.
• The vertical resolution for all three signals are as follows:
1. 525-line system (NTSC): 480 lines (number of visible lines)
2. 625-line system (PAL) : 576 lines (number of visible lines).
• In the standard, a line sampling rate of 13.5MHz for luminance and 3.375MHz for the two
chrominance signals.
6.3 PC VIDEO
• Number of multimedia applications that involve live video, use a window on the screen of a PC
monitor for display purposes.
Ex.: desk video telephony, videoconferencing, and also video-in-a-window.
• For multimedia applications that involve mixing live video with other information on a PC screen,
the line sampling rate is modified to obtain horizontal resolution like 640 (480 x pixels per line
with a 525-line PC monitor and 768 (576 x 4/3) pixels per with a 625-line PC monitor.
• To achieve the necessary resolution with a 525-line monitor, the line sampling rate is reduced from
13.5MHz to 12.2727MHz while for a 625-line monitor, the line sampling rate must be increased
from 13.5 MHz to 14.75MHz.
• It is to be remembered that all PC monitors use progressive scanning rather than interlaced
scanning.
*****
MODULE 2
DIGITIZATION PRINCIPLES
1. Explain the Signal Encoder design and Quantization operation of digitization principles in detail.
(10M) or Describe the function of signal encoder with the associated waveform. (8M)***
2. Explain the Signal decoder design.***
TEXT
3. Discuss different type of text data representation. (10M Dec 2024) (6M Jan 2020) *****
4. Define 3 types of text. Discuss the hypertext that enable integrated set of documents. (8M Dec
2012) (8M Dec 2010) *****
5. How is formatted text different from unformatted text. Discuss the origin of the term WYSIWYG.
(6M)
IMAGE
6. Explain an example of a digitized document is that produced by the scanner associated with a fax
machine.
What do you understand by the terms or explain the terms and give application of both color mixing:
I. Color gamut
II. Additive color mixing
III. Subtractive color mixing (8M)
7. Explain the following:
I. Aspect ratio***
II. Quantization interval. *** (4M Dec 2023)
8. With the help of architecture, Explain the Raster scan principles or Raster scan principles with TV
or monitor (7M July 2013) (8M Dec 2012) *****
9. with the help of a diagram, explain how a digital image produced by scanner or digital camera is
captured and stored within the memory of the component. (10M Aug 2022) ***
10. explain the following: Quantization intervals, Hypertext, Raster scan, Aspect ratio, 4:2:2 format
(10M July 2011)
AUDIO
11. Explain the principle operation of a PCM Speech Codec, with a block diagram also explain the
Compressor and Expander. (4M) (6M June 2012) (6M June 2010) (10M July 2013) (10M Aug
2022) *****
Or
With a neat block diagram, explain PCM signal encoding and decoding method. (10M dec 2023)
Or
With a neat diagram, explain the signal encoding and decoding using PCM principles. (10M)
12. With schematic Explain Audio Synthesizer. (4M) (6M July 2011) ***
VIDEO
13. With the help of a diagram, explain the principle of interlaced of scanning as used in most TV
broadcast applications. (10M Aug 2022) ***
14. Explain the 4:2:2 and 4:2:0 digitization formats. (10M) (8M Dec 2010) ******
15. With the aid of diagram, summarize following digitization formats. i)4:2:2 ii)SIF (4M) ***
16. With a neat diagram describe the following digitization formats: 4:2:2 and QCIF (10M Dec 2023)
Or
With a neat diagram describe the following digitization formats: 4:2:2 and QCIF. For each format state
temporal resolution, spatial resolution, bit rate and give example for each format (10M June 2012)
(10M June 2010) *****
PROBLEMS
17. Problem on aspect ratio ******
Derive the time to transmit the following digitized image at both 64kbps and 1.5Mbps
i) 640*480*8 VGA compatible image.
ii) 1024*768*24 SVGA compatible image. (6M Dec 2023) (4M June 2012) (4M June 2010) (6M Jan
2020)
18. Problem on Audio *****
Assuming the bandwidth of a speech signal is from 50 Hz through to 10 kHz and that of a music signal
is from 15 Hz through to 20kHz, derive the bit rate that is generated by the digitization procedure in
each case assuming the Nyquist sampling rate is used with 12 bits per sample for the music signal.
Derive the memory required to store a 10 minute passage of stereophonic music. (3M July 2013)
(4M July 2011) (4M Dec 2010) (4M Jan 2020)
19. Problem on CD-Quality audio *****
Assuming the CD-DA standard is being used, derive:
(i) the storage capacity of a CD-ROM to store a 60 minute multimedia title.
(ii) the time to transmit a 30 second portion of the title using a transmission channel of bit rate:
1. 64kbps
2. 1.5Mbps (4M Dec 2012) (10M Aug 2022)
20. Problem on 4:2:2 format *****
Derive the bit rate and the memory requirements to store each frame that result from the digitization of
both a 525-line and a 625 line system assuming a 4:2:2 format. Also find the total memory required
to store a 1.5 hour movie/video. (10M)