Unit Iv - Multimedia File Handling: Compression and Decompression
Unit Iv - Multimedia File Handling: Compression and Decompression
Unit Iv - Multimedia File Handling: Compression and Decompression
1
absolute data accuracy is not necessary.
The idea behind the lossy compression is that, the human eye fills in the missing information in the case of
video.
But, an important consideration is how much information can be lost so that the result should not affect. For
example, in a grayscale image, if several bits are missing, the information is still perceived in an acceptable
manner as the eye fills in the gaps in the shading gradient. Lossy compression is applicable in medical
screening systems, video tele-conferencing, and multimedia electronic messaging systems.
Lossy compressions techniques can be used alone o~' in combination with other compression methods in a
multimedia object consisting of audio, color images, and video as well as other specialized data types.
The following lists some of the lossy compression mechanisms:
Joint Photographic Experts Group (JPEG)
Moving Picture Experts Group (MPEG)
Intel DVI
CCITT H.261 (P * 24) Video Coding Algorithm
Fractals.
White Black
Run Code Run Code
Length Word Length Word
0 00110101 0 0000110111
1 000111 1 010
2 0111 2 11
3 1000 3 10
4 1011 4 011
5 1100 5 0011
6 1110 6 0010
7 1111 7 00011
8 10011 8 000101
9 10100 9 000100
\
10 00111 10 0000100 1
11 01000 10 0000100
11 01000 11 0000101
12 001000 12 0000111
13 000011 13 00000100
14 110100 14 00000111
15 110101 15 000011000
16 101010 16 0000010111
17 101011 17 0000011000
18 0100111 18 0000001000
19 0001100 19 0000 11 00 III
20 0001000 20 00001101000
21 0010111 21 00001101100
22 0000011 22 00000110111
23 0000100 23 00000101000
24 0101000 24 00000010111
25 0101011 25 00000011000
26 0010011 26 000011001010
27 0100100 27 000011001011
28 0011000 28 000011 00 11 00
29 00000010 29 000011001101
30 00000011 30 000001101000
31 00011010 31 000001101001
32 00011011 32 000001101010
33 00010010 33 000001101011
34 00010011 34 000011010010
35 00010100 35 000011 0 10011
For example, from Table 2, the run-length code of 16 white pixels is 101010, and of 16 black pixels
0000010111. Statistically, the occurrence of 16 white pixels is more frequent than the occurrence of 16
black pixels. Hence, the code generated for 16 white pixels is much shorter. This allows for quicker
decoding. For this example, the tree structure could be constructed.
36 00010101 36 000011010100
37 00010110 37 000011010101
38 000101 II 38 000011010110
9 00101000 39 000011 0 1 0 1 1 1
40 00101001 40 000001I01100
41 00101010 41 000001101101
42 00101011 42 000011011010
43 00101100· 43 0000 11 0 1 1011
44 00101101 44 000001010100
45 00000100 45 000001010101
46 00000101 46 000001010110
47 00001010 47 000001010111
48 00001011 48 000001100100
49 01010010 49 000001100101
50 010100II 50 000001010010
51 01010100 51 000001010011
52 01010101 52 000000100100
53 00100100 53 000000110111
The codes greater than a string of 1792 pixels are identical for black and white pixels. A new code indicates
reversal of color, that is, the pixel Color code is relative to the color of the previous pixel sequence.
Table 3 shows the codes for pixel sequences larger than 1792 pixels.
Run Length Make-up Code
(Black and White)
1792 00000001000
1856 00000001100
1920 00000001101
1984 000000010010
2048 000000010011
2112 000000010100
2176 000000010101
2240 000000010110
2304 000000010111
2368 000000011100
2432 000000011101
2496 000000011110
2560 000000011111
CCITT Group 3 compression utilizes Huffman coding to generate a set of make-up codes and a set of
terminating codes for a given bit stream. Make-up codes are used to represent run length in multiples of 64
pixels. Terminating codes are used to represent run lengths of less than 64 pixels.
As shown in Table 2; run-length codes for black pixels are different from the run-length codes for white
pixels. For example, the run-length code for 64 white pixels is 11011. The run length code for 64 black
pixels is 0000001111. Consequently, the run length of 132 white pixels is encoded by the following two
codes:
Makeup code for 128 white pixels - 10010
Terminating code for 4 white pixels - 1011
The compressed bit stream for 132 white pixels is 100101011, a total of nine bits. Therefore the
compression ratio is 14, the ratio between the total number of bits (132) divided by the number of bits used
to code them (9).
CCITT Group 3 uses a very simple data format. This consists of sequential blocks of data for each scanline,
as shown in Table 4.
Coding tree for 16 white pixels
EOL DATA FILL EOL DATA FILL EOL… DATA FILL EOL EOL EOL
LINE LINE LINE
1 2 n
Note that the file is terminated by a number of EOLs (End of. Line) if there is no change in the line [rom the
previous line (for example, white space).
TABLE 4: CCITT Group 3 1D File Format
Advantages of CCITT Group 3 ID
CCITT Group 3 compression has been used extensively due to the following two advantages: It
is simple to implement in both hardware and software .
It is a worldwide standard for facsimile which is accepted for document imaging application. This allows
document imaging applications to incorporate fax documents easily.
CCITT group 3 compressions utilizes Huffman coding to generate a set of make-up codes and a set
of terminating codes for a give bit stream.
CCITT Group 3 uses a very simply data format. This consists of sequential blocks of data for each
scanline.
3. CCITT Group 3 2D Compression
It is also known as modified run length encoding. It is used for software based imaging system and
facsimile.
It is easier to decompress in software than CCITT Group 4. The CCITT Group 3 2D scheme uses a "k"
factor where the
image is divided into several group of k lines. This scheme is based on the statistical nature of images; the
image data across the adjacent scanline is redundant.
If black and white transition occurs on a given scanline, chances are the same transition will occur within +
or - 3 pixels in the next scanline.
Necessity of k factor·
When CCITT Group 3 2D compression is used, the algorithm embeds Group 3 1 D coding between every k
groups of Group 3 2D coding, allowing the Group 3 1 D coding to be the synchronizing line in the event of
a transmission error.
Therefore when a transmission error occurs due to a bad communication link, the group 3 I D can be used to
synchronize and correct the error.
Data formatting for CClIT Group 3 2D
The 2D scheme uses a combination of additional codes called vertical code, pass code, and horizontal code
to encode every line in the group of k lines.
The steps for pseudocode to code the code line are:
(i) Parse the coding line and look for the change in the pixel value. (Change is found at al location).
(ii) Parse the reference line and look for the change in the pixel value. (Change is found at bl location). (iii) .
Find the difference in location between bland a 1: delta = b1- al
Advantage of CClIT Group 3 2D
The implementation of the k factor allows error-free
transmission . Compression ratio achieved is better than CClTT Group 3 1 D .
It is accepted for document imaging applications.
DisadvantageIt doesn’t provide dense compression
CCITT Group 4 2D compression
CClTT Group 4 compression is the two dimensional coding scheme without the k-factor.
In this method, the first reference line is an imaginary all-white line above the top of the image.The first
group of pixels (scanline) is encoded utilizing the imaginary white line as the reference line.
The new coded line becomes the references line for the next scan line. The k-factor in this case is the entire
page of line. In this method, there are no end-of-line (EOL) markers before the start of the compressed data
Figure below shows the components and sequence of quantization 5 * 8 Image blocks
ZigZag Sequence
Run-length encoding generates a code to represent the Count of zero-value OCT co-efficients. This process
of run-length encoding gives an excellent compression of the block consisting mostly of zero values.
Further empirical work proved that the length of zero values in a run can be increased to give a further
increase in compression by reordering the runs. JPEG came up with ordering the quantized OCT co-
efficients in a ZigZag sequence. ZigZag sequence the sequence in which the cells are encoded.
Entropy Encoding
Entropy is a term used in thermodynamics for the study of heat and work. Entropy, as used in data
compression, is the measure of the information content of a message in number of bits. It is represented as
Entropy in number of bits = log2 (probability of Object)
Huffman versus Arithmetic coding
Huffman coding requires that one or more sets of Huffman code tables be specified by the application for
coding as well as decoding. For arithmetic coding JPEG does not require coding tables.It able to adapt to the
image statistics as it encodes the image.
DC coefficient coding
Before DC coefficients are compressed the DC prediction is processed first.In DC prediction the DC
coefficient of the previous 8x8 block is subtracted from the current 8x8 block.
Two 8x8 blocks of a quantized matrix are shown in figure2.6. The Differential DC coefficient is delta
D=DCx - DCx-1.
AC coefficient coding
Each AC coefficient is encoded by utilizing two symbols symbol-1 and symbol-2. Symbol-1 represents two
piece of information called “run length” and “size”. Symbol-2 represents the amplitude of the AC
coefficient.
where L - approaches 0,
N(L) ~ number of stick L, and L is the length of the stick.
TIFF Structure
TIFF files consists of a header. The header consists of byte ordering flag, TIFF file format version number,
and a pointer to a table. The pointer points image file directory. This directory contains table of entries of
various tags and their information.
TIFF file format Header:
The next figure shows the IFD (Image File Directory) as its content. The IFD is avariable –length table
containing directory entries. The length of table depends on the number of directory entries in the table. The
first two bytes contain the total number of entries in the table followed by directory entrie. Each directory
entry consists of twelve bytes.The Last item in the IFD is a four byte pointer that points to the next IFD.
The byte content of each directory entry is as follows:
The first two byte contains tag number-tag ID.
The second two byte represent the type of data as shown in table3-1 below.
The next four bytes contains the length for the data type.
The final four bytes contain data or a pointer.
TIFF Tags
The first two bytes of each directory entry contain a field called the Tag ID.
Tag IDs arc grouped into several categories. They are Basic, Informational, Facsimile, Document storage
and Retrieval.
TIFF Classes: (Version 5.0)It has five classes
1. Class B for binary images
2. Class F for Fax
3. Class G for gray-scale images
4. Class P for palette color images
5. Class R for RGB full-color images.
The sub chunk contains a four-character ASCII string 10 to identify the type of data.
Four bytes of size contains the count of data values, and the data. The data structure of a chunk is same as
all other chunks.
RIFF chunk with two sub chunk:
The first 4 characters of the RlFF chunk are reserved for the "RIFF" ASCII string. The next four bytes
define the total data size.
The first four characters of the data field are reserved for form tyPe. The rest of the data field contains two
subchunk:
(i) fmt ~ defines the recording characteristics of the waveform.
(ii) data ~ contains the data for the waveform.
LIST Chunk
RlFF chunk may contains one or more list chunks.
List chunks allow embedding additional file information such as archival location, copyright information,
creating date, description of the content of the file.
RlFF MIDI FILE FORMAT
RlFF MIDI contains a RlFF chunk with the form type "RMID"and a subchunk called "data" for MIDI data.
The 4 bytes are for ID of the RlFF chunk. 4 bytes are for size 4 bytes are for form type 4
bytes are for ID of the subchunk data and 4 bytes are for the size of MIDI data.
RIFF DIBS (Device-Independent Bit Maps) .
DIB is a Microsoft windows standard format. It defines bit maps and color attributes for bit maps
independent of devices. DIEs are normally embedded in .BMP files, .WMF meta data files, and .CLP files.
DIB Structure
BITMAPINFOHEADER RGBQUAD PIXELS
A RIFF DIB file format contains a RIFF chunk with the Form Type "RDIB" and a subchunk called "data"
for DIB data.
4 bytes denote ID of the RIFF chunk
4 bytes refer size ofXYZ.RDI 4 bytes define Forum Type
4 bytes describe ID of the sub chunk data 4 bytes define size of DIB data.
RIFF PALETTE File format
The RIFF Palette file format contains a RIFF chunk with the Form Type "RP AL" and a subchunk called
"data" for palette data. The Microsoft Windows logical palette structure is enveloped in the RIFF data
subchunk. The palette structure contains the palette version number, number of palette entries, the intensity
of red, green and blue colours, and flags for the palette usage. The palette structure is described by the
following code segment:
Track chunk
The Track chunk is organized as follows:
.:. The first 4-character string is the identifier.
.:. The second 4 bytes contain track length.
4.2.5 TWAIN
A standard interface was designed to allow application to interface with different types of input
devices such as scanners, digital still cameras, and so on, using a generic TWAIN interface without
creating device- specific driver. The benefits of this approach areas follows
I. Application developers can code to a single TWAIN specification that allows application to
interface to all TWAIN complaint input devices.
2. Device manufactures can write device drivers for their proprietary devices and, by complying
to the TWAIN specification , allow the devices to be used by all TWAIN-compliant
applications
The Twain architecture defines a set of application programming interfaces (APls) and a protocol to
acquire data from input devices.
It is a layered architecture.
It has application layer, the protocol layer, the acquisition layer and device layer.
Application Layer: A TWAIN application sets up a logical connection with a device. TWAIN does not
impose any rules on the design of an application. However, it set guidelines for the user interface to select
sources (logical device) from a given list of logical devices and also specifies user interface guidelines to
acquire data from the selected sources.
The Protocol Layer: The application layer interfaces with the protocol layer. The protocol layer is
responsible for communications between the application and acquisition layers. The protocol layer
does not specify the method of implementation of sources, physical connection to devices, control of
devices , and other device-related functionality. This clearly highlights that applications are independent
of sources. The heart of the protocol layer, as shown in Figure is the Source Manager. It manages all
sessions between an application and the sources, and monitors data acquisition transactions.
The functionality of the Source Manager is as follows:
Provide a standard API for all TWAIN compliant sources
Provides election of sources for a user from within an application
Establish logical sessions between applications and sources, and also manages essions between
multiple applications and multiple sources
Act as a traffic cop to make sure that transactions and communication are routed to appropriate
sources, and also validate all transactions
Keep track of sessions and unique session identities
Load or unload sources as demanded by an application
Pass all return code from the source to the application
Maintain a default source
The Acquisition Layer: The acquisition layer contains the virtual device driver, it interacts directly
with the device driver. This virtual layerisalsocalledthesource.Thesourcecanbelocalandlogicallyconnected
to a local device, or remote and logically connected to a remote device(i.e.,a device ove rthe network).
The source performs the following functions:
~ Control of the device.
~ Acquisition of data from the device.
~ Transfer of data in agreed (negotiated) format. This can be transferred in native format or
another filtered format.
~ Provision of a user interface to control the device.
The Device Layer: The purpose of the device driver is to receive software commands and control the
device hardware accordingly. This is generally developed by the device manufacturer and shipped with
the device.
NEW WAVE RIFF File Format: This format contains two subchunks:
(i) Fmt (ii) Data.
It may contain optional subchunks:
(i) Fact
(ii) Cue points
(iii)Play list
(iv) Associated datalist.
Fact Chunk: It stores file-dependent information about the contents of the WAVE file.
Cue Points Chunk: It identifies a series of positions in the waveform data stream.
Playlist Chunk: It specifies a play order for series of cue points.
Associated Data Chunk: It provides the ability to attach information, such as labels, to sections of the
waveform data stream.
Inst Chunk: The file format stores sampled sound synthesizer's samples.
Sampling process
Sampling is a process where the analog signal is sampled over time at regular intervals to obtain the
amplitude of the analog signal at the sampling time.
Sampling rate
The regular interval at which the sampling occurs is called the sampling rate.
Digital Voice
Speech is analog in nature and is cOl1veli to digital form by an analog-to-digital converter (ADC). An ADC
takes an input signal from a microphone and converts the amplitude of the sampled analog signal to an 8, 16
or 32 bit digital value.
The four important factors governing the
ADC process are sampling rate resolution linearity and conversion speed.
Sampling Rate: The rate at which the ADC takes a sample of an analog signal. Resolution:
The number of bits utilized for conversion determines the resolution of ADC.
Linearity: Linearity implies that the sampling is linear at all frequencies and that the amplitude tmly
represents the signal.
Conversion Speed: It is a speed of ADC to convert the analog signal into Digital signals. It must be fast
enough.
VOICE Recognition System
Voice Recognition Systems can be classified into three types.
1.Isolated-word Speech Recognition.
2.Connected-word Speech
Recognition. 3.Continuous Speech
Recognition.
MIDI Interconnections
The MIDI IN port of an instrument receives MIDI ncssages to play the instrument's internal synthesizer.
The MIDI OUT port sends MIDI messages to play these messages to an external synthesizer. The MIDI
THRU port outputs MIDI messages received by the MIDI IN port for daisy-chaining external synthesizers.
Architecture
A video frame grabber is used to capture, manipulate and enhance video images.
A video frame grabber card consists of video channel multiplexer, Video ADC, Input look-up table with
arithmetic logic unit, image frame buffer, compression-decompression circuitry, output color look-up table,
video DAC and synchronizing circuitry.
Low Pass Filter: A low pass filter causes blurring of the image and appears to cause a reduction in noise.
High Pass Filter: The high-pass filter causes edges to be emphasized. The high-pass filter attenuates low-
spatial frequency components, thereby enhancing edges and sharpening the image.
Laplacian Filter: This filter sharply attenuates low-spatial-frequency components without affecting and
high-spatial frequency components, thereby enhancing edges sharply.
Frame Processing Frame processing operations are most commonly for geometric operations, image
transformation, and image data compression and decompression Frame processing operations are very
compute intensive many multiply and add operations, similar to spatial filter convolution operations.
Image scaling: Image scaling allows enlarging or shrinking the whole or part of an image.
Image rotation: Image rotation allows the image to be rotated about a center point. The operation can be
used to rotate the image orthogonally to reorient the image if it was scanned incorrectly. The operation can
also be used for animation. The rotation formula is:
pixel output-(x, y) = pixel input (x, cos Q + y sin Q, - x sin Q + Y cos Q)
where, Q is the orientation angle
x, yare the spatial co-ordinates of the original pixel.
Image translation: Image translation allows the image to be moved up and down or side to side. Again, this
function can be used for animation.
The translation formula is:
Pixel output (x, y) =Pixel Input (x + Tx, y + Ty) where Tx and Ty are the horizontal and vertical coordinates.
x, yare the spatial coordinates of the original pixel. Image transformation: An image contains varying
degrees of brightness or colors defined by the spatial frequency. The image can be transformed from spatial
domain to the frequency domain by using frequency transform.
Video Channel Multiplexer: It is similar to the video grabber's video channel multiplexer.
Video Compression and Decompression: A video compression and decompression processor is used to
compress and decompress video data.
The video compression and decompression processor contains multiple stages for compression and
decompression. The stages include forward discrete cosine transformation and inverse discrete cosine
transformation, quantization and inverse quantization, ZigZag and Zero run-length encoding and decoding,
and motion estimation and compensation.
Audio Compression: MPEG-2 uses adaptive pulse code modulation (ADPCM) to sample the audio signal.
The method takes a difference between the actual sample value and predicted sample value. The difference
is then encoded by a 4-bit value or 8-bit value depending upon the sample rate
Analog to Digital Converter: The ADC takes inputs from the video switch and converts the amplitude of a
sampled analog signal to either an 8-bit or 16-bit digital value.
Performance issues for full motion video:
During the capture the video hardware and software must be able to keep up with the output of the camera
to prevent loss of information. The requirements for playback are equally intense although there is no risk of
permanent loss of information. Consider the eg below,
It is a method of attaching multiple drives to a single host adapter. The data is written to the first drive first,
then after filling it, the controller, allow the data to write in second drive, and so on. Meantime Between
Failure (MTBF) = MTBF of single/drivel Total no. of drives.
RAID level 0 provides performance improvement. It is achieved by overlapping disk reads and writes.
Overlapping here means, while segment I is being written to drive 1, segment 2 writes can be initiated for
drive 2.
The actual performance achieved depends on the design of the controller and how it manages disk reads and
writes.
2.RAID Level 1 Disk Mirroring
The Disk mirroring causes two copies of every file to be written on two separate drives. (Data redundancy is
achieved).
These drives are connected to a single disk controller. It is useful in mainframe and networking systems.
Apart from that, if one drive fails, the other drive which has its copy can be used.
Performance: Writing
is slow.
Reading can be speeded up by overlapping seeks.
Read transfer rate and number ofI/O per second is better than a single drive. I/O
transfer rate (Bandwidth) = No. of drives x drive I/O transfer rate
I / Otransferr ate
No of I/O’s Per second = Averagesizeoftransfe r
controller
Segment 0 Segment 0
Host Adapter
.
organization of bit interleaving for RAID level2
It provides the ability to handle very large files, and a high level of integrity and reliability. It is good for
multimedia system. RAID Level 2 utilizes a hamming error correcting code to correct single-bit errors and
doublebit errors.
Drawbacks:
(i) It requires multiple drives for error correction (ii) It is an expensive approach to data redundancy. (iii) It
is slow.
Uses: It is used in multimedia system. Because we can store bulk of video and audio data.
5.RAID Level-4 Sector Interleaving: Sector interleaving means writing successive sectors of data on
different drives.
As in RAID 3, RAID 4 employs multiple data drives and typically a single dedicated parity drive. Unlike
RAID 3, where bits of data are Written to successive disk drives, an Ri\ID 4, the first sector of a block of
data is written to the first drive, the second sector of data is written to the secohd drive, and so on. The data
is interleaved at the data level.
RAID Leve1-4 offers cost-effective improvement in performance with data.
RAID Level-5 Block Interleaving: In RAID LevelS, as in all the other RAID systems, multiple drives are
connected to a disk array controller.
The disk array controller contains multiple SCSI channels.
A RAID 5 system can be designed with a single SCSI host adapter with multiple drives connected to the
single SCSI channel.
Unlike RAID Level-4, where the data is sector-interleaved, in RAID Level-5 the data is block-interleaved.
Host Adapter
1. CD-ROM
Physical Construction of CD ROMs:
It consists of a polycarbonate disk. It has 15 mm spindle hole in the center. The polycarbonate substrate
contains lands and pits.
The space between two adjacent pits is called a land. Pits, represent binary zero, and the transition from land
to pits and from pits to land is represented by binary one.
The polycarbonate substrate is covered by reflective aluminium or aluminium alloy or gold to increase the
reflectivity of the recorded surface. The reflective surface is protected by a coat oflacquer to prevent
oxidation. A CD-ROM consists of a single track which starts at the center from inside and spirals outwards.
The data is encoded on this track in the form of lands and pits. A single track is divided into equal length
sectors and blocks. Each sector or block consists of2352 bytes, also called a frame. For Audio CD, the data
is indexed on addressed by hours, rninutes, seconds and frames. There are 75 frames in a second.
2. Mini-Disk
Mini-Disk for Data is known as MD-Data. It was developed by Sony Corporation. It is the data version of
the new rewritable storage format. It can be used in three formats to support all users.
A premastered optical disk.
A recordable magneto-optical disk.
A hybrid of mastered and
recorded.
Its size is 2.5 inch. It provides large capacity. It is low cost. It is used in multimedia applications.A MD
demands as a replacement for audio cassette. A 2-1/2 inch MD-Data disk stores 140Mbytes of data and
transfer data at 150Kbytes/sec. the figure shows the format for MD-Data standard.
Possible Questions
2 - MARKS
1. What do you mean by compression and decompression.
2. State the types of compression.
3. What is Huffman Encoding?
4. Write short notes on Packpits Encoding?
5. Write down about CCITT group 3-I-D compression.
6. What is the necessity of k factor in CCITT Group-3 2D compression.
7. What are the steps involved for pseudo code to code line in data formatting for CCITT
Group-3 2D?
8. State the advantages of CCITT Group-3 2D.
9. Write short notes on YUV representation.
10. What is Discrete Cosine Transform?
11. Write short notes on Quantization.
12. What are the color characteristics.
13. What is predictive lossless encoding for?
14. What is the role of entropy encoding in data compression.
15. What is Macroblock in MPEG?
16. Define the term "Motion Compensation".
17. Write short notes on MPEG encoder.
18. What do you mean by vector quantization in MPEG standard.
19. What do you mean by audio compression.
20. Write short notes on fractal compression.
21. List the key formats available in Rich Text Format.
22. What is TIFF File Format?
23. What is TIFF tag ID?
24. What is full motion video?
25. What are the advantages of dye sublimation printer.
26. List the features of scanner.
27. What is ADC stands for?
28. Write short notes on histogram sliding. What is dithering?
29. What do you mean by RIFF-chunk
30. Define LIST chunk
31. Describe TWAIN architecture
32. How do scanner work.
33. Give some of the visual display technology standards
34. State the four important factors that governs ADC process.
35. State the three types of Voice Recognition system.
36. Write shor1 notes on MIDI.
37. What is digital camera? State its advantages.
38. What do you mean by Spatial Filter Processing.
39. Write short notes on disk spanning.
40. What is RAID?
41. State the uses of magnetic storage in multimedia.
42. Give brief notes on CD-ROM.
43. What are the three formats of minidisk?
44. What is juke box. Give other name used for juke box.
45. What are the four types of storage in cache organization for hierarchical storage systems.
16 Marks
1. Explain briefly about binary image compression schemes. (16)
2.(a) Explain the characteristics of color in detail. (10)
(b) State the relationship between frequency and wave length! in measuring radian energy.(6)
3.Write about JPEG in detail. (16)
4.(a) Explain DCT in detail. (12)
(b) Write short notes on zig zag sequence. (4)
5.State the requirements for full motion video compression in detail. (16)
6.What is MPEG? Discuss it in detail. (16)
7.Explain all data and file format standards in detail. (16)
8.Give a detailed description about voice recognition
(b) What is DIB stands for? (4)
9.Explain the different types of messages that are used with M1DIcommunication protocol.
(16) 10.Explain the TWAIN architecture with neat diagram. (16)
11.Explain some ofthe video and image display system technique in detail. (16)
12.Give short notes for the following standards: (i) MDA, (3), (ii) CGA (3), (iii) MGA (3); (iv) VGA (4),
(v) XGA (3).
14. Explain in detail about voice recognition system. (12)
(b Write down the formula for voice recognition accuracy and substitution error.
16. a) Describe hierrarchical storage management in detail.
(b) What is migration? Write short notes about it.
17. (i) What is Optical Disk Library? Explain it.
(ii) Discuss about Cache Management for storage system.