MPEG-1: Coding of Moving Pictures and Associated Audio For Digital Storage Media at Up To About 1,5 Mbit/s
MPEG-1: Coding of Moving Pictures and Associated Audio For Digital Storage Media at Up To About 1,5 Mbit/s
MPEG-1: Coding of Moving Pictures and Associated Audio For Digital Storage Media at Up To About 1,5 Mbit/s
Coding of moving pictures and associated audio for digital storage media at up to
about 1,5 Mbit/s
Part 1 addresses the problem of combining one or more data streams from the video
and audio parts of the MPEG-1 standard with timing information to form a single
stream as in Figure 1 below. This is an important fuction because, once combined into
a single stream, the data are in a form well suited to digital storage or transmission.
Figure 1 -- Prototypical ISO/IEC 11172 decoder.
Part 2 specifies a coded representation that can be used for compressing video
sequences - both 625-line and 525-lines - to bitrates around 1,5 Mbit/s. Part 2 was
developed to operate principally from storage media offering a continuous transfer
rate of about 1,5 Mbit/s. Nevertheless it can be used more widely than this because the
approach taken is generic.
A number of techniques are used to achieve a high compression ratio. The first is to
select an appropriate spatial resolution for the signal. The algorithm then uses block-
based motion compensation to reduce the temporal redundancy. Motion compensation
is used for causal prediction of the current picture from a previous picture, for non-
causal prediction of the current picture from a future picture, or for interpolative
prediction from past and future pictures. The difference signal, the prediction error, is
further compressed using the discrete cosine transform (DCT) to remove spatial
correlation and is then quantised. Finally, the motion vectors are combined with the
DCT information, and coded using variable length codes.
Figure 2 below illustrates a possible combination of the three main types of pictures
that are used in the standard.
Figure 2 -- Example of temporal picture structure.
Part 3 specifies a coded representation that can be used for compressing audio
sequences - both mono and stereo. The algorithm is illustrated in Figure 3 below.
Input audio samples are fed into the encoder. The mapping creates a filtered and
subsampled representation of the input audio stream. A psychoacoustic model creates
a set of data to control the quantiser and coding. The quantiser and coding block
creates a set of coding symbols from the mapped input samples. The block 'frame
packing' assembles the actual bitstream from the output data of the other blocks, and
adds other information (e.g. error correction) if necessary.
Figure 3 -- Basic structure of the audio encoder
Part 4 specifies how tests can be designed to verify whether bitstreams and decoders
meet the requirements as specified in parts 1, 2 and 3 of the MPEG-1 standard. These
tests can be used by:
Part 5, technically not a standard, but a technical report, gives a full software
implementation of the first three parts of the MPEG-1 standard. The source code is
not publicly available.