100% found this document useful (1 vote)
81 views59 pages

Video Theory

Video is the electronic capture and dissemination of still images in rapid succession to create the illusion of motion, synchronized with sound. Video can be either analog or digital. Analog video records light and sound as a fluctuating electrical wave, directly mirroring the original stimulus as a continuous signal. Digital video samples the analog wave and converts it to discrete numbers (zeros and ones), representing rather than directly copying the stimulus as a discontinuous signal. Digital video offers advantages like resistance to data distortion and the ability to manipulate images. The individual pictures that make up video are called frames, composed of electrically excitable pixels. Factors that define digital video include frame rate, color resolution, spatial resolution, and image quality.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
100% found this document useful (1 vote)
81 views59 pages

Video Theory

Video is the electronic capture and dissemination of still images in rapid succession to create the illusion of motion, synchronized with sound. Video can be either analog or digital. Analog video records light and sound as a fluctuating electrical wave, directly mirroring the original stimulus as a continuous signal. Digital video samples the analog wave and converts it to discrete numbers (zeros and ones), representing rather than directly copying the stimulus as a discontinuous signal. Digital video offers advantages like resistance to data distortion and the ability to manipulate images. The individual pictures that make up video are called frames, composed of electrically excitable pixels. Factors that define digital video include frame rate, color resolution, spatial resolution, and image quality.
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 59

Video

Prepared by: Haval Akrawi

1
What is Television?

Television is the electronic capture and dissemination


of still images in rapid enough succession to create
the illusion of motion, synchronized with sound.

What is Video?

Successive images, resulting in the appearance of


motion, created and stored electronically, either
analog or digital.
2
Analog
 A wave is recorded or used in its original form

 Light or sound are converted to a fluctuating electrical wave that is


directly recorded, usually to a magnetic tape medium, mirroring the
original stimulus.

 It produces an electrical copy of an original stimulus.

 Analog signal fluctuates exactly like the original stimulus.

 Analog signal is continuous.

3
Digital
 The analog wave is sampled at some interval, and then turned into
numbers that are stored in the digital device

 Light and sound are recorded not as an identical copy of the original
stimulus, but as discrete on-and-off pulses, zeros and ones, binary
digits.

 It is a representation rather than a copy.

 Discontinuous. Signal is sampled.

 Advantage: Resists data distortion and error in duplication. No


generational loss.

 Advantage: Allows for manipulation of sound and image.


4
Digital

5
Basic Image Formation
 The individual pictures that make up video are “frames.”

6
Basic Image Formation
 The individual pictures that make up video are “frames.”
 The frames are comprised of millions of electrically excitable “pixels”
(picture elements)

CRT (Cathode Ray Tube)

LCD (Liquid-Crystal Display)

7
LED (Light-Emitting Diode)
Basic Image Formation
 The individual pictures that make up video are “frames.”
 The frames are comprised of millions of electrically excitable “pixels”
(picture elements)
 For color, each pixel is comprised of three parts (red, blue, green)

8
Basic Image Formation
 The individual pictures that make up video are “frames.”
 The frames are comprised of millions of electrically excitable “pixels” (picture
elements)
 For color, each pixel is comprised of three dots (red, blue, green)
 Individual images are drawn by scanning along these pixels from left to right, top
to bottom

9
Basic Image Formation
 Scanning

High speed video showing the fluorescent screen of


CRT television being scanned by electron gun.
10

http://youtu.be/zVS6QewZsi4
Basic Image Formation
 Interlaced Scanning - Two fields comprise a frame.

11
Basic Image Formation

 Interlaced Scanning

12

Misaligned fields
Basic Image Formation
 Progressive Scanning (Computers and Most Digital Video)
 Refresh Rate: Frames scanned per second.

13
Basic Image Formation
 Progressive v Interlaced Scanning

14
Basic Image Formation
 Progressive v Interlaced Scanning

15
Basic Image Formation
Standard Television

 480i (480 horizontal lines visible, interlaced; 30 frames per second)

16
Basic Image Formation

Digital Television (DTV)


 Higher Picture Resolution
 Truer Color
 Wider Contrast Ratio

Four prominent systems:


 480p (progressive, 480 visible lines, 60 frames per second)
 720p (progressive, 720 visible lines, 60 frames per second,
High Definition Television [HDTV]
 1080i (interlaced, 60 fields/30 frames per second, High
Definition Television [HDTV]
 1080p (progressive, 60 frames per second, HDTV
17
Basic Image Formation

18
Basic Image Formation

720p 1080p 19
Basic Image Formation

20
Basic Image Formation

24p - 1080 lines of resolution, 24 frames per second.

 For use in conjunction with motion picture film


 Or to create a “film look.”

21
Introduction to Digital Video
• Video is a stream of data composed of discrete frames,
containing both audio and pictures
• Continuous motion produced at a frame rate of 15 fps or
higher
• Traditional movies run at 24 fps
• TV standard in USA (NTSC) uses ≈ 30 fps
With digital video, four factors have to be kept in mind.
# Frame rate
# Colour Resolution
# Spatial Resolution
# Image Quality
22
Frame Rate
The standard for displaying any type of non-film video is 30 frames per second
(film is 24 frames per second). Additionally these frames are split in half (odd
lines and even lines), to form what are called fields.

When a television set displays its analogue video signal, it displays the odd
lines (the odd field) first. Then is displays the even lines (the even field).

Each pair forms a frame and there are 60 of these fields displayed every second
(or 30 frames per second). This is referred to as interlaced video.

Fragment of the "matrix" sequence (2 After processing the fragment on the left by the FRC filter
frames) the frame rate increased 4 times
Colour Resolution
This second factor is a bit more complex. Colour resolution refers
to the number of colours displayed on the screen at one time.
Computers deal with colour in an RGB (red-green-blue) format,
while video uses a variety of formats. One of the most common
video formats is called YUV.

This test table was used to estimate the colour resolution.


First we determine the border when one of the colours on
the resolution chart disappears, and colour sharpness is found
on the scale on the right.
Spatial Resolution
The third factor is spatial resolution - or in other words, "How big is the picture?".
Since PC and Macintosh computers generally have resolutions in excess of 640 by 480,

The National Television Standards Committee ( NTSC) standard used in North America
and Japanese Television uses a 768 by 484 display.

The Phase Alternative system (PAL) standard for European television is slightly larger
at 768 by 576.

Spatial resolution is a parameter that shows how


many pixels are used to represent a real object
in digital form. Fig. 2 shows the same colour
image represented by different spatial
resolution. Left flower have a much better
resolution that right one

Fig. 1 Fig. 2
Image quality
The final objective is video that looks acceptable for your application.
For some this may be 1/4 screen, 15 frames per second (fps), at 8 bits
per pixel.
Other require a full screen (768 by 484), full frame rate video, at 24
bits per pixel (16.7 million colours).
Video Input Formats

 AVI  MPEG
 ActiveMovie  QuickTime
 Cinepak  RealVideo
 Indeo  Video for Windows
 motion-JPEG  XGA

27
Gamma Correction
• Gamma correction provides displaying an image accurately on a computer screen.
• Images which are not properly corrected can look either bleached out, or too dark.
• Trying to reproduce colors accurately also requires some knowledge of gamma.
• Varying the amount of gamma correction changes not only the brightness, but also
the ratios of red to green to blue.

Sample Input

Graph of Correction L' = L ^ (1/2.5)


Gamma Corrected Input

Monitor

Output 28
YUV is a colour space typically used as part of a colour image pipeline. The
Y component determines the brightness of the colour, the U and V
components determines the actual colour itself. Y ranges from 0 to 1 (or 0
to 255 in digital formats), and U and V range from -0.5 to 0.5 (or -128 to
127 in signed digital form, or 0 to 255 in unsigned form).

Y' value of 0 Y' value of 0.5 Y' value of 1


29
Color Models in Videos

 YUV
 YIQ
 YCbCr

30
YUV

 YUV coding used for PAL.


 It codes a luminance signal(for gamma correction) equal to y’.

31
YUV

 YUV coding used for PAL.


 It codes a luminance signal(for gamma correction) equal to y’.
 Chrominance refers to the difference between a color and a reference white
at a same luminance.

32
YUV

 YUV coding used for PAL.


 It codes a luminance signal(for gamma correction) equal to y’.
 Chrominance refers to the difference between a color and a reference white
at a same luminance.
 It can be represent by the color differences U , V.

33
YUV

 We go the (Y’,U,V) to (R,G,B) by inverting the matrix.


 Y’ is equal to the same value R’.(coz sum of coefficient is 1)
 For black & white image Chroma (UV) is zero.

34
YUV

CIL:

35
YIQ

 Y’IQ is used in NTSC.


 I for in-phase chrominance.
 Q for quadrature chrominance.

36
YIQ

 YIQ used in NTSC(National Television System Committee)

37
Y component

I component

Q component

38
YCbCr

 Y is the Luma component and Cb and Cr are the blue difference and red
difference Chroma component.
 Used for digital video encoding digital camera.
 YCbCr is used in JPEG and MPEG.

39
YCbCr

40
Color Space – Comparison
Color Color Primary Used for Pros and
space mixing parameters cons
RGB Additive Red, Easy but wasting
Green, Blue bandwidth
CMYK Subtractive Cyan, Magenta, Printer Works in pigment
Yellow, Black mixing
YCbCr additive Y(luminance), Video encoding, Bandwidth efficient
YPbPr Cb(blue chroma), digital camera
Cr(red chroma)
YUV additive Y(luminance), Video encoding Bandwidth efficient
U(blue chroma), for PAL
V(red chroma)

YIQ additive Y(luminance), Video encoding Bandwidth efficient


I(rotated from U), for NTSC
Q(rotated from V)
41
Video Signals(Types of Analog Video)

Video signals are organized into three different ways :


 Component video
 Three components: Y (luminance), U and V (color)
 Often use in production and post-production
 Composite video
 Combine three components into a signal
 Color component (U and V) is allocated half bandwidth as the luminance (Y)
 Often use in transmission
 S-video
 Separates the luminance from the two color (total two signals)

42
Component Video
 Make use of three separate signals for red green and blue
image plane.(Component Video)
 This kind of system has three wires (Connectors) for
connecting camera or other devices.
 Color signal not restricted to always RGB
 We can form three signal via a luminance-chrominance
transformation (YIQ or YUV)
 There are no crosstalk between three different channels.

43
Component Video

44
Composite Video
 Color(chrominance) and intensity(Luminance) signal are
mixed into single carrier wave.
 Chrominance is a composite of two color component(I and Q
or U and V).
 In NTSC TV, I and Q combine into composite chrome signal.
 When connecting to TV or VCR, composite video uses one wire
and video color signals are mixed, not sent separately.

45
S- Video

 S-Video uses two wires: One for luminance and other for and other for
composite chrominance.
 There is less crosstalk.

46
Digital Video Standards

 CCIR 601 (Rec. ITU-R BT.601)


 specifies the image format, and coding for digital television signals

Parameter Value

YUV encoding 4:2:2


Sampling frequency for Y (MHz) 13.5
Sampling frequency for U and V (MHz) 6.75
No of samples per line 720
No of levels for Y component 220
No of levels for U,V components 225
47
NTSC(National Television System Committee)

48
PAL
 PAL stands for Phase Alternate Lines

49
Perplexing
NTSC System
123 …

123 …
480 480

1 2 3 … 640 1 2 3 … 720
Analog to digital CCIR 601 standard
Pixels are square Pixels are not square
50
CCIR 601 Sampling

4 : 2 : 2 sampling 4 : 2 : 0 sampling 4 : 1 : 1 sampling


(co-site) (not co-site) (co-site)

Y samples

CB and CR samples
51
Introduction to Video Compression
 A video consist of time-ordered sequence of frames(Images).
 An obvious solution to video compression would be predictive coding based on
previous frame.
 Exploit spatial redundancy within frames (like JPEG: transforming, quantizing,
variable length coding)
 Exploit temporal redundancy between frames

52
Compression in the time domain
 difference between consecutive frames is often small
 remove inter-frame redundancy
 sophisticated encoding, relatively fast decoding

53
Difference Frames

 Differences between two frames can be caused by


 Camera motion: the outlines of background or
stationary objects can be seen in the Diff Image
 Object motion: the outlines of moving objects can be
seen in the Diff Image
 Illumination changes (sun rising, headlights, etc.)
 Scene Cuts: Lots of stuff in the Diff Image
 Noise

54
Motion Estimation

 Determining parameters for the motion descriptions


 For some portion of the frame, estimate its movement
between 2 frames- the current frame and the reference
frame
 What is some portion?
 Individual pixels (all of them)?
 Lines/edges (have to find them first)
 Objects (must define them)
 Uniform regions (just chop up the frame)

55
Motion Estimation

56
Motion Compensation
 Divide each frame into macroblocks of 16  16 pixels
 Predict where the corresponding macroblock in next frame
 Try all possible displacements within a limited range
 Choose the best match
 Construct difference frame by subtracting each macroblock from its
predicted counterpart
 Keep the motion vectors describing the predicted displacement of
macroblocks between frames

57
Picture Type
 I (intra) pictures
 Code without reference to other pictures
 Low compression rate
 P (predicted) pictures
 Code using motion compensated prediction from a past I or P picture
 Higher compression rate than I picture
 B (bidirectional-predicted) pictures
 Code bidirectional interpolation between the I or P picture which
preceded & followed them
 Highest compression rate

All are compressed using the MPEG version of JPEG compression


58
I P I P I

B B B B B B B B

01 02 03 04 05 06 11 12 13 14 15 16 21

Group of Pictures (GOP)

An MPEG sequence in display order

I P I P I

B B B B B B B B

01 04 02 03 11 05 06 14 12 13 21 15 16

An MPEG sequence in bitstream order (decode order) 59

You might also like