CCD Sensors and Camera Systems, Second Edition
CCD Sensors and Camera Systems, Second Edition
and DISPLAYS
Second edition
TABLE of CONTENTS
1. INTRODUCTION . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1
1.1. Solid State Detectors . . . . . . . . . . . . . . . . . . . . . . . . . . . 2
1.2. Imaging System Applications . . . . . . . . . . . . . . . . . . . . . . 3
1.2.1. General Imagery . . . . . . . . . . . . . . . . . . . . . . . . . 6
1.2.2. Machine Vision . . . . . . . . . . . . . . . . . . . . . . . . . . 7
1.2.3. Scientific Applications . . . . . . . . . . . . . . . . . . . . . . 8
1.2.4. Military Applications . . . . . . . . . . . . . . . . . . . . . . 8
1. 3. Configurations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
1.4. Image Quality . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1.5. Pixels, Datels, Disels, and Resets . . . . . . . . . . . . . . . . . . . 14
1. 6. References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 17
xiii
xiv CCD ARRAYS, CAMERAS, and DISPLAYS
5. CAMERAS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
5 .1. Camera Operation . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
5.2. Video Formats . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149
5.2.1. Video Timing . . . . . . . . . . . . . . . . . . . . . . . . . . 151
5.2.2. Broadcast/Non-broadcast Design . . . . . . . . . . . . . . 155
5.2.3. Component/Composite Signals . . . . . . . . . . . . . . . . 156
5.2.4. IRE Units . . . . . . . . . . . . . . . . . . . . . . . . . . . . 157
5.2.5. Digital Television . . . . . . . . . . . . . . . . . . . . . . . 159
TABLE of CONTENTS xv
APPENDIX o o o 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 371
Aol. Effective Focal Length 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 371
A020 f-Number 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 371
Ao3o Reference 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 372
INDEX 0 0 0 0 0 o 0 o o o 0 0 0 0 0 0 o 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 0 373
SYMBOL LIST
gamma numeric
wavelength
average wavelength
maximum wavelength
minimum wavelength
selected wavelength
wavelength of peak response
xviii
SYMBOL LIST xix
u,s display 1/e spot size in image space mm
UR rms value of random motion rms mm
UsP<lT display spot 1/e intensity size mm
UT component of three-dimensional noise model rms
Urn component of three-dimensional noise model rms
UTVH component of three-dimensional noise model rms
Uv component of three-dimensional noise model rms
d detector width mm
dec detector pitch mm
dccH horizontal detector pitch mm
dccv vertical detector pitch mm
dH horizontal detector size mm
do selected target dimension mm
dT target detail mm
dv vertical detector size mm
D observer to display distance m
Do optical diameter mm
DASH horizontal detector angular subtense mrad
DASv vertical detector angular subtense mrad
index numeric
spectral radiant exitance W/(m2-1-1m)
optical magnification numeric
spectral power W/1-1m
spectral photon exitance photons/(s-m2-1-1m)
SYMBOL LIST xxi
n index numeric
IlcATIIODE photons incident onto intensifier photocathode numeric
Do ARK number of dark electrons numeric
DDETECTOR number of photons incident onto detector numeric
n. number of electrons numeric
OEYE number of photons incident onto eye numeric
DIMAGE number of photons incident onto image plane numeric
llLENS number of photons incident onto lens numeric
OMCP number of photons incident on microchannel plate numeric
n,. number of photoelectrons numeric
llpe-8 background photoelectrons numeric
U,e-T target photoelectrons numeric
llPHOTON-CCD number on photons incident onto CCD numeric
DREAD pixels between active array and sense node numeric
llsCREEN photons incident onto intensifier screen numeric
DwiNDOW number of photons incident onto intensifier numeric
N index numeric
Nso Johnson's 50% probability cycles
NoETECTORs total number of detectors numeric
NH number of horizontal detectors numeric
NUNE number of raster lines numeric
NT equivalent number of cycles on target cycles
NTD! number of TDI elements numeric
NTRANS number of charge transfers numeric
NTV display resolution TVLIPH
Nv number of vertical detectors numeric
NwELL charge well capacity numeric
R range to target m
R, distance from lens to source m
R2 distance from lens to detector m
RAVE spectral averaged responsivity V/(JL-l-cm- 2) or DN/(JL-l-cm2)
R. average spectral responsivity A/W
R.(t-) spectral responsivity A/W
Rp peak responsivity A/W
RPHOTOMETRIC responsivity V/lux
Rq spectrally averaged quantum efficiency numeric
R.(t-) spectral quantum efficiency numeric
R. raster resolution lines/mm
RTVL horizontal display resolution TVLIPH
Rv responsivity A/lumen
RvERTICAL vertical display resolution lines
w target width m
WMONITOR monitor width m
X horizontal distance m
y vertical distance m
SYMBOL LIST XX111
The heart of the solid state camera is the solid state array. It provides the
conversion of light intensity into measurable voltage signals. With appropriate
timing signals, the temporal voltage signal represents spatial light intensities.
When the array output is amplified and formatted into a standard video format,
a solid state camera is created. Because CCDs were the first solid state
detectors, cameras are popularly called CCD cameras even though they may
contain charge injection devices (CIDs) or complementary metal-oxide-
semiconductors (CMOS) as detectors. These are solid state cameras.
The array specifications, while the first place to start an analysis, are only
part of the overall system performance. The system image quality depends on
all the components. Array specifications, capabilities, and limitations are the
basis for the camera specifications. Camera manufacturers cannot change these.
A well-designed camera will not introduce additional noise nor adversely affect
image quality provided by the array.
Because most sensors operating in the visible region use a CCD type
architecture to read the signal, they are popularly called CCD cameras. For
these devices, charge generation is often considered as the initial function of the
CCD. More explicitly, these cameras should be called solid state cameras with
a CCD readout.
INTRODUCTION 3
Hybrid arrays avoid some pitfalls associated with growing different materials
on a single chip and provide a convenient bridge between well-developed but
otherwise incompatible technologies. HgCdTe (sensitive to 8 to 12 p.m radiation)
is bump bonded to a CCD readout using indium as the contact and, as such, is
a hybrid array.
Recently, active pixels have been introduced. These devices are fabricated
with complementary-metal-oxide-semiconductor (CMOS) technology. The
advantage is that one or more active transistors can be integrated into the pixel.
As such, they become fully addressable (can read selected pixels) and can
perform on-chip image processing.
There are four broad applications: general imagery (which includes both
professional television broadcast and consumer camcorder systems), machine
vision, scientific, and military. Trying to appeal to all four applications,
manufacturers use words such as low noise, high frame rate, high resolution,
4 CCD ARRAYS, CAMERAS, and DISPLAYS
reduced aliasing, and high sensitivity. These words are simply adjectives with
no specific meaning. They only become meaningful when compared (i.e.,
camera A has low noise compared to camera B).
Table 1-1 lists several design categories. While the requirements vary by
category, a camera may be used for multiple applications. For example, a
consumer video camera often is adequate for many scientific experiments. A
specific device may not have all the features listed or may have additional
features not listed. The separation between professional broadcast, consumer
video, machine vision, scientific, and military devices becomes fuzzy as
technology advances.
Color cameras are used for professional television, camcorder, and film
replacement systems. With machine vision systems, color cameras are used to
verify the color consistency of objects such as printed labels or paint mixture
colors. While color may not be the primary concern, it may be necessary for
image analysis when color is the only information to distinguish boundaries.
While consumers demand color camera systems, this is not true for other
applications. Depending on the application, monochrome (black and white)
cameras may be adequate. A monochrome camera has higher sensitivity and
therefore is the camera of choice in low light level conditions.
Image enhancement applies to those systems that have a human observer and
helps the observer extract data. Some images belong to a small precious data set
(e.g., remote sensing imagery from interplanetary probes). The images must be
processed repeatedly to extract every piece of information. Some images are part
of a data stream that are examined once (e.g., real-time video) and others have
become popular and are used routinely as standards. These include the three-bar
or four-bar test patterns, Lena, and the African mandrill.
The camera cannot perfectly reproduce the scene. The array spatially
samples the image, and noise is injected b:y the array electronics. Spatial
sampling creates ambiguity in target edge location and produces moire patterns
when viewing periodic targets. While this is a concern to scientific and military
applications, it typically is of little consequence to the average professional
television broadcast and consumer markets.
Table 1-1
DESIGN GOALS
DESIGN I
GENERAL IMAGERY MACHINE VISION SCIENTIFIC MILITARY
CATEGORY
i
Image processing Menu-driven Application
Gamma correction Application specific
algorithms multiple options specific
Application specific
Image processing Real time not
Real time with emphasis on high Real time
time usually required
speed operation
Application specific -
Sensitivity High contrast targets
not necessarily an issue Low noise Low noise
(high signal-to- (noise not necessarily a
noise ratio) dominant design factor)
because lighting can be
controlled
operation operation
~
::a
g
~
~
~
Vl
6 CCD ARRAYS, CAMERAS, and DISPLAYS
Table 1-2
SYSTEM DESIGN CONSIDERATIONS
TRANSMISSION
ENVIRONMENT CAMERA DISPLAY
and STORAGE
Noise
Target size Frame rate Data rate
Hard copy
Target reflectance Sensitivity Type of storage
Soft copy
Distance to target Detector size Storage capacity
Resolution
Atmospheric transmittance Array format \'ideo compression
Lighting conditions Dynamic range
Color capability
Because solid state cameras have largely replaced image vacuum tubes, the
terminology associated with these tubes is also used with solid state cameras.
For example, compared to image vacuum tubes, solid state cameras have no
image burn-in, no residual imaging, and usually are not affected by
microphonics.
data much faster than the human. Many operations can be performed faster,
cheaper, and more accurately by machines than by humans. Machine vision
systems can operate 24 hours a day without fatigue. They operate consistently
whereas variability exists among human inspectors. Furthermore, these cameras
can operate in harsh environments that may he unfriendly to humans (e.g.,
extreme heat, cold, or ionizing radiation).
Low noise means low-dark current and low readout noise. The dark current
can be minimized by cooling the CCD. Long integration times can increase the
signal value so that the readout noise is small compared to the photon shot noise.
Although low light level cameras have many applications, they tend to be
used for scientific applications. There is no industry-wide definition of a "low
light level" imaging system. To some, it is simply a solid state camera that can
provide a usable image when the lighting conditions are less than 1 lux. To
others, it refers to an intensified camera and is sometimes called a low light
level television (LLLTV) system. An image intensifier amplifies a low light
level image so that it can be sensed by a solid state camera. The image-
intensifier/CCD camera combination is called an intensified CCD or ICCD. The
image intensifier provides tremendous amplification but also introduces
additional noise.
Solid state cameras are popular because of their ruggedness and small size.
They can easily be mounted on remotely piloted vehicles. They are replacing
wet-film systems used for mapping and photo interpretation.
INTRODUCTION 9
1.3. CONFIGURATIONS
Imaging systems for the four broad application categories may operate in a
variety of configurations. The precise setup depends on the specific
requirements. Figure 1-1 is representative of a closed circuit television system
where the camera output is continuously displayed. The overall image quality
is determined by the camera capability (which is based on the array
specifications), the bandwidth of the video format used (e.g., EIA 170, NTSC,
PAL, or SECAM), the display performance, and the observer. EIA 170 was
formerly called RS 170 and the NTSC standard is also called RS 170A.
Camera r- Transmitter ~
Link
Receiver r-- Display 1>
Observer
L----'
Camera
Video
Recorder
Display 1>
....__ ____, Observer
Figure 1-3. Imagery can be stored on video tape. However, the recorder'
circuitry may degrade the image quality.
For scientific applications, the camera output is digitized and then processed
by a computer (Figure 1-4). After processing, the image may be presented on
a monitor (soft copy), printed (hard copy), or stored. The digital image can also
be transported to another computer via the Internet, local area net, floppy disc,
or magnetic tape. For remote applications, the digital data may be stored on a
digital recorder (Figure 1-5). 14- 16
Perhaps the most compelling reason for adopting digital technology is the
fact that the quality of digital signals remains intact through copying and
reproduction unless the signals are deliberately altered. Digital signal
"transmission" was first introduced into tape recorders. Because a bit is either
present or not, multiple generation copies retain high image quality.
1>
.___ ___,~ Observer
Camera Computer
Digital Digital
Camera - Video Video Computer
Recorder Recorder
Figure 1-5. Digital systems can provide very high quality imagery.
Electronic bandwidth limitations may impose data compression
requirements. Data compression may alter the image but this alteration
may not be obvious on general imagery.
When it comes to displaying images (either hard copy or soft copy), the
range of the camera digitizer is often greater than the display device. A camera
may offer 12, 14, or 16 bits whereas the display may only be 8 bits. Here, a
look-up table is employed to match the camera output to the display. This may
be a simple linear relationship or specific expansion of a part of the camera's
gray scale. False color can also be introduced to enhance the displayed image.
False color is often useful for the human observer but serves no function in
machine vision systems.
12 CCD ARRAYS, CAMERAS, and DISPLAYS
While a machine vision system does not require a monitor (Figure 1-6), a
monitor is often used during system setup and for diagnostic evaluation. That is,
a computer algorithm compares an image to a stored standard. If the target does
not compare favorably to the standard, then the process is changed. For
example, this may mean sending a rejected item for rework or changing the light
intensity.
Camera
H Computer
Il __ Process
Control
Many formulas exist for predicting image quality. Each is appropriate under
a particular set of viewing conditions. These expressions are typically obtained
from empirical data in which multiple observers view many images with a
known amount of degradation. The observers rank order the imagery from worst
to best and then an equation is derived that relates the ranking scale to the
amount of degradation.
Early metrics were created for film-based cameras. Image quality was
related to the camera lens and film modulation transfer functions (MTFs). With
the advent of televisions, image quality centered on the perception of raster lines
and the minimum SNR required for good imagery. Here, it was assumed that
the studio camera provided a perfect image and only the receiver affected the
image quality.
INTRODUCTION 13
Many tests have provided insight into image metrics that are related to image
quality. Most metrics are related to the system MTF, resolution, or the signal-
to-noise ratio. In general, images with higher MTFs and less noise are judged
as having better image quality. There is no single ideal MTF shape that provides
best image quality.
Digital processing is used for image enhancement and analysis. Because the
pixels are numerical values in a regular array, mathematical transforms can be
applied to the array. The transform can be applied to a single pixel, group of
pixels, or the entire image. Many image processing designers think of images
as an array of numbers that can be manipulated with little regard to who is the
final interpreter.
With a solid state camera system, the lens, array architecture, array noise,
and display characteristics all affect system performance. Only an end-to-end
assessment will determine the overall image quality. For example, there is no
advantage to using a high quality camera if the display cannot produce a faithful
image. Often, the display is the limiting factor in terms of image quality and
resolution. No matter how good the camera is, if the display resolution is poor,
then the overall system resolution is poor. Only if the display's contrast and
spatial resolution are better than the camera will the camera image quality be
preserved.
where .:11 is the intensity difference between the target and its immediate
background. The factor k is a proportionality constant that depends on the
specific camera characteristics (aperture diameter, focal length, quantum
efficiency, etc.).
14 CCD ARRAYS, CAMERAS, and DISPLAYS
When the perceived SNR is above a threshold value, the target is just
perceived. Selecting a threshold SNR and solving the equation for ill provides
the minimum resolvable contrast (MRC). When the MRC is coupled with the
target description and atmospheric transmittance effects, the target range can be
predicted.
The overall system may contain several independent sampling systems. The
array spatially samples the scene, the computer may have its own digitizer, and
the monitor may have a limited resolution. A monitor "pixel" may or may not
represent a "pixel" in camera space. The designer and user must understand the
differences among the sampling lattices.
Electronic imaging systems are more complex and several sampling lattices
are present. The detector output represents a sampling of the scene. The detector
output is digitized and placed in a memory. After image processing, the data are
sent to a display medium. Although the display medium provides an analog
signal, it is typically digitally controlled.
Each device has its own minimum sample size or primary element. Calling
a sample a pixel or pel (picture element) does not seem bad. Unfortunately,
there is no a priori relationship between the various device pixels. The various
digital samples in the processing path are called pixels, datels, and disels
(sometimes called dixels). For analog systems, the minimum size is the resel
(Table 1-3). Each "-el" is mapped onto the next (Figure 1-7). However, there
is no standard definition for a resel. For optical systems it may be the Airy disk
or Rayleigh criterion. For electronic circuits, it is related to the bandwidth but
the definition has not been standardized. When a conversion takes place between
the analog and digital domains, the resel may be different from the digital
sample size. The analog signal rather than the digital sample may limit the
system resolution. In oversampled systems, the resel consists of many samples.
INTRODUCTION 15
Table 1-3
THE "-ELS"
ELEMENT DESCRIPTION
Pixel or pel
A sample created by a detector.
(picture element)
Resel
The smallest signal supported by an analog system.
(resolution element)
Observer
Figure 1-7. Each array is mapped onto the next. The number of
elements in each array may be different. Not every array exists in every
system.
For staring arrays, the total number of pixels is equal to the number of
detectors. The detector's spatial response is determined by the detector's size
(e.g., photosensitive area). If the extent is dH in the horizontal direction and the
optics focal length is fl, then the horizontal detector-angular-subtense (DASH) is
dH (1-2)
DAS --
H fl
16 CCD ARRAYS, CAMERAS, and DISPLAYS
(1-3)
d
PAS - CCH (1-4)
H fl
dccv (1-5)
PASv = fl
(1-6)
With a staring array that has a 100% fill factor, the PAS is equal to the DAS
(e.g., dH = dccH and dv = dccv).
If the Airy disk (a resel) is much larger than the PAS, then the optical resel
determines the system resolution. If the number of pixels across this resel is
large, the sampling theorem can be satisfied. If the electronic imaging system
output is in a digital format, then the number of datels (samples) equals the
number of pixels. If the camera's analog output is digitized, then the number of
datels is linked to the frame grabber's digitization rate. This number can be
much greater than the number of pixels. This higher number does not create
more resolution in terms of the "-els." However, the higher sampling frequency
may improve the overall system MTF.
After image processing, the datels are output to a display medium. For
monitors, each datel is usually mapped, one-to-one, onto each disel. Monitors
are often specified by the number of addressable pixels (defined as disels in this
text). Although the number of addressable pixels (disels) may be large for
cathode ray tube based monitors, the monitor resel may be limited by the
electron beam diameter. With most display media, the finite-sized spots of two
adjacent disels overlap to provide an analog (continuous) image.
Finally, the system designer must be aware of which subsystem limits the
overall system resolution. In some respects, the starting point for system design
should begin with the final interpreter of the data. The minimum "-el" should
be discernible by the interpreter to ensure maximum transfer of information.
However, the observer may not find this image aesthetically pleasing.
1.6. REFERENCES
Scenes in the visible and near infrared are illuminated by the sun, moon,
starlight, night glow, or artificial sources. Since both the target and its
background are illuminated by the same source, targets are detected when
differences in reflectance exist. The camera's output voltage depends on the
relationship between the scene spectral content and the spectral response of the
camera.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
18
RADIOMETRY and PHOTOMETRY 19
Spectral radiant sterance, Le, is the basic quantity from which all other
radiometric quantities can be derived. It contains both the areal and solid angle
concept 1 that is necessary to calculate the radiant flux incident onto a system. It
is the amount of radiant flux, (watts), per unit wavelength (micrometers)
radiated into a cone of incremental solid angle (steradians) from a source whose
area is measured in meters (Figure 2-1)
w (2-1)
Solid Angle oO
Source Area
aAs
Table 2-1
STANDARD SPECTRAL RADIOMETRIC UNITS
For Lambertian sources, the radiance is emitted into a hemisphere. Then the
spectral radiant exitance is related to the spectral radiant sterance by
Le<J. )
- Me<'J.) -----
w (2-2)
1t m 2 -~m-sr
and
w (2-4)
where the first radiation constant is c 1 = 3.7418 x 108 W-p.m4/m2, the second
radiation constant is ~ = 1.4388 x 104 p.m-K, and ).. is the wavelength expressed
in micrometers. The value T is also called the color temperature. Figure 2-2
illustrates Planck's spectral radiant exitance in logarithmic coordinates. Since a
photodetector responds linearly to the available power, linear coordinates may
provide an easier representation to interpret (Figure 2-3). Each curve has a
maximum at )..PEAK· Wien's displacement law provides )..PEAK T = 2898 p.m-K. A
source must have an absolute temperature above about 700 K to be perceived by
the human eye.
The spectral photon exitance is simply the spectral radiant exitance divided
by the energy of one photon (he/)..):
(2-5)
or
(2-6)
where his Planck's constant (h = 6.626 x 10-34 J-s), cis the speed of light (c =
3 x 108 m/s), and the third radiation constant is CJ = 1.88365 x 1027 photons-
p.m3/s-m2. Figures 2-4 and 2-5 provide the spectral photon exitance in
logarithmic and linear coordinates, respectively, for sources typically used for
CCD calibration. These curves have a maximum at )..PEAK T = 3670 p.m-K.
22 CCD ARRAYS, CAMERAS, and DISPLAYS
1E+08
1-
z
:$ 1E+06
ow
~~
....I<( 1E+04
~!::
1-><
ow 1E+02
w
D..
U)
1E+OO~---L---L-+~~~~~+-~~~~~
0.1 1 10 100
WAVELENGTH (pm)
+----1"'-----------------
U)
1.0E+17
0.1 1 10
WAVELENGTH (pm)
Figure 2-4. Planck's spectral photon exitance forT = 2856, 4870, and
6500 K. The maximum values occur at 1.29, 0. 75, and 0.56 p.m,
respectively. The units are photons/(s-m2-rm).
RADIOMETRY and PHOTOMETRY 23
4.0E+26
z
0
..... 3.0E+26
Ow
:z:u
ll.z
..Joe( 2.0E+26
~!::: 4870
..... x
ow 1.0E+26
w
ll.
en 2856
O.OE+OO
0.2 0.4 0.6 0.8 1.2
WAVELENGTH (pm)
2.3. PHOTOMETRY
where Mp(},) is the power expressed in units of WI p.m and KM is the luminous
efficacy for photopic vision. It is 683 lumens/W at the peak of the photopic
curve (A-:::::0.55 p.m) and 1746 lumens/W for the scotopic region at A-:::::0.505
p.m. Although both photopic and scotopic units are available, usually only the
photopic units are used (Table 2-2 and Figure 2-6). At very low light levels (less
than 5 x 10·3 lux), the eye's rods operate (scotopic response). For light levels
above 5 x 10·2 lux, the eye's cones (photopic response) respond.
24 CCD ARRAYS, CAMERAS, and DISPLAYS
Table 2-2
PHOTOPIC and SCOTOPIC EYE RESPONSE
w
en
z 0.8
0 Scotopic
D.
en 0.6
w
a:
w 0.4
~
1-
:5w 0.2
a: 0
0.35 0.4 0.45 0.5 0.55 0.6 0.65 0.7 0.75
WAVELENGTH (pm)
2.3.1. UNITS
Figure 2-7 illustrates the geometric relationship between the SI, CGS, and
English luminous incidance units. The numeric relationship is provide in Table
2-4.
A= 1m2
-----------1m -----------
Figure 2-7. Geometric relationship between the SI, CGS, and English
luminous incidance units. The solid angle is one steradian. (Not to
scale).
Table 2-3
STANDARD PHOTOMETRIC UNITS
(The SI units are recommended)
~
s
~-
Luminous incidance
(onto a target)
(also illuminance)
Ev
Lux (lx)
Lumen/m2
Phot(ph)
lumen/cm2
Footcandle (fc)
lumen/ft2
Nit
~
Luminous sterance Stilb (sb)
LV Lumen/m2-sr candela/ft2
(also brightness and luminance) cd/cm2
or cd/m2
8
\..)
\0
("''
RADIOMETRY and PHOTOMETRY 27
Table 2-4
CONVERSION BETWEEN SI, CGS, and ENGLISH UNITS
Natural lighting levels can vary by over 9 orders of magnitude (Table 2-5).
The minimum level is limited by night glow and the maximum level is provided
by the sun. At very low light levels (less than 5 x 10-3 lux), the eye's rods
operate (scotopic response). For light levels above 5 X 10-2 lux, the eye's cones
(photopic response) respond. Between these two values, both rods and cones are
operating and the eye's response is somewhat between the two values. This
composite response is called the mesopic response. Table 2-6 provides typical
artificial lighting levels.
Table 2-5
NATURAL ILLUMINANCE LEVELS
EYE AVERAGE LUMINOUS
SKY CONDITION
RESPONSE INCIDANCE (lux)
Twilight Photopic 10
Moonless, overcast
Scotopic 10·•
(night glow)
Table 2-6
TYPICAL ARTIFICIAL ILLUMINANCE LEVELS
AVERAGE LUMINOUS
LOCATION
INCIDANCE (lux)
TV studio 103
2.4. SOURCES
As shown in Figure 2-3 (page 22), the peak wavelength shifts toward the
blue end of the spectrum as the color temperature increases. This puts more
energy in the visible region. There is approximately 70% more luminous flux
available from a lamp operating at 3200 K than a light operating at 3000 K based
on blackbody curves. This is the basis for using a higher color temperature lamp
for higher illumination levels. However, the lifetime of tungsten halogen bulbs
decreases dramatically with increasing color temperature. It is far better to use
more lamps to increase the luminous flux than to increase the color temperature
and sacrifice lifetime.
Table 2-7
CIE RECOMMENDED ILLUMINANTS
CIE EFFECTIVE COLOR
DESCRIPTION
ILLUMINANT TEMPERATURE
Light from an
A 2856K
incandescent source
250
~
c;; 200
z
w
..... 150
~
w
;:: 100
.....
~ 50
w
a: 0
0.4 0.5 0.6 0.7 0.8
WAVELENGTH (pm)
Figure 2-8. Relative output of the standard 11luminants. Compare the shapes
to Figure 2-3, page 22.
Table 2-8
APPROXIMATE COLOR TEMPERATURE
APPROXIMATE COLOR
SOURCE
TEMPERATURE
Floodlights 3000K
'S
D.
.5
~l, 1 j1-.. -
==
(b)
0.3 0.4 0.5 0.6 0.7 0.3 0.4 0.5 0.6 0.7
Wavelength (~m)
(c) (d)
1.2E+09
1-
z 1.0E+09
:$
ow 8.0E+08
ill!~
..JOCS: 6.0E+08
Ql!!::
1->< 4.0E+08
ow
w 2.0E+08
a..
rn
O.OE+OO
0.4 0.45 0.5 0.55 0.6 0.65 0.7
WAVELENGTH (pm)
(2-8)
where the small angle approximation was used (valid when R/ ~ A 0 ) . TATM is
the intervening atmospheric transmittance. The number of on-axis photons
reaching the image plane is:
(2-9)
When the image size is much larger than the detector area (A1 ~ A0 ), the
source is said to be resolved or the system is viewing an extended source.
Equivalently, the detector is flood-illuminated. In most solid state camera
applications, the source is resolved. The value A0 is the effective sensitive
detector area. Microlenses (discussed in Section 3.11.1., Microlenses) and
optical anti-alias filters (discussed in Section 10.3.4., Optical Anti-alias Filter)
increase the effective area.
The number of photons incident onto the detector is simply the ratio of the
areas:
AD (2-10)
n DETECTOR = n IMAGE A
I
(2-11)
34 CCD ARRAYS, CAMERAS, and DISPLAY..'>'
Imaging System
Entrance Aperture
Area, Aa
Source
Detector
Area Ao
Area AS
~ R,
LqAoAD
(2-12)
nDETECTOR =
2 (1 M )2 TOP11CS TATM t /NT
fl + OP11CS
1 1
+ (2-13)
~ fl
or
(2-16)
where Rq()..) has units of electrons per photon and is simply the detector's
quantum efficiency and t1NT becomes is the integration time. As the source
moves to infinity, Movncs approaches zero.
where
'- 2 ('A)t A
t.n
pe
=JR ('A) 4 t.pEF2q-scENE
q
1t
( M
INT D T
)2 OPTICS
('A)T
ATM'
f'A)d'J... (2-18)
'-t 1+ OPTICS
36 CCD ARRAYS, CAMERAS, and DISPLAYS
The value GcAMERA contains both the array output conversion gain (units of
volts/electron) and the subsequent amplifier gain. Tables 2-5 and 2-6 provide
nominal values for Ee-scENE·
It would appear that by simply decreasing the f-number, 11pe or <lllpe would
increase. Often, lower f-number lens systems have more optical elements and
therefore may have lower transmittance. The value ToPTics/F 2 must increase to
increase 11pe or <l11pe. Lens systems may be described by the T-number:
F
T/# = (2-19)
JTOPT1CS
~============================================
Example 2-1
VISUAL THRESHOLD
After adapting to the dark at least 60 min, the eye can perceive a few
photons per second (absolute threshold). Here, the pupil dilates to about 8 mm
(A0 = 5 x I0-5 m2). When the eye is dark adapted, only the rods are functioning
and the eye's spectral response is approximately 0.38 ~-tiD to 0.66 ~-tm. The
number of photons per second reaching the eye from a Lambertian source is
0 6611
" m A M (>.. 1) A M (>.. 1)
A d>.. ~ A 1!..>.. (2-20)
nEYE - J
o.3811m
_____!!_
Rt
2
q
1t
'
S
_!?._
R1
2
q
1t
o'
S '
RADIOMETRY and PHOTOMETRY 37
where ilf. = 0.28 p.m. The peak scotopic eye response occurs at A, = 0.515 p.m.
The observer can just perceive this object at 640 K when he is in a completely
blackened room. As the ambient illumination increases, the pupil constricts (A,
decreases). Furthermore, the eye's detection capability depends on the ambient
lighting. As the light level increases, the target flux must also increase to be
perceptible.
The eye's ability to just discern intensity differences depends on the intensity
of the surrounding illuminance. At 0.01lux the object must provide 10 times
more flux than the background to be discerned by the eye. If the surround is at
700K, than the target must be at 750K to be perceptible. At 1000lux, the object
must provide 1000 times more flux then the surround. If the surround color
temperature is 1000 K, then the target must be heated to 1350 K to be
perceptible. Photometric units are not linearly related to color temperature.
2.6. NORMALIZATION
While the sensor may be calibrated with a standard illuminant, it may not
be used with the same illumination. Calibration with a standard illuminant is
useful for comparing camera responsivities under controlled conditions.
However, if the source characteristics in a particular application are significantly
different from the standard illuminant, the selection of one sensor over another
must be done with care. For example, street lighting from an incandescent bulb
is different from that of mercury-arc lights (compare Figure 2-8 with Figure 2-9,
pages 30 and 31, respectively). The only way to determine the relative
responsivities is to perform the calculation indicated in Equation 2-16.
The effect of ambient lighting is particularly noticeable when using low light
level televisions at nighttime. The spectral output of the sky (no moon) and
moon (clear night) are significantly different. Figure 2-13 illustrates the natural
night sky luminous incidence. Since an abundance of photons exists in the near
infrared (0. 7 to 1.1 p.m), most night vision systems (e.g., image intensifiers,
starlight scopes, snooper scopes, etc.) are sensitive in this region.
1- 2.0E+08
::J
D..
1- 1.5E+08
::J
0
w 1.0E+08
~
1-
5.0E+07
:3w
0:: O.OE+OO
0.2 0.4 0.6 0.8 1.2
WAVELENGTH (pm)
WAVELENGTH (pm)
A camera output depends on both the system spectral response and the color
temperature of the illuminating source (Equation 2-16, page 35). When the
camera spectral response is within the eye's response, then photometric units are
reasonable. When the camera's response is outside the eye's response curve,
photons contribute to signal but the photometry remains constant. Even though
two sources may provide the same luminous incidance, the camera output can
be quite different. Sometimes an infrared filter is used to restrict the wavelength
response. When the filter is added, the average responsivity becomes
'-2
J TIR-FILTER R.(J. . )M.(I.. ) d)..
AI A (2-21)
Rv = - - - - ' - - - - - - - - -
0.75 lumens
683 f V()..)M.(J.. ,T) d)..
0.38
For ideal photon detectors the spectral responsivity (expressed in A/W) is:
R.(J.) - __l_ R
Ap P
(2-22)
- 0 elsewhere
The ideal silicon detector response is used to estimate the relative outputs of
two systems: one with an ideal IR filter (transmittance is unity from 0.38 to 0. 70
p.m and zero elsewhere) and one without. The relative outputs are normalized
to that expected when illuminated with a CIE A illuminant (Table 2-9). The
sources are considered ideal blackbodies with exitances that follow Equation 2-4
(page 21). As the source temperature increases, the output increases but not
linearly with color temperature.
Table 2-9
RELATIVE OUTPUT
This is assuming ideal blackbodies and ideal Si spectral response.
A specific sensor may deviate significantly from these values.
Only photons within the visible region affect the number of available lumens.
For the system without the IR filter, the detector is sensing photons whose
wavelengths are greater than 0.7 p.m even though the number of lumens is not
affected by these photons (Figure 2-14). As the color temperature increases, the
flux available to the observer (lumens) increases faster than the output voltage.
As a result, responsivity, expressed in A/lumen, decreases with increasing color
temperature (Table 2-10). This decrease affects the wide spectral response
systems to a greater extent. Specifying the output in lumens is of little value for
an electro-optical sensor whose response extends beyond the eye's response.
Radiometric units are advised here.
The values in Tables 2-9 and 2-10 should be considered as illustrative. Solid
state detectors do not follow the ideal response (discussed in Section 4.1.1.,
Spectral Response). The IR filter cutoff also varies by manufacturer. Only a
detailed spectral response evaluation permits correct comparisons among
detectors. Since there is considerable variation in spectral response and type of
IR filter used, it is extremely difficult to compare systems based on responsivity
values only.
Photopic
4.0E+07 observer
w
0
z
<
1-
3.0E+07
x
w
..J 2.0E+07
~
1-
0 1.0E+07
w
D..
en O.OE+OO
0.4 0.6 0.8 1.2
WAVELENGTH (pm)
Figure 2-14. Spectral relationship between source characteristic, detector
response, and the photopic observer. As the color temperature increases,
the energy available to the observer increases rapidly. Therefore, the
number of lumens increases dramatically.
42 CCD ARRAYS, CAMERAS, and DISPLAYS
Table 2-10
RELATIVE RESPONSIVITY (A/lumen)
A specific sensor may deviate significantly from these values
Example 2-2
SOURCE SELECTION
A camera's signal-to-noise ratio is too small. The engineer can either change
the light source to one with a higher color temperature or add more lamps. What
are the differences?
=============================================~
RADIOMETRY and PHOTOMETRY 43
Example 2-3
RELATIVE VERSUS ABSOLUTE OUTPUT
If the relative spectral output of the two bulbs is the same (e.g., CIE A), the
responsivity is independent of the source flux. The bulb intensity should be
sufficiently high to produce a good signal-to-noise ratio. It should not be so high
that the detector saturates.
=============================================~
System output does not infer anything about the source other than that an
equivalent blackbody of a certain color temperature would provide the same
output. This is true no matter what output units are used (volts, amps, or any
other arbitrary unit). These units, by themselves, are not very meaningful for
system-to-system comparison. For example, .:1VCAMERA (and equivalently the
camera responsivity) can be increased by increasing the system gain. As such,
it is dangerous to compare system response based on only a few numbers.
However, since the human observer is not familiar with infrared images, the
imagery may appear slightly different compared to that obtained with the IR-
blocking filter.
2.8. REFERENCES
l. C. L. Wyatt, Radiometric System Design, Chapter:;, Macmillan Publishing Co., New York,
NY (1987).
2. D. Kryskowski and G. H. Suits, "Natural Sources," m Sources of Radiation, G. J. Zissis, ed.,
pp. 151-209, Volume 1, The Infrared and Electro-Optical Systems Handbook, J. S. Accetta and
D. L. Shumaker, eds., copublished by Environmental Research Institute of Michigan, Ann
Arbor, MI, and SPIE Press, Bellingham, WA (1993).
3. E. B. Noel, "Radiation from High Pressure Mercury 1\.rcs," Illumination Engineering, Vol. 36,
pg. 243 (1941).
4. The ONTAR Corporation, 129 University Road, Brookline, MA 02146-4532, offers a variety
of atmospheric transmittance codes.
5. F. E. Nicodemus, "Normalization in Radiometry," Applied Optics, Vol. 12(12), pp. 2960-2973
(1973).
6. Burle Electro-Optics Handbook, TP-135, pg. 73, Burle Industries, Inc., Lancaster, PA (1974).
Formerly known as the RCA Electro-Optics Handbook.
3
SOLID STATE ARRAYS
CCD refers to a semiconductor architecture in which charge is transferred
through storage areas. The CCD architecture has three basic functions: (a)
charge collection, (b) charge transfer, and (c) the conversion of charge into a
measurable voltage. A CCD can be attached to a variety of detectors. The basic
building block of the CCD is the metal-insulator-semiconductor (MIS) capacitor.
It is also called a gate. The most important MIS is the metal-oxide-
semiconductor (MOS). Because the oxide of silicon is an insulator, it is a natural
choice.
Because most sensors operating in the visible region use a CCD architecture
to move a charge packet, they are popularly called CCD arrays. However, the
charge injection device (CID) does not use a CCD for charge transfer. Rather,
two overlapping silicon MOS capacitors share the same row and column
electrode. Pixel readout occurs by sensing the charge transfer between the
capacitors at the pixel site.
Recently, active pixels have been introduced. These devices are fabricated
with complementary-metal-oxide-semiconductor (CMOS) technology. The
advantage is that one or more active transistors can be integrated into the pixel.
As such, they become fully addressable (can read selected pixels) and can
perform on-chip image processing.
45
46 CCD ARRAYS, CAMERAS, and DISPLAYS
from the simplified diagrams shown in this chapter. A specific device may not
have all the features listed or may have additional features not listed.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
3 .1. PHOTODETECTION
After photodetection, the stored charge is read out. The array architecture
often is optimized for a specific application. General imagery systems are
designed for interlace operation consistent with a standard video (e.g., EIA 170).
Research devices may operate in the progressive scan mode. This reduces some
image distortion when the object is moving. For scientific applications the output
may be in the slow scan mode, which simply means that the data read out rate
is lower than a standard frame rate (e.g., EIA 170).
SOLID STATE ARRAYS 47
3.1.1. PHOTOGATE
3.1.2. PHOTODIODE
(3-1)
After readout, the voltage is reset to its original value. If the number of
electrons is large, the voltage is reduced to zero and saturation occurs. Excess
photoelectrons will diffuse to neighboring wells and create blooming.
Photodiodes do not have an overlaying structure and therefore have higher
quantum efficiency than comparable photogates.
The hole accumulation diode (HAD) is a pinned photodiode that offers the
best features of the photogate and photodiode. It offers the high blue response
of the photodiode and large well capacity of the photogate.
Applying a positive voltage to the CCD gate causes the mobile positive holes
in the p-type silicon to migrate toward the ground electrode because like charges
repel. This region, which is void of positive charge, is the depletion region
(Figure 3-1). If a photon whose energy is greater than the energy gap is
48 CCD ARRAYS, CAMERAS, and DISPLAYS
/ Metal Electrode
Oxide Layer
I
1 Depletion I
Bulk Semiconductor
1 Region I
L _ _ _ _I (P-Type Silicon)
v1
v2 --~----------------.----------------
V3--~----~----~----+-----4-----.----
Si
• Polysilicon
Figure 3-2. Three-phase CCD. The gates must overlap for efficient
charge transfer.
SOLID STATE ARRAYS 49
Each gate has its own control voltage that is varied as a function of time.
The voltage is called the clock or clocking signal. When the gate voltage is low,
it acts as a barrier whereas when the voltage is high, charge can be stored.
Initially, a voltage is applied to gate 1 and photoelectrons are collected in well
1 (Figure 3-3b). When a voltage is applied to gate 2, electrons move to well 2
in a waterfall manner (Figure 3-3c). This process is rapid and the charge quickly
equilibrates in the two wells (Figure 3-3d). As the voltage is reduced on gate 1,
the well potential decreases and electrons again flow in a waterfall manner into
well2 (Figure 3-3e). Finally, when gate 1 voltage reaches zero, all the electrons
are in well 2 (Figure 3-3f). This process is repeated many times until the charge
is transferred through the shift register.
The CCD array is a series of column registers (Figure 3-4). The charge is
kept within rows or columns by channel stops or channel blocks and the
depletion regions overlap in one direction only. At the end of each column is a
horizontal register of pixels. This register collects a line at a time and then
transports the charge packets in a serial fashion to an output amplifier. The
entire horizontal serial register must be clocked out to the sense node before the
next line enters the serial register. Therefore, separate vertical and horizontal
clocks are required for all CCD arrays. This process creates a serial data stream
that represents the two-dimensional image.
Gate 1 Gate 2
2 2
I ~-
j
·-Ill
I N 111·-
<11-
Qj Qj c:
Q)
I ~ I ~ I t;.!!!
) __ j _ _ j c:
_a. 0
2 2 2
(c) {d)
Although any number of transfer sites (gates) per detector area can be used,
it generally varies from two to four. With a four-phase device, the charge is
stored under two or three wells depending upon the clock cycle. Figure 3-5
represents the steady state condition after the gates have switched and the charge
has equilibrated within the wells. It takes a finite time for this to happen. Figure
3-6 illustrates the voltage levels that created the wells shown in Figure 3-5. The
time that the charge has equilibrated is also shown. The clock rate is fcLocK =
1/tcwcK and is the same for all phases. The output video is valid once per clock
pulse (selected as T2) and occurs after the charge has equilibrated (e.g., Figure
3-3b or Figure 3-3f). Only one master clock is required to drive the array. Its
phase must be varied for the gates to operate sequentially. An anti-parallel
clocking scheme is also possible. Here v3 is the inverse of VI and v4 is the
inverse of Vz. That is, when VI is high, v3 is low. The anti-parallel clock
system is easier to implement because phasing is not required. With the anti-
parallel scheme, charge is always stored in two wells. For equal pixel sizes,
four-phase devices offer the largest well capacity compared to the two- or three-
phase systems. With the four-phase system, 50% of the pixel area is available
for the well.
52 CCD ARRAYS, CAMERAS, and DISPLAYS
Gates,
'
/ Stored Charge
.....!1.......!1......-r
.......!1.......!1.....
Ts
_ ___....,,
L.....-----1'
Equilibrated
Charge Times
/
c::::Jc::::Jc::::Jc::::Jc::::Jc::::Jc::::Jc::::Jc::::Jc::::J c::::J
~ ~ ~ ~ ~ ~ ~ ~ ~ ~ T3
__ll~--------------~n~-- Valid
Video
The three-phase system (Figure 3-7 and Figure 3-8) stores the charge under
one or two gates. Only 33% of the pixel area is available for the well capacity.
With equal potential wells, a minimum number of three phases are required to
clock out charge packets efficiently.
54 CCD ARRAYS, CAMERAS, and DISPLAYS
,.----i,...--------,----!1-- ~:
Gates ',..
tCLOCK
V1
l I
v2
I
v3
~ Equilibrated
c::::J c::::J c::::J c::::J c::::J c::::J c::::J c::::J Charge Times
T1 T2 T3 T4 Ts Ts T1 T2
Figure 3-8. Voltage levels for a three-phase system. The clock signals are
identical for all three phases but offset in time. T2 was selected as the valid
video time. Valid video exists only after the charge has equilibrated. (After
Reference 10).
The well potential depends on the oxide thickness underneath the gate. In
Figure 3-9, the alternating gates have different oxide thicknesses and therefore
will create different well potentials. With different potentials it is possible to
create a two-phase device. Charge will preferentially move to the right-hand side
of the pixel where the oxide layer is thinner. Assume initially that the wells
controlled by V2 are empty. This will occur after a few clock pulses. At time
T~> both clocks are low. When V2 is raised, the potential increases as shown at
time T2 • Because the effective potential increases across the gates, the charge
cascades down to the highest potential. Then V2 is dropped to zero and the
charge is contained under the V2 gate at time T3 • This process is repeated until
the charge is clocked offthe array. The voltage timing is shown in Figure 3-10.
While Figure 3-9 illustrates variations in the oxide layer, the potential wells can
also be manipulated though ion implantation. As with the four-phase device, the
two-phase device can also be operated in the anti-parallel mode: V2 is a mirror
image of V1 • That is, when V 1 is high, V2 is low.
56 CCD ARRAYS, CAMERAS, and DISPLAYS
Ga~~:
1--- Pixel --1--- Pixel --1
~-
I
·-
r / ) ·Cll
-
cu-
<D c:
.... .S!
'.E ~
tCLOCK
_....-- Equilibrated
E:::J Charge Times
T4
Jl~----~n~-------- Valid
Video
Figure 3-10. Voltage levels for a two-phase system. (After Reference 10).
SOLID STATE ARRAYS 57
The virtual phase device requires only one clock (Figure 3-11). Additional
charge wells are created by internally biased gates. Charge is stored either
beneath the active gate or the virtual gate. When V1 is low, the charge will
cascade down to the highest potential (which is beneath the virtual well). When
V 1 is applied at T2 , the active gate potential increases and the charge moves to
the active gate well.
v1
Gate"'
I I
G5 ~n )15 lvn J- Virtual Phase
~-
·- <0
C/l·-
l
<0-
T, ~
Q) c:
Q)
go
_a_
- Stored Charge
Figure 3-11. Charge transfer in a device with a virtual phase. The virtual
well is created by p- and n-material implants. These ions create a fixed bias
and therefore a well with fixed potential. By changing V 1, the active gate
potential can be lower than the virtual well (T 1) or higher (T2). This
represents one column. Rows go into the paper.
signal output and dynamic range at the expense of spatial resolution. Because it
increases the signal-to-noise ratio, binning is useful for low light level
applications for those cases where resolution 1s less important. Serial registers
and the output node require larger capacity charge wells for binning operation.
If the output capacitor is not reset after every pixel, then it can accumulate
charge.
Array architecture is driven by the application. Full frame and frame transfer
devices tend to be used for scientific applications. Interline transfer devices are
used in consumer camcorders and professional television systems. Linear arrays,
progressive scan, and time delay and integration (TDI) are used for industrial
applications. Despite an ever increasing demand for color cameras, black and
white cameras are widely used for many scientific and industrial applications.
Located next to each sensor is the transfer gate and then the CCD shift
register (Figure 3-12a). The shift register is also light sensitive and is covered
with a metal light shield. Overall pixel size is limited by the photodiode size.
For example, with a three-phase system, the pixel width is three times the width
of a single gate. For a fixed gate size, the active pixel width can be reduced
with a bilinear readout (Figure 3-12b). Because resolution is inversely related
to the detector-to-detector spacing (pixel pitch), the bilinear readout has twice
the resolution. For fixed pixel size, the bilinear readout increases the effective
charge transfer efficiency because the number of transfers is reduced by a factor
of 2 (discussed in Section 3.4., Charge Transfer Efficiency).
SOLID STATE ARRAYS 59
Photo
Diode
(a)
(b)
Figure 3-12. Linear array. (a) Simple structure and (b) bilinear readout.
The transfer gate clock pulse allows charge to move from the photodiode
into the CCD transfer register.
Figure 3-13 illustrates a full frame transfer (FFT) array. After integration,
the image pixels are read out line-by-line through a serial register that then
clocks its contents onto the output sense node (see Figure 3-4, page 50). All
charge must be clocked out of the serial register before the next line can be
transferred. In full frame arrays, the number of pixels is often based upon
powers of 2 (e.g., 512 X 512, 1024 X 1024) to simplify memory mapping.
Scientific arrays have square pixels and this simplifies image processing
algorithms.
60 CCD ARRAYS, CAMERAS, and DISPLAYS
Imaging
Area
Data rates are limited by the amplifier bandwidth and, if present, the
conversion capability of the analog-to-digital converter. To increase the effective
readout rate, the array can be divided into subarrays that are read out
simultaneously. In Figure 3-14, the array is divided into four subarrays. Because
they are all read out simultaneously, the effective clock rate increases by a
factor of 4. Software then reconstructs the original image. This is done in a
video processor external to the CCD device where the serial data are decoded
and reformatted.
Large area devices often allow the user to select one subarray for readout.
Thus, the user can trade frame rate against image size. This allows the user to
obtain high frame rates over the area of interest (discussed in Section 4.5.,
Frame Rates).
SOLID STATE ARRAYS 61
Subarray Subarray
Subarray Subarray
I I 'I '
I
I
I
I I I
'
I
'
I
I I
<J-11.------I I. . . __Readout
Readout
_
I
---.~
H>
Sense Sense
Node Node
and and
Amplifier Amplifier
Figure 3-14. A large array divided into four subarrays. Each subarray
is read out simultaneously to increase the effective data rate. Very large
arrays may have up to 32 parallel outputs. The subimages are
reassembled in electronics external to the device to create the full image.
A frame transfer (FT) imager consists of two almost identical arrays, one
devoted to image pixels and one for storage (Figure 3-15). The storage cells are
identical in structure to the light sensitive cells but are covered with a metal light
shield to prevent any light exposure. After the integration cycle, charge is
transferred quickly from the light sensitive pixels to the storage cells. Transfer
time to the shielded area depends upon the array size but is typically less than
500 p.s. Smear is limited to the time it takes to transfer the image to the storage
area. This is much less time than that required by a full frame device.
62 CCD ARRAYS, CAMERAS, and DISPLAYS
Image
Array
Shielded
Storage Array
l I I I Sense Node
1 Readout and Amplifier
Figure 3-15. Basic architecture for a full frame CCD. Frame transfer
devices can also have a split architecture similar to that shown in Figure
3-14.
The array may have a few dummy charge wells between the active and
storage areas. These wells collect any anticipated light leakage and their data are
ignored. Some arrays do not have the integral light shield. These arrays can
either be operated in the full frame mode (e.g., 512 x 512) or in the frame
transfer mode (e.g., 256 x 512). Here, it becomes the user's responsibility to
design the light shield for the frame transfer mode.
Example 3-1
SHIELD DESIGN
A camera manufacturer buys a full frame transfer array (512 x 512) but
wants to use it in a frame transfer mode (256 x 512). An engineer designs an
opaque shield that is placed 1 mm above the array. He is using f/5 optics. What
is the usable size of the array if the pixels are 20 pm square and the shield edge
is centered on the array?
SOLID STATE ARRAYS 63
When placed near the focal plane, the shield edge will create a shadow that
is approximately the distance from the focal plane (d = 1 mm) divided by the f-
number (F = 5). The full shadow is 200 p,m or 10 pixels. One-half of the
shadow falls on the active pixels and one-half illuminates the pixels under the
shield. This results in a usable array size of 251 x 512. An alignment error of
0.5 mm adds an additional25 pixels. If the alignment error covers active pixels,
the active area is reduced to 226 X 512. If the error covers shielded pixels, the
storage area is reduced to 226 x 512. Either way, the usable array is 226 x 512.
=================================================~
Example 3-2
USEFUL ARRAY SIZE
Sense Node
and Amplifier
A fraction of the light can leak into the vert teal registers. This effect is most
pronounced when viewing an ambient scene that has a bright light in it. For
professional television applications, the frame interline transfer (FIT) array was
developed to achieve even lower smear values (Figure 3-17).
SOLID STATE ARRAYS 65
--------
1
I
----L---
Shielded Shielded
Vertical Vertical
Pixel Transfer Transfer
Register Register
'\.
Active
Detectors
(a) (b)
Figure 3-18. Detailed layout of the 2:1 interlaced array. (a) The odd
field is clocked into the vertical transfer register and (b) the even field
is transferred. The vertical transfer register has four gates and charge is
stored under two wells. The pixel is defined by the detector center-to-
center spacing and it includes the shielded vertical register area. The
transfer gate is not shown. (After Reference 11).
Shielded Shielded
Vertical Vertical
Pixel Transfer Transfer
Register Register
'\
Active
Detectors
(a) (b)
While any architecture can be used to read the charge, the final video format
generally limits the selection. Although the charge is collected in rows and
columns, the camera reformats the data into a serial stream consistent with the
monitor requirements. This serial stream is related to a fictitious "scan" pattern.
Figure 3-20 illustrates the video lines and field-of-view of a standard format
imaging system. Figure 3-20a represents the "scan" pattern created by the
interlace device shown in Figure 3-18 whereas Figure 3-20b represents the
pseudo-interlace "scan" pattern of the device shown in Figure 3-19. Although
the dissected scene areas overlap in Figure 3-20b, the video timing is consistent
with the video format.
68 CCD ARRAYS, CAMERAS, and DISPLAYS
1
----- 2----- ...
3
}odd
••
----·484
••
____ .,.. }odd
485 (b)
(a)
CCD arrays do not scan the image, but it is convenient to represent the
output as if it were scanned. Progressive scan simply means the noninterlaced
or sequential line-by-line scanning of the image. This is important to machine
vision because it supplies accurate timing and has a simple format. Any
application that requires digitization and a computer interface will probably
perform better with progressively scanned imagery.
If a strobe light is used to stop motion, the image will appear only in the
field that was active during the light pulse. Because of image motion effects,
only one field of data can be used for image processing. This reduces the
vertical resolution by 50% and will increase vertical image aliasing.
Figure 3-23 illustrates four detectors operating in TDI mode. The charge
increases linearly with the number of detectors in TDI. Shot noise also
increases, but as the square root of the number of TDI elements, NTDI·
Theoretically, this results in an SNR improvement of v'NTm· The well capacity
limits the maximum number of TDI elements that can be used.
Multiple
Sensor
Nodes
Moving Object
Conveyor Belt
Figure 3-22. Typical TDI operation. The pixel clock rate must be
matched to the image velocity.
SOLID STATE ARRAYS 71
For this concept to work, the charge packet must always be in synch with
the moving image. If there is a mismatch between the image scan velocity and
pixel clock rate, the output is smeared and this adversely affects the in-scan
MTF (discussed in Section 10.3.3., TDI). The accuracy to which the image
velocity is known limits the number of useful TDI stages.
- Moving Ball on
Conveyor Belt
••
••
•
••
••
Figure 3-23. TDI concept. At time T 1, the image is focused onto the
first detector element and creates a charge packet. At T2 the image has
moved to the next detector. Simultaneously, the charge packet is clocked
down to the next pixel site. Here, the image creates additional charge
that is added to the stored packet (created by the first detector). After
four stages, the signal increases fourfold.
72 CCD ARRAYS, CAMERAS, and DISPLAYS
Although Figure 3-22 illustrates two readout registers, an array may have
multiple readout registers. Array sizes may be as large as 2048 by 32 TDI
elements. Additional readouts are required to avoid charge well saturation. The
multiple readout outputs may be summed in electronics external to the TDI
device or may be placed in a horizontal serial register for summing (Figure 3-
24).
t t t
I I I
"S "S "S "S
0
0 0 0
"C "C "C "C
<tl <tl <tl <tl
Ql Ql Ql Ql
a: a: a: a:
(ij (ij (ij (ij
0
'E
Ql
0
'E
Q)
·-e0
Q)
0
'E
Q)
> > > >
Figure 3-24. The array height must match the image height. To avoid
saturation, multiple vertical register readouts may be used.
The vertical readout register must clock out the data before the next line is
loaded into the register. The amplifier bandwidth or analog-to-digital conversion
time limits the pixel rate. For example, if the amplifier can support 107 pixels/s
and the register serves 500 vertical pixels, then the maximum (in-scan) line rate
is 20,000 lines/s. If the amplifier can only handle 2 x 106 pixels/sand the line
rate is maintained at 20,000 lines/s, then the vertical transfer register must be
divided into five registers - each connected to 100 pixels. These registers and
associated outputs operate in parallel (Figure 3-25).
SOLID STATE ARRAYS 73
Serial Readout
Image Motion Registers
... I
Vertical
Extent
of Image
- Nn>l-
Figure 3-25. The vertical direction may be divided into subarrays to
increase the line rate. The transfer register typically is a three-phase
device.
~=============================================
Example 3-3
TDI LINE RATE
The pixels in a TDI are 151'm square. The object is moving at 2 m/s and the
object detail of interest is 150 I'm. What is the lens magnification and what is
the pixel clock rate?
Each pixel must view 150 I'm on the target. Because the pixel size is 151'm,
the lens magnification is Mop-ncs = 15/150 = 1110. The TDI pixel clock rate is
(2 m/s)/150 I'm= 133Khz.
==============================================~
The early CCDs transferred electrons in channels that were at the surface.
Surface defects interfere with the transfer and create poor charge transfer
efficiency. These were SCCD or surface channel CCDs. By burying the channel
(BCCD), the charge was no longer affected by surface defects and this improved
the CTE significantly. Nearly all CCDs use buried channel technology today.
As the charge packet moves from storage site to storage site, a few electrons
are left behind. The ability to transfer all the charge is given by the CTE. As
the number of transfers increases, the charge transfer efficiency must also
increase to maintain signal strength.
The number of transfers across the full array is the number of pixels
multiplied by the number of gates. For example, for an array that is 1000 x
1000, the charge packet farthest from the sense node must travel 2000 pixels.
If four gates (four-phase device) are used to minimize cross talk, the farthest
charge packet must pass through 8000 wells (neglecting isolation pixels). The
net efficiency varies with the target location. If the target is in the center of the
array, then it only experiences one-half of the maximum number of transfers on
the array. If the target is at the leading edge, it is read out immediately with
virtually no loss of information. The loss of information is described by the CTE
MTF (discussed in Section 10.3.2., Charge Transfer Efficiency).
If many transfers exist (e.g., 8000) the transfer efficiency must be high
(e.g., 0.999995) so that the net efficiency is reasonable (0.99999SSOOO = 0.96).
Arrays for consumer applications typically have charge transfer efficiencies
greater than 0.9999 and CTEs approach 0.999999 for scientific grade devices.
SOLID STATE ARRAYS 15
where N is the number of transfers. For the first well, R = N. For the second
well, R = N - 1 and so on.
Figure 3-26 illustrates the output when a large number of pixels are
illuminated. The leading edge loses charge and decreases in value. For pixels
near the end of the pulse, the charge lost to trailing wells is replaced by the
charge acquired from leading wells. Here, the signal strength is maintained.
Finally, charge is lost to wells after the signal and the pulse width is stretched.
Ie adi ng trailing
- r-
r-
n (b)
Inn.
trailing
(a)
,
lr1
(c)
Figure 3-26. Input and output signals after many transfers. (a) Initially
generated charge, (b) charge distribution with low CTE, and (c) charge
distribution with high CTE.
76 CCD ARRAYS, CAMERAS, and DISPLAYS
Consider a very small spot of light that illuminates just one pixel and creates
no electrons. A well, N transfers away, will contain nN electrons. The fractional
amount of charge left is
(3-3)
The fractional amount of charge appearing in the first trailing well after N
transfers is
...J
0.999999
<
z 0.99999
(!) 0.8
Ci)
...J 0.6 0.9999
<
z
0 0.4
i=
0 0.2
~
u.. 0
0 1000 2000 3000 4000
NUMBER of TRANSFERS
Figure 3-27. Fractional signal as a function of transfer efficiency and
number of transfers.
...J
...J 0.4 0.999
~ 0.3
(!)Z
zQ
;;;it) 0.2
~~
1-u..
....en 0.1
a::
u::
1000 2000 3000 4000
NUMBER of TRANSFERS
Figure 3-28. Fractional signal for the first trailing charge well. With low
CTEs, as the number of transfers increases, more charge spills into the
second trailing well. The charge in the first trailing well decreases.
SOLID STATE ARRAYS 77
CTE is not a constant value but depends upon the charge packet size. For
very low charge packets, the CTE decreases due to surface state interactions.
This is particularly bothersome when the signal-to-noise ratio is low and the
array is very large (typical of astronomical applications). The CTE also
decreases near saturation due to charge spill effects.
The fraction of charge left behind also depends upon the clocking frequency.
At high frequencies, efficiency is limited by the electron mobility rate. That is,
it takes time for the electrons to move from one storage site to the next. Thus,
there is a tradeoff between frame rate (dictated by clock frequency) and image
quality (affected by CTE). Using smaller pixels partially overcomes the
limitation imposed by the electron mobility rate. However, a smaller pixel has
a smaller charge well capacity.
If the surface states significantly affect CTE, then the fat zero technique
helps. Fat zero puts a known amount of electrons into the well. These electrons
will fill the interface states and thereby improve the CTE for the photoelectrons.
Because the fat zero electrons partially fill the charge well, the remaining well
capacity for the photoelectrons is reduced.
I
Shift Register
Detectors I I I l 11 I I I I I I I I I
I Shift Register
Figure 3-29. For long arrays, the MTF can be increased by alternating
the readouts or by reading out subarrays (see Figure 3-14, page 61).
78 CCD ARRAYS, CAMERAS, and DISPLAYS
There are three methods to measure CTE: current, optical, and x-ray
injection. Because most devices do not have current injection terminals, the
optical and x-ray injection methods are used more often. With the optical
technique, either a row of detectors or a single detector is illuminated. The ratio
of the output of the first illuminated pixel to the average value of trailing pixels
is measured. The CTE is found using Equation 3-3. The optical technique
requires a spot 14 whose diameter is less than a pixel width. This can only be
achieved with carefully designed optics that are critically aligned. But x-rays 1
generate a known number of electrons in a localized area that is much smaller
than a pixel. The x-ray technique probably provides the highest accuracy.
r VREF
~Reset Ve
I Vout
The CCD output is proportional to the exposure, Lq(A)trNT· The output can
be increased by increasing the integration time and long integration times are
generally used for low light level operation. However, this approach is
ultimately limited by dark current leakage that is integrated along with the
photocurrent. With a large pixel (24 llm square), a dark current density of 1000
pA/cm2 produces 36,000 electrons/pixel/s. If the device has a well capacity of
360,000 electrons, the well fills in 10 s. The dark current is only appreciable
when t1NT is long. It is not usually a problem for consumer applications but is a
concern for scientific applications.
A critical design parameter is dark current noise reduction. There are three
main sources of dark current: (1) thermal generation in the depletion region, (2)
thermal generation and diffusion in the neutral bulk material, and (3) thermal
generation due to surface states. Dark current densities vary significantly among
manufacturers with values ranging from 0.1 nA/cm2 to 10 nA/cm2 in silicon
CCDs. Dark current due to thermally generated electrons can be reduced by
cooling the device. Surface state dark current is minimized with multi-phase
pinning 15 •16 (MPP).
80 CCD ARRAYS, CAMERAS, and DISPLAYS
where Tis the absolute temperature, &, is the band gap, and Er is the impurity
energy gap. With TDI, the pixels in a TDI column are summed and the dark
current increases:
JDADtiNTN1DI (3-8)
nDARK- q ---
Table 3-1
DARK CURRENT VALUES for T1 TC271
(Normalized to 100 pA/cm2 at 27°C)
TEMPERATURE DARK PERCENT CHANGE DOUBLING
(oC) CURRENT for lo C TEMPERATURE
(pA/cm2)
Window
Figure 3-31. Cross section of a two stage TEC integrated into a package.
A CCD array has a well size of 150,000 electrons, dark current density of
10pA/cm2 , and detector size 8~tm square. The advertised dynamic range is 80
dB. Is the dark current noise significant? Assume that the integration time is
1160 s (16.67 ms).
Assuming Poisson statistics, the dark current shot noise is ....{668 or 26 electrons
rms. The dynamic range is the charge well capacity divided by the noise floor.
With a dynamic range of 10,000:1, the noise tloor is 15 electrons. Because the
noise variances add, the total noise is:
Example 3-5
ARRAY COOLING
The dark current density decreases by a factor of 2 for every soc drop in
temperature. What should the array temperature be so that the dark current noise
is 5 electrons rms for the array described in Example 3-4?
The dark current noise must be reduced by a factor of 5/26 = 0.192 and the
dark current must be reduced by (0.1921 or 0.0369. The doubling factor
approximately provides
Then, TAMBIENT- TcooL = 38.1 oc. That is, the array temperature must be 38.1 oc
below the temperature at which the dark current density was originally
measured. Because the dark current is reduced by 0.0369, the time to saturation
increased by 1/0/0369 or 27.1 times to 102 s.
Although this varies by manufacturer, many use TAMBIENT = 20°C. Note that
this is an approximation only and Equation 3-7 should be used to determine the
required temperature. In reality, the user does not precisely select the
temperature but, if possible, cools the array well below the desired temperature.
The lowest possible temperature is limited by the cooling device.
=============================================~
Many arrays have extra photosites at the end of the array. These shielded
II II
photosites are used to establish a reference dark current level (or dark signal).
The number of dark elements varies with device and manufacturer and ranges
from a few to 25. The average value of the dark current pixels is subtracted
from the active pixels leaving only photogenerated signal. For example, if the
output of the dark pixels is 5 mV, then 5 mV is subtracted from the signal level
of each active pixel.
Figure 3-32a illustrates the output from a single charge well. Light leakage
may occur at the edge of the shield (see Example 3-1, page 62) and partially fill
a few wells. Because of this light leakage, a few pixels (called dummy or
84 CCD ARRAYS, CAMERAS, and DISPLAYS
isolation pixels) are added to the array. Figure 3-32b illustrates the output of
three dark reference detectors, three isolation detectors, and three active pixels.
The average value of the three dark pixels is subtracted from every active pixel
value.
Dark Current
-<D Vsignal
::I Cl
0.(1:1
Photo Electrons
s=
og Vout
0-+----.....__ _ _ _ __
(a)
Dark Current
Reference Pixels Active Pixels
Light Leakage
(b)
Figure 3-32. Typical output. (a) Single pixel and (b) active line with
isolation and dark pixels. The varying voltage on the active pixels is
proportional to the scene intensity. The dark pixels are covered with an
opaque mask. Light can leak around the mask and partially fill the isolation
pixel wells.
While this process is satisfactory for general imagery and machine vision
use, it may be unacceptable for scientific applications. Dark pixels have slightly
different dark current than the active pixels. Furthermore, the dark current value
varies from pixel to pixel. This variation appears as fixed pattern noise
(discussed in Section 4.2.6., Pattern Noise). Removing the average value does
not remove the variability.
In critical scientific applications, only the dark value from a pixel may be
removed from that pixel. That is, the entire array is covered (mechanical shutter
in place) and the individual pixel values are stored in a matrix. Then the shutter
is removed and the stored values are subtracted, pixel by pixel, as a matrix
subtraction. This method ensures that the precise dark voltage is subtracted from
each pixel. This, of course, increases computational complexity. Removal of
dark current values allows maximum use of the analog-to-digital converter
SOLID STATE ARRAYS 85
dynamic range. This method cannot remove the dark current shot noise. Dark
current shot noise is reduced by cooling.
Additional isolation pixels may exist between the readout register and the
sense node amplifier. These pixels reduce any possible interaction between the
amplifier temperature and first column pixel dark current. They also reduce
amplifier noise interactions.
When a well fills, charge spills over into adjacent pixels in the same column
resulting in an undesirable overload effect called blooming. Channel stops
prevent spill over onto adjacent columns. Overflow drains or antibloom drains
prevent spill over. The drain can be attached to every pixel or may only operate
on a column of pixels. Any photoelectron that is swept into the drain is instantly
removed. Thus, the location of the drain is important.
Figure 3-34 illustrates the potential well created by a lateral drain. When the
charge well fills over the drain barrier, charge flows into the drain. The barrier
height can be controlled by an external voltage so that variable integration times
are possible. Here, the antibloom drain is used for exposure control. The
exposure control can work at any speed. The drain, acting as an electronic
shutter, may be used synchronously or asynchronously. Asynchronous shuttering
is important for use in processes where the motion of objects does not match the
array frame timing or in high light level applications. The charge well is
drained, reset, and then photoelectrons are integrated for the desired time.
86 CCD ARRAYS, CAMERAS, and DISPLAYS
Transfer
Gate
Lateral Shielded
Anti-Bloom Transfer
Drain Register
Figure 3-33. Lateral antibloom drain. The transfer gate controls the
desired integration time.
White
Clip
Level
Figure 3-34. Lateral overflow drain. When the number of stored electrons
exceeds the saddlepoint, the excess electrons flow into the drain
response above the white clipping level. Each CCD manufacturer claims various
competitive methods of handling wide scene contrasts. These methods usually
involve manipulating the drain potential.
Interline transfer devices may have the drain at the top of a vertical shift
register. The excess charge cannot be drained while image information is in the
vertical shift registers (Figure 3-36). This limits the range of integration times
available. Figure 3-37 illustrates the variable integration afforded by the drain.
1 _White Clip~v~
--
::I
c.
::I
1 ""'-Knee
I
Point
0
I
I
Drain
Drain
To Sense Node
(a) (b)
Figure 3-36. Interline transfer device where the drain is at the top of the
vertical shift register. (a) Unwanted charge is sent to the drain and (b)
desired charge is read out.
Field Time
.. I
Transfer _ _ _ _ _ _ _ _ _ _ __..n.___ _.ll_
Gate
1+-----"Shutter"_ _ __..,+,...,_"Shutter" _I
Closed Open ~
3.9. CID
Column Select
The conversion of charge into voltage depends upon the pixel, readout line,
and amplifier capacitance. The capacitance is high because all the pixels on a
given row are tied in parallel. Therefore, when compared to a CCD, the charge
conversion is small, yielding a small signal-to-noise ratio. Because the readout
is nondestructive, it can be repeated numerous times. The multiple reads are
averaged together to improve the signal-to-noise ratio.
90 CCD ARRAYS, CAMERAS, and DISPLAYS
~
l..-1 (a) (b)
v1 > v2
I
> v3 *
(c)
\
Figure 3-40. CID pixel operation. (a) Integration, (b) readout, and (c)
injection.
SOLID STATE ARRAYS 91
Because each pixel sees a different capacitance, CIDs tend to have higher
pattern noise compared to CCDs. However, off-chip algorithms can reduce the
amount of pattern noise. With no charge transfer, CIDs are not sensitive to CTE
effects. Without multiple gates, CIDs have larger well capacities than
comparably sized CCDs. CIDs inherently have antibloom capability. Because
charge is limited to a single pixel, it cannot overflow into neighboring pixels.
Perhaps the greatest advantage of CIDs is random access to any pixel or pixel
cluster. Subframes and binned pixels can be read out at high frame rates.
3.10. CMOS
The main advantage of CMOS imagers is that they are compatible with
mainstream silicon chip technology. Because transistors are made by this
technology, on-chip processing is possible. 21 An active pixel sensor (APS) has
one or more active transistors integrated into the pixel (Figure 3-41).
Column Select
Row------------------------.---~~
Select _______________ _
I
(a)
~- ;ix~l/------------ Output
Column Select
Row
Select -----------------------~--___;~-
~--------------- -.
(b)
1 Voo
,.. I
L------------------
PiXel / Output
Figure 3-41. (a) Passive pixel device and (b) active pixel sensor. CCDs
and CIDs are regarded as passive pixel sensors. CIDs use photogates,
CCDs use either photogates or photodiodes, and CMOS devices typically
use photodiodes.
92 CCD ARRAYS, CAMERAS, and DISPLAYS
With the APS approach, highly integrated image sensors are possible. 22 By
placing processing on the chip, a CMOS camera is typically physically smaller
than a CCD camera which requires clocks, image reformatting, and signal
processing in separate hardware. A sophisticated APS array could create 23 a
"camera-on-a-chip." It is possible to build a frame transfer device where pixels
can be binned to enhance the SNR and provide variable resolution imaging. 24
CMOS devices tend to have higher dark currents due to the highly doped
silicon used. As such it does not appear that CVIOS sensors will replace CCDs
in low noise scientific applications. Because the active devices often take up real
estate, the area for the photosensor is reduced. This leads to reduced sensitivity
that can be partially offset by a microlens. Because each pixel has its own
amplifier, pattern noise is larger. However, more logic can be added to each
pixel and it seems possible to use on-chip signal processing that suppresses
FPN. 26 With charge limited to a single pixel, it cannot overflow into neighboring
pixels and create blooming as seen with CCDs.
While more integration can lead to the camera-on-a chip, production yield
will probably decrease as the number of transistors increase. Consider an array
that contains 640 x 480 pixels (307 ,200 pixels). With a CCD, there are 307,200
elements on the chip (neglecting dark pixels and the readout structure). With an
APS that has a photodiode and three transistors per pixels there are four times
more elements (1 ,228,800 total). If any one of the four elements fails, that pixel
is dead.
3 .11.1. MICROLENSES
Optical fill factor may be less than 100% due to manufacturing constraints
in full transfer devices. In interline devices, the shielded vertical transfer register
can reduce the fill factor to less than 20%. Micro lens assemblies (also called
microlenticular arrays or lenslet arrays) increase the effective optical fill-factor
(Figure 3-42). But it may not reach 100% due to slight misalignment of the
micro lens assembly, imperfections in the micro lens itself, nonsymmetric shielded
areas, and transmission losses. As shown by the camera formula (Equation 2-16,
page 35), the output is directly proportional to the detector area. Increasing the
optical fill factor with a microlens assembly increases the effective detector size
and, therefore, the output voltage.
The photosensitive area is below the gate structure and the ability to collect
the light depends upon gate thickness. The cone of light reaching the microlens
depends upon the f-number of the primary camera lens. Figure 3-42 illustrates
nearly parallel rays falling on the microlens. This case is encountered with high
f-number lens systems. Low f-number primary camera lenses increase the cone
angle and the effective fill-factor decreases with decreasing f-numberY
Microlenses are optimized for most practical f-numbers. As the array size
grows, off-axis detectors do not obtain the same benefit as on-axis detectors. 28
Detectors
(a)
Detectors
(b)
The subjective sensation of color can be created from three primary colors.
By adjusting the intensity of each primary (additive mixing), a full gamut
(rainbow) of colors is experienced. This approach is used on all color displays.
They have red, green, and blue phosphors that, when appropriately excited,
produce a wide spectrum of perceived colors. The CIE committee standardized
a color perception model for the human observer in 1931. The literature29 is rich
with visual data that form the basis for color camera design.
SOLID STATE ARRAYS 95
The "color" signals sent to the display must be generated by three detectors,
each sensitive to a primary or its complement. The primary additive colors are
red, green, and blue (R, G, B) and their complementary colors are yellow, cyan
and magenta (Ye, Cy, Mg). For high quality color imagery, three separate
detectors (discussed in Section 5. 3. 3., Color Correction) are used whereas for
consumer applications, a single array is used. The detectors are covered with
different filters that, with the detector response, can approximate the primaries
or their complements (Figure 3-43). A single array with filters is called a color
filter array (CFA).
1.0 .-----,----,--G-/""7""'i;\:-;---..,----,----,
f. ............................. j.......:/··1···············\+······························+-······························+················
0.9 ··················
i :: v
a: 0.2
1
............................................
/,\ 1\
I·······\\·············+································!······························!
0.1 ................................. 77 ··'S..._-, .... ~
0.0
0.4 0.5 0.6 0.7
Wavelength (~m)
(a)
0.9
0.8
3lc 0.7
0
c. 0.6
Ill
CD
a: 0.5
CD
>
:;:::
0.4
ca 0.3
Qj
a: 0.2
0.1
0.5 0.6
Wavelength (~)
(b)
Figure 3-43. Desired spectral response for the three primaries and their
complements. (a) The primaries and (b) their complements.
96 CCD ARRAYS, CAMERAS, and DISPLAl:S
Cy-G+B-W··R.
The arrangement of the color filters for a single array system is either a
stripe or a mosaic pattern (Figure 3-44). The precise layout of the mosaic varies
by manufacturer. One basic CFA patent30 was granted to Bryce E. Bayer at
Eastman Kodak in 1976.
R G B R G B G Ye G Cy G
R G B R G B G Ye G Cy G
R G B R G B G Ye G Cy G
R G B R G B G Ye G Cy G
Stripes
Cy w Ye G G Mg G Mg R G R G
Ye G Cy w Cy Ye Cy Ye G B G B
Cy w Ye G Mg G Mg G R G R G
Ye G Cy w Cy Ye Cy Ye G B G B
Mosaics
Figure 3-44. Representative stripe and mosaic arrays. Although shown
as a full frame device for clarity, they typically are interline transfer
devices. The layout depends upon the manufacturer's philosophy and
cleverness in reducing color aliasing.
SOLID STATE ARRAYS 97
For video applications, several rows or columns are devoted to dark current
and for light leakage. Therefore an array may be 650 X 492 but the light
sensitive part may be 640 x 480. It is manufacturer dependent whether to
specify the array size by the number of active pixels or the total number of
pixels (which includes the dark pixels).
Scientific array size tends to a power of 2 (e.g., 512 x 512, 1024 x 1024)
for easy image processing. There is a perception that "bigger is better" both in
terms of array size and dynamic range. Arrays may reach 8192 x 8192 with a
dynamic range of 16 bits. This array requires (8192)(8192)(16) or 1.07 Gbits
of storage for each image. Image compression schemes may be required if
storage space is limited. The user of these arrays must decide which images are
significant and through data reduction algorithms, store only those that have
value. Otherwise, he will be overwhelmed with mountains of data.
While large format arrays offer the highest resolution, their use is hampered
by readout rate limitations. For example, consumer camcorder systems operating
at 30 frames/s have a data rate of about 10 Mpixels/s. An array with 5120 x
5120 elements operating at 30 frames/s has a data rate of about 768 Mpixels/s.
Large arrays can reduce readout rates by having multiple parallel ports servicing
subarrays. Each subarray requires separate vertical and horizontal clock signals.
The tradeoff is frame rate (speed) versus number of parallel ports (complexity
of CCD design) and interfacing with downstream electronics. Because each
subarray is serviced by different on-chip and off-chip amplifiers, the displayed
image of the subarrays may vary in contrast and level. This is due to differences
in amplifier gains and level adjustments.
Table 3-2
ARRAY SIZE for STANDARD FORMATS
CAMERA STANDARDIZED ARRAY SIZE ARRAY
FORMAT (H XV) DIAGONAL
Although each manufacturer supplies a slightly different array size and pixel
size, nominal sizes for a 768 X 480 array are given in Table 3-3. With interline
transfer devices, approximately one-half of the pixel width is devoted to the
shielded vertical transfer register. That is, the detector active width is one-half
of the pixel width. Thus the active area of the pixel is rectangular in interline
transfer devices. This asymmetry does not appear to affect image quality in
consumer video products significantly.
Table 3-3
NOMINAL PIXEL SIZE for a 768 x 480 ARRAY
Sizes vary by manufacturer. Detector sizes are smaller.
NOMINAL PIXEL
CAMERA
SIZE
FORMAT
(H XV)
The decrease in optical format is related to cost. The price of CCD arrays
is mainly determined by the cost of processing semiconductor wafers. As the
chip size decreases, more devices can be put on a single wafer and this lowers
the price of each individual device. The trend of going to smaller devices will
probably continue as long as the optical and electrical performance of the
imagers does not change. However, smaller pixels reduce the charge well size.
For a fixed flux level and lens f-number, the smaller arrays have reduced
sensitivity.
3.12. REFERENCES
1. J. Janesick, T. Elliott, R. Winzemead, J. Pinter, and R. Dyck, "Sandbox CCDs," in Charge-
Coupled Devices and Solid State Optical Sensors V, M M. Blouke, ed., SPIE Proceedings Vol.
2415, pp. 2-42 (1995).
2. J. Janesick and T. Elliott, "History and Advancement of Large Area Array Scientific CCD
Imagers," in Astronomical Society ofthe Pacific Conference Series, Vol. 23, Astronomical CCD
Observing and Reduction, pp. 1-39, BookCrafters (1992).
3. M. J. Howes and D. V. Morgan, eds., Charge-Coupl1·dDevices and Systems, John Wiley and
Sons, New York, NY (1979).
4. C. H. Sequin and M. F. Tompsett, Charge Transfer Devices, Academic Press, New York, NY
(1975).
5. E. S. Yang, Microelectronic Devices, McGraw-Hill, NY (1988).
6. E. L. Dereniak and D. G. Crowe, Optical Radiation Detectors, pp. 186-269, John Wiley and
Sons, New York, NY (1984).
7. A. J. P. Theuwissen, Solid-State Imaging with Charge-Coupled Devices, Kluwer Academic
Publishers, Dordrecht, The Netherlands (1995).
8. M. Kimata and N. Tubouchi, "Charge Transfer Devices," in Infrared Photon Detectors, A.
Rogalski, ed., pp. 99-144, SPIE Press, Bellingham, WA (1995).
9. A. J.P. Theuwissen, Solid-State Imaging with Charge Coupled Devices, pp. 317-348, Kluwer
Academic Publishers, Dordrecht, The Netherlands (1995).
10. A. J. P. Theuwissen, Solid-State Imaging with Charge-Coupled Devices, pp. 54-66, Kluwer
Academic Publishers, Dordrecht, The Netherlands (1995).
11. A. J. P. Theuwissen, Solid-State Imaging with Charge Coupled Devices, pp. 161-165, Kluwer
Academic Publishers, Dordrecht, The Netherlands (1995).
12. H.-S. Wong, Y. L. Yao, and E. S. Schlig, "TDI Charge-Coupled Devices: Design and
Applications," IBM Journal Research Development, Vol. 36(1), pp. 83-106 (1992).
13. T. S. Lomheim and L. S. Kalman, "Analytical Modeling and Digital Simulation of Scanning
Charge-Coupled Device Imaging Systems," in Electro-Optical Displays, M. A. Karim, ed., pp.
551-560, Marcel Dekker, New York (1992).
14. G. Wan, X. Gong, and Z. Luo, "Studies on the Measurement of Charge Transfer Efficiency
and Photoresponse Nonuniformity of Linear Charge-coupled Devices," Optical Engineering,
Vol. 34 (11), pp.3254-3260 (1995).
15. J. Janesick, "Open Pinned-Phase CCD Technology," in EUV, X-ray, and Gamma Ray
Instruments for Astronomy and Atomic Physics, C. J. Hailey and 0. H. Siegmund, eds., SPIE
Proceedings Vol. 1159, pp 363-373 (1989).
16. J. Janesick, T. Elliot, G. Fraschetti, S. Collins, M. Blouke, and B. Corey, "CCD Pinning
Technologies," in Optical Sensors and Electronic Photography, M. M. Blouke, ed., SPIE
Proceedings Vol. 1071, pp. 153-169 (1989).
17. J. S. Campbell, "TC271 Characterization Report" in Area Array Image Sensor Products, Texas
Instruments Data Catalog, pp. B-43 to B-46 (1996).
18. J. R. Janesick, T. Elliott, S. Collins, M. M. Blouke, and J. Freeman, "Scientific Charge-
coupled Devices," Optical Engineering, Vol. 26(8), pp. 692-714 (1987).
19. C. Buil, CCD Astronomy, pp. 125-152, Willmann-Bell, Richmond, VA (1991).
20. S. Kawai, M. Morimoto, N. Mutoh, and N. Teranishi, "Photo Response Analysis in CCD
Image Sensors with a VOD Structure," IEEE Transacticms on Electron Devices, Vol. 40(4), pp.
652-655 (1995).
21. M. Schanz, W. Brockherde, R. Hauschild, B. Hosticka. and M. Schwarz, "Smart CMOS Image
Sensor Arrays," IEEE Transactions on Electron Devices, Vol. 44(10), pp. 1699-1704 (1997).
22. S. K. Mendis, S. E. Kemeny, and E. R. Fossum, "A 128 x 128 CMOS Active Pixel Sensor
for Highly Integrated Imaging Systems," IEEE IEDM Technical Digest, pp. 583-586 (1993).
SOLID STATE ARRAYS 101
Because the array is the basic building block of the camera, camera
terminology is often used for array specifications. Additional performance
measures exist at the system level. These include the noise equivalent irradiance
(NEI) and noise equivalent differential reflectance (NE.1p) and are discussed in
Chapter 6, Camera Performance.
The magnitude of each noise component must be quantified and its effect on
system performance must be understood. Noise sources may be a function of the
detector temperature. Predicted system performance may deviate significantly
from actual performance if significant 1/f noise or other noise is present. There
are a myriad of factors involved in system optimization. It is essential to
understand what limits the system performance so that intelligent improvements
can be made.
102
ARRAY PERFORMANCE 103
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
4.1. SIGNAL
The maximum output occurs when the charge wells are filled. The exposure
that produces this value is the saturation equivalent exposure (SEE). With this
definition, it is assumed that the dark current produces an insignificant number
of electrons so that only photoelectrons fill the well.
Sensitivity suggests something about the lowest signal that can be detected.
It is usually defined as the input signal that produces a signal-to-noise ratio of
one. This exposure is the noise equivalent exposure (NEE). The minimum signal
is typically given as equivalent electrons rms. It is only one of many
performance parameters used to describe system noise performance.
Figure 4-1 illustrates a signal transfer diagram. The shutter represents the
integration time, t1NT. The detector converts the incident photons, noETECTOR• into
electrons at a rate determined by the quantum efficiency. A voltage is created
when the electrons are transferred to the sense node capacitance. The signal
after the source follower amplifier is
Gq (4-1)
VsiGNAL - n, C •
where n, is the total number of electrons in the charge packet. It includes both
photoelectrons and dark current electrons. The output gain conversion, Gq/C,
is typically 0.1 p.V /e- to 10 p.V /e-. The source follower amplifier gain, G, is near
unity and, therefore, it is sometimes omitted from radiometric equations.
Amplifiers, external to the CCD device, amplify the signal to a useful voltage.
These amplifiers are said to be off-sensor or off-chip.
104 CCD ARRAYS, CAMERAS, and DISPLAYS
Shutter
(Integration Sense Node
Time) (Charge Off-Chip
"=~
Detector Conversion) Amplifier
VCAMERA
Digital
(Analog Output) Output
For an ideal material, when the photon energy is greater than the
semiconductor band gap energy, each photon produces one electron-hole pair
(quantum efficiency is one). However, the absorption coefficient in silicon is
wavelength dependent (Figure 4-2). Long wavelength photons are absorbed
deeper into the substrate than short wavelengths. Very long wavelength photons
may pass through the CCD and not be absorbed. Beyond 1.1 ~-tm, the absorption
is essentially zero because the photon energy is less than the band gap energy.
For a doping concentration of 10 17cm-3 , the depth at which 90% of the incident
photons are absorbed is given in Table 4-1. Because of reflections, the photon
incidance is higher. The values in Table 4-1 apply only to those photons that
enter the material.
ARRAY PERFORMANCE 105
1()4
1()3
102 ·1019fcm3
10 ·1018fcm3
s1017fcm3
1
0.4 0.5 0.6 0.7 0.8 0.9 1.0 1.1
Wavelength (l!m)
Any photon absorbed within the depletion will yield a quantum efficiency
near unity. However, the depletion region size is finite and long wavelength
photons will be absorbed within the bulk material. An electron generated in the
substrate will experience a three-dimensional random walk until it recombines
or reaches the edge of a depletion region where the electric field exists. If the
diffusion length is zero, all electrons created within the bulk material will
recombine immediately and the quantum efficiency approaches zero for these
photons. As the diffusion length approaches infinity, the electrons eventually
reach a charge well and are stored. Here, the quantum efficiency approaches
one.
106 CCD ARRAYS, CAMERAS, and DISPLAYS
Table 4-1
REPRESENTATIVE 90% ABSORPTION DEPTH
Wavelength (~tm) Depth (~tm)
0.40 0.19
0.45 1.0
0.50 2.3
0.55 3.3
0.60 5.0
0.65 7.6
0.70 8.5
0.75 16
0.80 23
0.85 46
0.90 62
0.95 150
1.00 470
1.05 1500
1.10 7600
>-
0
z
w 0.8
0 ideal
u::
u.. 0.6
silicon
w infinite diffusion
:Ill 0.4 length
::I
1-
z 0.2 zero diffusion
<C length
::I
a 0
0.4 0.6 0.8 1.2
WAVELENGTH (pm)
Ull )< i X )(
Figure 4-4. Idealized relative spatial response to long wavelength
photons.
~
c 50
Cll
·u-- 40
:ec
w ~ 30
_E 8....
:::J
....... 20 Front Illuminated
with UV Coating
c
ca
:::J
a OL-L-~~~d_~~~_L-~_L~_L_L~~~~
0.2 0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 1.1
Wavelength {llm)
~ 60 .-----------------------------------------,
g Front Illuminated
Cll 50
~ E' 4o
w ~ 30
_ Gi
E .......
a.. 20
:::J
c
ca 10
:::J
a oL-L-L-~~_L_L_L~~_J~L_L_L_L_~~~~
Devices with vertical antibloom drains have reduced red response. The
buried drain is in close proximity to where long photon absorption occurs. Any
electron that enters the drain is instantly removed.
Illuminating the array from the back side avoids the polysilicon problem and
increases the quantum efficiency below 0.6 p,m (Figure 4-6). Photons entering
the back side are absorbed in the silicon and diffuse to the depletion region.
However, short wavelength photons are absorbed near the surface and these
electron-hole pairs recombine before reaching a storage site in a thick wafer (see
Table 4-1). Therefore, the wafer is thinned to about 10 p,m to maintain good
spectral responsivity (Figure 4-7). In back-side thinned devices, the incident
photon flux does not have to penetrate the polysilicon gate sandwich structure
and interference effects are much easier to control. With a proper anti-reflection
coating, a quantum efficiency of approximately 85% is possible. However,
silicon has an index of refraction that varies with wavelength making it
somewhat difficult to manufacture an anti-reflection coating that is effective
across the entire spectrum. Coatings that optimize the response in the near IR
110 CCD ARRAYS, CAMERAS, and DISPLAYS
have reduced effectiveness in the visible (Figure 4-8). Lumogen reduces the
effectiveness of the anti-reflection coating. New anti-reflection coatings
overcome this problem and some offer UV response without the use of
lumogen. 6
80
---8
c
70 Back Illuminated
.
!
60 .... uv
Coating
>- 50
(,)
c
CD
·u 40
J
ffi
0
-
E
::J
c
I'll
::J
30
20
"' Front Illuminated
10 "'
0
0.3 0.4 0.5 0.6 0.7 0.8 0.9 1.0 1.1
Wavelength (J..Lm)
Silicon Silicon
Dioxide
Gate
Thinned
Silicon
Incoming
Light
A B
Figure 4-7. (a) Front illuminated and (b) back illuminated arrays. Owing to
their extremely complex and fragile design, back illuminated devices are
usually limited to scientific applications that require high quantum efficiency.
ARRAY PERFORMANCE 111
100
'#. 80
>
()
c:
Q)
60
·c:;
:0:::
.....
w 40
...c:E
:I
20
ca
:I
0
0
0.3 0.7 0.8 0.9 1.0 1.1
WAVELENGTH (Jim)
Figure 4-8. Effect of different anti-reflection coatings on back
illuminated CCDs. (From Reference 5).
CCD arrays for general and industrial applications are within a sealed
environment for protection. Light must pass though a window (may be glass or
quartz) to reach the array. In addition to reflection losses at all wavelengths, the
glass transmittance decreases for wavelengths below 0.4 JLm. For scientific
applications requiring high sensitivity, the glass can be coated with a broad band
anti-reflection coating. For UV applications, a quartz UV transmitting window
can be used.
4.1.2. RESPONSIVITY
Assume that the array is at a distance of R 1 from an ideal source whose area
is As. This radiometric setup is similar to that described in Section 2.5., Camera
Formula, page 33, except no lens is present. For convenience, radiant quantities
are used rather than photon flux. Assuming a Lambertian source,
n
1A A
= --S -
D
"J-
2
M (.A)
- R .,) t
e (~~. d'11. (4-2)
pe q 2 1t e lNT '
Rl AI
where R,(>-.) is the spectral response expressed in units of A/W. Note that the
detector incidance is E0 (A) = M0 (>-.)As/R?, Then
(4-3)
The array output voltage (after the source follow amplifier) is (G q ~)/C:
(4-4)
or
(4-5)
(4-6)
The value RAvE is an average response that has units of V/(J-cm-2) and the
quantity in the brackets has units of J/cm2 •
ARRAY PERFORMANCE 113
j M/'A)d'A
}.I
Note that Rq and RAvE are the for array package. Rq or RAvE is the detector
responsivity multiplied by the window transmittance. Usually the manufacturer
includes the window transmittance in the quoted responsivity.
Slope
-"~~ Saturation
--
::I
c.
::I Dark
0 Current
-----No Dark
Current
SEE
Exposure _ _____,~
The value RAvE is an average type responsh ity that depends upon the source
characteristics and the spectral quantum efficiency. While the source can be
standardized (e.g., CIE illuminant A or illuminant D 6500 ), the spectral quantum
efficiency varies by device (see Figure 4-5, page 108). Therefore extreme care
must be exercised when comparing devices solely by the average responsivity.
RSIGNAL(').. o) =
Gq ~A R ('A ) --
v
C he D '~ 0
Jfcm 2 (4-9)
G
= -A R ('A)
C D e o
or
RSIGNAL(').. 0) = G ~A D R q,0.)
C 1.24 0
-v-2 (4-10)
Jfcm
If the device has a digital output, the number of counts (integer value) is
. [ VCAMERA -2N
DN = mt -] , (4-11)
VMAX
Gq (4-12)
vMAX- GlcNWELL.
DN (4-13)
Jfcm2
This assumes that the analog-to-digital converter input dynamic range exactly
matches V MAx· This may not be the case in real systems. If an antibloom drain
is present, then the responsivity is defined up to the knee point (see Figure 3-35,
page 87) and Equation 4-11 is modified accordingly.
When using arrays for general video, the output is normalized to photometric
units (see Section 2.3.1., Units, page 25). It is convenient to express the array
l
average response as
0.75 11m
VSIGNAL - RPHOTOMETRIC ( 683 - -
AD
1
0.38
f M/'A) V('A) d).. ,
(4-14)
where RPHOTOMETRic has units of V /lux and the bracketed term has units of lux.
Combining Equations 4-4 and 4-14 yields
A A Az
G ____£___!!__
C R2
f L/'A) R,('A) tiNT d'A
RPHOTOMETRIC -
1 ).1
----0.7_s__m_ _ _ _ __
v (4-15)
11 lux
683 -
AD
1
0.38
f M ('A) V('A) d'A
p
116 CCD ARRAYS, CAMERAS, and DISPLAJ:S
GAD
c
f E.(J.. ) R.(). ) tiNT d)..
l v (4-16)
RPHOWMETRJC - ---0--=-.~s-.-
. m-- - - - -
lux
683 f Ee(J..) V(A.) d)..
0.38
The responsivity depends on the spectral response and the spectral content
of the illumination. The apparent variation in output with different light sources
was discussed in Section 2. 7., Normalization Issues, page 39. Selecting an array
based on the responsivity is appropriate if the anticipated scene illumination has
the same color temperature as the calibration temperature. That is, if the
photometric responsivity is measured with a CIE illuminant A and the scene
color temperature is near 2856 K, then selecting an array with the highest
photometric responsivity is appropriate. Other wise, the average photometric
responsivity is used for informational purposes only.
~=============================================
Example 4-1
MAXIMUM OUTPUT
The maximum output is the OGC multiplied by the well capacity or 420 mV.
If the dark current creates 5,000 electrons, then the maximum signal voltage is
reduced to (70,000-5,000)(0.006mV)= 390 mV. This signal may be amplified
by an off-chip amplifier.
ARRAY PERFORMANCE 117
Example 4-2
DIGITAL RESPONSIVITY
A device with a well capacity of 50,000 electrons has its output digitized
with a 12-bit analog-to-digital converter (ADC). What is the digital step-size
(digital responsivity)?
The step size is the well capacity divided by the number of digital steps.
50,000/2 12 = 12.2 electrons/DN. Equation 4-2 (page 112) provides the number
of electrons as a function of input exposure. The relationship between the ADC
maximum value and charge well capacity assumes an exact match between V MAx
and the ADC input range. If these do not match, an additional amplifier must
be inserted just before the ADC.
=============================================~
Example 4-3
RESPONSIVITY
From Table 3-3 (page 99), a 2/3-inch format pixel is approximately 11.4/lm
X 13.8/lm. If the interline transfer device has a 50% fill factor, then A 0 = 7.87
X 10-7 cm2 • The constant he is 1.99 X 10-25 J-m/photon. Using Equation 4-9,
6
06
RSlGNAL(}.)- (6 X 10-6 )( · X IQ- )(7.87 X 10-7 ) (0.8)- 11.4 X 106
• 1.99 x w-zs (4-17)
V
J-cm-2 -
(-;:V) ((J-m)fphoton
m ) (cm2) (electrons)
photons
118 CCD ARRAYS, CAMERAS, and DISPLAYS
~=============================================
Example 4-4
PHOTOMETRIC OUTPUT
If the color temperature does not change and only the intensity changes, then
the output is (250 p,V)(1000) = 250 mV. However, if the color temperature
changes, the output changes in a nonlinear manner. Similarly, two sources may
produce the same illuminance but VsiGNAL may be quite different. The color
temperature must always be specified when RPffoTaMETRic is quoted.
w 100
(/)
z
0 75
ll.
(/)
w
0::: 50
w
>
i= 25
:5
w
0::: 0
0.35 0.45 0.55 0.65 0.75 0.85 0.95
WAVELENGTH (pm)
Figure 4-10. Relative spectral responsivity of RsiGNAL(A).
Table 4-2
EQUATION 4-18 BRACKETED TERM
COLOR NO INFRARED WITH INFRARED
TEMPERATURE FILTER FILTER
Table 4-3
PHOTOMETRIC RESPONSIVITY (V flux)
responsivity and noise. The NEE is the exposure that produces a signal-to-noise
ratio of one. If the measured rms noise on the analog output is VNOISE• then the
NEE is calculated from the radiometric calibration:
When equated to electrons, NEE is simply the noise value in rms electrons. The
absolute minimum noise level is the noise floor and this value is used most often
for the NEE. Noise sources are discussed in Section 4.2., Noise. Although noise
is arms value, the notation rms is often omitted. When divided by the average
quantum efficiency, the noise equivalent signal (NES) is obtained:
The maximum signal is that input signal that saturates the charge well and
is called the saturation equivalent exposure (SEE). It is
VMAX J (4-21)
SEE-
RAVE
The well size varies with architecture, number of phases, and pixel size. The
well size is approximately proportional to pixel area (Table 4-4). Small pixels
have small wells. If an antibloom drain is present, the maximum level is taken
as the white clip level (see Figure 3-35, page 87). The maximum value of VMAx
is
Gq (4-22)
VMAx- G i - (NWEu- nDARK) •
c
For back-of-the-envelope calculations, the dark current is considered negligible
(i.e., the device is cooled or MPP is implemented). Then noARK ::::: 0.
ARRAY PERFORMANCE 121
Table 4-4
NOMINAL WELL CAPACITY for
TEXAS INSTRUMENTS FRAME TRANSFER DEVICES
(Varies by manufacturer and device type)
WELL CAPACITY
FORMAT
(electrons)
Dynamic range defined as the maximum signal (peak) divided by the rms
noise. If an antibloom drain is present, the knee value is used as the maximum
(See Figure 3-35, page 87). Expressed as a ratio, the dynamic range is
DRARRAY- SEE)
20log (- - dB. (4-24)
NEE
NWEu- nDARK
(4-25)
DRARRAY-
where < nsvs > is the system noise measured in electrons rms. The effects of
< nsvs > on dynamic range is further discussed in Section 4. 3., Array Signal-to-
Noise Ratio. Often, only the noise floor is considered in < nsvs > . The noise is
dependent upon amplifier design and is not related to charge well capacity.
122 CCD ARRAYS, CAMERAS, and DISPLAlS
While larger arrays have larger well capacities, this does not necessary mean
that the dynamic range increases by the same amount. As the pixel rate
increases, the amplifier bandwidth increases. Simultaneously, the amplifier noise
increases and the dynamic range decreases.
Sometimes the manufacturer will specify dynamic range by the peak signal
divided by the peak-to-peak noise. This peak-to-peak value includes any noise
spikes and, as such, may be more useful in determining the lowest detectable
signal in some applications. If the noise is purely Gaussian, the peak-to-peak is
assumed to be four to six times the rms value. This multiplicative factor is
author dependent.
If the system noise is less than the ADC's least significant bit (LSB), then
the quantization noise limits the system dynamic range. The rms quantization
noise, < nAoc > , is VLssl,[12 where VLss is the "oltage step corresponding to the
least significant bit. The dynamic range is 2N ,[12. For a 12-bit ADC, the
dynamic range is 14,189:1 or 83 dB. An 8-bit system cannot have a dynamic
range greater than 887:1 or 59 dB.
Example 4-6
ARRAY PARAMETERS
Maximum output:
The maximum output is RAvE SEE= 300mV.
Noise level:
The noise floor rms value is equal to the maximum output divided by the
dynamic range or 100 p.V rms.
Analog-to-digital Converter:
For a digital output, the ADC must have 12 bits (212 = 4096 levels) to span
the array's full dynamic range. The responsivity in DN is 2 12/SEE = 27.3
DN/(nJ/cm2) at Ap.
============================================~
ARRAY PERFORMANCE 123
4.2. NOISE
Many books and articles7•13 have been written on noise sources. The level of
detail used in noise modeling depends on the application. Shot noise is due to
the discrete nature of electrons. It occurs when photoelectrons are created and
when dark current electrons are present. Additional noise is added when reading
the charge (reset noise) and introduced by the amplifier (1/f noise and white
noise). If the output is digitized, the inclusion of quantization noise may be
necessary. Switching transients that are coupled though the clock signals also
appear as noise. It can minimized by good circuit design.
Figure 4-1 (page 104) illustrated the signal transfer diagram. The noise
transfer diagram is somewhat different (Figure 4-11). With different transfer
functions, both the optical and electronic subsystems must be considered to
maximize the system signal-to-noise ratio.
Photon Noise
Dark Current Noise
Fixed Pattern Noise Reset 1/1 Noise Iff Noise Quantization
Photo Responsle Non::::~::de Nolise On-Chip Amplifielr Nois:ff.Chip Amplifie\r Noise Nolise
Figure 4-11. The various subsystems are considered as ideal elements with
the noise introduced at appropriate locations.
Although the origins of the noise sources are different, they all appear as
variations in the image intensity. Figure 4-12a illustrates the ideal output of the
on-chip amplifier before correlated double sampling. The array is viewing a
uniform source and the output of each pixel is identical. Photon shot noise
produces a temporal variation in the output signal that is proportional to the
square root of the signal level in electrons. Each pixel output will have a slightly
different value (Figure 4-12b).
124 CCD ARRAYS, CAMERAS, and DISPLAYS
Reset
(a)
(b)
Photon
}
Noise
Dark Current
} Noise
(c)
} Reset Noise
(d)
(e)
l.
T Amplifier Noise
(f)
Ill Noise
(g)
Figure 4-12. Noise sources affect the output differently. (a) though (f)
represent the output after the on-chip amplifier. (g) is the output after the
sample-and-hold circuitry. The array is viewing a uniform source that
produces identical outputs (amplitude = A) for each pixel. (a) ideal output,
(b) photon shot noise, (c), dark current shot noise, (d) reset noise, (e)
amplifier noise, (f) amplifier 1/f noise, and (g) quantization noise. These
values change from frame-to-frame. Pattern noise is not shown but is a
variation that does not change significantly each frame. All these processes
occur simultaneously.
ARRAY PERFORMANCE 125
Under ideal conditions, each pixel would have the same dark current
and this value is subtracted from all pixels leaving the desired signal (see
Section 3.6., Dark Current, page 79). However, dark current also exhibits
fluctuations. Even after subtracting the average value, these fluctuations
remain and create fixed pattern noise (Figure 4-12c). The output capacitor
is reset after each charge is read. Errors in the reset voltage appear as output
signal fluctuations (Figure 4-12d). The amplifier adds white noise (Figure
4-12e) and 1/f noise (Figure 4-12t). Amplifier noise affects both the active
video and reset levels. The analog-to-digital converter introduces
quantization noise (Figure 4-12g). Quantization noise is apparent after image
reconstruction. Only photon shot noise and amplifier noise affect the amplitude
of the signal. With the other noise sources listed, the signal amplitude, A,
remains constant. However, the value of the output fluctuates with all the noise
sources. Because this value is presented on the display, the processes appear as
displayed noise. These noise patterns change from pixel-to-pixel and on the same
pixel from frame-to-frame.
Pattern noise refers to any spatial pattern that does not change significantly
from frame-to-frame. Dark current varies from pixel-to-pixel and this variation
is called fixed pattern noise (FPN). FPN is due to differences in detector size,
doping density, and foreign matter getting trapped during fabrication.
Photoresponse nonuniformity (PRNU) is the variation in pixel responsivity and
is seen when the device is illuminated. This noise is due to differences in
detector size, spectral response, and thickness in coatings. These "noises" are
not noise in the usual sense. PRNU occurs when each pixel has a different
average value. This variation appears as spatial noise to the observer. Frame
averaging will reduce all the noise sources except FPN and PRNU. Although
FPN and PRNU are different, they are sometimes collectively called scene
noise, pixel noise, pixel nonuniformity, or simply pattern noise.
Equivalently, the noise sources are RSSed (root sum of the squares) or added
in quadrature:
(4-26)
where (n2 ) is the noise variance for source i and v(n2;) = (n;) is the standard
deviation measured in rms units.
The noise model does not include effects such as banding and streaking
which are forms of pattern noise found in linear and TDI arrays. Banding can
occur with arrays that have multiple outputs serviced by different nonlinear
amplifiers (see Figure 3-14, page 61). Streaking occurs when the average
responsivity changes from column to column in TDI devices. 14 These effects are
incorporated in the three-dimensional noise model (discussed in Section 12.2. ,
Three-dimensional Noise Model) and are used to predict the performance of
military systems. Only a complete signal-to-noise ratio analysis can determine
which noise source dominates (discussed in Section 4.3., Array Signal-to-noise
Ratio).
ARRAY PERFORMANCE 127
Both photoelectrons and dark current contribute to shot noise. Using Poisson
statistics, the variance is equal to the mean:
(4-28)
The number of photoelectrons was given by Equation 4-2, page 112, and the
number of dark current electrons was given by Equation 3-6, page 80. These
values should be modified by the CTE:
Because the CTE is high (CTEN = 1), it is usually omitted from most equations.
However, CTE should be included for very large arrays. With TDI, the number
of photoelectrons and dark current electrons increases with the number of TDI
elements, NTDr:
(4-30)
While the dark current average value can be subtracted from the output to
provide only the signal due to photoelectrons, the dark current shot noise cannot.
Cooling the array can reduce the dark current to a negligible value and thereby
reduce dark current shot noise to a negligible level (see Section 3.6., Dark
Current, page 79).
The noise associated with resetting the sense node capacitor is often called
kTC noise. This is due to thermal noise generated by the resistance, R, within
the resetting FET. The Johnson noise current variance is
(4-31)
128 CCD ARRAYS, CAMERAS, and DISPLAYS
Because the resistance is in parallel with the sense node capacitor, the noise
equivalent bandwidth is .:lf = RC/4. Then
(4-32)
JkTC . (4-33)
(nREsE1) - - - e rms .
q
The value (nRESET) represents the uncertainty in the amount of charge remaining
on the capacitor following reset.
Amplifier noise consists of two components: 1/f noise and white noise. If
fKNEE is the frequency at which the 1/f noise equals the white noise, then the
amplifier noise density is
where S(fe1ec) is the total noise power spectral density from all sources. The
subscript elec is used to emphasize that this is electrical frequency (Hz).
H,y,(felec) is the frequency response of the system electronics. The variable
H(fe1ec) is used by electronic circuitry designers and its magnitude is identical to
the MTF. The noise equivalent bandwidth is that bandwidth with unity value that
provides the same total noise power. Assuming the noise is white over the
spectral region of interest [S(fe~ec) = SJ, ~felec is
The bandwidth ~felec applies only to those noise sources that are white and
cannot be applied to 1/f noise. Although common usage has resulted in calling
~felec the noise bandwidth, it is understood that it is a power equivalency.
If !cLocK is the time between pixels, the bandwidth for an ideal sampled-data
system is
fcLOCK (4-37)
2
V rms, (4-40)
where VLSB is the voltage associated with the least significant bit. For an ADC
with N bits, VLSs = VMAxf2N. When expressed in equivalent electrons,
C LSB _
(nADC) - --- -- e rms . (4-41)
GG1 q {fi
When the ADC is matched to the amplifier output, VMAx corresponds to the full
well. Then
(4-42)
Ideally, (nAoc) is less than the noise floor. This is achieved by selecting a high
resolution ADC (large N).
ARRAY PERFORMANCE 131
Example 4-7
TOTAL NOISE
The array dynamic is range 2500:1 and the ADC only has 256levels. Here, the
ADC noise dominates the array noise. If a 12-bit ADC is used (4096levels), the
LSB represents 12.2 electrons and (nAod = 3.5 electrons rms. The total noise
in electrons is
As with shot noise, CTEN should be added to the equation. Because CTEN = 1
for modest sized arrays, it is omitted from most equations. CTE should be
included for very large arrays. TDI devices, through their inherent averaging,
reduces PRNU by 1/NTDr· PRNU is usually supplied by the manufacturer. Off-
chip gain/level correction algorithms can minimize FPN and PRNU. For system
analysis, the corrected pattern noise value is used.
In principle, FPN and the noise floor can be reduced so that the system is
photon shot noise limited. This provides the theoretical limit:
However, all systems have some pattern noise. The plot of SNR versus
nonuniformity suggests a desired maximum level of acceptable nonuniformity .17
In Figure 4-13, the noise floor is considered negligible:
(4-47)
The total noise value (nsvs) increases when PRNU is excessive. Typically an
array is selected that has small PRNU. This value can be further reduced by off-
chip electronics. The final value should be sufficiently small so that it does not
contribute to the overall noise. Because cost increases as the PRNU decreases,
the optimum PRNU occurs at the knee of the curve. This occurs when the
photon shot noise is approximately equal to PRNU or U = 1/...[~. For worst
case analysis, the charge well capacity should be used for npe.
ARRAY PERFORMANCE 133
Ul 10000
I
CJ)
z
0
a::
.....
0
1000
w
....1
w
w
CJ) ....... - .. photon shot noise
0z 100~--------~~--------~----------~
0.01 0.1 10
PRNU (percent)
Figure 4-13. System noise as a function of PRNU for two signal levels.
The noise floor is considered negligible. The noise floor will increase
(nsvs) and the location of the knee. Shot noise is independent of PRNU
(horizontal dashed lines). The value 1/.J~ is 0.4% and 0.3% when 11pe
is 50,000 and 100,000, respectively. These values represent the "knee"
in the curve.
(4-48)
or
(4-49)
Either the rms noise or noise variance can be plotted as a function of signal
level. The graphs are called the photon transfer curve and the mean-variance
curve, respectively. Both graphs convey the same information. They provide18 •19
array noise and saturation level from which the dynamic range can be calculated.
134 CCD ARRAYS, CAMERAS, and DISPLAYS
For very low photon fluxes, the noise floor dominates. As the incident flux
increases, the photon shot noise dominates. Finally, for very high flux levels,
the noise may be dominated by PRNU. As the signal approaches well saturation,
the noise plateaus and then drops abruptly at saturation (Figure 4-14). As the
charge well reaches saturation, electrons are more likely to spill into adjoining
wells (blooming) and, if present, into the overflow drain. As a result, the
number of noise electrons starts to decrease. Figure 4-15 illustrates the rms
noise as a function of photoelectrons when the dynamic range is 60 dB. Here
dynamic range is defined as NwELd < nFLOOR > . With large signals and small
PRNU, the total noise is dominated by photon shot noise. When PRNU is large,
the array noise is dominated by U at high signal levels.
Dark current shot noise only affects those applications where the signal-to-
noise ratio is low (Figure 4-16). At high illumination levels, either photon shot
noise or pattern noise dominate. Because many scientific applications operate in
a low signal environment, cooling will improve performance. General video and
industrial cameras tend to operate in high signal environments and cooling will
have little effect on performance. A full SNR analysis is required before
selecting a cooled camera.
'iii 1000
s
II)
z 100
~
1-
0
w
..J
w
w
II)
0 1~------~------~------~----~
z
10 100 1000 10000 100000
PHOTO ELECTRONS
Figure 4-14. Photon transfer curve. At low input signals, the device
noise is dominated by the noise floor. As the signal increases, the noise
increases due to photon shot noise and pattern noise. As the well reaches
saturation, there is a drop in noise electrons due to spill over and well
capacity limitation.
ARRAY PERFORMANCE 135
enE 10000
.:.
(/)
z 1000
0 noise floor
a::
1- 100
0
w
..J
w 10
w
!/)
5
z
10 100 1000 10000 100000
PHOTOELECTRONS
(a)
enE 10000
.:.
!/) total noise
z 1000
0 noise floor
a::
1-
0
100
w
..J
w 10
w
!/)
5
z
10 100 1000 10000 100000
PHOTOELECTRONS
(b)
Figure 4-15. Photon transfer curves for (a) U = 2.5% and (b) U =
0.25%. The charge well capacity is 100,000 electrons and the noise
floor is 100 e- rms to produce a dynamic range of 60 dB. The noise floor,
photon shot noise, and PRNU have slopes of 0, 0. 5, and 1, respectively.
Dark noise is considered negligible. The drop at saturation is not shown
(see Figure 4-14).
136 CCD ARRAYS, CAMERAS, and DISPLAYS
Figure 4-16. Dark current shot noise effects. The noise floor is 10
electrons rms and PRNU is zero. (a) Dark current shot noise is 20e- rms
and (b) dark current shot noise is 2 e- rms. For large signals, photon shot
noise dominates the array noise. The drop at saturation is not shown
(Figure 4-14).
ARRAY PERFORMANCE 137
Figure 4-17 illustrates the mean-variance technique. In the shot noise limited
region,
(4-50)
and
G q)
2
2 (4-51)
VNOISE -
(
C nPE •
or the OGC is equal to the noise variance divided by the mean value. It is
simply the slope of the shot noise in Figure 4-17.
1E+08
w
0
zc( 1E+06
~ noise floor
c( 1E+04
>
w
en 1E+02
0
z 1E+OO
10 100 1000 10000 100000 1000000
PHOTOELECTRONS
1E+06
w
0
z
< 1E+04
ii: dark current
~
w
t/)
1E+02
noise floor
0
z 1E+OO
10 100 1000 10000 100000 1000000
PHOTOELECTRONS
Figure 4-17b. Mean-variance plots which corresponds to Figure 4-16a.
The slopes for noise floor, photon shot noise, and PRNU are 0, 1, and
2, respectively.
If both signal and noise are measured after the on-chip amplifier, the
measured photogenerated signal is VsiGNAL = G q npe/C and the noise is
Gq (4-53)
VNOISE - C (nsys) .
When VNOISE = 1, the signal is C/Gq. The shot noise intercept on the photon
transfer curve is 1/0GC (Figure 4-18). The OGC has units of ILV/electrons.
Similarly, if the values are collected after the ADC, when the shot noise is one,
the shot noise signal provides the output gain conversion in units of
DN/electrons.
- 10000
"'E
~ 1000
w
~
noise floor
100L---------------~~--------------
g 1/0GC
w 10
II) ••• • ·photon shot noise
0
z 1~~~~------~------~------
0.1 10 100 1000
SNR- (4-54)
(4-55)
The charge transfer efficiency factor, CTEN, should be added to r1pe in both the
numerator and denominator. For buried channel devices (the most popular)
CTEN is near one and is typically is omitted. Equation 4-55 suggests that the
maximum SNR occurs when nrE is a maximum. This occurs when the charge
wells are filled. However, as the charge wells fill, electrons can get close to the
surface even for buried channel devices. Thus as nrE approaches NwELL• CTE
can decrease. The best performance is a compromise between high SNR and
high CTE. High CTE is generally only a concern with large devices.
The SNR plots are similar to the photon transfer curves. At low flux levels,
the SNR increases with the signal (noise floor is constant). At moderate levels,
photon shot noise limits the SNR and at high levels, the SNR approaches 1/U.
In Figure 4-19, the SNR is expressed as a power ratio (in dB) which is 20
log(SNR). While the array may have a finite PRNU, it can be minimized off-
chip through an appropriate algorithm.
When photon shot noise dominates, the SNR depends only on the photon
flux level (SNR = ..fiLpe). In Figure 4-19, dynamic range is defined as
NwELd < nFwoR > . The instantaneous dynamic range, which includes all noise
sources, is the SNR. The theoretical maximum SNR is ..fNwELL· The SNR is
equal to the dynamic range only when the system is noise floor limited. This
occurs only for low signal values and small wells (less than about 10,000
electrons). For large well capacity arrays, the actual SNR can never reach the
value suggested by the dynamic range. Dynamic range is used to select an
appropriate analog -to-digital converter resolution (number of bits). This assures
that low contrast targets can be seen.
140 CCD ARRAYS, CAMERAS, and DISPLAYS
The array signal-to-noise ratio was derived for a large target that covers
many pixels. The noise rides on the signal. The camera SNR, which includes
all the noise sources, is provided in Section 6. 2. l. , Camera SNR. In Chapter 12,
Minimum Resolvable Contrast, the average noise associated with the target and
the background is considered.
60
,-
50
m 40
-
a:::
z
30
20
tn
10
0
10 100 1000 10000 100000
PHOTOELECTRONS
(a)
60
50
,m 40
-z
a::: 30
20
tn
10
0
10 100 1000 10000 100000
PHOTOELECTRONS
(b)
Figure 4-19. SNR as a function of noise for (a) U = 2.5% and (b) U =
0.25%. PRNU, photon shot noise, and the noise floor have slopes of 0,
10, and 20, respectively. Dark current shot noise is considered
negligible.
ARRAY PERFORMANCE 141
After a charge packet is read, the floating diffusion capacitor is reset before
the next charge packet arrives. The uncertainty in the amount of charge
remaining on the capacitor following reset appears as a voltage fluctuation (see
Section 4.2.2., Reset Noise, page 127).
Correlated double sampling (CDS) assumes that the same voltage fluctuation
is present on both the video and reset levels. That is, the reset and video signals
are correlated in that both have the same fluctuation. The CDS circuitry may be
integrated into the array package and this makes processing easier for the end
user. For this reason it is included in this chapter.
CDS may be performed 20 •21 in the analog domain by a clamping circuit with
a delay-line processor or by the circuit illustrated in Figure 4-20. When done in
the digital domain, a high speed ADC is required that operates much faster than
the pixel clock. Limitations of the ADC may restrict the use of CDS in some
applications. The digitized video and reset pulses are subtracted and then clocked
out at the pixel clock rate.
Valid
Video
I Cv
amplifier
(a) output
valid reset
(b)
valid video
For a fixed output clock rate, the time to read an array increases as the array
size increases. Very large arrays cannot be read in real time. Let the active
array size be m x n pixels. Let there be an additional NREAD pixels before the
sense node. Suppose the clock operates at fcLocK pixels/s. The minimum time to
clock out the array is
tARRAY
_ [mn + NREAD
IeweK
I (4-56)
(4-57)
ARRAY PERFORMANCE 143
The maximum frame rate is limited by NREAD· The actual frame rate may be less
due to electron mobility, on-chip amplifier bandwidth, or clock rate limitations.
For some CCDs and all CIDs and CMOS devices, subarrays (also called
windows) can be read out. As the subarray size decreases, the frame rate
increases, With increased frame rates, the integration time decreases and then
more illumination is required to maintain a specific signal-to-noise ratio.
w 200
1-
~
w 150
:::!!
~
LL 100
w
>
i= 50
:5
w
0:: 0
0 20 40 60 80 100
RELATIVE IMAGE SIZE
4.6. DEFECTS
Table 4-5
ARRAY DEFECTS
4.7. REFERENCES
1. J. Janesick, "Introduction to Scientific Charge-Coupled Devices," Short Course SC-70 presented
at the IS&T/SPIE's Symposium on Electronic Imaging: Science & Technology, San Jose CA
(1995). J. Janesick presents this course at numerous symposia.
2. R. H. Dyck, "VLSI Imagers," in VLSI Electronics: Microstructure Science, Volume 5, N. G.
Einspruch, ed., pp. 594, Academic Press, New York (1985).
3. R. H. Dyck, "Design, Fabrication, and Performance of CCD Imagers," in VLSI Electronics:
Microstructure Science, Volume 3, N. G. Einspruch, ed., pp. 70-71, Academic Press, New
York (1982).
4. A. J.P. Theuwissen, Solid-State Imaging with Charge-Coupled Devices, pp. 205-210, Kluwer
Academic Publishers, Dordrecht, The Netherlands (199:5).
5. The Leader in Spectroscopic Detection, pp. 33-34, Princeton Instruments Catalog, 3660
Quakerbridge Road, Trenton, NJ 08619 (April 1997).
6. The Leader in Spectroscopic Detection, pp. 29-30, Princeton Instruments Catalog, 3660
Quakerbridge Road, Trenton, NJ 08619 (April 1997).
7. C. H. Sequin and M. F. Tompsett, Charge Transfer De1ices, Academic Press, New York, NY
(1975).
ARRAY PERFORMANCE 145
Some agencies involved with standards include the Institute of Electrical and
Electronic Engineers (IEEE), Electronics Industry Association (EIA), EIA-Japan
(EIAJ), Joint Electron Devices Engineering Council (JEDEC), Society of
Motion Picture and Television Engineers (SMPTE), International Electro-
technical Commission (IEC), Technical Center of the European Broadcast Union
(EBU), Advanced Television Systems Committee (ATSC), Comite Consultatif
International des Radiocommunications (CCIR) (now ITU-R), and International
Telecommunications Union-Radiocommunications(ITU-R). Major manufacturers
such as Panasonic, Sony, and JVC influence the standards.
Because solid state devices are designed for specific applications, the array
and camera use the same terminology: progressive scan chips are used in
146
CAMERAS 147
progressive scan cameras, TDI arrays are used in TDI cameras, etc. Through
appropriate clocking, progressive scan cameras can also provide composite video
and be used with consumer displays. Scientific cameras can have both analog
and digital outputs. The corresponding sections in Chapter 3, Solid State Arrays,
should be read with the specific cameras listed here.
Both monochrome and color cameras are quite complex as evidenced by the
1478-page Television Engineering Handbook, K. B. Benson, ed., McGraw-Hill,
New York (1986). This chapter is an overview of solid state camera operation.
It highlights those features and specifications that are most often reported in
camera data sheets. Together with the preceding chapters, it provides the
necessary information to compare and analyze camera systems.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
Figure 5-1 illustrates a generic solid state camera. The lens assembly images
the light onto the detector array. General purpose consumer cameras typically
have an infrared blocking filter to limit the camera's spectral response to the
visible spectral band. Depending upon the application (military, industrial, or
scientific), the camera may or may not have this filter. Color cameras have a
spectral response limited to the visible region. Cameras may provide digital
image processing and these cameras have an internal ADC. Analog cameras will
not have an internal ADC. However, they still spatially sample the scene.
148 CCD ARRAYS, CAMERAS, and DISPLAYS
~
0 IE
Optics
_
CCD Array
Gain/Level
Matchedto
..______.ADC
ADC
ADC
Digitlzatlon
.
- Processing Correction
Digital Camera
DAC Reconstruction
Filter
f-- Output
Output
Figure 5-1. Generic functional block diagram. Not all components exist
in every camera. The gamma correction circuitry is appropriate when the
video is to be viewed on a CRT-based display. Analog cameras will not
have an internal ADC.
A single chip color camera typically has an unequal number of red, green,
and blue detectors resulting in different sampling frequencies for the different
colors (discussed in Section 8.3.2., Detector Array Nyquist Frequency). Because
the detectors are at discrete locations, target edges and the associated ambiguities
are different for each color. A black-to-white edge will appear to have colors at
the transition and monochrome periodic imagery may appear to have color.
Aliasing is extremely annoying in color systems. A birefringent filter placed
between the lens and the array reduces aliasing (discussed in Section 10. 3 .4. ,
Optical Anti-Alias Filter). Any method to reduce aliasing generally reduces the
overall MTF.
Correlated double sampling reduces reset noise and minimizes 1/f noise.
CDS circuitry may be integrated into the array package and this makes
processing easier for the end user.
Image processing or image reconstruction reformats the data into signals and
timing that are consistent with transmission and display requirements. The
advantage of a digital camera is that the output can be directly fed into an image
processor, machine vision system, or digital data capture system. If using an
analog output camera for machine vision, the video signal must pass through an
external analog -to-digital converter. These multiple conversions may modify the
imagery (discussed in Section 8.4., Aliasing in Frame Grabbers and Displays).
Thus, staying within the digital domain ensures maximum image fidelity.
(originally called EIA 170A or RS 170A). The timing for these two standards is
nearly the same. 1
In the NTSC and EIA 170 standards, a frame is composed of 525 scan lines
separated into two sequential fields (even and odd). Because time is required for
vertical retrace, only 485 lines are displayed. There are 242.5 active lines per
field. Because field 1 ends in a half-line, it is called the odd field. Figure 5-2
illustrates how the information is presented on a display. For commercial
television, the aspect ratio (horizontal to vertical extent) is 4:3.
Second Half
of Line 243
Line 244, Field 2-.__ ~/
Line 245, Field 2 ""'- r--:.::.. ..... .~ . ..-- f.-- Line 1, Field 1
_.... -:-./ I;v
;,..- Line 2, Field 1
Line 246, Field 2 \ I' ~::-~
..... Line 3, Field 1
f\-
=
... , '
.......
-- ./
'
''
''
Line 485, Field 2 \ .........
First Half of ''
Line 243
--------------
I\
- -- - ''
~·
~ Line 242, Field 1
Table 5-l
STANDARD VIDEO TIMING
FRAME LINES SCANNING TOTAL MINIMUM
ACTIVE
FORMAT RATE PER FREQIIENCY LINE ACTIVE LINE
LINES
(Hz) FRAME (kHz) TIME (}ts) TIME (}ts)
(5-1)
where fc is the video bandwidth cutoff frequency. The pixel length is defined as
the minimum distance between two impulses such that the two are still
discernible. This occurs when the maximum of one falls on the first minimum
(zero) of the next. The separation of these two vertical lines in time is Llt =
112fc. This is the smallest possible dimension in a linear system. For the video
standard, the number of resels ("pixels") per line is
The nominal4 video bandwidths of NTSC, PAL, and SECAM are 4.2, 5.5,
and 6 MHz, respectively. Table 5-2 lists the nominal number of horizontal
elements per line. The quoted resolution (number of horizontal pixels) varies in
the literature and depends upon the precise bandwidth and active line time
selected by the author. Multiplying the pixel number by the number of active
lines yields the elements per frame.
CAMERAS 153
Table 5-2
NOMINAL NUMBER OF VIDEO RESELS ("PIXELS")
Based upon electrical transmission bandwidth only
NOMINAL ACTIVE NUMBER of NUMBER
ELEMENTS
FORMAT BANDWIDTH LINE TIME HORIZONTAL of ACTIVE
PER FRAME
(MHz) ()<s) ELEMENTS LINES
The full bandwidth is somewhat wider than the nominal value but high
frequency components may be attenuated. For example, NTSC has a full
bandwidth of approximately 4.5 MHz. The total bandwidth allocated for sound,
chrominance, and luminance is 6, 8, and 8MHz for NTSC, PAL, and SECAM,
respectively. Using the allocated bandwidth, the maximum number of horizontal
pixels is 625 for NTSC and 827 for PAL and SECAM.
The number of pixels per line is often chosen so that the horizontal clock
rate is an integer multiple of the color subcarrier. For NTSC, 768 pixels equate
to four times the NTSC color subcarrier (see Section 3 .11.3., Number of
Detectors, page 97).
154 CCD ARRAYS, CAMERAS, and DISPLAtS
This larger number provides more sampling for each color and reduces color
aliasing. For single chip color cameras, resolution is more difficult to define.
The primaries or their complements spatially sample the scene unequally
(discussed in Section 8.3.2., Detector Array Nyquist Frequency). If the chip
contains 768 detectors horizontally, then the number of green, red, and blue
detectors is typically 384, 192, and 192, respectively. Because the eye is most
sensitive to luminance values and the green component represents luminance,
then the green resolution may be taken as the camera resolution. Most chips
contain more green detectors and therefore the resolution is more than just one-
third of an equivalent monochrome system. Through data interpolation, any
number of output data points can be created. This number should not be
confused with the camera spatial resolution. Single chip color cameras cannot
provide the resolution listed in Table 5-2.
To reduce cost, a camera may contain fewer detector elements. Here, the
full bandwidth of the video standard is not exploited. The imagery of these
cameras will be poorer compared to those cameras that fully use the available
bandwidth. Recall that the standard only specifies the timing requirements and
a maximum bandwidth. The timing ensures that the camera output can be
displayed on a standard display.
For the complete system, the horizontal resolution may be limited by the
detector array, bandwidth of the video standard, or by the display. Display
resolution is discussed in Section 7.5., Resolution. It is common for the display
to limit the overall system resolution.
For NTSC and EIA 170 compatible arrays, it is common to have about 480
detectors in the vertical direction. The image formatting algorithm inserts blank
lines (all black) to meet video standards (485 lines). Because many displays
operate in an overscan mode, the blank (black) lines are not noticed.
For higher vertical resolution, more lines are required. EIA343A (originally
RS 343A) was created 6 for high resolution closed circuit television cameras
(CCTV). Although the standard encompasses equipment that operates from 675
to 1023 lines, the recommended values are 675, 729, 875, 945, and 1023 lines
per frame (Table 5-3). The military primarily uses 875 lines/frame but it
requires "standards conversion" boxes that can record and present imagery on
EIA 170 compatible equipment. EIA 343A is still a monochrome standard.
Table 5-3
RECOMMENDED EIA 343A VIDEO FORMATS
SCANNING ACTIVE
ACTIVE LINE
LINES/FRAME FREQUENCY LINE
LINES TIME p.s
(kHz) TIME p.s
The EIA 170 standard was developed for the broadcasting of video signals
over limited bandwidth channels. The FCC dictated that the bandwidth cannot
exceed 6 MHz. To avoid flicker on the television receiver, a 2:1 interlace
concept was adopted. The NTSC standard was constrained to the same
bandwidth limitation.
Table 5-2 (page 153) listed the number of video resels supported by the
video standard. If the camera has more pixels, then the received image
resolution will be limited by the standard's bandwidth. If the camera has fewer
156 CCD ARRAYS, CAMERAS, and DISPLAYS
pixels, then the camera's resolution limits image quality. By using image
compression, high definition television cameras can effectively transmit more
resels than that suggested in Table 5-2. The number of transmitted resels
remains the same (constrained by bandwidth). It is through coding (e.g., image
compression) that the resels contain more information.
Because EIA 170 and NTSC are so pervasive, nearly all cameras are built
to these video standards. This permits saving imagery on off-the-shelf video
recorders and displaying it on just about any monitor. However, for non-
broadcast operation, it is not necessary to adhere to the 6 Mhz bandwidth
limitation. The only requirement is that the timing be identical to the standard
timing. A high resolution monitor (wide bandwidth) can display the imagery
from a camera that has more pixels and a wide bandwidth recorder can store
that imagery.
Departure from the standard video timing requires special monitors and
special recorders. Already digital recorder and digital monitors exist. Computer
monitors now operate in the progressive scan mode. To avoid flicker, the frame
rate exceeds 70Hz. The frame rate is determined by the video drive electronics
and is not generally user-selectable.
The term composite is used to denote a video signal that encodes luminance,
color, and timing information onto a single channel. It can be transmitted on a
single wire such as cable television. NTSC, PAL, and SECAM are composite
video signals. The signal must be decoded to create an image.
Component analog video (CAV) carries each part on a separate wire. RGB,
or their complementary values are carried on three separate cables. It is common
for the G to be a composite line and carry the sync signals. Component videos
CAMERAS 157
do not require encoding and therefore there is less chance that there will be
color or luminance cross coupling. This implies that component imagery will
produce a higher quality image.
Matrixing can convert the RGB into any color coordinate system. With the
ease of digital implementation, multiple outputs are available from many
cameras. That is, a camera may offer RGB, NTSC, and Y/C outputs to provide
flexibility to the user.
The Institute of Radio Engineers (IRE) assigned 140 IRE units to the peak-
to-peak composite signal (Figure 5-3). This provides 100 IRE units for the
active image signal and 40 IRE units for blanking. They assumed that the video
is one volt peak-to-peak when terminated by 750. Thus, each IRE unit is equal
to 1/140 = 7.14mV.
To avoid seeing the retrace, NTSC introduces a small setup. With most
systems, the video is usually clamped at the blanking level and the video signal
with setup does not contain a reference black signal. That is, the video does not
contain a signal to indicate the black level. This voltage only occurs when there
is a black object within the field-of-view. The blanking level (retrace) is
sometimes called "blacker-than-black" because it is 7.5 IRE units below the
black level. With the signal standardized to 1 V peak-to-peak, the setup reduces
the active video to 0.66 v.
158 CCD ARRAYS, CAMERAS, and DISPLAYS
100 IRE
~
\
Blanking/
Level t
40 IRE
Pure
Black
~ ...____Sync Pulse
I
Sync Tip
Figure 5-3. Definition of IRE units. The active video represents a gray
scale divided into 10 equal steps. NTSC assigns 7.5 IRE units to setup.
Table 5-4
VIDEO STANDARDS
VOLTS IRE UNITS
FORMAT Sync Synch
Setup Peak Blanking Setup Peak
tip tip
Composite
without -0.286 - 0.714 -40 0 - 100
setup
Non-
-0.300 - 0.700 -40 0 - 100
NTSC
The most popular sampling frequency is four times the subcarrier frequency
and is simply called 4f,c sampling. It is equal to 14.3 MHz for NTSC and 14.7
MHz for PAL. This allows the post-reconstruction filter to have a more gradual
cutoff with less ripple. This filter is easier to design and is less expensive.
Because the data exist at discrete frequencies, strict adherence to the Nyquist
criterion (discussed in Chapter 8, Sampling Theory) is not required.
5.2.6. HDTV/ATS
The current accepted meaning of high definition television (HDTV) is a
television system providing approximately twice the horizontal and vertical
resolution of present NTSC, PAL, and SEC AM systems. While the term HDTV
implies a receiver, the industry is equally concerned about transmission
standards. The video format must be standardized before building cameras and
receivers. The acronym HDTV is slowly being replaced with the advanced
television system or ATS.
typical living room where the sofa is nine feet from the television (discussed in
Section 7 .1., The Observer). The Japan Broadcasting Corporation (NHK)
provided the first HDTV demonstration in 1981. The aspect ratio was initially
5:3 but this was later changed to 16:9 to match the various aspect ratios used in
motion pictures.
In April 1994, the Grand Alliance approved two different resolutions for
HDTV (Table 5-5). Both formats have square pixels. The selected line rate
should simplify conversion from NTSC and PAL into HDTV. The lines per
frame of 1125 is 1517 times 525 and 9/5 times 625. Similarly, 787.5 is 3/2
times 525 but is approximately 100179 times 625lines. Thus, with the lower line
rate, it is more difficult to convert PAL into HDTV.
Doubling the resolution, increasing the aspect ratio, and separating color
from luminance requires an increased bandwidth of eight-fold over current
television requirements. With a nominal bandwidth of 4.2 MHz, the required
HDTV bandwidth is approximately 33.6MHz. A 10-bit system will have a data
rate of 336 Mbit/s. At this data rate, transmission losses in conventional wire
cable limit the maximum transmission distance to, perhaps, 1 km. This demands
image compression.
162 CCD ARRAYS, CAMERAS, and DISPLAYS
Table 5-5
GRAND ALLIANCE RECOMMENDED FORMATS
(Recommended in April 1994)
RESOLUTION LINES FRAME
SCAN
Pixels/line x PER RATE
TECHNIQUE
active lines FRAME (Hz)
24 Progressive
1920 X 1080
1125 30 Progressive
60 Interlaced
24 Progressive
60 Progressive
The FCC has allocated only a 6-MHz bandwidth for commercial television
channels. Thus, image compression somewhere between 50:1 and 75:1 is
required. The Grand Alliance recommends using MPEG-2 image compression.
Although current cameras are analog devices, 13 the Grand Alliance
recommended digital transmission.
Table 5-6
PICTURE FORMATS SUPPORTED by ATSC STANDARD
RESOLUTION
FRAME ASPECT SCAN
pixels/line x
RATE (Hz) RATIO TECHNIQUE
active lines
24 16:9 Progressive
1920 X 1080
30 16:9 Progressive
60 16:9 Interlaced
24 16:9 Progressive
1280 X 720 30 16:9 Progressive
60 16:9 Progressive
24 4:3 Progressive
30 4:3 Progressive
640 X 480
60 4:3 Progressive
60 4:3 Interlaced
Figure 5-4 illustrates a typical color camera functional block diagram. The
array output is processed to achieve the standard video signals. This section
describes the knee, color correction, gamma, and aperture correction.
Monochrome cameras tend to be used for scientific and industrial applications
and are described in Section 5 .4., Industrial/Scientific Cameras.
164 CCD ARRAYS, CAMERAS, and DISPLAYS
From
ceo
Array.---..,
R Color Gamma
G Encoder
Correction Correction
B Composite
1--...-,--..J Video
Sync
Figure 5-4. Typical color camera functional block diagram. Sync signals
and other control logic are not shown. The subsystem location and
design vary by manufacturer. Consumer cameras typically have a
composite output that is created by the encoder. Scientific and industrial
cameras may have a component output which is the signal before the
encoder. To increase versatility, many cameras can provide either
NTSC, RGB, or complementary color output. The RGB output provides
easy interfacing with a computer monitor.
Maximum Output
- ----
l
'5
c.
White Clip Level Variable
Slope
'5
0
Variable Location
4 I "
Input
Figure 5-5. The knee extends the dynamic range by compressing high
intensity signals. By decreasing the knee slope, the camera can handle
a larger scene dynamic range.
Figure 5-4 illustrated a typical color camera block diagram. The "color"
signals may have originated from a single CFA (see Section 3.11.2., Color
Filter Arrays, page 94) or three separate arrays as illustrated in Figure 5-6.
Because of high cost, three-chip cameras were used almost exclusively for high
quality broadcast applications. Now, three-chip color cameras are entering the
scientific market. Consumer camera designs, driven by cost, tend to use only
one chip. Although this is changing.
[~]- (5-3)
When used to improve color reproducibility (i.e., matching the camera spectral
response to the CIE standard observer), the process is called masking. 15 •16 When
manipulating signals into the NTSC standard, it is called matrixing.
166 CCD ARRAYS, CAMERAS, and DISPLAYS
Figure 5-6. A high quality color camera has three separate arrays. One
for each primary color or its complement. Light is spectrally separated
by coatings on the beam splitters and additional filters. The three-chip
camera is listed as CCD X 3 or 3CCD in sales brochures. Each color
samples the scene equally and this minimizes color moire patterns when
compared to a CFA. Although beam splitters are shown, most 3CCD
cameras use prisms for compactness.
Matrixing can be used to alter the spectral response and intensities of the
primaries to compensate for changes in scene color temperature. Intensity
changes are achieved simply by multiplying a primary by a constant. Spectral
changes are approximated by adding various ratios of the primaries together. For
example, if an object is viewed at twilight, the matrix can be adjusted so that the
object appears to be in full sun light (neglecting shadows). A scene illuminated
with a 2600-K source may appear somewhat yellow compared to a scene
illuminated with a 6500-K source. This occurs because a 2600-K source has less
blue than a 6500-K source. With white balance (matrixing) a white object can
appear to be illuminated by different blackbodies - usually ranging from 3200
to 5600K.
With CFAs, detectors sensitive to the same color are separated resulting in
a low fill factor. The optical anti-alias filter blurs the image and effectively
increases the optical area of the detector. The birefringent filter is placed
between the lens assembly and the detector array. Birefringent crystals break a
beam into two components: the ordinary and the extraordinary. This makes the
detector appear larger optically. While this changes the MTF (discussed in
Section 10.3.4., Optical Anti-Alias Filter), the sensitivity does not change. The
light that is lost to adjoining areas is replaced by light refracted to the detector.
With unequal sampling, monochrome edges may appear colored and small
objects may change color. Sophisticated image processing is required to
overcome the unequal sampling of the colors and to format the output into the
NTSC standard. The algorithm selected depends upon the CFA architecture (see
Figure 3-44, page 61). Algorithms may manipulate the ratios of blue-to-green
and red-to-green, 17 separate the signal into high and low frequency
components, 18 use a pixel shift processor, 19 or add two alternating rows
together20·21 to minimize color artifacts. The decoding algorithm is unique for
each CFA configuration. Clearly, each CFA design requires a new decoding
algorithm. Algorithm design is probably more of an art than a science because
the acceptability of color aliasing is not known until after the camera is built.
~=============================================
Example 5-1
SINGLE COLOR CHIP ALGORITHM
Figure 5-7 illustrates a typical CFA21 where rows are added to minimize
aliasing. The NTSC odd field consists of CCD rows 1 + 2, 3 + 4, etc., and the
NTSC even field consists of 2 + 3, 4 + 5, etc. Tables 5-7 and 5-8 provide the
outputs.
Table 5-7
SUMMING TWO ROWS
CCD ROW OUTPUT LINE
1+2 1 (odd field)
2 G Mg G Mg
3 Cy Ye Cy Ye
4
3[ Mg G Mg G
5 I Cy I Ye I Cy I Ye :
I __ .l _ _ L __I __ ..
J 4
Table 5-8
OUTPUT VALUES
TV
PIXEL 1 PIXEL 2 PIXEL 3 PIXEL 4
LINE
The NTSC signal requires equal R, G, and B information on each line and
this requires line interpolation. For example, pixel 3 on the first line does not
contain any red data. Red must be interpolated from adjoining pixels. The
algorithm selected depends upon the mathematics and the human eye response.
The selection depends upon the resultant imagery. That is, the interpolation
tends to be empirical.
=============================================~
CAMERAS 169
The luminous output of a cathode ray tube is related to its grid voltage by
a power law
(5-4)
If the camera has the inverse gamma (Figure 5-8), then the radiometric fidelity
is preserved at the display.
Vvideo
Analog ,.----
CCD
Array t--
Electronics
VScene Gamma
Correction
1fy
Video Display
Electronics -
VGrid CRT
y
'-----
~:\>
Observer
Camera Display
Figure 5-8. Gamma correction should only be used when the imagery is
directly presented on a CRT-based display.
Lo;splay
CCD Array Display
and ~
Electronics
Electronics
~------~~------/
Camera Image Processing System Display Drivers and Display
Although CRT displays have a range of gammas, NTSC, PAL, and SECAM
"standardized" the display gamma to 2.2. 2.8, and 2.8, respectively. However,
all three systems use 0.45 as the inverse. For NTSC, the monitor luminance is
proportional to the detector output and target intensity (2.2)(0.45) = 1. With
PAL and SECAM, the luminance still has 22 a residual gamma of 1.27. This
provides some contrast enhancement.
The gamma slope becomes very large for low values of input brightness. It
has therefore become accepted practice to limit the operation of the gamma
compensator to an arbitrary level. SMPTE Standard 240M was developed 23 for
high definition televisions but now is applied to many systems It has since been
replaced by ITU 709.
CAMERAS 171
where VscENE is the voltage before the gamma corrector and VvmEo is the voltage
after. The toe break intensity occurs at VscENE = 0.0228 and the toe slope is 4
(when LscENE < 0.0228). For gamma correction to be accurate, the black
reference level must be known. If gamma is inserted after the 7. 5 % setup, some
ambiguity exists in the black level resulting in some intensity distortion. While
gamma correction may contain errors, the human eye is very accommodating.
The consumer can also adjust the receiver contrast to optimize the image in
accordance with his personal preferences. However, this luxury is not available
for scientific applications.
60
-c
z 50
-
1-
:J 30
40
D..
1- 20
:J
0 10 SMPTE 240M
0
0 2 4 6 8 10
INPUT (ON)
Figure 5-10. Gamma output-input transformation recommended by
SMPTE 240M. Expansion of the first 10 levels of an 8-bit ADC system.
Data only exist at integer values. Compression of blacks is obvious.
172 CCD ARRAYS, CAMERAS, and DISPLAYS
Boost filters also amplify noise so that the signal-to-noise ratio remains
constant. For systems that are contrast limited, boost appears to improve image
quality. However, for noisy images, the advantages of boost are less obvious.
As a result, these filters are used only in high contrast situations (typical of
consumer applications) and are not used in scientific applications where low
signal-to-noise situations are often encountered.
Example 5-2
TDI CAMERA
M - ~ - y2 -
60 11m = 0 0236 (5-7)
oPncs R1 y1 (O.linch)(25,400 llmfinch) · '
where R 1 is the distance from the lens to the object and R2 is the lens to array
distance. The focal length is
(5-8)
simply state "no missing codes." This means that the DNL is less than one LSB.
However, 1.01 LSB DNL (fail specification) and 0.99 LSB DNL (pass
specification) probably look the same visually. Poor DNL can affect the
cosmetic quality of images.
z 8
e.
1-
7
::I 6
D.
1-
5
::I 4
_.
0 3
< 2
1-
(5 1
0 0
0 25 50 75 100
ANALOG INPUT (percent of full scale)
(a)
INL
best fit
z 8 ~
e.
1-
7
::I 6
D.
1-
5
::I 4
_.
0 3
< 2
1-
(5 1
0 0
0 25 50 75 100
ANALOG INPUT (percent of full scale)
(b)
The same problem may exist within a frame capture board. The user must
test the board for any nonlinearities. Because the frame capture board usually
resides within a computer, testing is relatively easy.
Although low light level televisions can be used for many applications, they
tend to be used for military and scientific applications. The term image
intensifier refers to a series of special imaging rubes28 that have been designed
to amplify the number of photons internally. These tubes increase the number
of incident photons by several orders of magnitude. The more popular second
generation (Genii) and third generation (Geniii) intensifier tubes use a
microchannel plate (MCP) for intensification. Third generation is similar in
design to second generation with the primary difference being higher quantum
efficiency.
When photons hit the photocathode, electrons are liberated and thereby
create an electron image. The electrons are focused onto the MCP and about
80% enter the MCP. The MCP consists of thousands of parallel channels (glass
hollow tubes) that are about 10 p,m in diameter. Electrons entering the channel
collide with the coated walls and produce secondary electrons. These electrons
are accelerated through the channel by a high potential gradient. Repeated
collisions provide an electron gain up to several thousand. Emerging electrons
strike the phosphor screen and a visible intensified image is created. The
intensifier is an analog device and an observer can view the resultant image on
the phosphor screen.
The intensifier output can be optically linked29 to a solid state array with
either a relay lens or a fiberoptic bundle (Figure 5-12). When coupled to a
CCD, it is called an ICCD. The intensified camera spectral response is
determined solely by the intensifier photocathode responsivity. Although Figure
5-13 illustrates Genii and Geniii spectral responsivities, a variety of
photocathodes are available. 30 The window material controls the short
wavelength cutoff whereas the photocathode determines the long wavelength
cutoff.
CAMERAS 177
Photo Cathode
Window
Screen Fiberoptic
Fiberoptic Taper
~
175
c:( 150
.§. 125
~ 100
>
i=
75
50
en
z 25
w
t/) 0
0.3 0.4 0.5 0.6 0.7 0.8 0.9
WAVELENGTH (pm)
Figure 5-13. Nominal spectral response of militarized (fielded) Genii
and Geniii image intensifier tubes. Geniii provides a better match to the
night sky illumination (see Figure 2-13, page 39). The responsivity
depends upon the window and photocathode material. Genii tubes use a
multialkali photocathode whereas Geniii tubes use GaAs photocathodes.
178 CCD ARRAYS, CAMERAS, and DISPLAYS
Although many phosphors are available for the intensifier output, a phosphor
should be selected whose peak emission is at the peak of the solid state array
spectral responsivity. A color camera is not a good choice because it has lower
quantum efficiency due to the color-separating spectral filters. Furthermore,
single chip color cameras have unequal sampling densities and this can lead to
a loss in resolution.
The intensified camera can provide high quality, real-time video when the
illumination is greater than 10-3 lux (corresponding to starlight or brighter). For
comparison, standard cameras (non-intensified l are sensitive down to about 1
lux. Thus, intensified cameras offer high quality imagery over an additional
three orders of magnitude in illumination. A three-order magnitude increase in
sensitivity implies that the MCP gain should be about a thousand. However, it
must be an order of magnitude larger to overcome the transmission loss in the
fiber optic bundle or relay lens. For very lo\\ illumination levels, it may be
necessary to increase the integration time and then real-time video is no longer
possible.
By simply changing the voltage on the image intensifier, the gain can be
varied. Thus, operation over a wide range of luminance levels ranging from
starlight to daylight is possible. Gating is probably the biggest advantage of the
intensified camera. Electronic shutter action is produced by pulsing the MCP
voltage. Gating allows the detection of low level signals in the presence of
interfering light sources of much greater energy by temporal discrimination.
Gate times as short as 5 ns FWHM (full width at half maximum) are possible.
This allows the detection of laser diagnostic pulses against very intense
continuous sources. Because full frame transfer or frame transfer CCDs are used
for ICCD applications, the inherent gating capability of the intensifier is used
as a shutter during transfer time. This eliminates image smear.
Because the detector size is typically smaller than the intensifier's resolution,
the intensifier's image must be minified. Fiber-optic bundles reduce camera size
and weight (lenses are relatively heavy). But fiberoptic coupling is complex
because it requires critical alignment of the fibers to the individual detectors.
The fiber area and pitch, ideally, should match the detector size and pitch for
maximum sensitivity and resolution. However, to alleviate critical alignment
issues, the fiber diameter is typically one-half the pixel pitch. This reduces
sensitivity somewhat.
state camera. The lens coupling represents a cost-effective way to add gating
capability to an existing camera. However, a lens-coupled system provide a
much lower throughput (2 x to 10 x lower) than a fiber coupled system.
There are four spatial samplers in the intensified camera: (1) sampling by
the discrete channels in the MCP, (2) the fiberoptic output window within the
intensifier, (3) the input to the fiberoptic minifier, and (4) the detector array. At
these interfaces, the image is re-sampled. Depending upon the location
(alignment) of the fiberoptic minifier, moire patterns can develop and the MTF
can degrade. In addition, manufacturing defects in the fiberoptic can appear31 as
a "chicken wire" effect.
Intensifiers are electron tubes. Tubes can bloom so that scenes with bright
lights are difficult to image. For example, it is difficult to image shadows in a
parking lot when bright street lights are in the field-of-view. Intensifiers are
subject to the same precautions that should be observed when using most
photosensitive coated pickup tubes. Prolonged exposure to point sources of
bright light may result in permanent image retention.
The intensifier specifications alone do not accurately portray how the system
will work. Only a detailed analysis will predict performance. Resolution of the
array will not describe system resolution if the image intensifier provides the
limiting resolution. While the array photoresponse may be quite uniform, the
overall nonuniformity is limited by the image intensifier. Standard military
image intensifier photocathodes have nonuniformities approaching ± 30%. Only
carefully selected intensifiers with low nonuniforrnity are used as part of the
intensified system.
5.5. REFERENCES
1. The timing requirements can be found in many books. See, for example, D. H. Pritchard,
"Standards and Recommended Practices," in Television Engineering Handbook, K. B. Benson,
ed., pp. 21.40-21.57, McGraw-Hill, New York, NY (1986).
2. D. H. Pritchard, "Standards and Recommended Practices," in Television Engineering
Handbook, K. B. Benson, ed., pp. 21.73-21.74, McGraw-Hill, New York, NY (1986).
180 CCD ARRAYS, CAMERAS, and DISPLAYS
The signal is modified by the spatial response of the optics and detector
whereas the noise originates in the detector and electronics. Noise characteristics
may be different from signal characteristics. System design requires optimizing
the signal-to-noise ratio.
This chapter covers maximum signal, minimum signal, dynamic range, and
the signal-to-noise ratio. It assumes that the image size is large compared to the
detector's area. Point source detection calculations are more difficult and can be
found in References 1 and 2. Camera metrics, depend in part, on the application
(see Section 5.2.2., Broadcast/Non-broadcast Design, page 155).
Image intensifier and array characteristics both affect the intensified camera
signal-to-noise ratio. ICCDs have a dynamic range that varies with the image
intensifier gain. This permits fabrication of general purpose ICCDs that operate
in lighting conditions that range from starlight to full sunlight.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of ContenTs.
182
CAMERA PERFORMANCE 183
6.1. NOMENCLATURE
G1 G2
Fixed Variable
VSIGNAL VCAMERA
The input can be measured at two different places. With the first
measurement, the lens focuses reflected scene illumination, PTEv-scENE• onto the
array where PT is the target reflectivity. Tables 2-5 and 2-6 (page 28) provide
typical values of Ev-scENE· Here, the camera's output depends upon the lens f-
number and transmittance (or equivalently the T-number). In the second
measurement, the faceplate illumination is used. For radiometric inputs, the
scene irradiance or faceplate irradiance is used. In Equation 4-2 (page 112), the
source was represented by Me and it is identical to PTEe-scENE· Because the
manufacturer has no control over the user's selection of lenses, the faceplate
illumination is often specified.
Array parameters are based upon the photon flux irradiating the detector,
Eq-FACEPLATE and the measured output VsiGNAL· The number of photons reaching
the detector is
(6-1)
(6-2)
(6-3)
The array outputs, nrE and < nsvs > , are derived theoretically or are calculated
from the measured values VsiGNAL or VcAMERA. Often only < nFwoR > is used for
<nsvs> to calculate DRARRAY· This provides the highest possible dynamic
range. If the camera does not introduce any additional noise, modify the noise
bandwidth, or minimize the maximum output, the camera SNR and DRcAMERA
will be identical to the array values.
E P T Ev-SCENE T T z (6-5)
v-FACEPLATE =
4 F 2 ( 1 + MOPTicsi OPTICS AIM UX.
On a clear day (TATM = 1), a white target (PT = I) which is far way (MoPT1cs =
0) provides
P T Ev-SCENE T (6-6)
Ev-FACEPLATE .. F OPTICS lux.
4 2
CAMERA PERFORMANCE 185
For back-illuminated targets (such as those used for calibration), the flux
emanating from the target is PTEv-scENE· The measurement of faceplate
illumination is usually performed with a calibrated lens. The value PTEv-scENE is
measured with a calibrated radiometer or photometer, VcAMERA is measured, and
then by using Equation 6-6 or 6-7 Ev-FACEPLATE is calculated.
The maximum and minimum signals depend on the spectral output of the
source and the spectral response of the detector. The source color temperature
is not always listed but is a critical parameter for comparing systems. Although
the CIE illuminant A is used most often, the user should not assume that this
was the source used by the camera manufacturer. Manipulation of the source
(CIE illuminant A, B, C, D 65 , or another unspecified source) and optical filter
characteristics can change the maximum and minimum signal values.
Based on signal detection theory, the minimum illumination would imply that
the signal-to-noise ratio is one. However, the definition of minimum illumination
is manufacturer dependent. Its value may be (a) when the video signal is, say,
30 IRE units, (b) when the SNR is one, or (c) when an observer just perceives
a test pattern. Because of its incredible temporal and spatial integration
capability, the human visual system can perceive SNRs as low as 0.05. This is
further discussed in Section 12.1.1., Perceived Signal-to-Noise Ratio. Therefore,
comparing cameras based on minimum illumination should be approached with
II II
care.
For many applications, the object is far away so that MoPTrcs approaches
zero. When measured at the camera output, the SNR is
(6-7)
where the subscript scene is used to indicate the input is the reflected scene flux.
Noise is inserted by the detector and by the amplifiers (see Figure 4-11, page
123). The electronics MTF, Hsvs(fe1ec), modifies the noise spectra. The noise
power spectral density, S(fe1ec), includes all the noise sources: shot noise, noise
floor, reset noise, pattern noise, quantization noise, amplifier white noise, and
amplifier 1/f noise. The subscript elec emphasizes that electrical frequency
applies to the amplifiers and subsequent electronic filters. The MTFs are
described in Chapter 10, System MTF.
(6-8)
AD
-4F Rp. 0) p -J_').. o> Mq-SCENi'A o>
2
TOP17CS(').. o> !l ')..
tiNT
(6-9)
SNRSCENE - ____:_=----------- - - - - - - - -
(nSYS)
CAMERA PERFORMANCE 187
The SNR is proportional to the detector area and integration time. Figure 6-2
provides the relative SNR as a function of these parameters. Binning is identical
to increasing the detector area. The output of four binned detectors is the same
as one detector four times larger. For fixed area, increasing the integration time
increases the SNR. For fixed integration time, binning increases the SNR.
Shorter integration times permit higher frame rates (See Section 4.5., Frame
Rates, page 142). Shorter integration times also decrease the SNR.
BIN SIZE
100 -10x10
<
w - 8x8
wa:: - 6x6
><
i=a:: - 4x4
oo
WI-
10 - 3x3
u.o
u.w
WI-
w
0
1
1 10 100
INTEGRATION TIME
Figure 6-2. Relative SNR as a function of bin size (detector area) and
integration time. Only shot noise is considered.
Although the theoretical maximum signal just fills the charge wells, the
manufacturer may limit the maximum output voltage to a lower stored-charge
value. Because the dynamic range is often expressed in decibels,
DRCAMERA-dB - 20 log (~
VMAX] (6-10)
N01se
188 CCD ARRAYS, CAMERAS, and DISPLAYS
(6-11)
Example 6-1
DIGITAL DYNAMIC RANGE
Adding noise powers, the total camera noise is related to the array noise and
ADC noise by
(6-12)
If an ADC is selected so that an LSB is equal to the array noise level, the ADC
increases the camera noise by 4%. The array dynamic range is 50,000/60 = 833
but the smallest available ADC has 10 bits or 1024 levels. This ADC devotes
1.23 bits to array noise. But if the LSB was one-quarter of the array noise level,
the camera noise would increase by only 1 %. The ADC dynamic range must be
greater than 3332:1. The smallest ADC that covers this range contains 12 bits.
=============================================~
CAMERA PERFORMANCE 189
The photon level that saturates the charge wells is the camera's maximum
input signal and VwELL = G 1 GqNwELL/C where G2 = 1. However, there is no
requirement that the camera manufacturer set the maximum output voltage equal
to VwELL· If the array has an antibloom drain, the maximum output may be
limited to the white clip level (see Figure 3-35, page 87).
1- 1.5
:::J
c.
1-
:::J
0
w
>
~
:5w
rx: 0.5
4 4.5 5 5.5 6 6.5 7
f-number
2
1-
::J
D..
1- 1.5
::J
0
w 1
>
i=
<C 0.5
..J
w
0:::
0
0.8 1 1.2 1.4 1.6 1.8 2
f-number
The minimum signal provided with gain "on" (G2 greater than one) is
usually calculated due to the difficulty of performing accurate low level
radiometric and photometric measurements. These values may be provided at
30%, 50%, or 80% video levels. That is, the illumination given produces an
output video that gives 30, 50, or 80 IRE units, respectively (see Section 5.2.4.,
IRE Units, page 157). Although a higher gain amplifier could provide 100%
video, the user can optimize the image by adjusting the gain and level of the
display. That is, the display's internal amplifiers can be used for additional gain.
In this context, the camera system consists of the camera and display.
If G2 is expressed in decibels,
(6-13)
( _ DRCAMERA-dB)
20
(6-15)
Mv-SCENe(SNR = 1) = Mv-SCENe(Max) ·10 •
~=============================================
Example 6-2
MINIMUM SIGNAL ESTIMATION
Using Equation 6-15, the noise signal is 1.16 X 10-3 lux. A signal-to-noise
of one provides an extremely noisy image. While a human observer can detect
some objects when the SNR is one, a machine vision system cannot.
=============================================~
Example 6-3
VIDEO SIGNAL ESTIMATION
The camera described in Example 6-2 provides a gain of 20 dB. What is the
SNR at 30% and 80% video?
Using Equations 6-14 and 6-15, at 30% video the faceplate illumination is
0.0195 lux and at 80% video, the faceplate illumination is 0.052 lux. The SNR
is 17 and 45, respectively. With a SNR of 17, the image will appear slightly
noisy but probably would be considered acceptable to most observers. However,
if no additional gain is used, the image will appear very dark on the display.
Here, the display gain must be increased to make the image viewable.
Depending upon the task, a machine vision system may need a higher SNR.
While computer programs are usually insensitive to absolute signal values, the
frame grabber's ADC may not provide enough levels for adequate image
192 CCD ARRAYS, CAMERAS, and DISPLAYS
processing. For example, if the ADC is an 8-bit system, 30% video only
provides digital numbers ranging from 0 to 77 .
=============================================~
Example 6-4
NOISE EQUIVALENT INPUT
photons (6-16)
s-m 2 -1-1m
When the SNR is unity, the number of photons is small and photon noise is
negligible (see Figure 4-16b, page 136). If the detector is cooled, dark current
is minimized. With PRNU and FPN correction, the minimum NEI is
NEI .. (6-17)
=============================================~
Example 6-5
NOISE EQUIVALENT REFLECTANCE
The NEI assumes that the target is against a black background. For real
scenes, both the target and background create signals. The target is only visible
when a sufficient reflectance difference exists. For detecting objects at long
ranges, the atmospheric transmittance loss must be included. The signal
difference between the target and its background is
CAMERA PERFORMANCE 193
}.
Both the target and background are illuminated by the same source. The
signals from both depend on the spectral reflectances that are taken as constants
over the spectral region of interest. Using the same simplifying assumptions as
before, solving for PT - p8 and calling this value the noise equivalent reflectance
difference (NE!J.p) provides
(6-19)
where T ATM is the total atmospheric transmittance and it depends on range. The
value TATM is usually expressed as TATM = exp(-uATMR). Then
Thus, NE!J.p depends on the source illumination, system noise, and the
atmospheric conditions. As the range increases, the target reflectance must
increase to keep the SNR at one. The maximum range occurs when the NE!J.p
is one. At this maximum range,
The variable Mq(>. )[pT(A)- p8 (A)] is a complicated function that depends on the
cloud cover, atmospheric conditions, sun elevation angle, surface properties of
the target, and the angle of the target with respect to the line-of-sight and sun
angle.
194 CCD ARRAYS, CAMERAS, and DISPLAYS
=================================================~
Example 6-6
MINIMUM PHOTOMETRIC SIGNAL ESTIMATION
A CCD array has a noise floor of 150 electrons rms. If placed into a
camera, what is the estimated minimum faceplate detectable signal?
The minimum signal occurs when the signal-to-noise ratio is one. That is,
npe = (nFLOoR). Several simplifying assumptions are required to obtain a back-of-
the-envelope expression for the minimum signal. If the CIE illuminant D6500 is
used, the spectral photon sterance is approximately constant over the spectral
region, ..:1>-. If the quantum efficiency is constant,
photons (6-22)
s-m 2 -1J.m
Because the energy of one photon is he/A, the faceplate spectral power is
(6-23)
Ev-FACEPLATE =
683
AD
f V(A_)Ep- FACEPLATE(A_)dA lux (6-24)
0.4 J.lm
or
h
E
V-FACEPLATE
(
..nFWOR)
A , t
c 68~
A
JV(A.)
0.1
,
dA. l
ux.
(6-25)
fl a II. /NT D 0.4 II.
CAMERA PERFORMANCE 195
From Figure 4-5 (page 108) TJ ::::0.3. For EIA 170 compatibility, t1NT = 1/30 s,
and the spectral bandwidth is <l}.. = (0. 7 - 0.4) p.m = 0. 3 p.m. The photometric
integral is approximately 0.2. The constant, he, is approximately 2 x 10'19 J-p.m.
Then
1.36 x w- 12 lux. (6-26)
Ev-FACEPUTE - - - - - -
AD
Assume a 50% fill factor interline transfer array is available. From Table 3-3
(page 99), a 1/z-inch format detector will have an active size of approximately
4.17 p.m X 10 p.m (A 0 = 41.7 X 10' 12 m2). Then the estimated minimum signal
is 3.3 X 10'2 lux.
While the assumptions are not unreasonable, the statement that the target can
be seen during twilight is an extrapolation. The spectral output of the D6500
source is substantially different than twilight with color temperature of
approximately 2000 K. Changing the spectral input increases the minimum
illumination. Note that these calculations were based solely upon SNR
considerations. The camera must have sufficient gain so that these signals can
be displayed.
=============================================~
Phosphor Screen
Photo Micro
Cathode Channel Plate \ Fiber-Optic ceo
~o~~~oE~~o~
Figure 6-5. Multiple photon-electron conversions occur in an ICCD.
While the image intensifier amplifies the image, the SNR may be reduced
due to quantum losses and additive noise. For moderate light levels, the ICCD
offers no advantage in SNR performance over the standard CCD camera. Its key
attribute is the ability to gate the scene temporally.
(6-27)
Using the same approximations as given in the preceding section, the number
of signal electrons generated by the photocathode is
nCATHODE =
(6-28)
where APIXEL is the projected area of a CCD pixel onto the photocathode.
Geometrically, it is the CCD detector area multiplied by the magnification of the
fiberoptic bundle. However, APIXEL depends on fiberoptic bundle characteristics
and microchannel pore size. It may be somewhere between 50% of the CCD
detector geometric projected area to several times larger. 3 As such, it represents
the largest uncertainty in signal-to-noise ratio calculations. The value A0 is the
center wavelength of the photocathode spectral responsivity and ~A is the
photocathode's spectral bandwidth.
CAMERA PERFORMANCE 197
The photons travel to the CCD via a fiberoptic bundle whose effective
transmittance is Tro· This is a composite value that includes the fiberoptic
transmittance and a coupling loss.
nPHOTON-CCD - TfonSCREEN • (6-3l)
or
Because the lens coupling is less efficient than the fiberoptic coupling, the
number of electrons produced is lower. The fiberoptic coupled system, with its
higher effective transmittance and better sensitivity, can detect lower signals.
The MCP amplifies these signals and adds excess noise so that
(n~cp) - kMCPG~cp(n;c-snor) • (6-35)
where kMcP is typically taken as 1.8 but varies with bias voltage. 5•6 The value
kMcP is also reported as an SNR degradation. The noise power is transformed
into photons by the phosphor screen
(6-36)
This value is converted to electrons by the CC D and the CCD noise is added
(n~cv) - fl~cv(n~CD-PHOTON) + (n~CD-DARK) + (n'i.woR) + (n;RNu) · (6-38 )
The total noise is
/2) -
\nCCD T)
2CCD ...22 2/2 )12
lfo T) SCREENkGMCP\nPC-SHOT + \nCCD-SHOT
)
(6-39)
Several limiting cases are of interest. When the gain is very high, the intensifier
noise overwhelms the CCD noise:
SNR- (6-41)
The SNR is independent of gain and CCD characteristics. Under high gain
conditions, an expensive, high quantum efficiency or thinned back illuminated
CCD is not necessary.
CAMERA PERFORMANCE 199
nCATHODE
SNR =
k
(6-42)
11cATHODEMi>..)A>.. t T A
p2 k TNT OPTICS PIXEL
4
High gain is used for photon counting applications. SNR, in this case, is limited
by the photocathode quantum efficiency, the excess noise factor, and the optical
system.
Under moderate gain situations where the CCD noise is greater than the
intensifier noise,
Comparing this with Equation 6-8 (page 186), the ICCD will have lower SNR
than a comparable standard camera. The advantage of the ICCD under these
conditions is the ability to gate the incoming photon flux temporally.
Example 6-7
IMAGE INTENSIFIER NOISE
_ 4 /k pz (nPC-DARK) h (6-44)
EBI = M ('
q "o) uA t TNT A PIXEL
A ,
T p otons.
11 CATHODE OPTICS
200 CCD ARRAYS, CAMERAS, and DISPLAYS
The ICCD will have the same equivalent background input under high gain
situations. Cooling the image intensifier reduces the EBI and it may be
negligible in gated applications. Under low gain situations, the CCD noise
dominates.
============================,=================~
Example 6-8
CCD versus ICCD
Under moderate gain situations, what is the relative SNR of the ICCD to the
CCD?
(6-45)
The P20 phosphor screen has a peak emission at 0.56 p.m with a conversion
efficiency of 0.063. An S20 bialkali photocathode's efficiency is 0.2 at 0.53 p.m.
The value ofTro is about 0.25. Combined, these factors give 0.00315. The MCP
gain must be greater than 110.00315 = 317 for the SNR1cco to be greater than
the SNRcco·
============================,=================~
6.4. REFERENCES
1. G. C. Holst, Electro-Optical Imaging System Performance, pp. 48-54, JCD Publishing (1995).
2. M. A. Sartor, "Characterization and Modeling of Microchannel Plate Intensified CCD SNR
Variations with Image Size," in Electron Tubes and Image Intensifiers, C. B. Johnson and B.
N. Laprade, eds., SPIE Proceedings Vol. 1655, pp. 74-84 (1992).
3. Mark Sartor, Xybion Corporation, private communication.
4. Y. Talmi, "Intensified Array Detectors," in Charge-Transfer Devices in Spectroscopy, J. V.
Sweedler, K. L. Ratzlaff, and M. B. Denton, eds., Chapter 5, VCH Publishers, New York
(1994).
5. I. Csorba, Image Tubes, pp. 120-124, Howard Sams, Indianapolis, Indiana (1985).
6. R. J. Hertel, "Signal and Noise Properties of Proximity Focused Image Tubes," in Ultrahigh
Speed and High Speed Photography, Photonics, and Videography, G. L. Stradling, ed., SPIE
Proceedings Vol. 1155, pp. 332-343 (1989).
7
CRT-BASED DISPLAYS
Displays are used for consumer television receivers, computer monitors,
scientific applications, and military applications. A display may be called a
monitor, visual display unit (VDU), video display terminal (VDT), or simply a
television receiver.
The major performance requirements are listed in Table 7-1. This table is
not meant to be all inclusive. It just indicates that different users have different
requirements. As displays are built for each user, they are specified in units
familiar to that user. For example, consumer displays will not have the MTF or
spot size listed. These parameters are provided for militarized displays.
Table 7-1
DISPLAY REQUIREMENTS
APPLICATION REQUIREMENTS
Television receivers have been available for over 50 years and the
technology has matured. They are considered adequate and little improvement
in cathode ray tube (CRT) technology is expected. Advances in television
displays currently focus on high definition television, flat panel displays, the
legibility of alphanumeric characters, and computer graphics capability. As a
result, display specifications are a mix of CRT terminology, video transmission
standards, alphanumeric character legibility, and graphics terminology.
201
202 CCD ARRAYS, CAMERAS, and DISPLAYS
system resolution is limited by the human eye. If the observer is too far from
the screen, not all of the image detail can be discerned.
In almost all cases, the output device sets the system dynamic range. Often,
the display is the limiting factor in terms of image quality and resolution. No
matter how good the camera is, if the display resolution is poor, then the overall
system resolution is poor.
Flat panel displays are an emerging technology with liquid crystal displays
(LCDs) prevalent in notebook (laptop) computers. LCDs require only a fraction
of the power required by cathode ray tubes and therefore lend themselves to
battery operated systems. However, CRTs will probably dominate the display
market due to low cost, high resolution, wide color gamut, high luminance, and
long life.
The commercially important flat panels are the light emitting diode (LED)
displays, AC plasma display panels (AC PDP). DC plasma display panels (DC
PDP), AC thin film electroluminescent (ACTFEL) display, DC thin film
electroluminescent (DCTFEL) displays, vacuum fluorescent displays (VFD), and
the liquid crystal displays (LCD).
LCD arrays can either be active matrix (AMLCD) or passive arrays. The
active matrix contains switching devices on each disel to maintain high
brightness and high contrast control. The liquid crystals may be either twisted-
nematic (TN) or supertwisted-nematic (STN). The details of these displays can
be found in References 1 through 6. Although this chapter discusses CRTs,
many concepts apply to all displays.
The observer is the starting point for display design. Design is based on both
perceptual and physical attributes (Table 7-2). For most applications, the design
is driven by perceptual parameters. These parameters are partially related to the
physical parameters. For example, sharpness is related to MTF. But the precise
relationship has not been quantified. While the physical parameters are important
to all applications, they tend to be quantified only for scientific and military
displays. Tables 7-1 and 7-2 illustrate different ways of specifying the same
requirement: the need for a high quality display.
Table 7-2
PERCEPTIBLE AND PHYSICAL PARAMETERS
PERCEPTUAL PHYSICAL
PARAMETERS PARAMETERS
Luminance
Resolution
Uniformity
Brightness Addressability
Contrast Gamma
Sharpness Color saturation
Color rendition Color accuracy
Flicker Color convergence (CRT)
Distortion (CRT)
Refresh rate
MTF
For monochrome systems, the optimum viewing distance occurs when the
raster pattern is barely perceptible. At closer distances, the raster pattern
becomes visible and interferes with image interpretation. That is, it is considered
annoying by most. For color CRTs, the observer must be far enough away so
that the individual dots of the three primaries are imperceptible. Otherwise, the
individual color spots become visible and the impression of full color is lost.
204 CCD ARRAYS, CAMERAS, and DISPLAYS
For design purposes, perceptibility occurs when targets are larger than 1 arc
minute. This corresponds to the normal visual acuity of 20/20. This design
criterion is quite good. An observer can adjust his distance accordingly. An
observer with poorer acuity (e.g., 20/40) will simply move closer to the screen
whereas someone with better acuity will probably move further away. Most
systems are observer-limited in that the observer cannot see all the detail
because he is typically too far from the display.
Television viewers sit about seven times the picture height or 4.2 times the
diagonal picture dimension. Knowing the dimensions of a typical living room
(sofa on one side, television on the other) suggests that consumer televisions
should have a diagonal picture size of 21 to 25 inches. These were the most
popular television receiver picture sizes. Sizes have changed reflecting changing
television viewing habits.
UJ
w
u.Z
o:J
o::c
Ww
m>
:::!c:e
::)-I
Zll.
UJ
c
0 20 40 60 80 100 120 140
Figure 7-2 illustrates a typical color cathode ray tube display. Although a
color display is illustrated, the block diagram applies to all CRT displays. The
input video signal is decoded into its components and then digitized. While
digital electronics may be used to process the signal, the CRT is an analog
device. Similarly, displays accepting digital inputs must convert the signal into
an analog voltage to drive the CRT. Numerous transformations occur to create
a visible image. The display matrix and display gain controls allow linear
mapping from input image to visible image. 7
Original
Image
or Data
.....----
Visible
Image
CRT
The display matrix provides the mapping between the standard video color
components and the actual colors emitted by the phosphors on the screen. This
color correction circuitry is an algorithm that can be represented as a matrix
(7-1)
206 CCD ARRAYS, CAMERAS, and DISPLAYS
With the NTSC format, the electron beams paint an image on the CRT
phosphor in the scan format shown in Figure 5-2, page 151. The color of the
visible image depends on the specific phosphors used. Electron beam density
variations, which represent scene intensity differences, create the visible image.
The output brightness of a CRT depends on the grid voltage and follows a
power law relationship. The slope of this curve, on a log-log scale, is the
display gamma (')'):
(7-2)
Gamma can range 8 from one to five. For design purposes, NTSC standardized
gamma at 2.2. PAL and SECAM standardized gamma at 2.8.
A color monitor has three electron beams that scan the CRT faceplate. Near
the faceplate is a metal mask that has a regular pattern of holes (Figure 7-3).
With the dot arrangement, it is called as shadow mask. With slots, it is called
an aperture grille. The three guns are focussed onto the same hole and the
emerging beams strike the CRT faceplate at different locations. At each location
there is a different phosphor. One fluoresces in the blue, one in the red, and one
in the green region of the spectrum. The arrangement is such that the beam
corresponding to the desired color will strike only the phosphor dot producing
that color. The electron guns emit only electrons and have no color associated
with them. Color is only created when the beam hits the phosphor even though
the guns are called red, green, and blue guns.
Hexagonal structure masks (the dots) are inherently more rigid than slots.
However, slots are easier to manufacture and therefore are usually found in
television receivers. The slots have a larger area and therefore aperture-grille
CRTs tend to be brighter than shadow-mask CRTs. With this brightness it is
said that the colors are richer. The distance between dots is less and more
difficult to perceive. Shadow-mask CRTs tend to provide sharper images.
Because computer monitors are viewed at a close distance, the dots are used
most often. The selection of slots over dots depends upon the application.
208 CCD ARRAYS, CAMERAS, and DISPLAYS
Metal Metal
Mask Mask
1\ r-- - r-- - /"""
~ ~n If1
'-- I
.__
lA B G R B G R[s B G R B G R B IG
'
Phosphorus
\I
Phosphorus
on Glass on Glass
Faceplate Faceplate
Figure 7-3. Color CRT mask patterns. (a) Dotted, (b) stripes, and (c)
slotted. Most consumer television receivers use slots whereas computer
displays use dots. The mask design varies with manufacturer.
Dots and stripes cannot be easily compared because the distance between
adjacent spots is different and depends upon orientation and the way measured.
Because there is ,.[3/2 difference, a 0.24-mm aperture-grille pitch is roughly
equal to a 0.28-mm shadow mask.
Figure 7-4 illustrates one possible arrangement of color dots. The dot pitch
is the center-to-center spacing between adjacent like-colored dots. The three dots
create a triad. Because each electron beam passes through the same hole, the
mask pitch is the same as the triad pitch. Table 7-3 relates the triad dot pitch to
the more common definitions of CRT resolution.
Table 7-3
RESOLUTION OF COLOR DISPLAYS
RESOLUTION TRIAD DOT PITCH
,..'2,'0~\\
\/- \
\888888
---000008
0.2Bmm 088888
---eeeeee
8@8088
Figure 7-4. Arrangement of phosphor dots on a CRT. There is one
shadow-mask hole for each triad. In 1997, most computer monitors had
a triad pitch of 0.28 mm. The triad is outlined.
By varying the ratio of red, green, and blue, a variety of effective color
temperatures can be achieved. The video signal is simply a voltage with no color
temperature associated with it. The display electronics can be adjusted so that
a white target appears as if illuminated by a source with color temperature range
somewhere between 3200 K and 10,000 K. Most displays are preset to either
6500 K or 9300 K. As the color temperature increases, whites appear to change
from a yellow tinge to a blue tinge. The perceived color depends on the adapting
illumination (e.g., room lights). Setting the color temperature to 9300K provides
aesthetically pleasing imagery and this setting is unrelated to the actual scene
colors (see Section 5.3.2., Color Correction, page 165).
7.2.3. HDTV
As the name implies, high definition television (HDTV) provides more
resolution than conventional television receivers The HDTV display has a 16:9
aspect ratio compared to the NTSC 4:3 aspect ratio. It is designed to be viewed
at three times the picture height. While the horizontal visual angle subtended by
NTSC systems is about 11 degrees, the HDTV subtends about 33 degrees. This
larger format gives the feeling of teleprescence - that you are there! The
operation of a CRT-based HDTV is the same as any other CRT.
I( r 2
(7-3)
L(r) - Ae -z- 05PDT)
As illustrated in Figure 7-5, the spot diameter is defined as the full width at
half-maximum (FWHM) and is:
Then
2
-4ln(2) (!.) (7-5)
L(r) = Ae s
CRT-BASED DISPLAYS 211
The advantage of assuming a Gaussian spot is that many resolution measures are
easily compared and the display MTF is easily specified (discussed in Section
10.7., CRT-based Displays).
1
~
en
z
w
1-
~ 0.5
w
>
~
~
w
0::
0
-3 -2 -1 0 1 2 3
NORMALIZED DISTANCE
Figure 7-5. The spot diameter is the full width at one-half maximum
intensity. The profile has been normalized to <TsPoT·
While the shrinking raster method is listed as a resolution test method, its
primary value is in determining the FWHM spot size. With the shrinking raster
method, every line is activated (Figure 7-6). If the lines are far apart, the
valleys between the lines are visible and the raster pattern is obvious. The lines
are slowly brought together until a flat field is encountered. That is, the entire
screen appears to have a uniform intensity. The resolution is the number of
displayed lines divided by the picture height of the shrunk raster:
X lines (7-6)
R =-
FF y em
212 CCD ARRAYS, CAMERAS, and DISPLAYS
1vcm
Shrunk Raster
Raster = X Lines
Figure 7-6. Shrinking raster resolution test method. The raster is shrunk
until the discrete lines can no longer be discerned.
As the lines coalesce, the MTF drops. Under nominal viewing conditions,
experienced observers can no longer perceive the raster when the luminance
variation (ripple) is less than 5%. The line spacing has been standardized9 to
2usPOT (Figure 7-7). Given a shrunk raster resolution of RFF lines/em, the line
spacing in the shrunk raster is 1/RFF em. Then •JsPoT = 112RFFcm and the spot
size is
235 (7-7)
S = FWHM = -
2RFF
While a flat field condition is desirable, the line spacing may be larger. The
shrinking raster methodology is applicable for CR Ts only and cannot be used for
flat panel displays because the "raster" is fixed in those displays.
output
j
raster hne center
Figure 7-7. When the line spacing is 2asroT, the spots appear to merge
and this produces a flat field. The peak-to-peak ripple is 3% and the
MTF associated with the residual raster pattern is 0.014. The output is
the sum of all the discrete line intensities.
7.4. DISELS
Monitor disels are not related to the camera disels. Each is generated by the
respective designs. Cameras with a standard video output may not have the
resolution suggested by the video standard. For example, a camera containing
an array of 320 x 240 pixels may have its output formatted into EIA 170 timing.
This standard suggests that it can support approximately 437 x 485 pixels (see
Table 5-2, page 153). The signal may be digitized by a frame grabber that
creates 1000 x 485 datels. The displayed image still represents 320 x 240
pixels. See Section 5.2.2., Broadcast/Non-broadcast Design (page 155), for a
discussion of pixels, datels, and disels.
214 CCD ARRAYS, CAMERAS, and DISPLAYS
A disel is the smallest imaginary area that can exist on a display (see Section
1.5., Pixels, Datels, Disels, and Resets, page 14). In the display literature,
"pixel" is used. If necessary, the reader should mentally insert pixel wherever
disel is written. The disel size indicates how many individual pieces of
information the system can carry. It does not indicate the extent to which the
detail will be resolved. The distance between adjoining disels is related to
resolution. As the distance decreases, disels overlap and the resolution becomes
poorer. This is further discussed in Section 7. 6 .. Addressability.
The vertical disel size is defined by the raster pitch. It is simply the display
height divided by the number of active scan lines. This is dictated by the video
standard. With an optimum system, the electron beam diameter and video
bandwidth are matched in the horizontal direction. For television receivers, the
number of monitor disels is matched to the video standard. These values were
given in Table 5-2, page 153.
Figure 7-9 illustrates the relationship between the shadow-mask pitch and the
beam profile. Assume the shadow-mask pitch is 0.28mm. If the 5% diameter
covers 2.5 holes, then, using Equation 7-9, asPOT = (0.28)(2.5)/4.9 = 0.143mm.
The spot size (Equation 7-4) isS= (2.35)(0.143) = 0.336mm. In Figure 7-9,
the spots are contiguous and each is considered a disel.
CRT-BASED DISPLAYS 215
"green" electron
beam profile
- - - - - -shadow mask
green image
red image
white image
,_ _ _ _ _ 5Diameter
percent
888
I
Disel
~
888
- - Disel - - - Disel - - - - - Disel -
The disel format is the arrangement of disels into horizontal and vertical
rows. Table 7-4 lists the common formats found in computer color monitors.
The format represents the CRT controller's ability to place the electron beam in
a particular location. The CRT is an analog device even though the format is
listed as discrete values. These monitors are said to be digitally addressed. The
disel format is (erroneously) called the resolution. The relationship between
resolution and format is discussed in Section 7.6., Addressability.
Disel density is usually specified in units of dots per inch. Table 7-5
provides the approximate relationship between "resolution" and disel density.
Confusion exists because dots also refer to the individual phosphor sites. In this
context, a "dot" is a disel. The horizontal disel density is the number of
horizontal disels (Table 7-4) divided by the honzontal display size.
CRT-BASED DISPLAYS 217
Table 7-4
FORMAT OF COLOR DISPLAYS
TYPE FORMAT
Table 7-5
TYPICAL DISEL DENSITY
DISEL DENSITY
"RESOLUTION"
(dpi)
Medium 50-70
Low <50
Figure 7-10 illustrates the "resolution" of three displays. For a fixed format,
as the display size increases, the disel size also increases to yield a decreasing
disel density. Plotted also is the human eye's resolution capability of 1 arc
minute per line expressed at a viewing distance three times the display height.
The 1280 x 960 display provides about the highest perceivable resolution. A
higher disel density will be beyond the eye's resolution limit and therefore
ineffective. While consumer receivers are viewed at 4.2 times the display height,
computer monitors are viewed at a closer distance. Figure 7-10 illustrates that
the observer is an integral component of perceived resolution. When monitors
are designed, a nominal viewing distance is assumed.
218 CCD ARRAYS, CAMERAS, and DISPLAYS
"resolution"
140 ultra-high
~- 120
ii)'fi
z r::: 100 high
w:.::: 3
c..!!! 80
...JGI
~ :6 60
a:- 40
20~~~~~~--+-~----~~--~-.~
10 12 14 16 18 20 22
DIAGONAL SCREEN SIZE (inches)
Figure 7-10. "Resolution" of three displays: (1) 640 x 480, (2) 1024 x
768, and (3) 1280 x 960. The observer's resolution is plotted for a
viewing ratio of three. The 1280 X 960 display is near the observer's
ability to see the disels.
7.5. RESOLUTION
Resolution suggests something about the highest spatial frequency that can
be seen. As an electron beam moves across the screen, its intensity can be
modulated. The maximum rate at which the beam can be modulated from full
off to full on defines the maximum number of (monitor) disels that can be
written in a row. The beam moves down vertically to the next line and again
travels left to right. The distance the beam moves down is the vertical extent of
a disel. While this can be any distance, it is usually selected so that the
individual raster lines cannot be seen. Resolution and addressability specify
monitor performance. They are combined into the resolution/addressability ratio
(RAR). The RAR provides guidance to monitor design.
Resolution metrics are intertwined with television standards such that they
are often presented together. Nearly all televisions are built to the same basic
design so that by specifying EIA 170 or NTSC, the television receiver bandwidth
and spot size is implied. In this sense, a video standard has a "resolution."
CRT-BASED DISPLAYS 219
The vertical resolution is limited by the number of raster lines. For example,
NTSC compatible monitors display 485 lines. However, to see precisely this
number, the test pattern must be perfectly aligned with the raster pattern.
Because the test pattern may be randomly placed, the pattern at Nyquist
frequency can have a zero output if 180° out-of-phase (discussed in Section
8.3.3., Image Distortion). To account for random phases, the Kell factor 12 is
applied to the vertical resolution to provide an average value. A value of 0.7 is
widely used:
RVERTICAL = (activescanlines) (Kellfactor) . (7-9)
Whereas the flat field condition exists when two adjacent lines cannot be
resolved, the TV limiting resolution is a measure of when alternate vertical bars
are just visible (on-off-on-off). The standard resolution test target is a wedge
pattern with spatial frequency increasing toward the apex of the wedge. It is
equivalent to a variable square wave pattern.
The measurement is a perceptual one and the results vary across the
observer population. The industry selected9 the limiting resolution as a bar
spacing of 1.18crsPoT· This result is consistent with the flat field condition. The
flat field condition was determined for two Gaussian beams whereas square
waves are used for the TV limiting resolution test. Suppose the Gaussian beams
in Figure 7-7 (page 213) were separated by an imaginary "black" beam. Then
the beams would be separated by just crsPoT but the intensity distribution would
be identical to Figure 7-7. The black beam does not contribute to the visible
image. For TV limiting resolution, the "on" lines are separated by (2)(1.18
crsPoT). Because bar targets are used, the resultant image is not a precise
Gaussian beam and a larger line separation is required (1.18crsPoT versus crsroT)
for the TV limiting resolution.
The flat field condition and high TV limiting resolution are conflicting
requirements. For high TV limiting (horizontal) resolution, crsroT must be small.
But raster pattern visibility (more precisely, invisibility) suggests that crsPoT
should be large. This is further discussed in Section 7.6., Addressability.
CRT-BASED DISPLAYS 221
Example 7-1
SPOT SIZE versus TV LIMITING RESOLUTION
The vertical extent of the 4:3 aspect ratio monitor is 8.4". With 485 lines,
each line is separated by 0.0173". The value CJsPoT is 0.0173/2 = 0.00865" and
the spot size (FWHM) is 0.0204".
===============================================~
7.5.4. MTF
TV limiting resolution occurs when the bar spacing is 1.18usPoT· Because a
cycle consists of two TV lines, the square wave fundamental frequency is
11(2.35usPoT) and the MTF is 0.028. Conversely, the display resolution is
selected as that spatial frequency that provides an MTF of 0.028. The MTF
equation is found in Section 10.6., CRT-based Displays. Any spatial frequency
could be used to specify resolution and the slightly different values are cited in
the literature. For commercial applications, the resolution is often specified by
the spatial frequency where the MTF is 10%.
But square waves are the most popular test patterns. A specification could
be: The CTF must be greater than 10% when viewing 320 line pairs. This is
equivalent to 640 disels in an on-off-on-off configuration). Note that by using
square wave targets, the system response is the contrast transfer function and not
the MTF (discussed in Section 9.5., Contrast Transfer Function). Careless usage
results in calling the system response to square waves, the MTF. This makes it
extremely difficult to compare specifications or to calculate performance from
published specifications.
222 CCD ARRAYS, CAMERAS, and DISPLAYS
When designing a simulation system, the distance to the display and display
size are selected first. These are based on physical constraints of the overall
system design. The simulation industry generally accepts resolution as that
spatial frequency (line pair) which provides a CTF of 10% at this viewing ratio.
Because the number of lines depends on what is seen visually, the simulation
industry uses units of optical line pairs (OLP)_ This translates into arc-minutes
per line pair.
7 .6. ADDRESSABILITY
There are two opposing design requirements. The first design requirement
is that the raster pattern be imperceptible to the observer. This is called the
adjacent disel requirement. If the display meets this requirement, the picture will
appear uniform and solid. It provides the flat field condition. Alphanumeric
characters will appear well constructed and highly legible. The second design
requirement is the alternating disel requirement. Here, individual lines (one disel
on and one disel off) should be highly visible.
RAR=-
s 2.36 a SPOT (7-11)
p p
where P is the disel pitch. For example, if a display is 27.5 em high and it
displays (addresses) 1024 lines, then Pis 27.5 cm/1024lines = 0.27 mm/line.
Assuming a 0.28-mm-wide spot then RAR is 0.28/0.27 = 1.04.
Before the RAR was defined, display design was based on the
experimentally determined flat field condition. The flat condition occurs when
P = 2crsPoT or RAR = 1.18. Here, the adjacent disel modulation is near zero -
the lines are not visible. The TV limiting resolution describes the alternating
disel situation. Now, P = 1.18crsPoT• RAR = 2, and the alternating disel
modulation is near zero (e.g., the lines are not discernible).
z 0.8
0 alternating
i= 0.6 disels
<
..J
::::) 0.4
c
0 0.2
::!:
0
0 0.5 1.5 2 2.5
RAR
Figure 7-11. Modulation based on Gaussian beam profiles for the
adjacent disel and alternating disels. A RAR of one is considered
desirable. The modulation of the alternating disel is associated with the
display's Nyquist frequency. The modulation of the adjacent disel is
associated with the display's sampling frequency. These are imaginary
frequencies because the CRT is an analog device. They represent the
ability to place a spot at the desired location on the CRT faceplate.
224 CCD ARRAYS, CAMERAS, and DISPLAYS
The alternating disel modulation represents one Gaussian beam on and one
off. This is different from one bar on and one bar off. Resolution tests are
performed with square waves. For an analog system, the contrast transfer
function is 417r times greater than the MTF (discussed in Section 9.5., Contrast
Transfer Function).
If the horizontal format supports 640 disels (see Table 7-4, page 217), then
320 cycles should be perceived. But the modulation is not specified at this
resolution. For high quality monitors, the 320 cycles are easily visible (RAR <
1). With average monitors, the 320 cycles are visible but with reduced
=
modulation (RAR 1). On a poorly designed monitor, the cycles will be barely
perceived (RAR > 1).
While any disel format can be specified, the RAR determines if the disels
are visible. For the configuration shown in Figure 7-9, (page 216), if the
shadow mask pitch is 0.28mm, then the disels will be separated by 0.336 mm
when the RAR is one. If using a 13-inch diagonal display, the width is 264mm
and 786 horizontal disels are viewable when the RAR = 1. More disels can be
displayed, but with reduced modulation (Table 7-6 and Figure 7-12). Thus the
disel format, by itself, does not fully characterize the display capability. For
optimized displays, it is reasonable to assume that the RAR is about one.
Table 7-6
ALTERNATING DISEL MODULATION
13-inch diagonal display and 0.28-mm shadow-mask pitch
5% intensity level illuminating 2.5 shadow-mask holes
DISELS RAR MODULATION
800 1.02 0.75
900 1.15 0.61
1000 1.27 0.47
ADDRESSABLE DISELS
ANSI!HFS 100-1988 requires that the adjacent disel modulation be less than
20%. This deviates significantly from the flat field condition (adjacent disel
modulation is 0.014). ISO 9241 recommends that the alternating disel
modulation be greater than 0.4 for monochrome and greater than 0.7 for color
systems. For monochrome systems, the RAR will vary from 0.8 to 1.31. For
color systems the RAR varies from 0.8 to 1.04. These standards were developed
to assure legibility of alphanumeric characters.
If the RAR is too high, then the output of two adjacent "on" lines is much
greater then the output of two separated lines due to the summing effect (Figure
7-13). For the flat field condition (Figure 7-7, page 213), the RAR is 1.18 and
the summed lines are 27% brighter than an individual line. If the RAR is very
large, the intensity of several adjacent lines "on" can be significantly brighter
than a single line "on". Thus, there is also a tradeoff in intensity with RAR.
Multi-sync monitors automatically adjust the internal line rate to the video
line rate. These monitors can display any format from 525 lines up to 1225
lines. Because the displayed image size remains constant, the vertical line
spacing changes. At the 1225 line format, the raster lines are close together and
form a uniform (flat) field. At the lower line rates, the individual raster lines are
separated and may be perceived. Because the same video amplifiers are used for
all line rates, the horizontal resolution is independent of the line rate.
226 CCD ARRAYS, CAMERAS, and DISPLAYS
That is, the RAR in the horizontal direction stays constant whereas the RAR in
the vertical direction changes with the number of lines. If the vertical RAR is
one at 1225 lines, then it must be 525/125 = 0.43 at 525 lines.
1.4
1-
:::J 1.3
c..
1-
:::J 1.2
0
1.1
c
w 1
:!
:! 0.9
:::J
U)
0.8
0.8 0.9 1.1 1.2 1.3 1.4
RAR
Figure 7-13. Summed output of contiguous lines "on". Ideally, the
intensity should remain constant whether one or multiple lines are "on".
~=============================================
Example 7-2
DISPLAY RESOLUTION
For a 4:3 aspect ratio display, the CRT vertical extent is 284 mm. The
estimated spot size is 284/800 = 0.355 mm and UsroT is S/2.35 = 0.335/2.35
=0.15 mm. If the number of displayed lines is NuNE• the RAR is
With 485 displayed lines (EIA 170), the RAR is 0.61. From Figure 7-11 the
raster pattern (adjacent disels) is quite visible. For EIA 343A with 809 active
lines, the RAR is at the desired value of one. At 946 active lines, the RAR
increases to 1.18. Here the alternating disel modulation starts to drop and the
vertical resolution starts to decrease.
============================================~
CRT-BASED DISPLAYS 227
The ability to see detail is related to the RAR. Figure 7-14 illustrates three
letters and the resultant intensity traces. The RAR must be near one so that a
reasonable contrast ratio exists between "on" and "off" disels. With reasonable
contrast, the inner detail of the character is seen and the character is legible.
Similarly, with a reasonable contrast ratio, adjacent letters will appear as
separate letters. With characters, the alternating disel pattern is called one stroke
separated by a space.
Slit
Photometer
.
Scan
Path
Slit Photometer
Scan Path Luminance
(Within) (Between)
Figure 7-14. The RAR must be near one so that the inner details of
characters are visible and the double width appears solid. Similarly with
RAR near one, two adjacent characters appear separate. Each dot
represents a disel. Lower RARs will make the characters appear very
sharp. However, the raster pattern will become visible and the characters
will not appear solid.
7. 9. CONTRAST
The air-glass interface and the phosphor reflect ambient light. This reduces
contrast between disels and reduces the display's color gamut. As a result, most
displays are measured in a darkened room where the reflected light is less than
1 % of the projected light. Typically the contrast is measured with a
checkerboard pattern consisting of 16 alternating rectangles. The average value
of the eight white rectangles and the average value of the eight black rectangles
are used to calculate the contrast ratio.
CRT-BASED DISPLAYS 229
Figure 7-15. Complex symbols are more legible when the display
modulation is high. Each dot represents a hole in the shadow mask. Top:
modulation= 0.5 at the Nyquist frequency. Bottom: modulation= 0.1
at the Nyquist frequency. When viewed at 10 feet, the individual dots
are no longer perceptible. At this distance, the top characters appear
sharp (RAR = 1. 25) and the bottom ones appear fuzzy (RAR = 1. 8). By
courtesy of the Mitre Corporation.
230 CCD ARRAYS, CAMERAS, and DISPLAYS
where LwHrrE and LsLAcK represent the luminances from the white and black
rectangles, respectively. The reflections are the same regardless if a disel is on
or off. About 4 percent of the ambient light (LAMBIENT) is reflected at the air-
glass interface. The CRT phosphor is an excellent Lambertian reflector with a
typical reflectivity of 70%. "Black" displays minimize these reflections and
thereby increase the perceived contrast.
The contrast ratio for typical displays is about 20: 1 under normal ambient
lighting conditions. This decreases as the ambient light increases. The eye
operates in the reverse way. As the ambient lighting increases, more contrast is
necessary to see target-background differences. Thus, as the displayed contrast
decreases in bright light, the eye requires more contrast. This explains why
television imagery is so difficult to see in sunlight and why it appears so bright
in a darkened room.
7.10. REFERENCES
1. L. E. Tannas Jr., ed., Flat panel Displays and CRTs, Van Nostrand, New York, NY (1985).
2. J. A. Castellano, Handbook of Display Technology, Academic Press, New York, NY (1992).
3. S. Sherr, Electronic Displays, Second edition, Wiley and Sons, New York, NY (1993).
4. J. C. Whitaker, Electronic Displays, McGraw-Hill, New York, NY (1994).
5. J. Peddie, High-Resolution Graphics Display Systems, Windcrest/McGraw-Hill, NY (1994).
6. L. W. MacDonald, ed., Display Systems, Design and Applications, Wiley, NY (1997).
7. W. Cowan, "Displays for Vision Research," in Handbook of Optics, M. Bass, ed., Second
edition, Vol. 1, Chapter27, McGraw-Hill, New York. NY (1995).
8. T. Olson, "Behind Gamma's Disguise," SMPTE Journal, Vol. 104(7), pp. 452-458 (1995).
9. L. M. Biberman, "Image Quality," in Perception of Displayed Information, L. M. Biberman,
ed., pp. 13-18, Plenum Press, New York (1973).
10. P. G. J. Barten, "Spot Size and Current Density Distributions of CRTs," in Proceedings of the
SID, Vol. 25(3), pp. 155-159 (1984).
11. A. Miller, "Suppression of Aliasing Artifacts in Digitally Addressed Shadow-mask CRTs,"
Journal of the Society of Information Display, Vol. 3(3), pp 105-108 (1995).
12. S.C. Hsu, "The Kell Factor: Past and Present," SMPTE Journal, Vol. 95, pp. 206-214(1986).
13. G. M. Murch and R. J. Beaton, "Matching Display Resolution and Addressability to Human
Visual Capacity," Displays, Vol. 9, pp. 23-26 (Jan 19R8).
8
SAMPLING THEORY
Sampling is an inherent feature of all electronic imaging systems. The scene
is spatially sampled in both directions due to the discrete locations of the
detector elements. This sampling creates ambiguity in target edges and produces
moire patterns when viewing periodic targets. Aliasing becomes obvious when
(a) the image size approaches the detector size and (b) the detectors are in a
periodic lattice (the normal situation). Spatial aliasing is rarely seen in
photographs or motion pictures because the grains are randomly dispersed in the
photographic emulsion.
If the detector size and spacings are different in the horizontal and vertical
directions, sampling effects will be different in the two directions. This leads to
moire patterns that are a function of the target orientation with respect to the
array axis. Although sampling is two-dimensional, for simplicity sampling
effects will be presented in one-dimension.
231
232 CCD ARRAYS, CAMERAS, and DISPLAYS
After aliasing, the original signal can never be recovered. It would appear
that aliasing is an extremely serious problem. However, the extent of the
problem depends upon the final interpreter of the data.
Figure 8-2. A raster scan system creates moire patterns when viewing
wedges or star bursts.
SAMPLING THEORY 233
Sampled data systems are nonlinear and do not have unique MTFs. 1-6 The
"MTF" depends on the phase relationships of the scene with the sampling
lattice. Superposition does not hold and any MTF derived for the sampler
cannot, in principle, be used to predict results for general scenery. To account
for the nonlinear sampling process, a sample-scene MTF is added (discussed in
Section 10.10., Sampling Effects). As the sampling rate increases, the MTF
becomes better defined. As the sampling rate approaches infinity, the system
becomes an analog system and the MTF is well defined. Equivalently, as the
detector size and detector pitch decrease, signal fidelity increases. Reference 7
provides an in-depth discussion on sampling effects. This chapter highlights
those results.
The symbols used in this book are summarized in the Symbol List (page
xviii) and it appears after the Table of Contents.
234 CCD ARRAYS, CAMERAS, and DISPLAYS
Three conditions must be met to satisfy the sampling theorem. The signal
must be band-limited, the signal must be sampled at an adequate rate, and a low-
pass reconstruction filter must be present. When any of these conditions are not
present, the reconstructed analog data will not appear exactly as the originals.
After digitization, the data reside in data ;mays (e.g., computer memory
location) with nonspecific units. The user assigns units to the arrays during
image reconstruction. That is, the data are read out of the memory in a manner
consistent with display requirements. Because monitors are analog devices, the
data are transformed into an analog signal using a sample-and-hold circuit. This
creates a "blocky" image and can be removed by a reconstruction filter
(discussed in Section 10.6.4., Post-Reconstruction Filter). If the original signal
was oversampled and if the post-reconstruction filter limits frequencies to fN,
then the reconstructed image can be identical to the original image.
SAMPLING THEORY 235
LL
1-
:e
w
>
~
<
...J
w
0::::
fH
RELATIVE FREQUENCY
(a)
LL
1-
:e
w
2!
1-
:3w
0::::
RELATIVE FREQUENCY
(b)
LL
1-
:e
w
2!
1-
<
...J
w
0::::
(c)
Figure 8-3. Sampling replicates frequencies at nfs ±f. (a) Original band-
limited signal, (b) frequency spectrum after sampling, and (c) when fs =
2fH, the bands just adjoin.
236 CCD ARRAYS, CAMERAS, and DISPLAYS
8.2. ALIASING
As the sampling frequency decreases, the first side band starts to overlap the
base band. It is the summation of these power spectra that create the distorted
image (Figure 8-4). The overlaid region creates distortion in the reconstructed
image. This is aliasing. Once aliasing has occurred, it cannot be removed.
Within an overlapping band, there is an ambiguity in frequency. It is impossible
to tell whether the reconstructed frequency resulted from an input frequency of
fo or nfs ± fo.
All frequency components above fN are folded back into the base band so
that the base band contains
To avoid aliasing in the electronic domain, the signal may be passed though
a low-pass filter (anti-aliasing filter). These filters cannot remove the aliasing
that has taken place at the detector. They can only prevent further aliasing that
might occur in the downstream analog-to-digital converter. Figure 8-5 illustrates
the cutoff features of an ideal filter. The ideal filter passes all frequencies below
fN and attenuates all frequencies above fN. The ideal filter is, of course,
unrealizable. 10 It also can produce undesirable effects such as ringing.
::!!
::;)
0:::
t-
o
w
ll.
UJ
0:::
w
3:
0
ll.
RELATIVE FREQUENCY
Figure 8-4. Aliasing alters both the signal <md noise base band spectra.
In a real system, the signal and noise are not necessarily band-limited
and some aliasing occurs.
SAMPLING THEORY 237
sample
frequency
0.8
LL 0.6
1-
:::! 0.4
0.2
0
0 0.25 0.5 0.75 1.25 1.5 1.75 2
Nyquist
frequency
0.8
LL 0.6
!i 0.4
0.2
0~------~----~-------+------~
0 0.25 0.5 0.75
Figure 8-6. Signals above Nyquist frequency are aliased down to the
base band. The area bounded by the MTF above fN may be considered
as an "aliasing" metric. 11
Nyquist Sample
frequency frequency
0.8
LL 0.6
1-
::E 0.4
0.2
0
0 0.25 0.5 0.75 1.25 1.5 1.75 2
NORMALIZED FREQUENCY
Figure 8-7. MTF representation of an undersampled system as a function
of f/fN. This restrictive representation erroneously suggests that there is
no response above fN.
SAMPLING THEORY 239
Sections 8-1 and 8-2 described sampling in generic terms. The variable fwas
used to denote frequency. There are four different frequency domains that can
exist with an electro-optical imaging system (discussed in Section 10 .1.,
Frequency Domains). Imagery at the detector plane is described by image-space
spatial frequency. The horizontal and vertical variables are represented by ui and
vi, respectively. They have units of cycles/rom.
The detector MTF cannot exist by itself. Rather, the detector MTF must also
have the optical MTF to make a complete imaging system. In the horizontal
direction, the MTF of a single rectangular detector is
(8-2)
Idetector! petectorl
(a)
(b)
Figure 8-8. A detector spatially integrates the signal. Figure 8-1 (page
231) represents the output of extremely small detectors. (a) Detector
width is one-half of the center-to-center spacing and (b) detector width
is equal to the center-to-center spacing. (a) Typifies an interline transfer
CCD array in the horizontal direction (see Figure 3-16, page 64) and (b)
is representative of frame transfer arrays. The heavy lines are the
detector output voltages. As the phase changes, the output also changes.
Figure 8-9 illustrates the detector MTF in one dimension. The MTF is
equal to zero when ui = k/dH. The first zero (k = 1) is considered the detector
cutoff, uiD. Any input spatial frequency above detector cutoff will be aliased
down to a lower frequency. It is customary to plot the MTF only up to the first
zero (Figure 8-1 0). The optical system cutoff limits the absolute highest spatial
frequency that can be faithfully imaged. The detector cutoff may be either higher
(optics-limited system) or lower (detector-limited system). Most CCD-based
imaging systems are detector-limited (discussed in Section 10.4., Optics-Detector
Subsystem).
SAMPLING THEORY 241
0.8
LL 0.6
1-
::!! 0.4
0.2
0
0 0.5 1.5 2 2.5 3
NORMALIZED SPATIAL FREQUENCY
Oo8
LL 006
1-
::!! Oo4
Oo2
0
0 Oo25 Oo5 Oo75
NORMALIZED SPATIAL FREQUENCY
Microlenses increase the effective detector size but do not affect the Nyquist
frequency. By increasing the detector area, the detector cutoff decreases. With
microlenses, the array in Figure 8-11b can optically appear as the array in
Figure 8-lla.
Most single chip color filter arrays have an unequal number of red, green,
and blue detectors. A typical array designed for NTSC operation will have 768
detectors in the horizontal direction with 384 detectors sensitive to green, 192
sensitive to red, and 192 sensitive to the blue region of the spectrum. Suppose
the arrangement is G-B-G-R-G-B-G-R. The "green" detectors will have a center-
to-center spacing as shown in Figure 8-11b.
The spacing of the "blue" and "red" detectors is twice the "green" detector
spacing. This produces a "blue" and "red" array Nyquist frequency that is one-
half of the "green" array Nyquist frequency (Figure 8-12). Other detector
layouts will create different array Nyquist frequencies. The "color" Nyquist
frequencies can be different in the horizontal and vertical directions. These
unequal array Nyquist frequencies create color aliasing in single chip cameras
that is wavelength specific. Black-and-white scenes can appear as green, red, or
blue imagery. 12 A birefringent crystal, inserted between the lens and the array,
changes the effective sampling rate and minimizes color aliasing (discussed in
Section 10.3.4., Optical Anti-Alias Filter).
SAMPLING THEORY 243
rn
1++1
dCCH
1 Sampling:
0.8
LL 0.6
1-
:::t 0.4
0.2
0
0 0.25 0.5 0.75 1
NORMALIZED SPATIAL FREQUENCY
(a)
D1+- D
+I
D
dCCH
1 , Sampling
0.8
LL 0.6
1-
:::t 0.4
0.2
0
0 0.25 0.5 0.75 1
NORMALIZED SPATIAL FREQUENCY
(b)
1r--:----.:
0.8 'lii
·s .
C"~·
LL 0.6 >- r::: '
1- Z111
: ::::J I
1 Nyquist Sampling
LL frequency frequency
1- 0.8 without
:::IE
..J 0.6
<(
0
0.4
~
w 0.2
row summing
>
0
0 0.25 0.5 0.75
The square wave (bar pattern) is the most popular test target. It is
characterized by its fundamental frequency only. The expansion of a square
wave into a Fourier series clearly shows that it consists of an infinite number of
sinusoidal frequencies. Although the square wave fundamental frequency may
be oversampled, the higher harmonics may not. During digitization, the higher
order frequencies will be aliased down to lower frequencies and the square wave
will change its appearance. There will be intensity variations from bar-to-bar and
the bar width will not remain constant.
N= (8-3)
Figure 8-14 illustrates Nasa function of f0 /fN. As f0 /fN approaches one, the beat
frequency becomes obvious. Lomheim et. al. 13 created these beat frequencies
when viewing multiple bar patterns with a CCD camera.
en
w
10
..J
(.) 8
>
(.)
u. 6
0
0:: 4
w
m 2
:::iE
~
z 0
0.5 0.6 0.7 0.8 0.9
NORMALIZED FREQUENCY
Figure 8-14. Number of input frequency cycles required to see one
complete beat frequency cycle as a function of f0 /fN.
246 CCD ARRAYS, CAMERAS, and DISPLAYS
Figures 8-15 through 8-18 illustrate an ideal staring array output when the
optical MTF is unity for all spatial frequencies and therefore represents a worst
case scenario. That is, all higher harmonics that would have been attenuated by
the optical system are available for aliasing by the detector. For these figures,
dH/dccH is one. The individual detectors sample the square wave in the same
manner as illustrated in Figure 8-8b (page 240). Section 10.4., Optics-Detector
Subsystem describes how the optical MTF modifies the imagery.
If u.,/u;N= 0.952, the beat frequency is equal to 9.9 cycles of the input
frequency (Figure 8-15). Although a square wave consists of an infinite number
of frequencies, it is characterized by its fundamental, u.,. The target must contain
at least 10 cycles to see the entire beat pattern. A 4-bar pattern may either be
replicated or nearly disappear depending on the phase.
When U0 /U;N is less than about 0.6 (Figure 8-16), the beat frequency is not
obvious. Now the output nearly replicates the input but there is some slight
variation in pulse width and amplitude. In the region where u.,/u;N is
approximately between 0.6 and 0.9, adjacent bar amplitudes are always less than
the input amplitude (Figure 8-17).
SAMPLING THEORY 247
When U /U;N is less than about 0.6, a 4-bar pattern will always be seen
0
(select any four adjoining bars in Figure 8-16). When llofU;N < 0.6, phasing
effects are minimal and a phase adjustment of ± 1/2d in image space will not
affect an observer's ability to resolve a 4-bar target. In the region where f0 /fN
is approximately between 0.6 and 0.9, 4-bar targets will never look correct
(Figure 8-17). One or two bars may be either much wider than the others or one
or two bars may be of lower intensity than the others. This phasing effect is a
measurement concern and is not normally included in modeling.
Figure 8-16. Ideal staring system output when U0 1U;N = 0.522. The value
N = 0.54 and dH/dccH = 1. The light line is the input and the heavy line
is the detector output. The output nearly replicates the input when U0 1U;N
< 0.6.
Figure 8-17. Ideal staring system output when U0 /U;N = 0. 811. The value
N = 2.14. The light line is the input and the heavy line is the detector
output. The output never looks quite right when U0 /u;N is between 0.6
and 0.9.
248 CCD ARRAYS, CAMERAS, and DISPLAYS
Figure 8-18. Ideal staring system output when U 0 /uiN = 1.094. The value
N = 5. 8. The light line is the input and the heavy line is the detector
output. By selecting the appropriate phase, the output appears to
replicate an input 4-bar pattern.
Systems can detect signals with spatial frequencies above cutoff but cannot
faithfully reproduce them. For example, patterns above the Nyquist frequency
are aliased to a frequency below Nyquist and a 4-bar pattern may appear as a
distorted 3-bar pattern.
The phasing effects shown in Figures 8-15 through 8-18 "disappear" when
the target is moving. Here, each frame provides a different image. For each
target whose spatial frequency is above Nyquist frequency, one, two, or three
bars are visible depending on the relative phase. As the pattern moves, a
different set of bars is visible. The eye blends these individual patterns so that
you perceive four bars. This prompted some researchers 14 to develop a dynamic
test in which the 4-bar target moves.
SAMPLING THEORY 249
Example 8-1
SYSTEM CUTOFF
A staring array consists of detectors that are 10 p,m x 10 p,m in size. The
detector pitch is 15 p.m. The focal length is 15 em. The aperture diameter is 3
em and the average wavelength is 0.5 p,m. What is the system cutoff?
System cutoff is the smaller of the optical cutoff, detector cutoff, or Nyquist
frequency. In image space, the optical cutoff is uic = D0 /(}..fl) = 400 cycles/mm.
The detector cutoff is uic = 1/dH = 100cycles/mm. The detector pitch provides
sampling every 15 p,m for an effective sampling rate at 66.7 cycles/mm. Because
the Nyquist frequency is one-half the sampling frequency, the system cutoff is
33.3 cycles/mm.
Because object space is related to image space by the lens focal length, the
optical cutoff is 30 cycles/rnrad, detector cutoff is 15 cycles/rnrad, sampling is
10 cycles/rnrad, and the Nyquist frequency is 5 cycles/rnrad.
Because the MTF remains zero for all frequencies higher than the optical
cutoff, any scene frequency above uic will appear as a uniform blob with no
modulation. Spatial frequencies higher than the detector cutoff or the Nyquist
frequency (but less than ud will be aliased down to lower frequencies.
=============================================~
For many applications, the camera output is an analog signal that is then
re-sampled by a frame grabber (frame capture device). If the frame grabber's
sampling rate is sufficiently high and the analog signal is band-limited, the
reconstructed signal will replicate the analog signal in frequency, amplitude, and
pulse width.
The variations in pulse width can only be seen on a monitor because the frame
grabber data reside in a computer memory. Note that output pulse widths are
equally spaced. They just do not match up with the input signal.
.. - ...
Figure 8-19. Frame grabber output after sample-and-hold circuitry when
uju;N = 0.952. The light line is the input and the heavy line is the
output. Here, u;N is the Nyquist frequency created by the frame grabber
clock rate. The output is representative of what is seen on a monitor.
Some frame grabbers have an internal anti-alias filter. This filter ensures that
the frame grabber does not produce any additional aliasing. The filter cutoff is
linked to the frame grabber clock and is not related to the camera output. Once
aliasing has occurred in the camera, the frame grabber anti-alias filter cannot
remove it. If the filter does not exist, the frame grabber may create additional
aliasing. 15
Aliasing is rarely discussed with displays, but under certain conditions can
occur. Many displays employ digital techniques to match the electronic signal
to the CRT characteristics. Ideally, the internal analog-to-digital converter
should have a low-pass filter to eliminate possible aliasing. Although good
engineering design recommends that the low-pass filter be included, cost
considerations may eliminate it. With standard video formats, the bandwidth is
limited by the video format and there is no need for the low-pass filter. If a
nonstandard bandwidth is used such as that required by a high resolution CCD
camera, these monitors may exhibit some aliasing. This becomes noticeable as
the array size increases. For example, a high quality monitor can display 1000
x 1000 pixels. If the array consists of 4000 x 4000 elements, the display may
alias the image as depicted in Figure 8-1 (page 231). There is often a loss of
resolution when changing to a lower resolution format.
SAMPLING THEORY 251
8.5. REFERENCES
1. S. K. Park and R. A. Schowengerdt, "Image Sampling, Reconstruction and the Effect of
Sample-scene Phasing," Applied Optics, Vol. 21(17), pp. 3142-3151 (1982).
2. S. E. Reichenbach, S. K. Park, and R. Narayanswamy, "Characterizing Digital Image
Acquisition Devices," Optical Engineering, Vol. 30(2), pp. 170-177 (1991).
3. W. Wittenstein, J. C. Fontanella, A. R. Newberry, and J. Baars, "The Definition of the OTF
and the Measurement of Aliasing for Sampled Imaging Systems," Optica Acta, Vol. 29(1), pp.
41-50 (1982).
4. J. C. Felz, "Development of the Modulation Transfer Function and Contrast Transfer Function
for Discrete Systems, Particularly Charge Coupled Devices," Optical Engineering, Vol. 29(8),
pp. 893-904 (1990).
5. S. K. Park, R. A. Schowengerdt, and M. Kaczynski, "Modulation Transfer Function Analysis
for Sampled Image Systems," Applied Optics, Vol. 23(15), pp. 2572-2582 (1984).
6. L. deLuca and G. Cardone, "Modulation Transfer Function Cascade Model for a Sampled IR
Imaging System," Applied Optics, Vol. 30(13), pp. 1659-1664 (1991).
7. G. C. Holst, Sampling, Aliasing, and Data Fidelity, JCD Publishing, Winter Park, FL (1998).
8. C. E. Shannon, "Communication in the Presence of Noise," Proceedings of the IRE, Vol. 37,
pp. 10-21 (January 1949).
9. A. B. Jerri, "The Shannon Sampling Theorem - Its Various Extensions and Applications: A
Review," Proceedings of the IEEE, Vol. 85(11), pp. 1565-1595 (1977).
10. G. C. Holst, Sampling, Aliasing, and Data Fidelity, pp. 77-78, JCD Publishing, Winter Park,
FL (1998).
11. G. C. Holst, Sampling, Aliasing, and Data Fidelity, pp. 293-320, JCD Publishing, Winter
Park, FL (1998).
12. J. E. Greivenkamp, "Color Dependent Optical Prefilter for the Suppression of Aliasing
Artifacts," Applied Optics, Vol. 29(5), pp 676-684 (1990).
13. T. S. Lomheim, L. W. Schumann, R. M. Shima, J. S. Thompson, and W. F. Woodward,
"Electro-Optical Hardware Considerations in Measuring the Imaging Capability of Scanned
Time-delay-and-integrate Charge-coupled Imagers," Optical Engineering, Vol. 29(8), pp. 911-
927 (1990).
14. C. M. Webb, "MRTD, How Far Can We Stretch It?" in Infrared Imaging Systems: Design,
Analysis, Modeling, and Testing V, G. C. Holst, ed., SPIE Proceedings Vol. 1689, pp. 297-307
(1994).
15. G. C. Holst, Sampling, Aliasing, and Data Fidelity, pp. 154-156, JCD Publishing, Winter
Park, FL (1998).
9
LINEAR SYSTEM THEORY
Linear system theory was developed for electronic circuitry and has been
extended to optical, electro-optical, and mechanical systems. It forms an
indispensable part of system analysis. For modeling purposes, electronic imaging
systems are characterized as linear spatial-temporal systems that are shift-
invariant with respect to both time and two spatial dimensions. Although space
is three-dimensional, an imaging system displays only two dimensions.
Electrical filters are different from spatial filters in two ways. They are
single-sided in time and must satisfy the causality requirement that no change in
the output may occur before the application of an input. Optical filters are
double-sided in space. Electrical signals may be either positive or negative,
whereas optical intensities are always positive. As a result, optical designers and
circuit designers often use different terminology.
252
LINEAR SYSTEM THEORY 253
least significant bit, then the system can be treated as quasi-linear over a
restricted region.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
Let the response to two inputs, f 1(t) and fz(t), be g 1(t) and gz(t):
(9-2)
For a linear system, the response to a sum of inputs is equal to the sum of
responses to each input acting separately. For any arbitrary scale factors, the
superposition principle states
(9-3)
When passing the signal through another linear system, the new operator
provides:
(9-4)
254 CCD ARRAYS, CAMERAS, and DISPLAYS
L VIN
1 1
(t )<'i(t - t ) 11t
1
'
(9-5)
t'--~
where o(t - t') is the Dirac delta or impulse function at t = t' and viN(t') is the
signal evaluated at t'.
object
Dirac delta
vou·J.t) = L 1
h{viN(t )<'i(t-t )11t
1 1
} • (9-6)
t 1--co
Comparing Equation 9-5 to Equation 9-3, for each t', v1N(t') replaced ai and o(t
- t') replaced fi(t). That is, the input has been separated into a series of
functions a 1f 1(t) + a2fz(t) + ....
LINEAR SYSTEM THEORY 255
(9-7)
where * indicates the convolution operator (other texts may use *or ®). The
system impulse response is h{o(t)}. In Figure 9-2, each impulse response is
given for each t'. The individual impulse responses are added to produce the
output. This addition is the definition of superposition.
Figure 9-2. The linear operator, h{ }, transforms each input Dirac delta
into an output impulse response. The sum of the impulse responses
creates the output. Although shown separated for clarity, the impulse
responses have infinitesimal separations.
The convolution theorem of Fourier analysis states that the Fourier transform
of the convolution of two functions equals the product of the transform of the
two functions. That is, multiple convolutions are equivalent to multiplication in
the frequency domain. Using frequency domain analysis is convenient because
multiplications are easier to perform and easier to visualize than convolutions.
Then
(9-9)
256 CCD ARRAYS, CAMERAS, and DISPLAYS
The variable f denotes electrical frequency (Hz). Upper case variables represent
the Fourier transform of the signal (lower case variables). Let the input be an
impulse. The Fourier transform of an impulse response is
(9-10)
That is, all frequencies are present. Because the input to the system contains all
frequencies, the output must be a function of the system only. Any variation in
phase or amplitude must be due to the system itself. The output is
i(x,y) = JJo(x ,y 1 1
) h{ &(x- x 1) o(y- y 1)} dx 1 dy 1 • (9-14)
LINEAR SYSTEM THEORY 257
It is symbolically represented by
If all frequencies are present in the image (i.e., an impulse), O(u,v)= 1, then
l(u,v) = H(u,v). The relationship between spatial frequency (u,v) and electrical
frequency (f) is further discussed in Section 10.1., Frequency Domains.
The electronic imaging system response consists of both the optical response
and the electronic response: h(x,y,t). Time and spatial coordinates are treated
separately. For example, optical elements do not generally change with time and
therefore are characterized only by spatial coordinates. Similarly, electronic
circuitry exhibits only temporal responses. The detector provides the interface
between the spatial and temporal components, and its response depends on both
temporal and spatial quantities. The conversion of two-dimensional optical
information to a one-dimensional electrical response assumes a linear
photodetection process. Implicit in the detector response is the conversion from
input photon flux to output voltage.
The optical transfer function (OTF) plays a key role in the theoretical
evaluation and optimization of an optical system. The modulation transfer
function (MTF) is the magnitude and the phase transfer function (PTF) is the
phase of the complex-valued OTF. In many applications, the OTF is real-valued
and positive so that the OTF and MTF are equal. When focus errors or
aberrations are present, the OTF may become negative or even complex valued.
258 CCD ARRAYS, CAMERAS, and DISPLAYS
Symbolically
and
With appropriate scaling, the electronic frequencies can be converted into spatial
frequencies. This is symbolically represented by f-+u. The electronic circuitry
is assumed to modify the horizontal signal only (although this depends on the
system design). The combination of spatial <md electronic responses is some-
times called the system OTF:
and
(9-21)
While individual lens elements each have their own MTF, the MTF of the lens
system is not usually the product of the individual MTFs. This occurs because
one lens may minimize the aberrations created by another.
The system MTF and PTF alter the image as it passes through the circuitry.
For linear-shift-invariant systems, the PTF simply indicates a spatial or temporal
shift with respect to an arbitrarily selected origin. An image where the MTF is
drastically altered is still recognizable, whereas large PTF nonlinearities can
destroy recognizability. Modest PTF nonlinearity may not be noticed visually
except those applications where target geometric properties must be preserved
(i.e., mapping or photograrnmetry). Generally, PTF nonlinearity increases as the
spatial frequency increases. Because the MTF is small at high frequencies, the
nonlinear-phase-shift effect is diminished.
VMAX - VMIN AC
MODULATION - M - - (9-23)
VMAX + VMIN DC '
where VMAX and VMIN are the maximum and rrummum signal levels,
respectively. The modulation transfer function is the output modulation produced
by the system divided by the input modulation as a function of frequency:
(9-24)
The concept is presented in Figure 9-4. Three input and output signals are
plotted in Figures 9-4a and 9-4b, respectively, and the resultant MTF is shown
in Figure 9-4c. As a ratio, the MTF is a relative measure with values ranging
from zero to one.
260 CCD ARRAYS, CAMERAS, and DISPLAYS
The MTF is a measure of how well a system will faithfully reproduce the
input. The highest frequency that can be faithfully reproduced is the system
cutoff frequency. If the input frequency is above the cutoff, the output will be
proportional to the signal's average value with no modulation.
iu. 0.5
f,
SPATIAL FREQUENCY
(c)
Figure 9-4. Modulation transfer function. (a) Input signal for three
different spatial frequencies, (b) output for the three frequencies, and (c)
the MTF is the ratio of output-to-input modulation.
LINEAR SYSTEM THEORY 261
If the system MTF is known, the image for any arbitrary object can be
computed. First, the object is dissected into its constituent spatial frequencies
(i.e., the Fourier transform of the object is obtained). Next, each of these
frequencies is multiplied by the system MTF at that frequency. Then, the
inverse Fourier transform provides the image.
o(x) ~ -1 + -2 ~ 1 sm
L.-- . (- nx)
21t - n ~ 1, 3, 5, ···, (9-25)
2 1t n do
where do is the period of the square wave. The fundamental frequency U 0 is lld0 •
Note that the peak-to-peak amplitude of the fundamental is 417r times the square
wave amplitude. In the frequency domain, the square wave provides discrete
spatial frequencies of Un = 1/d0 , 3/d0 , 5/d0 , ••• , with amplitudes 217r, 2/37r,
2/57r, ··· , respectively.
Let a circular optical system image the square wave. The MTF for a
circular, clear aperture, diffraction-limited lens is
- 0 elsewhere .
The optical cutoff in image space is uic = D0 /(A. fl), where Do is the aperture
diameter, 'A is the average wavelength, and fl is the focal length.
262 CCD ARRAYS, CAMERAS, and DISPLAYS
Taking the inverse Fourier transform provides the resultant image. Equivalently,
(9-28)
+ MTF0PTICS(3u
0
)[~
31t
sin(61t u 0
x)] + ....
If llo is greater than U;c/3, only the fundamental of the square wave will be
faithfully reproduced by the optical system. Here, the square wave will appear
as a sine wave. Note that the optical MTF will reduce the image amplitude. As
U0 decreases, the image will look more like a square wave (Figure 9-5). Because
the optics does not pass any frequencies above U;c. the resultant wave form is
a truncation of the original series modified by the optical MTF. This results in
some slight ringing. This ringing is a residual effect of the Gibbs phenomenon.
Note that with imaging systems, the intensity is always positive. As the input
frequency approaches U;c (Figure 9-5d), the modulation is barely present in the
image. However, the average scene intensity approaches 0.5. That is, the scene
detail can no longer be perceived at or above u,c· The scene did not disappear.
r l r\
~ ~
(i) (i)
z z
w
1-
~
J
w
1-
~ J
(a) (b)
~ ll
~
(i)
z
w
1-
~
(c) (d)
Figure 9-5. A square wave modified by a circular aperture as a function
of x/d0 • As the square-wave fundamental frequency increases, the edges
become rounded. It will appear as a sinusoid when U0 ~ U;c/3. (a) U0 =
U;c/40, (b) U0 = U;c/10, (c) U0 = U;c/3, and (d) U0 = 9u;cl10.
LINEAR SYSTEM THEORY 263
CTF(u) ~ ~ I MTF(u)
MTF(3u 0 ) + MTF;5u) _ .... I (9_29)
3
or
t
MTF CTF
T
.i. 1
Figure 9-6. AC components of the CTF and MTF. The CTF is usually
equal to or greater than the MTF.
264 CCD ARRAYS, CAMERAS, and DISPLAYS
CTF(5u)
5
(9-31)
CTF(7u 0 ) CTF(llu 0 )
+ ------"- + + irregular terms I
7 11
or
MTF(u) (9-32)
The value r is the number of different prime factors in k. The value m is the
total number of primes into which k can be factored. The value Bk = 0 for r <
m.
For low spatial frequencies, the output nearly replicates the input (Figure 9-
7a). For frequencies near 1/3 cutoff, the output starts to appear sinusoidal (Figure
9-7b). Above 1/3 cutoff (Figure 9-7c), the MTF amplitude is 11'14 times the CTF
amplitude.
LINEAR SYSTEM THEORY 265
Figure 9-7. Input and output wave forms for a linear system. (a) Very
low frequency signals are faithfully reproduced, (b) mid-spatial
frequencies tend to look like sinusoids, and (c) input square waves with
spatial frequencies above 1/3 uMTF=o appear as sinusoidal outputs.
The relationship between the CTF and MTF was developed for analog
optical systems and is appropriate for wet-film photographic cameras and image
intensifiers. It can also be applied in the scan direction for analog scanning
devices (vidicons and displays).
For sampled data systems, where aliasing can occur, the relationship should
be used cautiously. CTF measurements are appropriate tests for system
performance verification. Conversion to the MTF that can have significant
mathematical complexity when aliasing is present.
Sampling is present in all solid state imaging systems due to the discrete
location of the detectors. Phasing effects between the detector and the location
of the target introduce problems at nearly all spatial frequencies. The interaction
between the bar frequency and the system sampling frequency produces sum and
difference frequencies that produce variations in amplitude and bar width (see
Section 8.3.3., Image Distortion, page 245).
266 CCD ARRAYS, CAMERAS, and DISPLAYS
The CTF is well behaved (in the sense that the bar width remains constant)
when the bar spatial frequency is proportional to the system Nyquist frequency,
u0 - (9-43)
k
where k is an integer. These spatial frequencies provide the in-phase CTF. The
CTF can be used to assess system perform<mce at the spatial frequencies
selected. It is usually specified only at the Nyquist frequency (k = 1).
9.6. REFERENCES
1. J. W. Coltman, "The Specification of Imaging Properties by Response to a Sine Wave Input,"
Journal of the Optical Society of America, Vol. 44(6), pp. 468-471 (1954).
2. I. Limansky, "A New Resolution Chart for Imaging Systems," The Electronic Engineer, Vol.
27(6), pp. 50-55 (1968).
3. G. C. Holst, Testing and Evaluation of Infrared Imaging Systems, pp. 238-246, JCD
Publishing, Winter Park, FL (1993).
10
SYSTEM MTF
The equations describing array (Chapter 4) and camera (Chapter 6)
performance assume that the image is very large and it illuminates many
detector elements. That is, the object contained only very low spatial frequencies
and that the system MTF is essentially unity over these spatial frequencies.
The optics and detector MTFs modify the signal spatial frequencies. These
frequencies may also be affected by electronic filter MTFs although these
electronic filter functions are typically designed to pass the entire image
frequency spectrum without attenuation.
But noise originates at the detector and within the electronic circuitry. It is
modified by the electronic MTFs only. Detector array noise is fixed but analog
amplifier noise increases as the electronic bandwidth increases. The camera SNR
is maximized when the electronic bandwidth passes the signal with minimal
degradation. Wider bandwidth electronics will increase the noise but not the
signal and thereby reduce the SNR.
(10-1)
where (Ilsys) includes shot, reset, amplifier, quantization, and fixed pattern noise
sources (see Section 4.2., Noise, page 123). The total amplifier noise depends
upon the electronic bandwidth and that depends on the electronic frequency
response associated with the readout signal chain.
A solid state camera consists of optics, an array of pixels, and a signal chain
with electronic frequency response characteristics. The complete imaging system
includes the display:
MTFSYS - MTFOP17CSMTFDETECWRMTFSIGNAL-CHAINMTFDISPLAY 0
<10-2)
267
268 CCD ARRAYS, CAMERAS, and DISPLAYS
MTFHvs is the human visual system MTF which includes both the eye optical
MTF and neural processing. When MTFPERcErvEo is coupled to system noise,
system responsivity, and the eye integration capabilities, it is possible to predict
the minimum resolvable contrast (described in Chapter 12, Minimum Resolvable
Contrast). MRC is used by the military to predict target detection and
recognition ranges for a given probability of success.
Target
D~allb ~---
1 I
I___ JI
1... dr ...1
1. . .~--
. 2 dr ------..1
Because solid state arrays are undersampled, the highest frequency that can
be reproduced is the array Nyquist frequency, uiN· As a result, it is common
practice to evaluate Equation 10-1 at 11o = uiN· This dictates the smallest target
feature that can be faithfully reproduced. The value uiN should be added to all
MTF curves (see Section 8.3.2., Detector Array Nyquist Frequency, page 242).
The MTFs presented in this chapter apply to all staring arrays. Only the
optics and detector array are sensitive to the scene spectral and spatial content.
The remaining subsystem MTF components only modify the electrical signals.
These signal chain MTFs apply to all imaging systems independent of the system
spectral response.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
There are four different locations where spatial frequency domain analysis
is appropriate. They are object space (before the camera optics), image space
(after the camera optics), display, and observer (at the eye) spatial frequencies.
Simple equations relate the spatial frequencies in all these domains.
Figure 10-2 illustrates the spatial frequency associated with a bar target. Bar
patterns are the most common test targets and are characterized by their
fundamental spatial frequency. Using the small angle approximation, the angle
subtended by one cycle (one bar and one space) is d,/R~> where d0 is the spatial
extent of one cycle and R 1 is the distance from the imaging system entrance
aperture to the target. When using a collimator to project the targets at
apparently long ranges, the collimator focal length replaces R 1• Targets placed
in the collimator's focal plane can be described in object space.
Image
cycles (10-4)
mrad
A similar equation exists for the vertical object-space spatial frequency. The
object-space domain is used by the military for describing system performance.
where ui is the inverse of one cycle in the focal plane of the lens system.
Although used interchangeably, line-pairs suggest square wave targets and cycles
suggest sinusoidal targets. To maintain dimensionality, if U00 is measured in
cycles/mrad then the focal length must be measured in meters to obtain
cycles/mm.
(NH- l)dCCH + dH
fe - ui Hz. (10-7)
tH-UNE
SYSTEM MTF 271
While digital filters operate on a data array, the effective sampling rate is related
to the timing of the readout clock:
1
f.s- - - Hz. (10-8)
e TCLOCK
After the digital-to-analog converter, the serial stream data rate and therefore
the filter design is linked to the video standard. The active line time, tvmEo-LINE•
(Table 10-1) and array size provide the link between image space and video
frequencies:
"v _
J,
HFOV·jl
U; Hz . (10-9)
tVIDEO-UNE
Video standards were discussed in Section 5.2.1., Video Timing (page 151). For
staring arrays,
(10-10)
Table 10-1
STANDARD VIDEO TIMING
TOTAL MINIMUM ACTIVE
FORMAT LINE TIME LINE TIME (JLS)
(JLS) (VIDEO-LINE
The field-of-view and monitor size link image spatial frequency to the
horizontal and vertical display spatial frequencies:
HFOV ·jl cycles
ud = u. (10-11)
WMoNnvR I mm
and
Both the monitor width, WMoNITOR• and height, HMoNITOR• are usually measured
in millimeters. When the monitor aspect ratio is the same as the FOV ratio (the
usual case),
HFOV VFOV
(10-13)
1 cycles
ueye
deg (10-14)
2 tan-1 (_!_ _1 )
2 Dud
Here, the arc tangent is expressed in degrees. Because image detail is important,
the small angle approximation provides
!.·ycles (10-15)
ueye = 0.01745Dud
deg
Wmonitor
H
Monitor
Table 10-2
TRANSFORM PAIR UNITS
SUBSYSTEM
SIGNAL VARIABLES FREQUENCY VARIABLES
RESPONSE
The optical anti-aliasing filter, which is used to blur the image on a single
chip color filter array, is also part of the lens system. Its performance is best
described by the way it affects the detector MTF. Therefore, its description is
included in the detector section.
I
I U. u.
MTForncs<u;) - - cos- (-' ) - - '
n; ";c ";c f ·-1u. 2
1 ·( - ' )
U;c
- 0 elsewhere .
The image-space optics cutoff is uic = D0 /(}, fl) = 1/(FA.), where D0 is the
aperture diameter and fl is the effective focal length. The f-number, F, is equal
to fl/D 0 .
Figure 10-4 illustrates MTF0 .,.ncs as a function of u/uic· Because the cutoff
frequency is wavelength dependent, Equation 10-16 and Figure 10-4 are only
valid for noncoherent monochromatic light. The extension to polychromatic light
is lens specific. Most lens systems are color corrected (achromatized) and
therefore there is no simple way to apply this simple formula to predict the
MTF. As an approximation to the polychromatic MTF, the average wavelength
is used to calculate the cutoff frequency:
(10-17)
SYSTEM MTF 275
For example, if the system spectral response ranges from 0.4 to 0.7 p.m, then
A.AvE = 0.55JLm.
For most systems operating in the visible spectral region, the optical system
may be considered near diffraction limited and in focus. Approximations for
aberrated and defocused optical systems are in Reference 2. The MTFs for
optical systems with rectangular apertures3 or telescopes with a central
obscuration (Cassegrainian optics) are also available. 4
0.8
u. 0.6
1-
::::E 0.4
0.2
10.3. DETECTORS
where sinc(x) = sin(7rx)/(7rX) and d" and dv are the photosensitive detector sizes
in the horizontal and vertical directions, respectively. Note that arrays are often
specified by the pixel size. With 100% fill factor arrays, the pixel size is equal
to the detector size. With finite fill factor arrays, the photosensitive detector
dimensions are less than the pixel dimensions. A microlens or an anti-alias filter
may optically modify the effective size (discussed in Section 10.3.4., Optical
Anti-Alias Filter).
Although the detector MTF is valid for all spatial frequencies from - oo to
+ oo, it is typically plotted up to the first zero (called the detector cutoff) which
occurs at U;o = 1/d" (Figure 10-5). Because d" may be different than dv, the
horizontal and vertical MTF can be different.
276 CCD ARRAYS, CAMERAS, and DISPLAYS
A phase shift occurs when ui > 1/dH and v, > lldv. Figure 10-5 illustrates
the MTF without regard to the array Nyquist frequency. The analyst must add
the array Nyquist frequency to the curve (see Section 8.3.2, Detector Array
Nyquist Frequency, page 242). The MTF of virtual phase detectors is a complex
expression. 5
0.8
LL 0.6
1-
:::E 0.4
0.2
0~-------+--------+--------+------~
0 0.25 0.5 0.75
NORMALIZED SPATIAL FREQUENCY
where aAss is the spectral absorption coefficient. Figure 4-2 (page 105)
illustrated the absorption coefficient in bulk p-type silicon. The factor L 0 is the
depletion width and L(u) is the frequency-dependent component of the diffusion
length (LmFF) given by
(10-20)
The diffusion length typically ranges between 50 ~-tm and 200 J.tm. The depletion
width is approximately equal to the gate width. For short wavelengths (A < 0.6
~-tm), aAss is large and lateral diffusion is negligible. Here, MTFmFFusroN
approaches unity. For the near IR (A > 0.8 ~-tm), the diffusion MTF may
dominate the detector MTF (Figure 10-6). By symmetry, MTFmFFusroN(u) =
MTFmFFusroN(v;) and L(u) = L(v;).
1
u. 0.6 urn
1- 0.8
::E
z 0.6
0.8 urn
0
u;
:::> 0.4 1.06 urn
u.
u. 0.2
c
0
0 10 20 30 40 50
SPATIAL FREQUENCY (cy/mm)
Figure 10-6. MTFmFFusroN as a function of wavelength. The value L0 =
10 ~-tm and LmFF = 100 J.tm. As the diffusion length increases, the MTF
decreases. For comparison, a 20-~-tm detector will have a cutoff at 50
cycles/mm.
278 CCD ARRAYS, CAMERAS, and DISPLAYS
where (3 is the region over which the detector responsivity is flat (Figure 10-7).
At the boundary between pixels, the trapezoidal response is 0.5 indicating equal
probability of an electron going to either pixel.
I I I I
I....,....__D_e_te-c-to-r-1••1 ,. Detector • I,. Detector • I
I I I I
I I I I
I I I I I I
1,.. Center~l
!Center Spacing ! ~-_D~~~~~~n
l )
Silicon Dioxide
I L 5 <::
I
p:= --
1 I
100%
t
Relative
Response
0
1,..
dec
Figure 10-7. Trapezoidal approximation for long -wavelength photon
absorption. Most of the charge goes to the charge well nearest to the
photon absorption site.
SYSTEM MTF 279
(10-22)
where NTRANs is the total number of charge transfers from a detector to the
output amplifier and E is the charge transfer efficiency for each transfer. It is the
number of detectors multiplied by the number of CCD phases. When NTRANs(E-
1) is small, then MTF(U;N) = 1 - 2NTRANs(l -E).
This MTF depends upon the number of charge transfers. If the image is
located next to the readout, the number of transfers is small. But if the image
is at the extreme end of the array, the charge must be transferred across the
entire array and through the serial readout register. For an average response,
NTRANs should be one-half the maximum number of transfers. This is
mathematically equivalent to (MTFcTE)'h.
~=============================================
Example 10-1
Number of Transfers
What is the maximum number of transfers that exist in an array that contains
1000 x 1000 detector elements?
For a charge packet that is furthest from the sense node, the charge must be
transferred down one column (1000 pixels) and across 1000 pixels for a total of
2000 pixels. The number of transfers is 4000, 6000, and 8000 for two-phase,
three-phase, and four-phase devices, respectively.
=================================================~
280 CCD ARRAYS, CAMERAS, and DISPLAYS
0.999999
u. 0.6
1-
:::1! 0.4
0.2
0~--~~-----+----~~----+-----~
0 2000 4000 6000 8000 10000
NUMBER OF TRANSFERS
Figure 10-8. MTFcTE at Nyquist frequency for several values of transfer
efficiency. As the number of transfers increases, the efficiency must also
increase to ensure that the MTF is not adversely affected.
10.3.3. TDI
TDI operation was described in Section 3.3.6., Time Delay and Integration,
page 70. It is essential that the clock rate match the image velocity so that the
photogenerated charge packet is always in sync with the image. Any mismatch
will blur the image. Image rotation with respect to the array axes affects both
the horizontal (TDI direction) and vertical (readout direction) MTFs
simultaneously. For simplicity they are treated as approximately separable -
MTF(ui,vi) :::::: MTF(u)MTF(vi).
If the TDI direction is horizontal, velocity errors degrade'~ the detector MTF
by
where NTm is the number of TDI elements. The value dERROR is the difference
between the expected image location and the actual location at the first pixel
(Figure 10-9). After Nm1 elements, the image is displaced NmrdERROR from the
desired location. MTFm1 is simply the MTF of an averaging filter for Nm1
samples displaced dERROR one relative to the next. As an averaging filter, the
effective sampling rate is 1/dERROR·
SYSTEM MTF 281
.IPGJGJD
I
----..1 I
I
I
I
I
•
1...... I I I
dERROR I I
_.I ~ I I
I
2dERROR • I
----..1I I
1.._. I
3dERROR •
I
1... ...
4dERROR
where .:lV is the relative velocity error between the image motion and the charge
packet "motion" or "velocity." The charge packet "velocity" is dccH/t1NT where
t1NT is the integration time for each pixel.
Figure 10-10 portrays the MTF for two different relative velocity errors for
32 and 96 TDI stages. As the number of TDI stages increases, dERRoR must
decrease to maintain the MTF. Equivalently, the accuracy of knowing the target
velocity must increase as Nmr increases. Because the image is constantly
moving, a slight image smear occurs 10 because the photogenerated charge is
created at discrete times (e.g., at intervals of lcwcK). This linear motion MTF
(discussed in Section 10.5.1., Linear Motion) is often small and therefore is
usually neglected.
282 CCD ARRAYS, CAMERAS, and DISPLAYS
0.005
.... "' .. "' . . . "' . -. "' . "' ...... .. ..
. .. . "' .......... 0.01
0.9 ...
0.005
u..
1- 0.8
::::!
0.7
0.01 ·. "-,
0.6 - 1 - - - - + - - - - - - - - - - + - - - - - 1
0 0.1 0.2 0.3 0.4 0.5
NORMALIZED SPATIAL FREQUENCY
(10-25)
where MTFm1(vi) is simply the MTF of an averaging filter for Nm1 samples
displaced dccHtan(8) one relative to the next. Figure 10-11 illustrates the vertical
MTF degradation due to image rotation for 32 and 96 stages. The spatial
frequency is normalized to dccHVi where dccH is the horizontal detector pitch.
TDI offers the advantage that the signal-to-noise ratio increases by the
square root of the number of TDI stages. This suggests that many TDI detectors
should be used. The disadvantage is that the scan velocity must be tightly
controlled as the number of TDI stages increases.
SYSTEM MTF 283
0.5
··---- ... :·-----
... ...... ___ _.... __ . ..
0.8 ... . . .
0.5
LL 0.6
1-
::iil 0.4
0.2
0+-----~------~------~----~------~
0 0.1 0.2 0.3 0.4 0.5
NORMALIZED SPATIAL FREQUENCY
The filter design depends upon the CFA design. If detectors sensitive to like
colors are next to each other, the filter may not be required. If the detectors
alternate, then an appropriately designed filter will make the detectors appear
twice as large. Similarly, if the detectors are spaced three apart, then the filter
will make the detector appear three times greater. This larger value (the
effective detector size) is used in Equation 10-18 (page 275).
284 CCD ARRAYS, CAMERAS, and DISPLAYS
Detector Effective
Width Width
1.. ...1 1.. -----..1
Optical
Anti-Alias
Filter
I
6
{a) (b)
Figure 10-12. The birefringent filter increases the effective optical area
of the detector. (a) No filter and (b) filter. Rays that would fall between
detectors are refracted onto the detector.
Multiple crystals are used to change the effective detector size both vertically
and horizontally. Because the spacing depends upon the CF A design, there is no
unique anti-alias filter design. Clearly, if the detectors are unequally spaced, the
Nyquist frequency for each color is different. Similarly, with different effective
sizes, the MTFs are also different for each color. Complex algorithms are
required to avoid color aliasing (see Example 5-1, page 167) and to equalize the
MTFs. The anti-alias filter is essential for CFAs to reduce color moire patterns
and can be used in any camera to reduce aliasing.
1
0.8
...:IE
LL 0.6
0.4
0.2
0
0 0.25 0.5 0.75
NORMALIZED SPATIAL FREQUENCY
(a)
Nyquist frequency
1
0.8
...:IE
LL 0.6
0.4
0.2
0
0 0.25 0.5 0.75
NORMALIZED SPATIAL FREQUENCY
(b)
1 Nyquist frequency
0.8
...
LL
:iil! 0.4
0.6
0.2
For many solid state cameras, the system MTF is dominated by the optics
and detector MTFs. Here, the electronic MTFs maximize the signal. They are
approximately unity over the spatial frequencies of interest. With electronic
zoom, the effects of the display and observer MTFs can be minimized.
Therefore, it is worthwhile to treat the optics and detector MTFs as a subsystem
assembly and to examine this subsystem MTF.
Assume that the detector array has square detectors with 100% fill factor (dH =
dv = dccv = dccH =d). When dAIRY is larger than the detector size, the system
is said to be optics-limited. Here, changes in the optics MTF (i.e., Do or fl)
significantly affect MTFsYSTEM· If dAIRY < d, the system is detector-limited and
changes in the detector size affect MTFsYSTEM at uiN· Figure 10-14 illustrates
MTFoPTicsMTFoETECTOR at Nyquist frequency as a function of d/F>.. It is the ratio
of the optical cutoff to the detector cutoff: uic/uiD = d/F>..
optics detector
0.7 limited , limited detector
0.6
0.5 system
LL 0.4
1-
:::!! 0.3
0.2
0.1
0
0 2 4 6 8 10
d/FA
Figure 10-14. MTF 0 pncsMTFoETEcroR at Nyquist frequency as a function
of d!F>.. The fill factor is 100% (d = d 11 = dv = dccH= dccv). The
vertical line indicates where the Airy disk diameter is equal to the
detector size. The best MTF that can be achieved occurs when
MTFoPTics is negligible compared to MTFoETECTOR· Equivalently,
MTFopncsMTFoETECTOR == MTFoETECfOR = 0.637.
SYSTEM MTF 287
Smaller chips make smaller cameras (we live in a world were smaller is
presumed more desirable). However, to maintain the MTF the f-number must
decrease. For visible systems (A =
0.5 p.m), the transition from the optics-
limited to detector-limited case occurs when F = d/1.22. If d = 10 p.m, then F
must be greater than 8.2 to enter the optics-limited condition (Figure 10-15).
~
< 0
0 5 10 15 20
f-number
Figure 10-15. Airy disk diameter as a function off-number when )\ =
0.5 p.m. The nominal pixel sizes for the various video formats are from
Table 3-3 (page 99).
With wet-film cameras, in most cases, more scene detail can be discerned
by enlarging the print. This occurs because the silver granules in the film are
very small compared to the Airy disk. That is, the camera system is almost
always optics-limited. Enlarging the image provides more detail. With solid state
cameras, this is not true. Electronic zoom (which is analogous to print
enlargement) provides more detail only up to the point where individual pixels
can be seen. There is no more detail available.
288 CCD ARRAYS, CAMERAS, and DISPLAYS
The output voltage (for square detectors) is given by the simple camera
formula:
(10-27)
Decreasing the focal length, increases the MTF at uiN· It also increases the
camera output (improves sensitivity). But decreasing the focal length increases
the scene spatial frequencies relative to the detector response. That is, the
system's response moves from being detector-limited to being optics-limited.
Thus, a tradeoff exists between sensitivity and image quality (as defined by the
MTF).
While parametric analyses are convenient, absolute values are lost. The
MTF at Nyquist frequency is important. However, it is even more important to
know the relationship between the scene spatial content and the actual value of
the Nyquist frequency. In object space, as the focal length decreases, the
absolute value of the Nyquist frequency also decreases. It is instructive to see
the effects of d/Fft. on computer-generated imagery. Details of the simulation are
provided in Reference 13. Figure 10-16 illustrates the original scene.
When d/E\ = 0.5, MTF(uN) = 0 and the optics have low-pass filtered the
scene. Here, no aliasing can occur. Figure 10-17a provides the MTFs and
Figure 10-17b illustrates the resultant imagery which includes a Butterworth
post-reconstruction filter where N = 10 (discussed in Section 10.6.4., Post-
Reconstruction Filter). While the low-pass filtering action of the optics removed
all aliasing, it also significantly attenuated the in-band frequency amplitudes.
1
0.8
u. 0.6
1-
::1 0.4
0.2
0
0 0.25 0.5 0.75 1
NORMALIZED SPATIAL FREQUENCY
(a)
(b)
Figure 10-17. (a) MTF0 PT1cs and MTFoETECTOR when d/FA. = 0.5 and (b)
reconstructed image including a Butterworth filter where N = 10. The
optics has significantly blurred the image and no aliasing exists. Created
by the System Image Analyzer software. 14
290 CCD ARRAYS, CAMERAS, and DISPLAYS
Figures 10-18 and 10-19 illustrate the cases where d!F>- = 2.44 and 10,
respectively. Note that the MTFs illustrated in Figures 10-18a and 10-19a do not
include the reconstruction filter whereas the imagery includes a 10th-order
Butterworth reconstruction filter.
1
0.8
LL 0.6
1-
:E 0.4
0.2
0
0 0.25 0.5 0.75 1
NORMALIZED SPATIAL FREQUENCY
(a)
(b)
Figure 10-18. (a) MTFoPTics and MTFoETE( TOR when d/FA = 2.44 and (b)
reconstructed image including a Butterworth filter where N = 10.
SYSTEM MTF 291
Careful examination of Figure 10-19b will reveal that the edges are sharper
compared to Figure 10-18b but aliasing has increased. However, the effect is not
dramatic when using a high order Butterworth reconstruction filter. At Nyquist
frequency, MTF 0 PT1csMTFoETECToR increased from 0.472 to 0.592. In terms of
imagery, whether d/FA = 10 is better than d/FA = 2.44 depends on the scene
spectral content and the reconstruction filter used. The amount of aliased signal
is scene-dependent and the amount of aliasing accepted depends on personal
preference.
1
0.8
u.. 0.6
1-
:::! 0.4
0.2
0
0 0.25 0.5 0.75 1
NORMALIZED SPATIAL FREQUENCY
(a)
(b)
Figure 10-19. (a) MTFoPTics and MTFoETECTOR when d/FA = 10 and (b)
reconstructed image including a Butterworth filter where N = 10.
292 CCD ARRAYS, CAMERAS, and DISPLAYS
10.5. MOTION
where aL is the distance the target edge moved across the image plane. It is
equal to vR..::lt where vR is the relative image velocity between the sensor and the
target. For scientific and machine vision applications, the exposure time is equal
to the detector integration time. But observer's eye blends many frames of data
and the eye integration time should be used. Although the exact eye integration
time is debatable, the values most often cited are between 0.1 and 0.2s.
Figure 10-20 illustrates the OTF due to linear motion as a fraction of the
detector size, d. Recall that the MTF is the magnitude of the OTF. Linear
motion only affects the MTF in the direction of the motion. As a rule-of-thumb,
when the linear motion causes an image shift less than about 20% of the detector
size, it has minimal effect on system performance.
1
0.8
0.6
LL 0.4
1-
0 0.2
0
-0.2
-0.4
0 0.2 0.4 0.6 0.8
NORMALIZED SPATIAL FREQUENCY
With high frequency motion, it is assumed that the image has moved often
during the integration time so that the central limit theorem is valid. The central
limit theorem says that many random movemenl s can be described by a Gaussian
distribution. The Gaussian MTF is
(10-30)
where uRis the rms random displacement in millimeter. Figure 10-22 illustrates
the random motion MTF. As a rule-of-thumb, when the rms value of the
random motion is less than about 10% of d (detector size), system performance
is not significantly affected. Jitter is considered equal in all directions and is
included in both the horizontal and vertical system MTFs. High frequency
motion blurs imagery (Figure 10-23).
lab=O
1
0.1
0.8
Ll. 0.6
1-
:IE 0.4
0.2
0
0 0.2 0.4 0.6 0.8
NORMALIZED SPATIAL FREQUENCY
(a) (b)
Figure 10-23. Random motion blurs imagery. (a) Original image, (b)
two-dimensional random motion. Generated 14 by SIA.
Ideally, H(fe) is a filter with MTF of unity up to the array Nyquist frequency
and then drops to zero. This filter maximizes the SNR by attenuating out-of-
band amplifier noise and passing the signal without attenuation. The ideal filter
is unrealizable but can be approximated by Nh-order Butterworth filters:
1
H LOWPASS(je ) - ---;:::========
2N (10-31)
1 f.
+ [ f.3dB ]
where fe3dB is the frequency at which the power is one-half or the amplitude is
0. 707. As N ~ co, HwwrAss(fe) approaches the ideal filter response with the
cutoff frequency of feJdB (Figure 10-24).
296 CCD ARRAYS, CAMERAS, and DISPLAYS
0.8
u. 0.6
1-
:E 0.4
0.2
Figure 10-24. Butterworth filters for N = I, 5, 10, and 50. The value
fe3dB = 0.5.
The sampling process replicates the frequency spectrum at nf.s (see Section
8.1., Sampling Theorem, page 234). Similarly, digital filter response is
symmetrical about the Nyquist frequency and repeats at multiples of the
sampling frequency. The highest frequency of interest is the Nyquist frequency.
Digital filters process data that reside in a memory. The units assigned to the
filter match the units assigned to the data arrays. With one-to-one mapping of
pixels to datels, the filter sampling frequency is the same as the array sampling
frequency. With this mapping, each filter coefficient processes one pixel value.
Digital filters can be two-dimensional.
SYSTEM MTF 297
There are two general classes of digital filters 16 : infinite impulse response
(IIR) and finite impulse response (FIR). Both have advantages and
disadvantages. The FIR has a linear phase shift whereas the IIR does not. IIR
filters tend to have excellent amplitude response whereas FIR filters tend to have
more ripple. FIR filters are typically symmetrical in that the weightings are
symmetrical about the center sample. They are also the easiest to implement in
hardware or software.
Figure 10-25 illustrates two FIR filters. The digital filter design software
provides the coefficients, Ai. The central data point is replaced by the digital
filter coefficients as they operate on the neighboring data points. The filter is
then moved one data point and the process is repeated until the entire data set
has been operated on. Edge effects exist with any digital filter. The filter
illustrated in Figure 10-25a requires seven inputs before a valid output can be
achieved. At the very beginning of the data set, there are insufficient data points
to have a valid output at data point 1, 2, or 3. The user must be aware of edge
effects at both the beginning and the end of his data record. In effect, this states
that edges cannot be filtered.
The following MTF equations are only valid where there are no edge
effects. For FIR filters where the multiplicative factors (weightings) are
symmetrical about the center, the filter is mathematically represented by a cosine
series (sometimes called a cosine filter). With its frequency response transposed
to electrical frequency, an FIR filter with an odd number of samples (also called
taps) provides
N-1
2 (10-32)
L
k-0
(10-33)
298 CCD ARRAYS, CAMERAS, and DISPLAYS
i l i l t i t Multiplier
7~!
"'!'
'y
Output
(a) Odd Number
D D D D Pixel Values
"'\ 1/
C? Output
(b) Even Number
Figure 10-25. Symmetrical digital filters. (a) 7-tap (odd number) filter
and (b) 4-tap (even number) filter.
The sum of the coefficients should equal unity so that the HoFILTER(fe = 0) is one:
(10-34)
Although the above equations provide the filter response in closed form, the
response of a real filter is limited by the ability to implement the coefficients.
The smallest increment is the LSB.
SYSTEM MTF 299
With an averaging filter, 9 all the multipliers shown in Figure 10-25 are
equal. When NAvE datels are averaged, the equivalent MTF is
(10-35)
10.6.3. SAMPLE-AND-HOLD
After the digital-to-analog converter (DAC), the analog signal changes only
at discrete times. The sample-and-hold circuitry within the DAC extends the data
into a continuous analog signal. The MTF of a zero-order sample-and-hold when
transposed into image space is:
Nyquist frequency
0.8
LL 0.6
1-
:::E 0.4
0.2
0~-------+--------+--------+------~
0 0.25 0.5 0.75
After the digital-to-analog conversion, the image still appears blocky due to
the discrete nature of the conversion. The ideal post-reconstruction filter
removes all the higher order frequencies (Figure 10-27) such that only the
original smooth signal remains. The output is delayed but this is not noticeable
on most imagery.
input
Ideally, H(fv) is a filter with MTF of unity up to the array Nyquist frequency
and then drops to zero. The ideal filter is unrealizable but can be approximated
by N1h-order Butterworth filters:
2 (10-38)
1+ f ]
_v_ N
( fv3dB
where fv3dB is the frequency at which the power is one-half or the amplitude is
0.707. As N-+ oo, HwwPAss(fv) approaches the ideal filter response with the
cutoff frequency of fv3dB. Real filters will have some roll-off and fvJdB should be
sufficiently large so that it does not affect the in-band MTF. The post-
reconstruction filter can be of the same functional form as the anti-alias filter
(see Section 10.6.1., page 295) but serves a different purpose. It removes the
spurious frequencies created by the sampling process.
Example 10-2
POST-RECONSTRUCTION FILTER
For EIA 170, tuNE is 52.09 J.tS. The Nyquist frequency in the video domain
is
~ - ..!. 640
= 6.14 Mhz (10-39)
VN 2 52.09 X 10-6
The bandwidth necessary to transmit the detector data must exceed 6.14 MHz
to avoid roll-off effects. This is a greater bandwidth than allowed by the EIA
170 standard. However, the camera output only has to conform to EIA 170
timing so that the information can be displayed on a conventional monitor.
===============================================~
10.6.5. BOOST
The boost amplifier can either be an analog or a digital circuit with peaking
that compensates for any specified MTF roll-off. The MTF of a boost filter, by
definition, exceeds one over a limited spatial frequency range. When used with
all the other subsystem MTFs, the resultant MTFsvSTEM is typically less than one
for all spatial frequencies. Excessive boost can cause ringing at sharp edges. The
boost algorithm should enhance signal amplitudes with frequencies below the
Nyquist because reconstruction usually limits the response to the Nyquist
frequency.
SYSTEM MTF 303
A variety of boost circuits are available. 17 One example is the tuned circuit
with response
1
(10-40)
where Q is the quality factor and equal to boost amplitude when fv = fsoosT·
Although illustrated in the video domain, boost can be placed in any part of
the circuit where a serial stream of data exists. By appropriate selection of
coefficients, digital filters can also provide boost. MTFsoosT can also
approximate the inverse of MTFSYSTEM (without boost). The combination of
MTFsoosT and MTFsYsTEM can provide unity MTF over spatial frequencies of
interest. This suggests that the reproduced image will precisely match the scene
in every spatial detail.
The display MTF is a composite that includes both the internal amplifier and
the CRT responses. Implicit in the MTF is the conversion from input voltage to
output display brightness. Although not explicitly stated, the equation implies
radial symmetry and the MTFs are the same in both the vertical and horizontal
directions. That is, the MTF is considered separable and MTFcRT(~,vd) =
MTFcRT(ud)MTFcRT(vd) and MTFcRT(vct) = MTFcRT(Uct). While this may not be
precisely true, the assumption is adequate for most calculations.
The amplifier bandwidth is not a number that is often useful to the user. We
must assume that the manufacturer selected an adequate bandwidth to provide
the display resolution. For example, a 1600 x 1200-disel display must have a
wider bandwidth than a 1280 X 1024-disel display by a factor of approximately
1600/1280= 1.25. MTFcRT includes both the electronics MTF and spot diameter.
For an infinitesimal input, the spot diameter is affected by the video bandwidth.
That is, the minimum spot diameter is the equivalent point spread function.
304 CCD ARRAYS, CAMERAS, and DISPLAYS
where <TsroT is the standard deviation of the Gaussian beam profile and S is the
spot diameter at FWHM. These parameters must have units of millimeters
because the frequency has units of cycles/mm. The spot diameter is
S - J81n(2) a SPOT - 2.35 a SPOT. (10-42)
If the spot size is not specified, then usPoT can be estimated from the TV
limiting resolution. With the TV limiting resolution test, an observer views a
wedge pattern and selects that spatial frequency at which the converging bar
pattern can no longer be seen. The display industry has standardized the bar
spacing to 2.35usroT·
Figure 10-28 illustrates MTFoETEcroR and MTF 01sPLAY· The display spot size
is independent of the detector size and the display resolution is independent of
the number of detector elements. Matching the display resolution (number of
pixels) to the number of detector elements does not uniquely specify the display
MTF as suggested by Equation 10-44. For example, suppose an array contained
1000 elements horizontally. A monitor that can resolve 1000 pixels may not
provide good imagery if the resolution/addressability ratio (RAR) is large.
Equation 10-44 is only valid when RAR :::::: 1. The RAR was described in Section
7.6., Addressability (page 222).
Many consumer CCD cameras match the number of vertical detectors to the
number of active video lines. With consumer televisions, NTV :::::: 0. 7Nv. Thus,
the video standard uniquely defines the CRT MTF. For 100% fill factor arrays,
2
-2" 2(dccv
- - u .) (10-45)
.. e 1.645 I
SYSTEM MTF 305
0.8
u.. 0.6
1-
::::!! 0.4
0.2
0
0 0.2 0.4 0.6 0.8
NORMALIZED SPATIAL FREQUENCY
~=============================================
Example 10-3
DISPLAY MTF
Equations 10-43 or 10-44 are used to calculate image MTF when the input
is a sinusoid. The Nyquist frequency is used most often when specifying system
response. For EIA 170, Nv = 485. Inserting dvui = 0.5 into Equation 10-45
provides
2 2( 0.5 485 )2 (10-46)
M 'T'F ( ) = -
.L CRT' UiN e
1t 2.35 N7V
For a 100% fill factor array, the detector MTF is 0.63 at uiN· Figure 10-29
illustrates the detector and display MTF at Nyquist frequency. The display
resolution must increase to 1400 TVLIPH for display MTF to be 0.90 at the
Nyquist frequency. This suggests that most imaging systems are display
resolution limited. This result is based upon the assumption that the beam profile
is Gaussian. A different shaped beam (i.e., flat-topped), will provide a different
relationship between TV limiting resolution and MTF.
306 CCD ARRAYS, CAMERAS, and DISPLAYS
0.7 detector
0.6
0.5
LL 0.4
1-
:2 0.3
0.2
0.1
0
0 200 400 600 800 1000 1200
============================================~
The HVS MTF should only be used to understand the relationship between the
imaging system MTF and the eye response. It is not used to describe sensor
performance. When MTFrERCEivEo is combined with system noise, system
responsivity, and the eye integration capabilities, it is possible to predict the
minimum resolvable contrast (described in Chapter 12, Minimum Resolvable
Contrast). The eye's contrast sensitivity is also included in all image quality
metrics (discussed in Chapter 11, Image Quality).
where LMAx and LMIN are the maximum and minimum luminances, respectively.
The eye's detection capability depends upon the visual angle subtended by the
target size at the observer. As shown in Figure 10-30, in the absence of noise,
the eye's modulation threshold (minimum perceivable modulation) is
characteristically J -shaped. It is labeled as M1 in the visual psychophysical
literature. The eye is most sensitive to spatial frequencies 20 that range between
3 and 5 cycles/deg at typical ambient lighting levels.
zc
0...1 0.1
i=O
:5i}i
::JW
ca: 0.01
OJ:
:::tl-
0.001
0.1 1 10 100
SPATIAL FREQUENCY (cy/deg)
Schulze22 recommended
MTFHVS(Jeye) - 2.71 (e -O.ll38f,,. - e-0.32Sf.,.) • (10-50)
MTFHVS(f.eye ) = I'
sin2 2:. /,/.ye
2 1 PEAK
]1/3] •
(10-52)
0.8
u. 0.6
1-
::!!! 0.4
0.2
0
0 5 10 15 20
SPATIAL FREQUENCY (cy/deg)
Figure 10-31. Various mathematical equations that approximate the HVS
MTF.
Insufficient data exist to say with certainty which eye model is best.
Because of this uncertainty, all analyses, no matter which eye model is used,
must only be used for comparative performance purposes.
Two scenarios are possible: (1) the observer is allowed to move his head and
(2) the head is fixed in space. Because the eye's detection capability depends
upon the angular subtense of the target, head movement may provide different
results than if the head is fixed. In laboratory testing, the distance to the monitor
is not usually specified nor limited in any way. To maximize detection capability
(stay on the minimum of the contrast threshold curve), an observer
subconsciously moves toward the monitor to perceive small targets and further
away to see larger targets.
equal detection capability for all spatial frequencies such that the eye's contrast
sensitivity approaches a constant. This results in a nearly constant MTF that is
called the "non-limiting eye MTF" or MTFHvs = 1. Equation 10-53 is used in
those situations where the head is fixed, such as when a pilot is strapped into a
cockpit seat.
MTF theory will provide an optimistic result of fiber optic coupled ICCDs. It
does not include the CCD mismatch or the chicken wire effects. The MTF
associated with lens coupling is probably more accurate. The MTF of an ICCD
is MTF1NTENSIFIER muitiplied by the appropriate MTFs presented in this chapter.
0.8
LL 0.6
1-
::::E 0.4
0.2
0
0 5 10 15 20 25 30 35
SPATIAL FREQUENCY (cy/mm)
~ cos(_5_ e] ,
uiN
(10-56)
where 8 is the phase angle between the target and the sampling lattice. For
example, when u; = uiN• the MTF is a maximum when 8 = 0 (in-phase) and a
zero when e = 7r/2 (out-of-phase).
MTFMEvw/u) ~cos -
1t -
U; .
(
2 U;s
l (10-57)
(10-58)
Figure 10-33 illustrates the difference between the two equations. Because these
are approximations, they may be considered roughly equal over the range of
interest (zero to the Nyquist frequency). MTFrHASE can be applied to both the
vertical and horizontal MTFs.
312 CCD ARRAYS, CAMERAS, and DISPLAYS
0.8
LL 0.6
......
:E 0.4
0.2
0
0 0.25 0.5 0.75 1
Frequencies above uiN are aliased down to lower frequencies. Signal aliasing
is ignored in most analyses and has been ignored here. A complete end-to-end
analysis should include both aliased noise and aliased signal36 • These aliased
components lirnit37 the extent to which a sampled image can be sharpened.
Image reconstruction may include data removal (decimation) or data expansion
(interpolation). The effective sampling frequency and aliasing can appear to
change using different image processing algorithms. 3840
10.11. REFERENCES
1. See, for example, R. J. Farrell and J. M. Booth, Design Handbook for Imagery Interpretation
Equipment, Reprinted with corrections, Report D 180- 19063-1, Boeing Aerospace Company,
Seattle, WA, February 1984.
2. G. C. Holst, Testing and Evaluation of Infrared Imaging Systems, pp. 106-108, JCD
Publishing, Winter Park, Fl (1993).
3. G. C. Holst, Testing and Evaluation of Infrared Imaging Systems, pp. 195-200, JCD
Publishing, Winter Park, Fl (1993).
4. G. C. Holst, Testing and Evaluation of Infrared Imaging Systems, pp. 104-105, JCD
Publishing, Winter Park, Fl (1993).
SYSTEM MTF 313
27. FUR92 Thermal Imaging Systems Performance Model, Analyst's Reference Guide, Document
RG5008993, Ft. Belvoir, VA, January 1993.
28. G. M. Williams, Jr., "A High Performance LLLTV CCD Camera for Nighttime Pilotage," in
Electron Tubes and Image Intensifiers, C. B. Johnson and B. N. Laprade, eds., SPIE
Proceedings Vol. 1655, pp. 14-32 (1992).
29. I. P. Csorba, Image Tubes, pp. 79-103, Howard W. Sams, Indianapolis, IN (1985).
30. S. K. Park and R. A. Schowengerdt, "Image Sampling, Reconstruction and the Effect of
Sample-scene Phasing," Applied Optics, Vol. 21(17). pp. 3142-3151 (1982).
31. J. C. Feltz and M. A. Karim, "Modulation Transfer Function of Charge-coupled Devices,"
Applied Optics, Vol. 29(5), pp. 717-722 (1990).
32. J. C. Feltz, "Development of the Modulation Transfer Function and Contrast Transfer Function
for Discrete Systems, Particularly Charge-coupled Devices," Optical Engineering, Vol. 29(8),
pp. 893-904 (1990).
34. L. de Luca and G. Cardone, "Modulation Transfer Function Cascade Model for a Sampled IR
Imaging System," Applied Optics, Vol. 30(13), pp. 1659-1664 (1991).
35. F. A. Rosell, "Effects of Image Sampling," in The Fundamentals of Thermal Imaging Systems,
F. Rosell and G. Harvey, eds., p. 217, NRL Report 8311, Naval Research Laboratory,
Washington D.C. (1979).
36. S. K. Park, "Image Gathering, Interpolation, and Restoration: A Fidelity Analysis," in Visual
Information Processing, F. 0. Huck and R. D. Juday, eds., SPIE Proceedings Vol. 1705, pp.
134-144 (1992).
37. S. K. Park and R. Hazra, "Image Restoration Versus Aliased Noise Enhancement," in Visual
Information Processing Ill, F. 0. Huck and R. D. Juday, eds., SPIE Proceedings Vol. 2239,
pp. 52-62 (1994).
38. I. Ajewole, "A Comparative Analysis of Digital Filters for Image Decimation and
Interpolation," in Applications of Digital Image Processing VIII, A. G. Tescher, ed., SPIE
Proceedings Vol. 575, pp. 2-12 (1985).
38. W. F. Schrieber and D. E. Troxel, "Transforma[ion Between Continuous and Discrete
Representations of Images: A Perceptual Approach,·· TEEE Transactions on Pattern Analysis
and Machine Intelligence, Vol. PAMA-7(2), pp. 178-186 (1985).
39. R. W. Schafer and L. R. Rabiner, "A Digital Signal Processing Approach to Interpolation,"
Proceedings of the IEEE, Vol. 61(8), pp. 692-702 (1973).
40. A. H. Lettington and Q. H. Hong, "Interpolator for Infrared Images," Optical Engineering,
Vol. 33(3), pp. 725-729 (1994).
11
IMAGE QUALITY
Our perception of good image quality is based upon the real-world
experiences of seeing all colors, all intensities, and textures. An imaging system
has a limited field-of-view, limited temporal and spatial resolutions, and presents
a two-dimensional view of a three-dimensional world. In the real-world our eyes
scan the entire scene. Not all of the available information is captured by an
imaging system. Furthermore, an imaging system introduces noise and the loss
of image quality due to noise can only be estimated. Cameras sensitive to
wavelengths less than 0.4 JI-m and greater than 0.7 JI-m provide imagery that we
cannot directly perceive. The quality of the ultraviolet or infrared imagery can
only be estimated because we do not know how it really appears.
Many formulas exist for predicting image quality. Each is appropriate under
a particular set of viewing conditions. These expressions are typically obtained
from empirical data in which multiple observers view many images with a
known amount of degradation. The observers rank the imagery from worst to
best and then an equation is derived which relates the ranking scale to the
amount of degradation.
If the only metric for image quality was resolution, then we would attempt
to maximize resolution in our system design. Many tests have provided insight
into image quality metrics. In general, images with higher MTFs and less noise
are judged as having better image quality. There is no single ideal MTF shape
that provides best image quality. For example, Kusaka 1 showed that the MTF
that produced the most aesthetically pleasing images depended on the scene
content.
315
316 CCD ARRAYS, CAMERAS, and DISPLAYS
The metrics suggested by Granger and Cupery, Shade, and Barten offer· 6
additional insight on how to optimize an imaging system. Granger and Cupery
developed the Subjective Quality Factor (SQF): an empirically derived
relationship using individuals' responses when viewing many photographs. Shade
used photographs and included high quality TV images. Barten's approach is
more comprehensive in that it includes a variety of display parameters. It now
includes 7 contrast, luminance, viewing ratio, number of scan lines, and noise.
While Barten has incorporated the sampling effects of flat panel displays,?
no model includes the sampling (and associated aliasing) that takes place at the
detector. Until this aliasing is quantified, no metric will predict image quality
fqr sampled data systems. Only an end-to-end system approach can infer overall
image quality.
There are potentially two different system design requirements: (1) good
image quality and (2) performing a specific task. Sometimes these are
equivalent, other times they are not. All image quality metrics incorporate some
form of the system MTF. The underlying assumption is that the image spectrum
is limited by the system MTF. Equivalently, it is assumed that the scene
contains all spatial frequencies and that the displayed image is limited by system
MTF - a reasonable assumption for general imagery. While good image quality
is always desired, a military system is designed to detect and recognize specific
targets (discussed in Chapter 12, Minimum Resolvable Contrast). Optimized
military systems will have high MTF at the (assumed) target frequencies and
other spatial frequencies are considered less important. Computer monitors are
usually designed for alphanumeric legibility.
Often, the display is the limiting factor in terms of image quality and
resolution. No matter how good the electronic tmaging system is, if the display
resolution is poor, then the overall system resolution is poor. If the system is
optics-limited, then a high resolution display does not offer any "extra" system
resolution. A high resolution display just ensures that all the information
available is displayed. System resolution may be limited by the HVS. If the
observer is too far from the screen, not all of the image detail can be discerned.
A large number of metrics are related to image quality. Most are based on
monochrome imagery such as resolution, MTF, and minimum resolvable
contrast. CRT resolution was presented in Section 7.5., Resolution (page 218).
Color reproduction and tonal transfer issues, which are very important to color
cameras, are not covered here.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
Table 11-1
SUBSYSTEM MEASURES of RESOLUTION
SUBSYSTEM RESOLUTION METRIC
Rayleigh criterion
Optics Airy disk diameter
Blur diameter
Detector-angular-subtense
In 1tantaneous-field-of-view
Detectors Effective-instantaneous-field-of-view
Detector pitch
Pixel-angular-subtense
Electronics Bandwidth
Table 11-2
RESOLUTION MEASURES for ANALOG SYSTEMS
RESOLUTION DESCRIPTION
Ability to distinguish two
Rayleigh criterion
adjacent point sources
Diffraction-limited diameter
Airy disk
produced by a point source
Table 11-3
RESOLUTION MEASURES for SAMPLED DATA SYSTEMS
RESOLUTION DESCRIPTION
Instantaneous field-of-view Angular region over which the detector senses radiation
Analog measures of spatial resolution may be determined by: (1) the width
of a point source image, (2) the minimum detectable separation of two point
sources, (3) the spatial frequency at which the MTF drops to a defined level, or
(4) the smallest detail resolved by an observer. These measures assume that the
system output image is a replica of the input scene (a linear-shift-invariant
system).
When viewing aerial imagery that contains a test pattern (such as the U.S.
Air Force 1951 standard 3-bar target), an image analyst determines the smallest
discernible cycle on the ground. This cycle width (bar plus space) is the GRD.
It includes the system MTF and possible degradation by the atmosphere and is
320 CCD ARRAYS, CAMERAS, and DISPLAYS
The smallest elements that can be created by a digital subsystem are the
pixel, datel, and disel. These are created by the detector array, ADC, and the
display media, respectively. The resel is the smallest element that can be created
by an analog system. The optics, analog electronics, and video standard (or
transmission link) each have their own resel. System resolution is a convolution
of all -els.
II II
have been in existence so long. They were developed to describe the resolution
of specific subsystems and do not provide an end-to-end performance metric.
Nor do they include aliasing. Aliasing effects are discussed in Reference 15.
Table 11-4 provides various optical resolutions. As analog metrics, these are
various definitions of a resel. Diffraction measures include the Rayleigh criterion
and the Airy disk diameter. The Airy disk is the bright center of the diffraction
pattern produced by an ideal optical system. In the focal plane of the lens, the
Airy disk diameter is
A.
dAIRY - 2.44-.fl - 2.44A.F . (11-2)
Do
degradation requires a high resolution sensor. Both the human eye and
photographic film visually have this resolution. However, with electronic
imaging systems, the detectors are often too large to see the degradation because
the detectors are often larger than the blur diameter.
Table 11-4
OPTICAL RESOLUTION METRICS (in image space)
DEFINITION
RESOLUTION DESCRIPTION
(usual units)
Considering the Airy disk, the wavelength dependence suggests that the
resolution increases as the wavelength decreases (smaller is better). However,
this is only true when the detector size is much smaller than the Airy disk size.
For most electronic imaging systems the reverse is true. The detector is the
same size or larger than the disk diameter (see Section 10.4., Optics-Detector
Subsystem, page 286). This means that the Airy disk is not adequately sampled
and therefore the resolution afforded by the optics cannot be exploited. Systems
designed for star detection are usually optics-limited (optical resel larger than a
pixel), whereas for general imagery the systems are usually detector-limited
(pixel larger than the optical resel)
Detector arrays are often specified by the number of pixels and detector
pitch. These are not meaningful until an optical system is placed in front of the
array. Table 11-5 provides the most common resolution metrics expressed in
object-space units. The detector-angular-subtense (DAS) is often used by the
military to describe the resolution of systems when the detector is the limiting
subsystem. If the detector's element horizontal and vertical dimensions are
different, then the DAS in the two directions is different. Note that the PAS is
different than the DAS (see Section 1.5., Pixels, Datels, Disels, and Resels,
page 14). Only with 100% fill factor arrays are they the same. Spatial sampling
rates are determined by the pitch in image space or the PAS in object space.
322 CCD ARRAYS, CAMERAS, and DISPLAYS
Table 11-5
DETECTOR ELEMENT RESOLUTION MEASURES (in object space)
DEFINITION
RESOLUTION DESCRIPTION
(usual units)
The detector horizontal cutoff is lion = fl/dH = 1/DAS in object space. With
detector arrays, signal fidelity is limited to the Nyquist frequency. Here, the
effective-instantaneous-field-of-view (EIFOV) offers an alternate measure of
resolution (Figure 11-1). The value uEIFov is the spatial frequency at which the
MTF is 0.5. Calling 1/DAS the cutoff provides a misleading representation.
Recall that the detector can respond to frequencies above 1/DAS (see Figure 8-
9, page 241).
0.8 Nyquist
frequency
LL 0.6
1-
::E 0.4
0.2
0~----~------P-----~--~--P-----~
0 0.1 0.2 0.3 0.4 0.5
llt:Jmv
NORMALIZED SPATIAL FREQUENCY
For high data rate systems such as line-scanners, the electronic bandwidth
may limit system response. For electronic circuits, resolution is implied by its
bandwidth. The minimum pulse width is approximately
..
1
N-- (11-3)
MINIMUM 2 BW ,
Usually systems are designed to have high MTF, and Figure 10-14 (page
286) illustrated the MTF at Nyquist frequency. However, the value of Nyquist
frequency is a measure of resolution; 100% fill factor staring arrays have the
highest Nyquist frequency.
Other measures also exist such as the limiting resolution. This is the spatial
frequency at which the MTF is, say, 5%. Figure 11-2 illustrates two systems
that have the same limiting resolution. Which system is selected depends on the
specific application. System A is better at high spatial frequencies, and system
B is better at low spatial frequencies. Equivalently, if the scene contains mostly
low frequencies (large objects), then system B should be used. If edge detection
is important (high spatial frequencies), then system A should be considered. This
clearly indicates the difficulty encountered when using resolution exclusively as
a measure of system performance.
URES
Spatial Frequency
Figure 11-2. The MTFs of two different imaging systems with the same
limiting resolution, uREs·
IMAGE QUALITY 325
w
en
z
0
a.
en
w
0::
w
>
~w
0::
SPATIAL FREQUENCY
Figure 11-3. The spatial frequency of the intersection of the film
threshold modulation and MTFsYSTEM is a measure of resolution.
Conceptually, for spatial frequencies above uRES• the film requires more
modulation than that provided by the system. Below uREs the film can easily
discern the scene modulation. Because many measures were in support of aerial
imagery, the intersection was also called the aerial image modulation (AIM).
With the advent of electronic imaging systems, the film TM was replaced with
the HVS's threshold modulation (discussed in Section 11.7., MTFA).
The value REQ cannot be directly measured and is a mathematical construct that
expresses overall performance. As REQ decreases, the resolution "improves"
(smaller is better). As an approximation, the system resolution, REQ-sYs• may be
estimated from the component equivalent resolutions, R;. by
..; 2 (11-6)
REQ-SYS .. R1 + R22 + ... + R,.2 •
Shade's approach using the square of the MTF emphasized those spatial
frequencies at which the MTF is relatively high. It appears to be a good measure
for classical systems in which the MTF is decreasing (such as a Gaussian
distribution). The equivalent resolution approach assumes that the system is
completely analog and it ignores sampling effects. Therefore, REQ becomes a
resel. It probably should not be used (nor should any other image quality metric)
to compare systems with significantly different MTFs. Nor should it be used
with systems that can create significant aliasing.
Table 11-6
ONE-DIMENSIONAL EQUIVALENT RESOLUTIONS
SUBSYSTEM MTF REQ
2(
-cos
11
-I(- - -u, 1-( ~J)
"t ) 1.845AF
Optics "tc ";c
r
1 - - whenN- 1
1tf.3dB
N1h-order
low-pass filter
1 (f.
+ f.3dB
1
- - whenN- co
2f.3dB
CRT-based
e -2x a~POTu;
2
2{it OSPOT
display
REQ-SYS - d ( 1.~45r + 1 .
(11-7)
328 CCD ARRAYS, CAMERAS, and DISPLAYS
1-z
ffiQ
....li- d
e(::;) 1
~cl
::;)(/)
Ow
wa:
0~----~------~------~------~----~
0 2 3 4 5
d!F>.
Figure 11-4. Equivalent resolution normalized to d as a function of
d/F)\. The vertical line indicates where the Airy disk diameter is equal
to the detector size. The best resolution occurs when d/F)\ approaches
infinity. Here the optics MTF approaches unity.
Example 11-1
PIXELS and DATELS
An imaging system consists of 512 x 512 detector elements. Assume that the
fill factor is 100% and the camera's analog output is digitized at either 1024
samples/line or 2048 samples/line. What is the resolution?
The external analog-to-digital converter creates either two or four datels for
each pixel. In both cases, the pixel size remains the same. Increasing the
sampling rate does not change the system's resolution but improves the
repeatability of all measurements by reducing phasing effects that might occur
in the analog-to-digital converter. Any image processing algorithm that operates
on this higher number must consider the sensor resolution.
IMAGE QUALITY 329
Cameras with a standard video output may not have the resolution suggested
by the video standard. For example, a camera containing an array of 320 x 240
pixels may have its output formatted into EIA 170 timing. The signal may be
digitized by a frame grabber that creates 640 x 480 datels. The image size is
still 320 x 240 pixels.
===============================================~
Example 11-2
DISELS and PIXELS
Each pixel is mapped onto four datels that are then mapped one-to-one onto
disels. Here, the imaging system determines the system resolution, not the
number of datels or disels. High quality monitors only ensure that the image
quality is not degraded. Electronic zoom cannot increase resolution. But
decimation or minifying may reduce resolution. Because image processing is
performed on datels, the image analyst must be made aware of the system
resolution.
============================================~
Example 11-3
SYSTEM EQUIVALENT RESOLUTION
This CCD camera contains 480 horizontal detectors. Each detector is 15 ~tm
on 15 Jtm centers (100% fill factor). The optical aperture is 1 inch and the focal
length is 5.6 inches (F = 5.6). The analog electronics is wideband and it does
not significantly affect the resolution. Camera output is in the standard EIA 170
format with a bandwidth of 4.2 MHz and tvmEO-LINE = 52.45 JtS. The frame
grabber provides 640 horizontal samples and a standard consumer CRT-based
monitor is used.
330 CCD ARRAYS, CAMERAS, and DISPLAYS
Assuming A.AvE ""0.5/Lm, REQ for the optics is 1.845A.AvEID0 = 0.0363 mrad.
The DAS is d/fl or 0.105 mrad. Here, d/(FA.) = aD 0 /A. = 5.33. That is, the
optical/detector subsystem is detector-limited. The HFOV is
Nydy 6
480 (15 x 10 ~~-
m)
HFOV- - 50.6 mrad . ( 11 _8)
.fl 3
(5.6 in ) ( 25.4 x 1o- : )
With an active line time of 52.45/Ls, the video bandwidth in object space is
6
/,
BW _ 52.45x10- 4 _ xl06 _ 435 cycles .
tVIDEO-UNE
2 (11-9)
o-VIDEO - HFOV 50.6 X 10" 3 mrad
Assuming a sharp cutoff, REQ "" 112f3ds = 0.115 mrad. The system equivalent
resolution is
Here, the EIA 170 bandwidth significantly affects the resolution. Neglecting
phasing effects, each frame grabber datel represents approximately
(0 .159)(480)/640 = 0.119 mrad. These are overlapping resolution elements
where the center-to-center spacing is (0.105)(480)/640 = 0.0787 mrad.
11.7. MTFA
The area bounded by the system MTF and the HVS's modulation threshold
(M1) is called the modulation transfer function area (MTFA) (Figure 11-Sa). As
the MTFA increases, the perceived image quality appears to increase. The
spatial frequency at which the MTF and M 1 intersect is a measure of the overall
resolution. However, the resolution is not unique because M1 depends on the
display viewing distance. The HVS's inhibitory response (see Figure 10-31,
page 309) is omitted from the MTFA approach. Mathematically,
fo
For low noise, general imagery, the HVS's contrast sensitivity has little
effect on the MTFA. The integrand in Equation 11-13 can be replaced with just
the system MTF. Here, the high spatial frequencies are weighted too heavily
compared to the low spatial frequencies and Shade's approach appears better.
Modulation
Threshold
(a)
High Noise
Modulation
Threshold
Low Noise
Modulation
Threshold
(b)
Figure 11-5. MTFA. The area between system MTF and M1 up to the
intersection is a measure of image quality. The spatial frequency at the
intersection is a measure of limiting resolution. The area and limiting
resolution depend on the noise level within the imagery. (a) Low noise
and (b) elevated noise.
The SQF is an adequate image quality metric for low noise imagery when
the illumination is fixed at a moderate level. As with the MTFA and Shade's
equivalent pass band, the SQF is intended for general imagery only. It cannot
be used for specific applications such as the legibility of alphanumeric
characters.
Table 11-7
SUBJECTIVE QUALITY FACTOR
SQF SUBJECTIVE IMAGE QUALITY
0.92 Excellent
0.80 Good
0.75 Acceptable
0.50 Unsatisfactory
0.25 Unusable
334 CCD ARRAYS, CAMERAS, and DISPLAYS
The multiplication overcomes the theoretical objection raised with the MTFA
where the CSF was subtracted from the MTF. MTFsvs includes all the
subsystem MTFs up to and including the display. Barten modeled the contrast
sensitivity function as
where
(11-17a)
and
where feye has units of cycles/degree, LAvE is the average luminance in candelas
per square meter, and w is the angular display size as seen by the observer. The
factor l/ln(2) allows SQRI to be expressed in just-noticeable-difference (JND)
units. Both average luminance level and display size affect the CSF. This CSF
is a good approximation over five orders of magnitude of luminance levels and
for angular display sizes ranging from 0.5° to 60°.
IMAGE QUALITY 335
Contrast loss due to reflected ambient lighting effectively reduces the display
MTF. If .::lL is the additional luminance due to reflected ambient-light, then the
modulation depth is reduced by the factor:
LAVE (11-18)
T) DISPlAY -
LAVE + 11L
This factor is the same for all spatial frequencies of interest and the perceived
effect on image quality is incorporated into the SQRI by multiplying MTFSYS by
V'YIDISPLA Y •
The display parameters dictate the limits of integration. The value fMIN is that
spatial frequency that provides one-half cycle across the display and fMAx is the
maximum spatial frequency of the display and this is related to display
addressability (see Section 7.6., Addressability, page 222). For example, if the
number of active lines is NAcnvE• then
(Kellfactor) NACTIVE
(11-19)
!MAx - DFov ·
2
Noise increases the contrast sensitivity function. Because the SQRI must
decrease with noise, Barten added noise as an additional factor to the HVS
MTF:
(11-20)
336 CCD ARRAYS, CAMERAS, and DISPLAYS
The value Mn can be calculated from the noise power spectral density presented
to the eye. 5 Barten predicted the quality of imagery when both static (pattern
noise) and dynamic (temporally varying) noise are present. He considered both
one-dimensional and two-dimensional noise sources. His approach encompasses
actual sensor operation where different noise components may exist in the
horizontal and vertical directions. As such, he considered some of the three-
dimensional noise components used for military applications (discussed in
Section 12.2., Three-Dimensional Noise Model).
11.10. REFERENCES
1. H. Kusaka, "Consideration of Vision and Picture Quality - Psychological Effects Induced by
Picture Sharpness," in Hu11Uln Vision, Visual Processin!?. and Digital Display, B. E. Rogowitz,
ed., SPIE Proceedings Vol. 1077, pp. 50-55 (1989).
2. E. M. Granger and K. N. Cupery, "An Optical Merit Function (SQF) which Correlates With
Subjective Image Judgments," Photographic Science and Engineering, Vol. 16, pp. 221-230
(1972).
3. 0. H. Shade, Sr., "Image Gradation, Graininess, and Sharpness in Television and Motion
Picture Systems," published in four parts in SMPTE Journal: "Part I: Image Structure and
Transfer Characteristics," Vol. 56(2), pp. 137-171 (1951); "Part II: The Grain Structure of
Motion Pictures - An Analysis of Deviations and Fluctuations of the Sample Number," Vol.
58(2), pp. 181-222 (1952); "Part III: The Grain Structure of Television Images," Vol. 61(2),
pp. 97-164 (1953); "Part IV: Image Analysis in Photographic and Television Systems," Vol.
64(11), pp. 593-617 (1955).
4. P. G. Barten, "Evaluation of Subjective Image Quality with the Square-root Integral Method,"
Journal of the Optical Society of America. A, Vol. 17( 10), pp. 2024-2031 (1990).
5. P. G. Barten, "Evaluation of the Effect of Noise on Subjective Image Quality," in Hu11Uln
Vision, Visual Processing, and Digital Display II, J. P. Allenbach, M. H. Brill, and B. E.
Rogowitz, eds., SPIE Proceedings Vol. 1453, pp. 2-15 (1991).
6. P. G. Barten, "Physical Model for the Contrast Sensitivity of the Human Eye," in Hu11Uln
Vision, Visual Processing, and Digital Display III, B. E. Rogowitz, SPIE Proceedings Vol.
1666, pp. 57-72 (1992).
7. P. G. Barten presents short courses at numerous symposia. See, for example, "Display Image
Quality Evaluation," Application Seminar Notes, SID International Symposium held in Orlando,
FL (May 1995). Published by the Society for Information Display, Santa Ana, CA, or "MTF,
CSF, and SQRI for Image Quality," IS&T/SPIE's Symposium on Electronic Imaging: Science
and Technology, San Jose, CA (February 1995).
8. G. C. Holst, Electro-Optical 111Ulging System Performance, pp. 218-228, JCD Publishing,
Winter Park, FL (1996).
9. Air Standardization Agreement: "Minimum Ground Object Sizes for Imaging Interpretation,"
Air Standardization Co-ordinating Committee report AIR STD 101111 (31 December 1976).
10. Air Standardization Agreement: "Imagery Interpretabiltty Rating Scale," Air Standardization
Co-ordinating Committee report AIR STD 101111 (10 July 1978).
11. K. Riehl and L. Maver, "A Comparison of Two Common Aerial Reconnaissance Image Quality
Measures," in Airborne Reconnaissance XX, W. G. Fishell, A. A. Andraitis, A. C. Crane, Jr.,
and M. S. Fagan, eds., SPIE Proceedings Vol. 2829, pp. 242-254 (1996).
12. J. M. Irvine, "National Imagery Interpretability Rating Scales (NIIRS): Overview and
Methodology," in Airborne Reconnaissance XXI, W. C. Fishell, ed., SPIE Proceedings Vol.
3128, pp. 93-103 (1997).
IMAGE QUALITY 337
13. G. C. Holst, Sampling, Aliasing, and Data Fidelity, pp. 311-317, JCD Publishing, Winter
Park, FL (1998).
14. J. C. Leachtenauer, W. Malila, J. Irvine, L. Colburn, and N. Salvaggio, General Image-Quality
Equation: GIQE," Applied Optics, Vol. 36(32), pp. 8322-8328 (1997).
15. G. C. Holst, Sampling, Aliasing, and Data Fidelity, pp. 308-311, JCD Publishing, Winter
Park, FL (1998).
16. L. M. Beyer, S. H. Cobb, and L. C. Clune, "Ensquared Power for Obscured Circular Pupils
with Off-Center Imaging," Applied Optics, Vol. 30(25), pp. 3569-3574 (1991).
17. A. J. den Dekker and A. van den Bos, "Resolution: A Survey," Journal of the Optical Society
A, Vol. 14(3), pp. 547-557 (1997).
18. See, for example, W. J. Smith, Modern Optical Engineering, 2"" Edition, McGraw-Hill, New
York (1990).
19. J. M. Lloyd, Thermal Imaging, p. 109, Plenum Press, New York (1975).
20. H. L. Snyder, "Image Quality and Observer Performance," in Perception of Displayed
Information, L. M. Biberman, ed., pp. 87-118, Plenum Press, New York, NY (1973).
12
MINIMUM RESOLVABLE CONTRAST
Most image quality metrics were developed for high contrast imagery. Here,
image quality is closely linked with the MTF. High MTF cameras provide better
image quality and therefore provide better resolution under optimum conditions.
While most resolution metrics are applicable to high contrast targets, the relative
rating of systems may change for low contrast scenes. A high MTF system with
excessive noise will not permit detection of low contrast targets.
MRC was originally developed for vidicons and low light level televisions.
The approach taken here extends the thermal imaging system performance model
(popularly called FLIR92) methodology 3 to visible systems. FLIR92 calculates
the minimum resolvable temperature (MRTl where the target-background
338
MRC 339
Within the real world, there is a probability associated with every parameter.
The target contrast is not one number but a range of values that follow a diurnal
cycle. The atmospheric transmittance is not fixed but can change in minutes.
There appear to be an overwhelming set of combinations and permutations.
Therefore, only a few representative target contrasts and a few representative
atmospheric conditions are selected and the performance range is calculated for
these conditions.
The Nyquist frequency limits the highest frequency that can be reproduced.
The specific manner that the Nyquist frequency is included in the model greatly
affects range performance. This is a result of the assumptions used and is not
necessarily representative of actual system performance. While no model is
perfect, models are still excellent for comparative analyses.
The symbols used in this book are summarized in the Symbol List (page
xviii) which appears after the Table of Contents.
The original Night Vision Laboratory static performance model 5 and FLIR92
use the Komfeld-Lawson eye model (see Section 10.8., Human Visual Response,
page 306). These models were validated by the U.S. Army when detecting tank-
sized targets at modest ranges. Insufficient data exist to say with certainty which
of the eye models available is best. Because of this and other model
uncertainties, range predictions cannot be placed on an absolute scale. All
analyses, no matter which eye model is used, must be used only for comparative
performance purposes.
340 CCD ARRAYS, CAMERAS, and DISPLAYS
The human visual system "filters" enhance the displayed SNR. The
perceived signal-to-noise ratio, referred back to the array output, is
MTF lln 1
SNR - k SYS pe • (12-1)
P (nSYS) (eye spatial filter) (eye temporalfilter)
MTFsys includes the camera, electronics, display, and HVS MTFs (see Chapter
10, System MTF, page 267). All calculations are performed with object-space
units (e.g., cycles/mrad). The value .::1{\e is the difference in the number of
photoelectrons generated by the target and its immediate background. The value
k depends on the optical aperture diameter, focal length, detector spectral
quantum efficiency, etc.
For nearly all imaging systems, square waves (bar targets) are used for
system characterization. Using the CTFsys (See Section 9.5., Contrast Transfer
Function, page 263), the perceived signal-to-noise ratio is
where k' is a constant that includes k and the conversion from MTF to CTF.
MRC 341
0.001 +-----+------+----+-----+
0.1 10 100 1000
(nSYS)(eyespatial.filter)(eyetemporal.filter) . ( _)
~nMimMUM-SNRTil~~-------------------------- 12 3
k1 CTFSYS
The system noise may consist of many components and the HVS may
integrate each component differently. The function (3i is the "filter" that
interprets noise source (I\):
or
(12-5)
342 CCD ARRAYS, CAMERAS, and DISPLAYS
(12-6)
Following FLIR92 terminology, 3 (n1} is the shot noise. The other noise terms
represent all other noise sources. The three-dimensional noise model quantifies
each (n?}.
In Section 6.2.1., Camera SNR (page 186), a target SNR was calculated.
The background light level was assumed to be zero so that the photon shot noise
was associated with the target only. For target detection, the target exists on a
positive background. Both the target and the background regions provide photons
that the camera converts to photoelectrons. Let ~-T and ~-s be the number of
photoelectrons associated with the target and background, respectively. There
is shot noise associated with both. The "average" shot noise variance is
(12-8)
MRC 343
Historically, the target was always considered darker than the background.
However, the target can be brighter than the background and there does not
appear to be any difference between the detectability of objects that are of
negative or positive contrast.
Substituting Equations 12-7 and 12-8 into Equation 12-6 and calling the
minimum detectable contrast the MRC, provides
2
1t MRC+2
MRC- SNR - - - -
TH 8 kMTFsYS 2npe-B
Because MRC appears on both sides of the equation, it can be solved either
iteratively or as a quadratic equation. The MRC is a family of curves dependent
on the background level (Figure 12-2). MRC approaches a maximum value of
unity as the system MTF approaches zero. Thus, MRC is linked to MTF but the
resolution (i.e., the spatial frequency where the contrast approaches one)
depends on the background.
0.8
0 0.6
0:::
::!!i 0.4
increasing
0.2 light level
0~~~====~~--------~--~
0 0.2 0.4 0.6 0.8
Perceived resolution depends on the size of the target, ambient lighting, and
target-to-background contrast. These issues were not considered when discussing
display resolution because they are relatively constant.
For signals embedded in noise, the situation is different. The HVS is then
very sensitive to the signal-to-noise ratio. Slight variations affect target visibility.
As illustrated in Figure 12-3, the signal increases faster than the accompanying
photon shot noise. Given a threshold of perceptibility, as the SNR increases,
more detail can be seen because more signals rise above the threshold. Thus,
camera "resolution" depends on background light level. This "resolution" is
strictly due to the HVS's response and is not related to hardware performance.
In contrast, hardware resolution is typically specified as that spatial frequency
where the hardware MTF is, say, 10% and is measured under relatively high
signal-to-noise ratio conditions.
9 increasing
0:: 8 signal
~ 7
c s~-
w 5
~ 4 r----/
w threshold
~ ~,-----~~--~~~~~~----------
~ ~1-----~~-----+--~======~~~~~--
0 0.2 0.4 0.6 0.8 1
NORMALIZED SPATIAL FREQUENCY
Figure 12-3. As the target signal increases, the SNR increases. For a
fixed threshold, increasing the signal appears to increase the resolution.
10 100
RELATIVE INTENSITY
The noise is divided into a set of eight components that relate temporal and
spatial noise to a three-dimensional coordinate system (Figure 12-5). This
approach allows full characterization of all noise sources including random
noise, pattern noise, streaks, rain, 1/f noise, and any other artifact. Analyzing
the noise in this manner has the advantage of simplifying the understanding of
complex phenomena by breaking it down into a manageable set of components.
The method simplifies the incorporation of complex noise factors into model
formulations.
H , , ," , ,
, ,,
2
v ,
,
m
,,
-
n
Figure 12-5. Three-dimensional noise model coordinate system
illustrating data set NTVH·
Table 12-1
THREE-DIMENSIONAL NOISE DESCRIPTORS
3-D NOISE
DESCRIPTION
COMPONENT
Depending on the system design and operation, any one of these noise
components can dominate. The origins of these components are significantly
different and the existence and manifestation depend on the specific design of the
imaging system. Not all of the noise components may be present in every
imaging system.
Assuming the noise components are independent, the total system noise is
(12-10)
348 CCD ARRAYS, CAMERAS, and DISPLAYS
Figure 12-8. Image with dominant horizontal banding (high uTV or CTy).
350 CCD ARRAYS, CAMERAS, and DISPLAYS
With all systems, the most prevalent noise source is assumed to be temporal
noise (uTVH). For staring arrays, the next most important noise source is <TvH·
Table 12-2 lists the FLIR92 recommended 3 values for PtSi staring arrays
operating in the 3 to 5/Lm range. Other systems using different detectors may
perform differently. These values were obtained empirically.
For scanning systems, temporal row noise (uTV) and fixed row noise (uv) are
the important contributors. Table 12-3 lists the FLIR92 default values
appropriate for scanning systems with HgCdTe detectors operating in the 8 to
12/Lm region. These values apply to all linear arrays (vertically oriented) with
horizontal scanning.
Table 12-2
DEFAULT VALUES for STARING SYSTEMS
RELATIVE NOISE DEFAULT VALUE
<1vHf<1TvH 0.40
<1TV/<1TVH 0
<Tyf<TTVH 0
<1TH/<1TVH 0
<1H/<1TVH 0
Table 12-3
DEFAULT VALUES for SCANNING SYSTEMS
RELATIVE NOISE DEFAULT VALUES
<1yH/<7TVH 0
<1TV/<1TVH 0.25
<Tyf<TTVH 0.25
<Tmf<TTVH 0
<1Hf<1TVH 0
MRC 351
1t2 1 MRC + 2 K ( )
MRC tu ) - SNR ----- H Uob
(12-11)
11' ob TH 8 kMTFJu 0 b) 2npe-B
and
The functions KH(U00) and Kv(V00) are summary noise factors that combine the
three-dimensional noise model components with the HVS's spatial and temporal
integration capabilities. The variable MTFH(u00) is the product of all the
horizontal MTFs and MTFv(V00) is the product of all the vertical MTFs (see
Chapter 10, System MTF, page 267). In the original development of the MRC
model, only the HVS spatial and temporal integration factors were considered.
Later models included the HVS MTF.
352 CCD ARRAYS, CAMERAS, and DISPLAYS
(12-13)
and
(12-14)
For staring systems, the default (see Table 12-2) MRCs become
and
For scanning systems (applies to linear arrays) the default (see Table 12-3)
MRCs are
and
E V (E'r"H +_!).
1i' +_!E
4
T
4
(12-18)
MRC 353
The values Er, EH, and Ev, respectively, represent the temporal, horizontal
spatial and vertical spatial integration afforded by the HVS interpreter. The
model assumes that the HVS temporally integrates perfectly and continuously
over the HVS integration time. The noise is considered uncorrelated from frame-
to-frame so that it adds in quadrature. When the information update rate, FR, is
high enough, and if there is signal correlation between adjacent frames:
(12-19)
Table 12-4
MRC EQUATION NOMENCLATURE
MODEL
MEASUREMENT BAR ORIENTATION
NOMENCLATURE
The noise filters and noise power spectrum were described in Section 6.2.1.,
Camera SNR (page 186). The noise spectrum is modified by the electronic filters
that occur after the insertion of noise. Because noise is assumed to originate in
the detector, MTFoPTics• MTFMonoN• and MTFoETECTOR are not part of the noise
filters. As appropriate, the remaining MTFs provided in Chapter 10, System
MTF (page 267), are components of the noise filters. For equation brevity, the
horizontal and vertical noise filters are listed as HNF-H(~b) and HNF-v(vob),
respectively. As appropriate, f., fv, and ud are expressed in equivalent object-
space units (u0 b). Because electrical signals are typically read out in a serial data
354 CCD ARRAYS, CAMERAS, and DISPLAYS
stream (consistent with video standard timing), the noise appears in the
horizontal direction only. The variable vobis the transposed variable vd and also
the vertical digital filter response if present.
For horizontal MRC, the bars are oriented vertically and the bar aspect
ratio, TAsPECT• appears in the vertical summary noise factors. For a bar target
with fundamental spatial frequency U 0 , the summary noise factors are
and
E fu )
V-II'- o
=
jl
~
dCCVJIH
0
fv )12 sinc2
NF-V' ob
(T
ASPECTVob
2 uo
l
dv
ob
. (12-21)
For the vertical MRC, the bar aspect ratio is in the horizontal summary
noise factor:
and
(12-23)
In the early literature, only the horizontal MRC was calculated and only
random temporal noise was modeled. With no other noise sources,
where Er. EH, and Ev are given by Equations 12-19, 12-20, and 12-21,
respectively.
MRC 355
If MTFs are near unity and the noise is white [S(u00) :::: 1] over the spatial
frequencies of interest, then
(12-25)
and
(12-26)
Then
1t2 SNRTH
MRC(u,)-- ----==--
8 kMTF(u,)
The original SNR formulation 16 contained an (a/A) 112 factor where "a" is the
target area at the image plane and "A" is the effective area of the photosensitive
device. The MRC is the inverse of the SNR and therefore should contain
(A/a) 112 • This is identical to the last square-root term in Equation 12-27. "A" is
the photosensitive area that contributes to noise. With vidicons, the "A"
represented the entire image surface. With solid state arrays, it is the effective
pixel size, dccHdccvlfl2 • The target size is (1!u )(TAsPEcr/uo). Equation 12-27
0
reduces to the basic MRC equation used by most researchers 17 •18 prior to the
introduction of the three-dimensional noise model.
The FLIR92 documentation19 states: SNRm and the eye integration time, t,,
are most often the model parameters used to "tune" an MRTD prediction to a
set of measurements. These values are treated as constants that are adjusted so
that the predicted values match laboratory data.
The value te is somewhat nebulous and its value depends on the specific task
at hand and ambient light level. FLIR92 recommends changing te according to
the available light level (Figure 12-9).
4
~~
...,
~~
.......
~
b--.....
0
-5 -4 -3 -2 -1 0 2 3
LOG BRIGHTNESS
Figure 12-9. HVS integration values as a function of light level. FLIR92
documentation recommends using te = 0. I s for laboratory MRT
predictions. The same technique can be applied to MRC.
While many arguments exist for selecting the "correct" values, SNRTH and
te are simply constants. Both SNRTH and te have a range of acceptable values.
No matter what value is selected for t0 , consistency throughout the model must
be maintained. The value te is also used in the linear motion MTF (see Section
10.5.1., Linear Motion, page 292).
(a) (b)
(c) (d)
0.8
0 0.6
0::
:::E 0.4
0.2
oL---~~~~--L_ ________
v1 f 20 u1
RELATIVE FREQUENCY
The Johnson criterion provides the link between the target angular subtense
and the spatial frequency scale on the MRC graph. The target's inherent
contrast, C0 , is modified by the atmospheric conditions (attenuation and path
radiance). When the target contrast is reduced to the MRC curve, the
performance range is obtained. Depending on the Johnson criterion selected, this
becomes the detection, recognition, or identification range.
However, real systems have automatic gain circuitry that may be activated
by the path radiance signal. If so, the effect is to reduce system gain. Path
radiance will partially fill the charge wells.
For most applications, the target and background are illuminated by the same
source (sun, moon, etc.). The apparent contrast, CR, is the inherent contrast
modified by the atmosphere. Dropping the wavelength notation for equation
brevity,
C _ !J.L_
-aA,..,R[ ]L
e p T - p B SCENE _ C - - - -
0 L -a.,.,R L L o L (12-29)
e pB SCENE + A 1M 1+ A 1M e a A1JIR
p bLSCENE
where LscENE is the ambient illumination that irradiates the target and
background.
360 CCD ARRAYS, CAMERAS, and DISPLAYS
1
c- c ·-- (12-30)
R o1 + SGR(eaATMR- 1)
The SGR is approximately 0.2/p 8 for a clear day (LsKv/LscENE 0.2) and 1/p8 =
for an overcast day (LsKviLscENE =
1). When the SGR is one, the received
contrast is simply the inherent contrast reduced by the atmospheric
transmittance:
(12-31)
where CTH is the modulation threshold at which 50% of the observers would see
the target. The use of the scattering cross section in this definition rather than
the extinction coefficient implies that absorption of particles at visual
wavelengths is small enough to ignore. This view is probably justified except in
cases of polluted air. Koschmieder set CTH to 0.02 and evaluated aATM at "A=
0.555 ttm.Then
_ 3.912 R
'1' • TA7M e-a,~.ruR - e RYJs (12-33)
~ ransmtttance - -
MRC 361
(12-35)
Table 12-5
INTERNATIONAL VISIBILITY CODE
SCATTERING
DESIGNATION VISIBILITY
COEFFICIENT' (TATM
Dense fog 0-50 m > 78.2km- 1
The Johnson25 methodology has become known as the equivalent bar pattern
approach. Observers viewed military targets through image intensifiers and were
asked to detect, decide the orientation, recognize, and identify the targets. Air
Force tri-bar charts with bars the same contrast as the scaled models were also
viewed and the maximum resolvable bar pattern frequency was determined. The
number of bars per critical object dimension was increased until the bars could
just be individually resolved. In this way, detectability was correlated with the
sensor's threshold bar pattern resolution (Table 12-6). These results became the
foundation for the discrimination methodology used today. Critical dimension
refers to the target dimension that must be discerned for detection, recognition,
or identification. In the early literature, the minimum dimension was called the
critical dimension.
Table 12-6
JOHNSON'S RESlfLTS
(From Reference 25)
CYCLES ACROSS
DISCRIMINATION
MEANING CRITICAL TARGET
LEVEL
DIMENSION
An object is present
Detection 1.0 ± 0.025
(object versus noise).
Although Johnson used 6.4 cycles across the critical target dimension for
identification, studies with thermal imaging systems at the Night Vision
Laboratory suggested that eight cycles. were more appropriate for identification.
Orientation is a less popular discrimination level. Because today's standards are
based on Johnson's work, they are labeled as the Johnson criteria though they
are not the precise values recommended by him.
MRC 363
Table 12-7
COMMON DISCRIMINATION LEVELS
One-dimension Two-dimension
TASK DESCRIPTION
Nso N,o
Table 12-8
DISCRIMINATION CUMULATIVE PROBABILITY
PROBABILITY of
MULTIPLIER
DISCRIMINATION
1.00 3.0
0.95 2.0
0.80 1.5
0.50 1.0
0.30 0.75
0.10 0.50
0.02 0.25
0 0
P(N) = (12-36)
where
Figure 12-12 illustrates the full TTPF for the three levels of discrimination
as identified by Johnson. Because of the variability in the population, some
people can only detect the target, others will recognize it, and still a smaller
portion will identify the target. This leads to the wide variations seen in field test
results.
MRC 365
~
..J
0.8
iii 0.6
<C
m 0.4
0
0:: 0.2
ll..
5 10 15
NUMBER OF CYCLES
(a)
1
~ 0.8
::J
iii 0.6
<C
lXI 0.4
~
ll..
0.2
2 4 6 8 10
NUMBER OF CYCLES
(b)
N _ JATARGET f = he f (12-38)
T R 2D R 2D"
366 CCD ARRAYS, CAMERAS, and DISPLAYS
Consider a target that is 2 X 8 m and six cycles are required across the critical
dimension. Then the conversion from spatial frequency to range is
fH
R- - . 4 ~ (12-39)
6 f i2D - --
6 J2D '
where six cycles represent 95% probability of recogmt10n for the two-
dimensional model. The intersection of the MRC curve and the target apparent
contrast, CR (using Equation 12-31, page 360), versus range curve is the range
at which the target can be discerned according to the discrimination level
selected (Figure 12-13).
0.1
0
0:::
:! 0.01
0.001
0 2 4 6 8
SPATIAL FREQUENCY (cy/mrad)
(a)
0.1
0
0:::
:! 0.01
0.001
0 2 3 4 5 6
RANGE (km)
(b)
apparent target
contrast
0.1
... j .. ..
T
H
0.01
l_
w --1
f20
SPATIAL FREQUENCY
TTPF
p •.......• p
0.5 0.5
OL---------------~-- o~~----~----------
R Nr
RANGE NUMBER OF CYCLES
In Chapter 8, Sampling Theory, page 231, it was stated that the highest
frequency that could be faithfully reproduced is the Nyquist frequency. Solid
state cameras are undersampled systems and are limited to a spatial frequency
dictated by the detector center-to-center spacing. In the development of the MRC
equation, sampling effects were ignored.
Nyquist limit
0.8
~ 0.6
:::2 0.4 ',
~ 0.8
:J
iii 0.6
<C
m 0.4
Nyquist
0
0::: 0.2 limit
ll.
0
0 0.5 1.5 2
RELATIVE RANGE (km)
12.7. REFERENCES
1. 0. H. Shade, Sr., "Image Gradation, Graininess, and Sharpness in Television and Motion
Picture Systems," published in four parts in SMPTE Journal: "Part I: Image Structure and
Transfer Characteristics," Vol. 56(2), pp. 137-171 (1951); "Part II: The Grain Structure of
Motion Pictures - An Analysis of Deviations and Fluctuations of the Sample Number," Vol.
58(2), pp. 181-222 (1952); "Part III: The Grain Structure of Television Images," Vol. 61(2),
pp. 97-164 (1953); "Part IV: Image Analysis in Photographic and Television Systems," Vol.
64(11), pp. 593-617 (1952).
2. F. A. Rosell and R. H. Willson, "Performance Synthesis of Electro-Optical Sensors," Air Force
Avionics Laboratory Report AFAL-TR-72-229, Wright Patterson AFB, OH (1972).
3. FUR92 Thermal Imaging Systems Performance Model, User's Guide, NVESD document
UG5008993, Fort Belvoir, VA (1993).
4. J. D' Agostino and C. Webb, "3-D Analysis Framework and Measurement Methodology for
Imaging System Noise," in Infrared Imaging Systems: Design, Analysis, Modeling, and Testing
II, G. C. Holst, ed., SPIE Proceedings Vol. 1488, pp. 110-121 (1991).
5. W. R. Lawson and J. A. Ratches, "The Night Vision Laboratory Static Performance Model
Based on the Matched Filter Concept," in The Fundamentals of Thermal Imaging Systems, F.
Rosell and G. Harvey, eds., NRL Report 8311, pp. 159-179, Naval Research Laboratory,
Washington, D.C. (1979).
6. S. Daly, "Application of a Noise Adaptive Contrast Sensitivity Function in Image Data
Compression," Optical Engineering, Vol. 29(8), pp. 977-987 (1990).
7. H. Pollehn and H. Roehrig, "Effect of Noise on the Modulation Transfer Function of the Visual
Channel," Journal of the Optical Society of America, Vol. 60, pp. 842-848 (1970).
8. A. Van Meeteren and J. M. Valeton, "Effects of Pictorial Noise Interfering With Visual
Detection," Journal of the Optical Society of America A, Vol. 5(3), pp. 438-444 (1988).
370 CCD ARRAYS, CAMERAS, and DISPLAYS
9. L. Scott and J. D' Agostino, "NVEOD FLIR92 Thermal Imaging Systems Performance Model,"
in Infrared Imaging Systems: Design, Analysis, Modeling, and Testing III, G. C. Holst, ed.,
SPIE Proceedings Vol. 1689, pp. 194-203 (1992).
10. System Image Analyzer software is available from JCD Publishing, 2932 Cove Trail, Winter
Park, FL 32789.
11. T. S. Lomheim and L. S. Kalman, "Analytical Modelmg and Digital Simulation of Scanning
Charge-CoupledDeviceimagingSystems," inElectro-OpticalDisplays, M.A. Karim, ed., pp.
551-560, Marcel Dekker, New York (1992).
12. J. A. Ratches, W. R. Lawson, L. P. Obert, R. J. Bergemann, T. W. Cassidy, and J. M.
Swenson, "Night Vision Laboratory Static Performance Model for Thermal Viewing Systems,"
ECOM Report ECOM-7043, p. 34, Fort Monmouth, NJ (1975).
13. J. M. Lloyd, Thermal Imaging Systems, pp. 182-194, Plenum Press, NY (1975).
14. F. A. Rosell, "Laboratory Performance Model," in The Fundamentals of Thermal Imaging
Systems, F. Rosell and G. Harvey, eds., NRL Report 8311, pp. 85-95, Naval Research
Laboratory, Washington, D.C. (1979).
15. R. L. Sendall and F. A. Rosell, "Static Performance Model Based on the Perfect Synchronous
Integrator Model," in The Fundamentals of Thermal Irrwging Systems, F. Rosell and G. Harvey,
eds., NRL Report 8311, pp. 181-230, Naval Research Laboratory, Washington, D.C. (1979).
16. F. A. Rosell, "Video, Display, and Perceived Signal-to-Noise Ratios," in The Fundamentals
of Thermal Imaging Systems, F. Rosell and G. Harve}, eds., NRL Report # 8311, pp. 49-83,
Naval Research Laboratory, Washington, D.C. (1979)
17. W. W. Frame, "Minimum Resolvable and Minimum Detectable Contrast Prediction for Vidicon
Cameras," SMPTE Journal, Vol. 104{1), pp. 21-26 (l 985).
18. W. W. Frame, "Minimum Resolvable and Minimum Detectable Contrast Prediction for
Monochrome Solid-State Imagers," SMPTE Journal, Vol. 96(5), pp. 454-459 (1987).
19. FUR92 Thermal Imaging Systems Performance Model. Analyst's Reference Guide, NVESD
documentRG5008993, pp. ARG-12- 13, Fort Belvoir. VA (1993).
20. J. A. Ratches, W. R. Lawson, L. P. Obert, R. J. Bergemann, T. W. Cassidy, and J. M.
Swenson, "Night Vision Laboratory Static Performance Model for Thermal Viewing Systems,"
ECOM Report ECOM-7043, p. 56, Fort Monmouth, ~J (1975).
21. F. A. Rosell, "Psychophysical Experimentation," in 7he Fundamentals of Thermal Imaging
Systems, F. Rosell and G. Harvey, eds., NRL Report # 8311, p. 225, Naval Research
Laboratory, Washington, D.C. (1979).
22. G. C. Holst, Electro-Optical Imaging System Performance, Chapters 19, 20, and 21, JCD
Publishing, Winter Park, FL (1995).
23. L. Levi, Applied Optics, pp. 118-124, Wiley and Sons (1980).
24. W. E. K. Middleton, Vision Through the Atmasphere, University of Toronto Press (1958).
25. J. Johnson, "Analysis of Imaging Forming Systems," in Proceedings of the Image Intensifier
Symposium, pp. 249-273, Warfare Electrical Engineering Dept., U.S. Army Engineering
Research and Development Laboratories, Ft. Belvoir, VA (1958). This article is reprinted in
Selected Papers on Infrared Design, R. B. Johnson and W. L. Wolfe, eds., SPIE Proceedings
Vol. 513, pp. 761-781 (1985).
26. L. Scott and R. Tomkinson, "An Update on the C 2NVEO FLIR90 and ACQUIRE Sensor
Performance Model," in Infrared Imaging Systems: Design, Analysis, Modeling, and Testing
II, G. C. Holst, ed., SPIE Proceedings Vol. 1488, pp. 99-109 (1991).
27. J. D. Howe, "Electro-Optical Imaging System Performance Prediction," in Electro-Optical
Systems Design, Analysis, and Testing, M. C. Dudzik, ed., p. 92. This is Volume 4 of the
Infrared and Electro-Optical Systems Handbook, J. S Accetta and D. L. Shumaker, eds.,
copublished by Environmental Research Institute of Michigan, Ann Arbor, MI, and SPIE Press,
Bellingham, WA (1993).
APPENDIX
I
P,
Figure A-1. The optical system can be considered as a single thick lens.
P1 and P2 are the principal surfaces. The effective focal length is
measured from the second principal surface. The clear aperture limits
the amount of scene radiation reaching the detector. Although shown as
planes, the principal surfaces are typically spherical.
A.2. f-NUMBER
The radiometric equations were derived from plane geometry and paraxial
ray approximations. When using solid angles, the image incidance is
proportional to sin2(U') where U' is the maximum angle subtended by the lens.
For paraxial rays, the principal surface is considered a plane. This
representation is shown in most textbooks.
371
372 CCD ARRAYS, CAMERAS, and DISPLAYS
Lens design theory 1 assumes that the principal surface is spherical - every
point on the surface is exactly a focal-length distance away (Figure A-2). This
"sine corrected" lens has zero spherical aberratwn. Thus, sin2(U') is equal to
11(4F2). The numerical aperture is another mea~ure of the energy collected by
the optical system. When the image is in air !refractive index of unity) the
numerical aperture is
1 (A-1)
NA - sinU1 -
2F
Since the largest angle is 1r/2, the smallest theoretical value for F is Vz. This
theoretical limit on F is not obvious from the radiometric equation.
P1
Not all optical systems have spherical principal planes. Often we assume the
principal plane is flat. For hand-drawn optical layouts, the lenses are
approximated by thin lenses and paraxial rays are used (large f-numbers). Here,
=
sin(O) 0 and Equation A-1 reduces to F = fl/D. Then sin2(U') is 11(4f2+ 1).
Whether 1/(4F2 + 1) or 1/4f2 is used in the radiometric equations depends upon
the f-number definition. If the analyst calculates the f-number from the effective
lens diameter and focal length, then 11(4f2+ 1) should be used. If the optical
designer provides the f-number, then the analyst must consult with him to ensure
that the appropriate factor is used. For large f-numbers, the factors are
approximately equal.
A.3. REFERENCE
1. W. J. Smith, Modem Optical Engineering, second ediuon, pp. 142-145, McGraw-Hill, New
York (1990).
INDEX
Absorption coefficient 104, 277 Camcorder 6
Active pixel sensor 91 Camera formula 33
Addressability 222 Camera operation 147
Adjacent disel requirement 222 Camera performance 182
Advanced television system 160 Cathode ray tube 205
Aerial image modulation 325 Causality 252
Airy disk 16, 257, 286, 320 CAV 156
Aliasing 232, 236 CCIR 150
display 250 CDS 141
frame grabber 250 Center-to-center spacing 16
Alternating disel requirement 222 Channel block 49
Amplifier noise 128 Character recognition 227
Analog-to-digital converter 174 Charge aggregation 57
Anti-aliasing filter 236 Charge conversion 78
Anti-parallel clock 51 Charge grouping 57
Antibloom drain 85 Charge injection device 89
Aperture correction 172, 302 Charge transfer efficiency 74, 279
APS 91 Chicken wire 179
Array CID 89
frame interline transfer 64 CIE 29
frame transfer 61 Clock 49
full frame transfer 59 Cluster defect 144
interline transfer 63 CMOS 91
linear 58 Color correction 165
TDI 70 Color filter array 95, 242, 283
Atmospheric transmittance 360 Color temperature 21, 31, 209
ATS 160 display 32
Column defect 144
Back illuminated 109 Component analog video 156
Bayer 96 Composite video 156
Beat frequency 245 Configuration 9
Bilinear readout 58 Contrast 228
Binning 57, 58, 187, 299 Contrast transfer function 263
Birefringent crystal 283 Contrast transmittance 359
Blackbody 21 Convolution 255, 257
Blanking 151 Cooling 80
Blur diameter 320 Correlated double sampling 141
Boost filter 302 CTE 74
Brightness 26 Cutoff
Broadcast standard detector 240, 275
NTSC 150 optical 261, 27 4
PAL 150
SECAM 150 Dark current 79
Buried channel 74 Dark current noise 79, 127
Butterworth filter 295 Dark pixel 83
DAS 15, 321
Calibration source 29 Date) 14
373
374 CCD ARRAYS, CAMERAS, and DISPLAYS
MTF Normalization 37
display 303 NTSC 149
eye 306 Nyquist frequency 231,242
Gaussian 294
HVS 306 Object space 269
image intensifier 311 Observer 203
lens 261 Observer space 269
linear motion 292 Optical anti-alias filter 283
median 312 Optical resolution 320
non-limiting 311 Optical transfer function 257
optical anti-alias 283 Optics-limited 287, 321
optics 274 Organic phosphor 109
random motion 294 Out-of-phase 246, 312
Output gain conversion 78, 137
sample and hold 299 Output structure 78
sample-scene 312 Overflow drain 85
separable 268 Oversampling 232
system 286
TDI 280 PAL 150
tuned circuit 303 PAS 16
Multi-pinned phasing 82 Path radiance 359
Pel 14
National Imagery Interpretability Phas~
Rating Scale 320 four 51
NEE 103 three 53
NIIRS 320 cWO 55
Noise virtual 57
1/f 128 Phase reversal 292
ADC 130 Phase transfer function 257
amplifier 128 Phosphor 109, 207
dark current 127 Photodiode 47
excess 198 Photogate 47
fixed pattern 131 Photometric conversion 118
Johnson 127 Photometry 23
kTC 127 Photon standard 102
photoresponse nonuniformity Photon transfer 102, 133
131 Photopic 23
quantization 130 Photo response nonuniformity 131
reset 127 Pitch 16, 242
shot 127 1riad 208
summary factors 351 Pixel 14
three-dimensional 346 Pixel trap 144
white 128 Pixel-angular-subtense 16, 322
Noise equivalent bandwidth 128, 129 Planck's blackbody law 21
Noise equivalent exposure 119 Point defect 144
Noise equivalent input 192 Point spread function 257
Noise equivalent reflectance Poisson statistics 127
difference 193 Polysilicon 107
Noise equivalent signal 120 Post- reconstruction filter 300
Non-destructive readout 89 PRNU 131
INDEX 377
Professional broadcast television 6 Shrinking raster 211
Progressive scan 68, 156, 172, 205, Shutter speed 35
299 Signal-to-noise ratio 139, 340
Pseudo-interlacing 66, 299 Size
video chip 99
Quantization noise 130 Sky
Quantum efficiency 104, 111 night 39
Sky-to-ground ratio 360
Radiance 20 Smear 60, 68
Radiative transfer 19 SNR 186
Radiometry 18 SNRru 355
Range prediction 359, 365 Source
Raster 203 calibration 29
Rayleigh criterion 320 Spatial frequency 269
Recorder Spectral response 107
digital 11 Square wave 262
Relay lens 176, 195 Square-root integral 334
Reset 14, 320 Square-wave response 263
Reset noise 127 Sterance 20, 26
Resolution 218, 222, 317 Streaking 73
detector 321 Strobe light 68
electronic 323 Subarray 60
limiting 324 Subcarrier frequency 160
optical 320 Subjective quality factor 332
perceived 344 Summary noise factor 351
Shade 326 Super pixeling 57, 299
TV limiting 220 Superposition 255, 261
Resolution/addressability ratio 223, Surface channel 74
304 Synchronous integrator 345
Responsivity 40, 111
Ringing 262, 302 T-number 36
RS 170 149 Taps 297
RS 170A 150 Target transfer probability function
363
Sample-and-hold 299 TDI 70, 173, 280
Sample-scene 312 TEC 80
Sampling 231, 368 Temperature
Sampling theorem 234 color 21, 31
Saturation equivalent exposure 120 Thermoelectric cooler 80
Scan pattern 67 Three-dimensional noise 346
Scene 186 Threshold modulation 325
Scene! 14 Time delay and integration 70
Scientific application 8 Transmittance
Scotopic 23 atmospheric 360
SECAM 150 Tuned circuit 345
SEE 103 TVLIPH 220
Setup 157 Two-dimensional MRC 358
Shade's equivalent resolution 325
Shades of gray 227 Undersampling 232
Shield 61
Shot noise 127 Velocity error 280
378 CCD ARRAYS, CAMERAS, and DISPLAYS
Well capacity 48
White noise 128
Wien's law 21
Zoom 287