0% found this document useful (0 votes)
87 views

Chapter1 Introduction Image Processing

This document provides an introduction to digital image processing. It defines image processing as the manipulation and analysis of information contained in images. Some examples of image processing applications include medical imaging, facial recognition, and optical character recognition. The key aspects covered are: - Images are digitized through sampling and quantization, converting a continuous signal to a discrete matrix of numbers (pixels). - Digital image processing uses mathematical methods to manipulate and analyze these pixel values without losing the intuitive nature of images. - Common purposes of image processing include image enhancement, pattern recognition, data reduction, image synthesis, and data compression.

Uploaded by

vada1980
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
87 views

Chapter1 Introduction Image Processing

This document provides an introduction to digital image processing. It defines image processing as the manipulation and analysis of information contained in images. Some examples of image processing applications include medical imaging, facial recognition, and optical character recognition. The key aspects covered are: - Images are digitized through sampling and quantization, converting a continuous signal to a discrete matrix of numbers (pixels). - Digital image processing uses mathematical methods to manipulate and analyze these pixel values without losing the intuitive nature of images. - Common purposes of image processing include image enhancement, pattern recognition, data reduction, image synthesis, and data compression.

Uploaded by

vada1980
Copyright
© © All Rights Reserved
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 8

Chapter 1

Introduction
I
mages are everywhere. No wonder, since we as human beings rely on the images
we perceive with our eyes more than any other sensory stimulus. Almost all of the
information we digest comes to us in the form of an image; whether we look at a
photograph, watch television, admire a painting, or read a book, it all makes use of im-
agery. Images are so natural to us, that we go to great lengths to convert almost any kind
of information to images. For example: the TV weather forecast shows the temperature
distribution in some geographical area as an image with different colors representing
different temperatures, medical scanners can show human metabolism activity as an
image where bright spots indicate high activity, etc. Moreover, our vision is usually the
most efcient of our senses: consider, for example, a computer keyboard. The function
of each key is represented by a small image (a character). We could also have identied
each key by a specic relief texture, but it would be far less efcient. We could even try
to give each key a specic smell, but it is easy to imagine (!) the trouble we would have
typing.
We are also adept at a number of image processing tasks. For example, the focusing of our
eyes: when we look at something, the rst image our eyes send to the brain is probably
out of focus. The brain then tries to correct this by adjusting the eye lens, whereupon
a new image is sent from the eyes to the brain, and so on. This feedback process is so
fast that we arent even aware of it. Another example is stereo vision: our eyes send
two two-dimensional images to the brain, but the brain is able to merge these into one
three-dimensional image virtually instantaneously.
The science of image processing combines this natural way humans use images with the
science of mathematics. This provides a unique blend, since images and image process-
ing are described in a rigorous mathematical way without loss of the intuitive character
of imagery.
1.1 Image processing: denition and examples
Image processing can be dened as
10 Introduction
The manipulation and analysis of information contained in images.
This denition is of course very broad, and includes a wide range of natural and arti-
cial processes, from the use of a pair of glasses to automatic analysis of images trans-
mitted by the Hubble telescope. Simple forms of image processing can be found all
around us; examples include:
the use of glasses or contact lenses
brightness, contrast, etc. controls on a television or monitor
taking (and developing) a picture using a photocamera
natural examples: reection of scenery on a water surface, distortions of scenery
in the mist, a fata morgana, etc.
Examples of the use of advanced image processing include:
forensic science: enhancement of images of video surveillance cameras, automatic
recognition and classication of faces, ngerprints, DNA codes, etc. from images
industry: checking of manufactured parts, application to CAD/CAM
information processing: reading of handwritten and printed texts (frequently
referred to as OCR; optical character recognition), scanning and classication of
printed images
Alarge number of applications of image processing can be found in the medical sciences
using one or more medical images of a patient, e.g.,
Visualization. For example: before we can make a 3D visualization of a three
dimensional object (such as the head in gure 1.1), we often need to extract the
object information rst from two-dimensional images.
Computer aided diagnosis. For example: in western countries it is common to
regularly make breast radiographs of females above a certain age in order to detect
breast cancer in an early stage. In practice, the number of images involved is
so large that it would be very helpful to do part of the screening by means of
automated computer image processing.
Image segmentation, i.e., the division of an image into meaningful structures. For
example: the division of a brain image into structures like white brain matter,
grey brain matter, cerebrospinal uid, bone, fat, skin, etc. An example can be seen
in gure 1.2. Segmentation is useful in many tasks, ranging from improved visu-
alization to the monitoring of tumor growth.
Image registration (also called image matching), i.e., the exact alignment of two or
more images of the same patient, which is necessary if the information contained
in these images is to be combined in a meaningful new image. An example is
shown in gure 1.3.
1.1 Image processing: denition and examples 11
Figure 1.1 Example of extraction of information by image processing: the head visualized in
three dimensions on the left was extracted from two dimensional images such as the one on the
right.
Figure 1.2 Example of segmentation: the brain on the left was segmented into six structures
(indicated by different grey values in the right image) by an automatic algorithm.
12 Introduction
Figure 1.3 Example of two medical images that have been brought into registration: The upper
left and lower right quarters belong to the rst image, the lower left and upper right belong to
the second image. Visualizations like these enable the viewer to make use of the information
contained in the separate images simultaneously.
1.2 The purpose of image processing
Image processing applications can have many purposes. Most of the time, the purpose
will be in one or more of these categories:
Image enhancement, e.g., the reduction of noise or sharpening of images.
Pattern recognition, e.g., the automatic detection of a certain shape or texture in
images.
Reduction of data to information that is more easily handled or interpreted, e.g.,
the reduction of an image to an easier image, to a set of objects or features, or to
a set of measurements.
Image synthesis, e.g., reconstructing a three-dimensional scene from two-dimen-
sional photographs.
Combination of images. When images of two different modalities (types) are
made from the same scene, combining them involves registration, and, after that,
often data reduction and image synthesis.
Data compression. To reduce the size of computer les containing images and
speed up image transmission across a network, data compression is often a must.
1.3 The digital image
In this book, we concern ourselves only with digital image processing, and not analog
processing. The reason for this is that analog processing requires special hardware (elec-
trical circuits), which makes building a special image processing application a difcult
1.3 The digital image 13
00 00 01 01 01 03 01 00
01 00 46 4a 45 7f 01 01
00 00 4a 5b 4c 63 00 01
00 04 4c 4c 4c 4c 4d 00
00 3d 4c 4c 4b 4c 49 00
00 0a 4d 4c 4d 4d 17 00
00 00 48 50 4d 57 01 00
00 00 01 02 10 00 01 01
Figure 1.4 Example of sampling and quantization. The left image shows an original scene with
continuous intensity values everywhere. After sampling using 88 locations (middle image), the
image has real intensity values at specic locations only. After quantization (right image), these
real values are converted to discrete numbers (here: hexadecimal values).
task. Moreover, the use of analog hardware is rapidly becoming obsolete in many image
processing areas, since it can often be replaced by digital hardware (computers) which
is much more exible in its use.
But what exactly is a digital image? Obviously, we start with some kind of imaging
device such as a videocamera, a medical scanner, or anything else that can convert a
measure of a physical reality to an electrical signal. We assume this imaging device
produces a continuous (in space) electrical signal. Since such an analog signal cannot
directly be handled by digital circuits or computers, the signal is converted to a dis-
cretized form by a digitizer. The resulting image can then directly be used in digital
image processing applications.
The digitizer performs two tasks, known as sampling and quantization (see gure 1.4).
In the sampling process, the values of the continuous signal are sampled at specic
locations in the image. In the quantization process, the real values are discretized into
digital numbers. After quantization we call the result a digital image. So this answers
the question at the beginning of this section: a digital image is nothing more than a
matrix of numbers. Each matrix element, i.e., a quantized sample, is called a picture
element or a pixel. In the case of three-dimensional images this is named a volume element
or voxel.
We can indicate the location of each pixel in an image by two coordinates (x, y). By con-
vention, the (0, 0) pixel (the origin) is in the top left corner of the image, the x axis runs
from left to right, and the y axis runs from top to bottom (see gure 1.5). This may take
a little getting used to, because this differs from the conventional mathematical notation
of two-dimensional functions
1
, as well as from conventional matrix coordinates
2
.
1
Where the origin is at the bottom left, and the y axis runs from bottom to top.
2
Where the origin is the top left, the x axis runs from top to bottom, and the y axis from left to right.
14 Introduction
0 1 2 3 4 5
0
1
2
3
4
5
pixel (0, 0)
x
y
Figure 1.5 Coordinate conventions for images. Pixel (3, 1) is marked in grey.
If a digital image is nothing more than a matrix of numbers, a very sceptic person might
say that digital image processing is nothing more than a collection of mathematical
algorithms that operate on a matrix. Fortunately, reality is not nearly as boring as this
sounds, because in practice we will seldomly use the matrix representation shown in
gure 1.4. Instead, we work with the middle image from gure 1.4 which is in fact
the same matrix, but with intensity values assigned to each number but which usually
makes much more sense to a human being. Throughout this book, you will nd that
image processing algorithms will be formulated as mathematical operators working on
pixel values or pixel matrices, but the results of these algorithms will also be displayed
using images.
1.3.1 Mathematical representation of images
In this book, we will use either functions or matrices to mathematically represent an
image. Sometimes, it is also useful to use another graphical representation than the
standard image where a pixel value is represented by a certain luminance. For ex-
ample, we may represent pixel values as height, like in the example below. With the
mathematical representations, we will always use the image coordinate convention, so
the following are equivalent:
f(x, y) = xy for (x, y) {0, 1, 2, 3} {0, 1, 2}
M =

0 0 0 0
0 1 2 3
0 2 4 6

1.4 Image processing and related disciplines 15

Further on in this book, we will come across images with more than two dimensions.
The mathematical representations will be logical extensions of the ones above for two-
dimensional images. For example, a movie can be regarded as a time series of two-
dimensional images (frames), so a pixel value at location (x, y) of frame t can be repre-
sented by, e.g., f(x, y, t). Another example is given by some medical images that have
three spatial dimensions, and are therefore represented as, e.g., f(x, y, z). If we want to
visualize such three-dimensional images we often use volume renderings or simply dis-
play the three-dimensional image one two-dimensional slice at a time (see gure 1.1).
1.3.2 Display of images
When printing or displaying an image, its pixel values must be converted to grey values
on the paper or screen. There are many possible ways to do this, but the most common
three are:
Find the lowest and highest pixel value occurring in the image and represent these
as black and white respectively. Map the other pixel values between these ex-
tremes in a linear fashion.
Find the lowest and highest pixel value that the image type supports (even if they
do not actually occur in the image) and represent these as black and white. Map
the other pixel values between these extremes in a linear fashion. For example, if
the image type supports integer pixel values in the range {0, 1, 2, ..., 255}, then we
map 0 to black and 255 to white, even if 0 or 255 do not occur in the image at all.
Use a palette. A palette is a table that explicitly describes the grey value for each
pixel value.
1.4 Image processing and related disciplines
There are many disciplines that are closely related or have a large overlap with im-
age processing, such as computer vision, image understanding, and scene analysis. Frankly,
there is no real consensus on the borders between them, nor even on their denitions.
16 Introduction
You could argue that they are all part of image processing
3
, given the broad denition
we gave to image processing. A -debatable- graph showing relations of image process-
ing and other disciplines can be seen in gure 1.6.
Figure 1.6 The relation of image processing and various connected disciplines.
Here, the various disciplines are dened by their actions regarding images and mod-
els, where the model is usually a compact mathematical structure describing a phe-
nomenom using its essential parameters. For example, the image could be a photograph
of a face, and the model a graph of facial feature points such as the corners of the eyes,
the tip of the nose, etc, and the model parameters could be the distances between these
points. A typical computer vision task would be to nd the feature points in the image
and match them to the model, e.g., for face recognition. A typical computer graphics
task would be to generate an image of a face given the facial model and a set of dis-
tances. A typical image processing task would be to produce a new image by somehow
enhancing the photograph, e.g., by improving the contrast, red eye removal, etc.
Although the disciplines mentioned in the graph all overlap to some extent, such over-
lap is small in the case of both computational geometry and computer graphics. This is
not the case for image processing, computer vision, and image analysis, and attempting
to nd the borders between these areas often leads to articial results. In this book, we
will therefore not attempt it and simply consider all three to be image processing.
3
But this upsets some people.

You might also like