DIP Unit 1
DIP Unit 1
Digital Image Processing (DIP) refers to the use of computer algorithms to perform image
processing on digital images. It involves manipulating and analyzing images to enhance
their quality or extract useful information.
Newspaper industry
In the 1920s, the Bartlane cable picture transmission system allowed pictures to be sent
between London and New York in less than three hours.
Space exploration
In 1964, the Space Detector Ranger 7 used DIP to send lunar photos back to Earth. These
photos were geometrically corrected, noise reduced, and had their gradation transformed.
This was a major success, allowing for the computerized mapping of the moon’s surface.
The rapid development of digital imaging was made possible by the introduction of MOS
integrated circuits in the 1960s and microprocessors in the early 1970s. These
developments, along with advances in computer memory storage, display technologies,
and data compression algorithms, made digital imaging possible.
❖ Medical Imaging: Used in X-rays, CT scans, MRIs, and ultrasounds for diagnostics,
disease detection, and treatment planning. Image processing techniques help
enhance, segment, and analyze medical images.
❖ Remote Sensing: Utilized in satellite and aerial imaging for environmental
monitoring, land-use mapping, agricultural assessment, and disaster management.
DIP helps to analyze and interpret large-scale spatial data.
❖ Forensics: In crime scene investigations, DIP is applied to enhance images from
surveillance, analyze fingerprints, identify patterns, and improve low-quality images
for evidence collection.
❖ Astronomy: Used in processing images from telescopes and space missions to
study celestial bodies, detect exoplanets, analyze the structure of galaxies, and
identify cosmic phenomena.
❖ Industrial Automation and Quality Control: Employed in automated inspection
systems in manufacturing to identify defects, check measurements, and ensure
product quality using computer vision.
❖ Biometrics: Applied in facial recognition, fingerprinting, iris scanning, and other
biometric identification techniques for security and access control.
❖ Robotics and Autonomous Systems: DIP enables robots and self-driving cars to
interpret visual information from their surroundings for navigation, obstacle
avoidance, and object recognition.
❖ Digital Photography and Multimedia: Used for image enhancement, compression,
and restoration in photo editing software, as well as in the creation of visual effects
in media and entertainment.
❖ Agriculture: Employed in precision farming for crop monitoring, pest detection, and
yield estimation through drone and satellite imagery analysis.
❖ Document Processing: In Optical Character Recognition (OCR) and text extraction,
DIP is used to convert scanned documents into editable and searchable formats.
1. Image Acquisition: Image acquisition is the first step of the fundamental steps of
DIP. In this stage, an image is given in the digital form. Generally, in this stage, pre-
processing such as scaling is done.
2. Image Enhancement: Image enhancement is the simplest and most attractive area
of DIP. In this stage details which are not known, or we can say that interesting
features of an image is highlighted. Such as brightness, contrast, etc…
3. Image restoration: Image restoration is the stage in which the appearance of an
image is improved.
4. Colour image processing: Colour image processing is a famous area because it
has increased the use of digital images on the internet. This includes colour
modeling, processing in a digital domain, etc….
5. Wavelets and Multi-Resolution Processing: In this stage, an image is represented
in various degrees of resolution. Image is divided into smaller regions for data
compression and for the pyramidal representation.
6. Compression: Compression is a technique which is used for reducing the
requirement of storing an image. It is a very important stage because it is very
necessary to compress data for internet use.
7. Morphological Processing: This stage deals with tools which are used for
extracting the components of the image, which is useful in the representation and
description of shape.
8. Segmentation: In this stage, an image is a partitioned into its objects.
Segmentation is the most difficult tasks in DIP. It is a process which takes a lot of
time for the successful solution of imaging problems which requires objects to
identify individually.
9. Representation and Description: Representation and description follow the output
of the segmentation stage. The output is a raw pixel data which has all points of the
region itself. To transform the raw data, representation is the only solution. Whereas
description is used for extracting information’s to differentiate one class of objects
from another.
10. Object recognition: In this stage, the label is assigned to the object, which is based
on descriptors.
11. Knowledge Base: Knowledge is the last stage in DIP. In this stage, important
information of the image is located, which limits the searching processes. The
knowledge base is very complex when the image database has a high-resolution
satellite.
Quantization: Assigning discrete values to the intensity levels of the sampled image. This
step reduces the infinite range of intensity values to a finite range.
Connectivity: How pixels are connected to each other, which is important for defining
regions and boundaries.
Distance Measures: Calculating the distance between pixels, which is used in various
image processing algorithms.
Nonlinear Operations: Operations where the output is not a linear function of the input,
such as thresholding and morphological operations.
These fundamentals form the basis of digital image processing, enabling a wide range of
applications from medical imaging to satellite imagery.
Brightness Adaptation:
The ability of the eye to adjust to different levels of light intensity, meaning the perceived
brightness of an image depends on the surrounding luminance.
Contrast:
The difference in intensity between adjacent areas of an image, which significantly impacts
how details are perceived.
Color Perception:
How the human eye interprets different wavelengths of light, including color mixing and the
ability to distinguish between colors.
Spatial Resolution:
The brain’s tendency to group visual elements based on proximity, similarity, continuity,
closure, and other factors, which can sometimes lead to misinterpretations.
Eye Structure:
Understanding the anatomy of the eye, including the lens, retina, and photoreceptor cells
(rods and cones), is important to comprehend how light is captured and processed.
Radio Waves:
Infrared (IR):
Visible Light:
Ultraviolet (UV):
X-rays:
Gamma Rays:
Image Sensing
Image sensing involves capturing images using various types of sensors. The choice of
sensor depends on the application and the type of electromagnetic energy being detected.
Here are the main types of sensors:
1. Single Sensor:
Used for capturing images one pixel at a time. This method is often used in
applications requiring high precision, such as scientific imaging.
2. Line Sensor:
Consists of a row of sensors that capture one line of an image at a time. Commonly
used in scanners and some types of industrial inspection systems.
3. Array Sensor:
A 2D array of sensors that captures the entire image at once. This is the most
common type used in digital cameras and smartphones.
Image Acquisition
Image acquisition involves converting the captured image into a digital form that can be
processed by a computer. This process includes several steps:
1. Illumination:
The scene is illuminated using a light source. The type of illumination can vary,
including visible light, infrared, X-rays, etc., depending on the application.
2. Energy Reflection or Transmission:
The energy from the illumination source is either reflected off or transmitted through
the objects in the scene. For example, visible light is reflected off objects, while X-
rays pass through the body in medical imaging.
3. Sensing:
The reflected or transmitted energy is captured by the sensor. The sensor converts
this energy into an electrical signal.
4. Digitization:
The electrical signal is converted into a digital form using an analog-to-digital
converter (ADC). This involves sampling the signal at discrete intervals and
quantizing the signal into discrete levels.
5. Image Formation:
The digitized signal is processed to form a digital image. This image can then be
stored, displayed, or further processed.
Image Sampling
Sampling refers to the process of converting the continuous spatial coordinates of an
image into discrete values. Essentially, it involves selecting a grid of points (pixels) from the
continuous image.
Spatial Resolution: The number of pixels used to represent an image. Higher resolution
means more pixels and finer detail.
Example: If you have a continuous image, sampling it at regular intervals will give you a grid
of pixels. For instance, a 1000x1000 pixel image has 1,000,000 sampling points.
Image Quantization
Quantization involves converting the continuous range of intensity values (amplitudes) of
the sampled image into discrete levels.
Intensity Levels: The number of distinct values that a pixel can have. For example, an 8-bit
image can have 256 different intensity levels (0-255).
Example: In a grayscale image, quantization would map the continuous range of gray
shades to a finite number of discrete levels, such as mapping all shades between 0 and
255.
1. Neighbors:
• 4-Neighbors (N4): The pixels that are directly adjacent to a given pixel in the
horizontal and vertical directions. For a pixel at coordinates (x, y), its 4-
neighbors are at (x+1, y), (x-1, y), (x, y+1), and (x, y-1).
• 8-Neighbors (N8): The pixels that are adjacent to a given pixel in both
horizontal, vertical, and diagonal directions. This includes the 4-neighbors
plus the diagonal neighbors at (x+1, y+1), (x-1, y-1), (x+1, y-1), and (x-1, y+1).
2. Connectivity:
• 4-Connectivity: Two pixels are 4-connected if they are neighbors in the 4-
neighborhood.
• 8-Connectivity: Two pixels are 8-connected if they are neighbors in the 8-
neighborhood.
• m-Connectivity (Mixed Connectivity): Combines 4-connectivity and 8-
connectivity to avoid ambiguities in certain situations.
3. Distance Measures:
• Euclidean Distance: The straight-line distance between two pixels. For pixels
at (x1, y1) and (x2, y2), the Euclidean distance is given
by:d=(x2−x1)2+(y2−y1)2
• City Block Distance (Manhattan Distance): The distance measured along
axes at right angles. For pixels at (x1, y1) and (x2, y2), the City Block distance
is:d=∣x2−x1∣+∣y2−y1∣
• Chessboard Distance: The maximum of the absolute differences in the
horizontal and vertical directions. For pixels at (x1, y1) and (x2, y2), the
Chessboard distance is:d=max(∣x2−x1∣,∣y2−y1∣)
4. Adjacency:
• Pixels are adjacent if they share a common edge or corner. Adjacency is
defined based on the type of connectivity (4-connected or 8-connected).
These relationships are essential for various image processing tasks, such as
segmentation, edge detection, and morphological operations.
Linear Operations
Linear operations are those where the output is a linear function of the input. These
operations satisfy the principles of superposition and homogeneity.
Superposition:
If ( f(x) ) and ( g(x) ) are two inputs, and ( a ) and ( b ) are constants, then a linear operation ( T
) satisfies:
T[af(x)+bg(x)]=aT[f(x)]+bT[g(x)]
Homogeneity:
Nonlinear Operations
Nonlinear operations do not satisfy the principles of superposition and homogeneity. The
output is not a linear function of the input, making these operations more complex but
often more powerful for certain tasks.
Applications
❖ Linear Operations: Often used for tasks like noise reduction, edge detection, and
frequency domain analysis.
❖ Nonlinear Operations: Useful for tasks that require shape analysis, noise removal,
and image segmentation.