0% found this document useful (0 votes)
200 views25 pages

CVR Unit 2

Uploaded by

B Sri Ram Vishnu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
200 views25 pages

CVR Unit 2

Uploaded by

B Sri Ram Vishnu
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 25

UNIT-2

Linear Filters: Linear filters, commonly used in digital signal processing and
image processing, operate by applying a linear transformation to an input signal
or image to produce an output.
They are designed to emphasize or reduce certain aspects of the signal or image,
such as smoothing, sharpening, edge detection, or noise reduction.
A linear filter processes each point in a signal (or pixel in an image) as a weighted
sum of its neighbours.
The output of a linear filter is directly proportional to the input: for any inputs
x1 and x2 with scalars a and b H(a.x1 + b.x2) = a.H(x1) + b.H(x2), where H is the
filter operation.
Types of Linear Filters:
1. Low-Pass Filters: These filters allow low-frequency components to pass
while attenuating high-frequency components. They are used for
smoothing or blurring an image, reducing noise or minor detail.
2. High-Pass Filters: These filters pass high-frequency components while
attenuating low-frequency components, emphasizing edges and fine
details.
3. Band-Pass Filters: These filters allow a specific range of frequencies to
pass while blocking frequencies outside that range, which is useful for
applications that need to isolate certain frequency bands.

Convolution Operation

 Linear filtering is typically implemented via convolution, which involves a


kernel (filter) that defines the filter's characteristics.
 In a 1D signal, convolution is defined as:
𝑀

𝑦[n] = ∑ 𝑥 [𝑛 − 𝑘 ]. ℎ[𝑘 ]
𝑘=−𝑀

where x is the input signal, h is the filter kernel, and y is the output.
 In a 2D image, convolution for a pixel located at (i, j) is:
𝑀 𝑁

𝑦[i, j] = ∑ ∑ 𝑥 [𝑖 − 𝑚, 𝑗 − 𝑛]. ℎ[𝑚, 𝑛]


𝑘=−𝑀 𝑘=−𝑁

Where h[m, n] is the kernel, typically a small matrix.

Limitations:

 Linear filters may blur edges or fine details when smoothing.


 They are less effective in handling non-Gaussian noise or non-linear
transformations in data, as they lack adaptability.

Advantages:

 Linear filters are computationally simple and can be efficiently


implemented with convolution.
 They preserve the structure of the signal or image without introducing
complex non-linear artifacts.

Applications of Linear Filters:

 Noise Reduction: Smoothing filters, like the Gaussian filter, reduce noise
by averaging pixel values.
 Edge Detection: High-pass and edge-detecting filters, such as Sobel and
Laplacian, highlight transitions, helping in object detection and boundary
extraction.
 Image Sharpening: High-frequency components are emphasized,
enhancing fine details.
 Data Smoothing in Signal Processing: Low-pass filters are applied to
reduce random noise in time-series data, useful in speech and audio
processing.

Convolution: Convolution is a mathematical operation used extensively in signal


processing, image processing, and machine learning, especially in convolutional
neural networks (CNNs).

 It combines two functions (or signals) to produce a third function that


represents how the shape of one is modified by the other.
 Convolution can be applied in one-dimensional (1D), two-dimensional
(2D), or higher-dimensional data.
 Convolution is a process that merges two functions or signals. In a simple
sense, it “slides” one function over another, calculating their overlap at
each position.
 In digital signal and image processing, convolution is usually performed
between an input signal (or image) and a kernel (or filter) to modify the
original signal in specific ways (e.g., smoothing, sharpening).

Applications of Convolution:

 Image Processing: Convolution is used for tasks like blurring, sharpening,


and edge detection.
 Signal Processing: Convolution filters are applied to extract or suppress
specific frequency components in audio or other time-based signals.
 Convolutional Neural Networks (CNNs): Convolution is the foundation of
CNNs, where it enables the network to learn and recognize patterns like
edges, shapes, and textures, essential for computer vision tasks.
Shift-Invariant Linear Systems: Shift-Invariant Linear Systems, also known as
Linear Time-Invariant (LTI) systems, are fundamental in signal processing,
control systems, and communications.

 These systems are defined by two key properties: linearity and time (or
shift) invariance.
 They are mathematically predictable, stable under certain conditions, and
have straightforward responses to inputs, making them essential in
system analysis and design.
Advantages of Shift-Invariant Linear Systems:

 Predictability: The system’s response to any input can be predicted based


on its impulse response, making LTI systems easy to analyse and model.
 Efficiency: Convolution in LTI systems is computationally efficient,
especially in the frequency domain using the Fourier or Laplace
transforms.
 Stability and Causality: Well-designed LTI systems ensure stable and
causal behaviour, which is essential in real-world applications.
 Versatility in Analysis: LTI systems can be analysed using various
mathematical tools, enabling versatile applications across multiple
domains.

Applications of Shift-Invariant Linear Systems:

 Signal Processing: LTI systems are used to design filters that remove
noise, enhance certain frequencies, or separate signal components.
 Control Systems: LTI models allow engineers to predict system behaviour,
design stable control systems, and ensure desired performance.
 Communications: In communication systems, LTI filters manage
frequency bands, extract signals from noisy environments, and equalize
channels.
 Image Processing: LTI properties enable convolution-based filters to
smooth, sharpen, or detect edges in images consistently.

Spatial Frequency: Spatial Frequency refers to the rate of change in intensity or


variation of patterns in a spatial domain, typically measured across distances in
images or physical spaces.

 In the context of image processing and signal analysis, spatial frequency


describes how rapidly pixel values (intensities) change across an image.
 It is a crucial concept in fields like image analysis, optics, and computer
vision, as it relates to the level of detail and texture in an image.
 Spatial frequency measures the rate at which brightness (or colour)
changes in an image. High spatial frequency corresponds to rapid changes
(e.g., edges, fine textures), while low spatial frequency indicates gradual
changes (e.g., large areas of smooth colour).
 It is expressed in units like cycles per distance (e.g., cycles per millimetre)
in physical measurements, or cycles per pixel in digital images. Each
"cycle" represents a complete repetition of a pattern within the specified
distance.
Types of Spatial Frequency:

1. Low Spatial Frequency: Low spatial frequency components in an image


represent large-scale variations in intensity, such as broad, flat regions or
slow changes in tone. These components capture general shapes and
structures but not fine details. For instance, the overall shape of a large
object in an image without much detail.
2. High Spatial Frequency: High spatial frequency components represent
fine details, such as edges, textures, and sharp transitions. These are
regions where pixel values change quickly over short distances, capturing
intricate details or noise in an image.

Properties of Spatial Frequency:

 Orientation: Spatial frequency can have a directional component. For


instance, vertical or horizontal lines in an image have different
orientations, which correspond to specific directions in the frequency
domain.
 Phase and Amplitude: Each frequency component has a phase, indicating
the position of the frequency in the image, and an amplitude, showing the
strength of that frequency. Together, these determine the appearance of
structures in the reconstructed image.
 Bandwidth: Bandwidth refers to the range of spatial frequencies present
in an image. High-bandwidth images contain a wide range of frequencies,
representing both fine and coarse details, while low-bandwidth images
may only have broad, general patterns.

Fourier Transform: The Fourier Transform (FT) is a mathematical operation that


transforms a signal from its original domain (often time or spatial) to a
representation in the frequency domain.

 This transformation decomposes a signal into its constituent frequencies,


providing insight into the amplitude, frequency, and phase of the
components within the signal.
 The Fourier Transform is foundational in fields like signal processing,
image analysis, audio engineering, physics, and telecommunications.

Properties of Fourier Transforms: Fourier Transforms exhibit a number of


important properties that simplify analysis and computation, particularly in
linear systems:
 Linearity: The Fourier Transform of a linear combination of functions is
the same linear combination of their individual transforms.
 Time and Frequency Shifting: Shifting a function in time results in a phase
shift in the frequency domain. Shifting a function in frequency results in
modulation (multiplying by a complex exponential) in the time domain.
 Convolution Theorem: The Fourier Transform of the convolution of two
functions is the product of their Fourier Transforms. This property is
essential in signal processing, where convolution is used in filtering.
 Duality: If F(t) has a Fourier Transform f(w) then f(t) has a Fourier
Transform F(-w).
 Scaling: Scaling a function in time (compressing or expanding it) inversely
scales it in frequency, affecting its bandwidth.
 Parseval’s Theorem: The total energy in the time domain is equal to the
total energy in the frequency domain. This is critical for signal power
analysis.

Types of Fourier Transforms:

1. Continuous Fourier Transform (CFT): Used for continuous signals,


providing a continuous spectrum of frequencies.
2. Discrete Fourier Transform (DFT): Used for discrete signals, often
sampled data, providing a discrete spectrum of frequencies. It’s widely
applied in digital signal processing and is implemented through the Fast
Fourier Transform (FFT) algorithm for computational efficiency.
3. Fourier Series: For periodic signals, Fourier Series represents a signal as a
sum of sines and cosines. Fourier Transforms generalize this concept to
non-periodic signals.
4. Short-Time Fourier Transform (STFT): Analyses localized time-frequency
information, useful for non-stationary signals where frequency content
changes over time.

Applications of Fourier Transform:

 Audio and Music Processing: Identifying frequency components allows


for equalization, compression, noise reduction, and synthesis.
 Image Processing: Enhances images by filtering certain spatial
frequencies; for example, using Fourier Transforms in MRI, CT scans, and
satellite imagery analysis.
 Telecommunications: Fourier analysis aids in encoding and decoding
signals, as well as analysing bandwidth requirements.

Sampling: Sampling is the process of converting a continuous-time signal into a


discrete-time signal by measuring its amplitude at regular intervals, or
“samples.”

It’s a fundamental concept in digital signal processing, allowing analog signals


(e.g., sound, light, temperature) to be represented digitally for processing,
storage, and transmission.

Key Components in Sampling:

 Sampling Rate (Frequency): This is the rate at which samples are taken
from the continuous signal. Measured in samples per second, or Hertz
(Hz).
 Sampling Interval (Period): The reciprocal of the sampling rate, it’s the
time between consecutive samples, denoted as Ts = 1/fs
 Quantization: After sampling, the analog values are often quantized,
meaning they’re rounded to the nearest level in a discrete set of values
(usually due to limited resolution in digital systems).
Types of Sampling:

1. Uniform Sampling: The signal is sampled at regular, fixed intervals. This is


the most common form of sampling and is used in audio, image
processing, and general signal processing applications.
2. Non-Uniform Sampling: The sampling intervals vary over time. Non-
uniform sampling can be used in cases where certain parts of a signal are
more critical than others.
3. Random Sampling: Samples are taken at random intervals, often used in
applications where a random distribution is beneficial, like compressed
sensing or sparse data acquisition.
4. Adaptive Sampling: Sampling frequency varies depending on the signal’s
characteristics, with higher sampling rates applied in regions with rapid
changes, providing an efficient way to handle signals with varying
frequency components.

Applications of Sampling:

 Audio Processing: Analog audio signals are sampled to create digital audio
files, which are then stored and processed. Common sampling rates for
audio are 44.1 kHz (CD quality) and 48 kHz (professional audio).
 Image Processing: An image is sampled both horizontally and vertically.
Higher sampling rates (i.e., more pixels) result in better image quality.
 Telecommunications: Analog signals are sampled to create digital signals,
allowing them to be transmitted and processed more effectively in digital
communication systems.
 Medical Imaging: In devices like MRI or CT scanners, signals are sampled
to create digital images of the body.

Aliasing: Aliasing is a phenomenon in signal processing where different signals


become indistinguishable from each other when sampled, resulting in distortion
or artifacts in the reconstructed signal.

 This typically occurs when the sampling rate is too low to accurately
capture the high-frequency content of a continuous signal, causing higher
frequencies to "fold" into the lower frequencies and create misleading
representations.
 When a continuous signal is sampled, the discrete samples capture
information about the signal’s frequency content.
 However, if the sampling rate is insufficient (below twice the highest
frequency of the signal), certain high-frequency components will overlap
or "fold back" into lower frequencies, distorting the signal.
 This effect makes it challenging to distinguish the original frequencies
from the alias frequencies, causing errors in signal analysis and
reconstruction.

Types of Aliasing:

1. Temporal Aliasing: Occurs in time-domain signals like audio and video. It


causes high-frequency components to show up as lower frequencies in
the time domain. For instance, in audio, this results in frequency
distortion, while in video, it can cause visual artifacts.
2. Spatial Aliasing: Happens in image processing when sampling spatial data
(like pixels in a photo). This can produce a jagged, blocky appearance
known as the "moiré effect." A common example is when a high-
frequency pattern, like a striped shirt, appears to have unwanted bands
or patterns on a camera.

Filters as Templates: "Filters as Templates" refers to the use of filters in image


processing, computer vision, and signal processing to detect specific patterns,
shapes, or features within data.

 Filters act as "templates" that match certain characteristics in the input


signal, such as edges, textures, or repetitive patterns, allowing for
selective enhancement or detection of these features.
 By adjusting the filter design, we can shape it to respond optimally to
certain spatial or frequency components, making it an essential tool for
image analysis and feature detection.

Types of Template Filters: Filters used as templates can be broadly divided into
categories based on the feature they’re designed to match:

 Edge Detectors: These filters emphasize edges or transitions in intensity


within an image. Examples include the Sobel and Prewitt filters, which
detect gradients along specific directions.
 Corner Detectors: These filters detect corners and intersections in an
image, which are useful for feature matching and tracking. An example is
the Harris corner detector.
 Pattern or Shape Detectors: Filters designed to recognize specific shapes
or patterns, such as circles, lines, or blobs, within an image.
 Frequency Filters: Some filters target specific frequency components to
emphasize or suppress patterns that occur at certain scales (e.g., high-
frequency details or low-frequency background information).
Applications of Filters as Templates: Filters as templates are widely used for
feature detection, pattern recognition, and enhancement in various
applications, including:

 Edge Detection: Detecting edges in an image using templates like Sobel


or Canny filters, where the filter emphasizes areas of rapid intensity
change, indicating edges.
 Shape Detection: Matching specific shapes like circles, lines, or other
geometric patterns using specialized templates to detect objects or
boundaries within an image.
 Texture Analysis: Detecting textures or repetitive patterns, such as brick
or weave patterns, which might have distinct spatial frequencies that
specific filters can target.
 Image Recognition and Matching: Using filters as templates for object or
face detection, where a filter shaped like a face, eye, or other features can
identify similar patterns in the image.

Edge Detection:

Noise: n image processing, signal processing, and communication systems,


"noise" refers to unwanted or random variations that interfere with the desired
signal, degrading its quality and often obscuring important features.

Noise can arise from a variety of sources, including environmental interference,


sensor limitations, electronic circuitry, or transmission errors, and it can appear
in different forms, such as random grainy patterns in images or static in audio
signals.

Sources of Noise: Noise can enter a system through various stages in signal
processing, with primary sources being:

 Environmental Interference: External factors, such as electromagnetic


interference from electronic devices, solar radiation, or atmospheric
disturbances, can introduce noise.
 Sensor Limitations: Physical limitations in imaging sensors, microphones,
or other sensors lead to inherent noise (like thermal noise) in the captured
signal.
 Electronic Circuitry: Random fluctuations in electrical circuits, particularly
in amplifiers and other components, generate noise. Thermal and
Johnson-Nyquist noise are common examples.
 Data Compression and Transmission: Lossy compression and errors in
data transmission add noise, causing data to lose quality and fidelity.

Noise Characteristics: Noise can be described and analysed by its statistical


properties, such as mean, variance, and distribution. Some key characteristics
include:

 Mean and Variance: Mean measures the average noise level, while
variance measures the spread (or power) of noise. In Gaussian noise,
mean is zero, and variance determines the noise intensity.
 Probability Distribution: Defines the likelihood of various noise
amplitudes occurring. Common distributions include Gaussian, Poisson,
and uniform.
 Power Spectral Density (PSD): Describes how noise power is distributed
across different frequencies. White noise has a constant PSD across all
frequencies, while coloured noise has frequency-dependent
characteristics.
 Temporal and Spatial Properties: Noise can vary over time (temporal
noise) or across space (spatial noise), such as in a video or image sequence
where noise fluctuates frame by frame.

Estimating Derivatives: Estimating derivatives is an important concept in


numerical analysis, particularly in image processing, signal processing, and other
computational fields. Derivatives provide critical information about the rate of
change of a function and are essential for tasks like edge detection in images,
detecting trends in data, and performing optimizations.
Applications of Derivative Estimation:

 Image Processing: First derivatives are commonly used to detect edges,


while second derivatives help in locating corners and contours. Gradient
magnitudes and directions indicate edge strength and orientation, which
are essential for object detection and image segmentation.
 Signal Processing: Derivatives highlight changes in signals, identifying
trends, peaks, and rapid transitions (e.g., ECG waveform analysis).
 Physics and Engineering: Derivatives represent quantities like velocity
and acceleration, essential for motion analysis in mechanics.
 Optimization: Derivatives are used to find maximum and minimum values
of functions, critical in machine learning, economics, and engineering.

Detecting Edges: Edge detection is a fundamental task in image processing and


computer vision that involves identifying points in an image where there is a
significant change in brightness or intensity.

These changes in intensity represent boundaries between objects and


backgrounds, different regions, or textures in an image.

Edge detection simplifies the analysis of images by reducing data and preserving
structural properties, helping in tasks like:

 Object Detection: Identifying objects based on their boundaries.


 Image Segmentation: Dividing an image into meaningful regions based on
edges.
 Feature Extraction: Extracting essential features for tasks such as facial
recognition or tracking.
 Scene Interpretation: Determining object locations and relationships
within an image.

Challenges in Edge Detection:

 Noise Sensitivity: Edges in noisy images can be harder to detect


accurately. Techniques like Gaussian smoothing are often used to
mitigate noise before edge detection.
 False Edges: Textures and minor variations in pixel intensity can produce
false edges.
 Complex Shapes and Gradients: Gradual changes in intensity are harder
to detect with simple methods; multi-scale methods or advanced
algorithms like Canny are more effective.
 Choice of Parameters: Many edge detectors require parameter tuning
(e.g., threshold values in Canny), and the best values can vary with image
content.

Performance Metrics for Edge Detection:

 Accuracy: The ability of the detector to identify true edges and avoid false
detections.
 Localization: The accuracy of edge placement, ensuring edges are
detected close to their true positions.
 Sensitivity to Noise: How well the method performs in the presence of
noise, which can produce false edges.
 Computational Efficiency: How quickly the method processes an image,
essential in real-time applications.

Texture: Texture refers to patterns, structures, or variations in pixel intensities


or colours within an image that create a sense of surface or structure.

 Textures are usually characterized by repeated spatial patterns, variations


in intensity, or colour that give the appearance of a material’s surface.
 Texture is an important visual cue used in identifying objects, recognizing
patterns, segmenting images, and analysing surfaces.
 Texture is the perceived surface quality or "feel" of an object, represented
visually in an image through repetitive patterns or structures.
 In images, textures often exhibit regularities in patterns, spatial
frequencies, and statistical properties.

Approaches to Representing Texture: There are several methods for


representing texture in images, each suited to different applications and levels
of computational complexity. The most common methods fall into four broad
categories: statistical, structural, model-based, and transform-based methods.

1. Statistical Methods: Statistical methods analyse the distribution and


relationships of pixel intensities within an image, providing features to
characterize texture without needing precise knowledge of its structure.

 Gray-Level Co-occurrence Matrix (GLCM): GLCM is a matrix that


represents the frequency of specific pixel intensity pairs occurring at a
certain distance and orientation. It captures spatial relationships between
pixels, giving information about texture’s roughness, contrast, and
directionality.
 Local Binary Patterns (LBP): LBP is a simple and efficient texture
descriptor that labels each pixel based on the relationship between its
intensity and that of its neighbours. It converts the local area around a
pixel into a binary pattern by comparing the centre pixel to its surrounding
pixels.

2. Structural Methods: Structural methods describe texture based on


identifiable structures, shapes, or repeated patterns in an image. They are more
explicit but work best with textures that have clearly defined regularity or
repetition.

 Primitive and Placement Model: Defines textures by identifying basic


patterns or primitives (e.g., lines, dots) and their placement rules.
Patterns are then characterized by the spatial arrangement, orientation,
and frequency of these primitives.
 Textons: Textons are fundamental micro-patterns, like spots, edges, or
blobs, that make up a texture. These are used to classify textures by
comparing arrangements of textons in an image.

3. Model-Based Methods: Model-based methods represent texture by creating


a model that can generate or simulate the texture’s appearance.

 Markov Random Fields (MRF): MRF is a probabilistic model that


represents the dependencies of pixels based on neighbouring pixels. It
captures the local interactions of pixels and is particularly effective for
capturing subtle textural patterns and spatial dependencies.
 Fractal Models: Fractals represent textures that exhibit self-similarity or
scale invariance. Fractal models are useful for natural textures like
mountains, trees, or clouds, where patterns recur across different scales.

4. Transform-Based Methods: Transform-based methods analyze textures by


examining the frequency domain rather than the spatial domain, often using
transformations to reveal patterns in the spatial frequency of an image.

 Fourier Transform: The Fourier Transform decomposes an image into its


frequency components, allowing for analysis of textures based on their
spatial frequencies. Textures with periodic patterns have strong peaks in
the frequency domain, while random textures have more dispersed
frequency representations.
 Gabor Filters: Gabor filters are tuned to specific frequencies and
orientations, making them highly effective for capturing textures at
various scales and directions. Applying Gabor filters produces a set of
feature maps that describe the texture’s frequency and orientation
characteristics.

Analysis and Synthesis Using Oriented Pyramids: Analysis and Synthesis Using
Oriented Pyramids is a technique in image processing used to analyse and
reconstruct images by decomposing them into multiple levels, each capturing
different orientations and spatial scales.

 Oriented pyramids are an extension of the traditional Gaussian or


Laplacian pyramid, but instead of decomposing only by scale, they also
incorporate orientations, making them particularly useful for capturing
directional features in textures and images.
 This method has applications in a variety of fields, including texture
analysis, image compression, object recognition, and image synthesis, as
it allows for the representation of images in terms of both frequency
(scale) and direction (orientation) information.

Key Concepts of Oriented Pyramids:

1. Multi-Resolution Analysis:

 Traditional image pyramids (like Gaussian and Laplacian pyramids) create


a sequence of images that progressively reduce in resolution.
 Each level in the pyramid represents the image at a coarser resolution,
capturing lower-frequency components of the image.

2. Multi-Orientation Analysis:

 In addition to scale, oriented pyramids decompose images according to


direction.
 This decomposition is achieved by applying directional filters, each tuned
to a specific orientation (e.g., horizontal, vertical, and diagonal).
 Multi-orientation analysis enables the oriented pyramid to capture
directional patterns, which are essential for analysing and synthesizing
textures, edges, and other features with strong orientation
characteristics.
3. Representation in Terms of Scales and Orientations:

 By representing the image in both scales and orientations, oriented


pyramids provide a more comprehensive image representation.
 Each level of the pyramid is further divided into orientation subbands,
where each subband captures specific directional details at a particular
resolution.

Structure of Oriented Pyramids: An oriented pyramid is typically constructed


through the following steps:

1. Decomposition:

 The image is filtered and down sampled to create multiple levels, similar
to a standard pyramid.
 At each level, the image is also decomposed into different orientations
using directional filters (e.g., Gabor filters or steerable filters), resulting in
orientation subbands.
 These subbands capture information at different directions, such as edges
and textures aligned with each specific orientation.
 This structure results in a pyramid where each level has multiple
subbands, each corresponding to a particular orientation.

2. Low-Pass and High-Pass Components:

 Similar to other pyramids, oriented pyramids separate low-frequency


(smooth) components from high-frequency (detail) components at each
level.
 The low-pass component is down sampled and carried to the next level of
the pyramid, while the high-pass component captures details specific to
that level’s orientation subbands.
 This stepwise decomposition enables the pyramid to retain information
across scales, allowing analysis and synthesis across the full image
resolution.

3. Filters for Orientation:

 Common filters for oriented pyramids include steerable filters and Gabor
filters, both of which are effective in isolating specific orientations.
 Steerable Filters: These can be "steered" or rotated to match different
orientations, making them versatile in capturing specific directional
information.
 Gabor Filters: These filters are also tuned to particular frequencies and
orientations and are commonly used in texture analysis due to their
resemblance to human visual receptive fields.

Analysis Using Oriented Pyramids: In analysis, the oriented pyramid is used to


extract information from an image by decomposing it into components that
reveal the image's scale- and direction-based structure.

1. Feature Extraction:

 By analysing different subbands at each level, features corresponding to


edges, textures, and structures in specific orientations and scales can be
extracted.
 These features are valuable for tasks like object recognition, texture
classification, and edge detection, where both scale and orientation play
a role.

2. Pattern Recognition:

 Oriented pyramids help recognize patterns that exhibit orientation and


frequency characteristics, aiding in identifying objects or regions with
specific directional textures.
 In medical imaging, for instance, oriented pyramids can help detect edges
and textures that may indicate specific tissue types or abnormalities.

3. Multi-Resolution Analysis:

 Because oriented pyramids capture multiple resolutions, they enable the


examination of fine details and coarse structures simultaneously.
 This multi-resolution feature helps in analysing complex images with
varying degrees of detail, such as natural scenes or textured materials.

Synthesis Using Oriented Pyramids: Synthesis involves reconstructing an image


using the information captured in the oriented pyramid. The goal is to recreate
the original image, or a modified version, using the multi-scale and multi-
orientation data from the pyramid decomposition.

1. Reconstruction:

 To reconstruct the image, each orientation subband is upsampled and


combined with the low-pass component at each level.
 This process is repeated up the pyramid, with each level's high-pass
details and orientation subbands added back in to recreate the finer
details progressively.
 By combining the low-frequency and orientation-specific high-frequency
components, a complete image can be synthesized.

2. Texture Synthesis:

 Oriented pyramids are highly effective for texture synthesis, as textures


often have strong directional patterns.
 By adjusting the values in specific orientation subbands, textures with
desired characteristics can be created, which is valuable in applications
like graphic design and 3D modelling.

3. Image Compression:

 In image compression, oriented pyramids enable the retention of


important structural details while discarding redundant information.
 Compression is achieved by selectively retaining components that contain
essential details, reducing the amount of data needed to represent the
image.

Advantages of Oriented Pyramids:

 Directional Sensitivity: The oriented pyramid captures details in specific


orientations, providing a better representation of edges, textures, and
structures compared to standard pyramids.
 Multi-Resolution Representation: By decomposing the image into
multiple scales, oriented pyramids enable analysis and synthesis at
different levels of detail.
 Efficient Reconstruction: Due to the structured decomposition and the
retention of essential features, oriented pyramids facilitate efficient and
accurate image reconstruction.
 Versatility in Applications: Oriented pyramids are widely used in texture
synthesis, pattern recognition, image compression, and image
enhancement tasks.
Synthesis by Sampling Local Models: Synthesis by Sampling Local Models is a
technique used primarily in texture synthesis and image generation, where new
images or textures are created by sampling and replicating local structures from
an existing image.

 Instead of recreating the texture or image from scratch, this method


samples local neighbourhoods or patches in the source image and
combines them in a way that captures the original’s appearance.
 This approach is based on the assumption that textures or images can be
generated by replicating and blending these local patches in new
configurations.

Key Concepts in Synthesis by Sampling Local Models:

1. Local Models:

 Local models represent small, self-similar patches in a texture or image


that contain the fundamental structural information of the texture.
 For textures that are stochastic or random, local patches appear similar
to each other, while in structured textures, the model captures repeating
patterns or features.

2. Sampling and Neighbourhood Matching:

 The process of synthesis by sampling involves selecting patches from the


source image based on a neighbourhood matching criterion.
 When synthesizing a new texture, each pixel or patch in the new image is
generated by matching it to similar patches in the original image, ensuring
the new patch fits the existing structure.
 Matching criteria often involve similarity measures like mean squared
error or other distance metrics that compare texture and pattern
information.

3. Overlap and Blending:

 In patch-based synthesis, overlapping regions are crucial for seamless


texture transitions.
 Overlapping helps avoid visible seams and create a continuous texture.
 A common blending technique, such as minimum error boundary cut, can
be used to blend overlapping areas, minimizing noticeable transitions
between patches.
4. Sampling Patterns:

 Sampling patterns affect the quality and randomness of the synthesized


texture.
 For example, seamless tiling ensures that textures can repeat without
visible boundaries, making them ideal for backgrounds in applications like
video games or computer graphics.
 Hierarchical synthesis can also be used, where large patterns are
generated first, followed by filling in smaller details, ensuring consistency
in scale.

Steps in Synthesis by Sampling Local Models

1. Initialize the Output Texture:

 Start with an initial patch or sample, which can be a single pixel or a small
patch taken from the source image.
 This patch serves as the foundation, from which the texture is
progressively synthesized.

2. Iterative Patch Selection:

 For each new patch, find a region in the source texture with the highest
similarity to the neighbourhood of the region being synthesized.
 The selection process involves comparing neighbourhoods using a
similarity metric (like mean squared error) to find the best match.

3. Place and Blend the Patch:

 Once a suitable patch is found, place it in the new texture, ensuring


overlap with existing regions.
 Use blending or boundary cut techniques to minimize visible seams in the
overlap area, allowing the patches to merge smoothly.

4. Repeat the Process:

 Repeat patch selection and placement until the entire target texture or
image is filled.
 If required, hierarchical structures can be synthesized by applying this
process from larger to smaller patches, or by progressively filling in
smaller details.
Types of Textures and Models for Sampling

1. Stochastic Textures:

 Stochastic textures have random but statistically similar patterns, like


grass or sand.
 For these, small patches suffice since the lack of strong structure allows
for a broader selection of potential matches.

2. Structured Textures:

 Structured textures have regular patterns or geometrical structures, like


bricks or tiles.
 For structured textures, larger patches are often more effective, as they
help capture the repeated patterns in the correct alignment.

3. Procedural Models:

 In some cases, procedural models may be used as the source texture,


where the synthesis method samples from a generative model instead of
a fixed texture.
 This is often employed for synthesizing textures on the fly, such as in 3D
rendering and virtual environments.

Shape from Texture: Shape from Texture is a technique in computer vision and
image processing used to infer the 3D shape or surface geometry of an object
from variations in texture observed in a 2D image.

 The approach leverages the way textures distort on surfaces due to


perspective, scale, and orientation to estimate the depth, orientation, and
curvature of surfaces within a scene.
 This concept is grounded in the fact that as a surface recedes into the
distance or curves away, the texture elements (or texels) appear to
change in size, orientation, and density due to foreshortening and
perspective effects.
Key Concepts in Shape from Texture:

1. Texture Gradients:

 Texture gradients refer to changes in the texture pattern across a surface.


 Gradients in size, density, orientation, and spacing provide cues about the
surface’s shape and depth.

2. Perspective Distortion:

 As objects move further away in perspective, texture elements appear to


shrink and become more compact due to foreshortening.
 By analysing the rate and nature of this distortion, algorithms can
estimate the surface’s tilt, curvature, and orientation.

3. Spatial Frequency and Scale:

 The spatial frequency of textures (i.e., the number of texture elements


per unit area) changes with distance.
 High-frequency textures (small, densely packed features) appear more
compact when observed at an angle or from afar, indicating surface slant
and depth.
 By analysing how texture scales across the surface, algorithms can infer
3D shape information.

Techniques in Shape from Texture:

1. Texture Gradient Analysis:

 This method involves analysing the distribution and pattern of texture


gradients, such as size, orientation, and density, to understand surface
curvature.
 Gradients are calculated and then translated into shape estimations,
assuming the texture follows a known pattern.

2. Statistical Texture Analysis:

 Statistical models examine variations in texture properties, like density


and distribution, to estimate surface shape.
 Features such as average size, aspect ratio, and alignment of texels are
analysed to deduce surface orientation and depth.
3. Fourier Transform-Based Analysis:

 Fourier transforms can be used to analyze the frequency components of


textures on surfaces.
 By analysing how frequency patterns vary across an image, one can infer
surface curvature and orientation based on frequency compression in
particular directions.

4. Markov Random Fields (MRFs):

 MRFs model spatial relationships between neighbouring texels, making


them well-suited to describe texture distortions caused by changes in
surface shape.
 By training MRFs on sample textures, systems can recognize shape-
induced distortions and estimate 3D surface geometry.

5. Shape-from-Texture with Machine Learning:

 Machine learning models, especially convolutional neural networks


(CNNs), can be trained on large datasets of textures applied to surfaces of
known shapes.
 These models learn to recognize complex patterns in texture variation
that correspond to particular surface curvatures and orientations, making
shape-from-texture estimations highly accurate.

Practical Steps in Shape from Texture:

1. Pre-processing: Enhance texture features by removing noise and


normalizing contrast to ensure that variations are due to the surface
rather than image quality. Edge detection may also be applied to outline
texture elements.
2. Texture Gradient Calculation: Compute gradients in size, orientation, and
density across the texture. Analyse gradient fields to assess how texture
elements are distributed, with regions of higher gradient intensity
indicating changes in surface orientation.
3. Estimate Surface Orientation and Depth: Use the texture gradients to
estimate slant and tilt angles for each area. Integrate these local shape
estimates to construct a global 3D shape representation of the object.
4. Refinement: Refine the shape estimate by combining it with other visual
cues, such as shading or stereo vision if available, for a more accurate
reconstruction.

You might also like