Watershed Transform
Watershed Transform
Related terms:
The most widespread variant employs the gradient image (recall § 4.2.6) as the basis
for the watershed transform. Gradient magnitude, however, is strongly sensitive to
image noise. Therefore, appropriate filtering is essential. There are many variants
how the watershed transform may be used as a basis for a general segmentation
approach. The “over-segmentation” problem may be solved by some criteria for
merging regions. The user must be provided with some facilities to influence the
merging process. We describe the approach introduced by Hahn and Peitgen [2000]
and later refined by Hahn and Peitgen [2003].
Merging Basins The decomposition of an image into regions is the basis for merging
them. In the metaphorical sense of a landscape, catchment basins are merged at
their watershed locations by flooding them. While some regions merge early (with
low flooding level), other regions are merged later (see Fig. 4.15). In order to support
interactive merging, Hahn and Peitgen [2003] introduced a merge tree. This tree
consists of the original catchment basins as leafs and of intermediate nodes that
represent merging events. A merging event is characterized by the nodes that are
merged and by the flood level that is necessary for merging. As a first step, a certain
amount of flooding may be applied (“pre-flooding” which may already be sufficient
for segmenting the target structure [Hahn and Peitgen, 2000]).
Figure 4.15. Principle of the watershed transform where the intensity values define
hills and basins. For segmentation purposes, basins may be flooded in order to
combine corresponding regions. As result, a merge tree arises consisting of merge
events (From: [Hahn and Peitgen, 2003]).
Figure 4.16. The image data is considered as a landscape and flooded to merge
small basins. To prevent merging, a watershed is erected between an include and
an exclude point specified to the user.(Courtesy of Horst Hahn, Fraunhofer MEVIS
Bremen)
A traditional way to obtain regional features is to use prior knowledge, that is,
predefined ROIs, which summarizes all voxel-wise features in each predefined ROI.
However, this method is inappropriate in the case of using multiple templates for
complementary representation of brain images, since in this way ROI features from
multiple templates will be very similar (we use the volume-preserving measurement
to calculate the template-specific morphometric pattern of tissue density change
within the same ROI w.r.t. each different template). To capture different sets of
distinctive brain features from different templates, a clustering method (Fan et al.,
2007) is adopted for adaptive feature grouping. Since clustering will be performed
on each template space separately, the complementary information from different
templates can be preserved for the same subject image. As indicated in Fan et
al. (2007), the clustering algorithm can improve the discriminative power of the
obtained regional features, and reduce the negative impacts from registration errors.
Let denote a voxel-wise tissue density value at voxel u in the kth template for the
ith training subject, i [1, N]. The ROI partition for the kth template is based on
the combined discrimination and robustness measure, DRMk(u), computed from
all N training subjects, which takes into account both feature relevance and spatial
consistency as defined below:
(9.1)
where Pk(u) is the voxel-wise Pearson correlation (PC) between tissue density set
and label set {yi [−1, 1], i [1, N]} (1 for AD and −1 for NC) from all N training
subjects, and Ck(u) denotes the spatial consistency among all features in the spatial
neighborhood (Fan et al., 2007).
Watershed segmentation is then performed on each calculated DRMk map for ob-
taining the ROI partitions for the kth template. Note that, before applying watershed
segmentation, we use a Gaussian kernel to smooth each map DRMk, to avoid any
possible oversegmentation, as also suggested in Fan et al. (2007). As a result, for
example, we can partition the kth template into totally Rk nonoverlapping regions, ,
with each region rlk owning voxels. It is worth noting that each template will yield
its own unique ROI partition, since different tissue density maps (of same subject)
are generated in different template spaces.
Fig. 9.4 shows the partition results obtained from the same group of images reg-
istered to the two different templates. It is clear that the obtained ROIs are very
different, in terms of both their structures and discriminative powers (as indicated
by different colors). Those differences will naturally guide the subsequent steps of
feature extraction and selection, and thus provide the complementary information
to represent each subject and also improve its classification.
Fig. 9.4. Watershed segmentation of the same group of subjects on two different
templates. Color indicates the discriminative power learned from the group of
subjects (with the hotter color denoting more discriminative regions). Upper row:
two different templates. Lower row: the corresponding partition results.
Each regional feature is then normalized to have zero mean and unit variance, across
all N training subjects. Finally, from each template, M (out of Rk) most discriminative
features are selected using their PC. Thus for each subject, its feature representation
from all K templates consists of M × K features, which will be further selected for
classification. Fig. 9.5 shows the top 100 regions selected using the regional feature
aggregation scheme, for the same image registered to two templates (as shown in
Fig. 9.4). It clearly shows the structural and discriminative differences of regional
features from different templates.
Fig. 9.5. Illustration of the top 100 regions identified using the regional feature
aggregation scheme, where the same subject is registered to two different templates.
The axial, sagittal, and coronal views of the original MR image of the subject after
warping to each of the two different templates are displayed. Color indicates the
discriminative power of the identified region (with the hotter color denoting more
discriminative region). Upper row: image registered to template 1. Lower row: image
registered to template 2. (For the definitions of both hetero-M and homo-M, please
refer to Section 9.2.4.3.)
8.6 Proposed Method for Red Blood Cell Detection and Seg-
mentation
The controlled watershed method with morphological operation is robust and
efficient for early segmentation algorithms as reported by Refs. [13,18]. It is
fundamentally different from conventional segmentation tools as it uses edge de-
tectors to connect to pixel component extractors. The solution to rectify the under-
and over-segmentation problems is by utilizing watershed segmentation to divide
images into unique regions based on their regional minima. For overlap blood cell
images, watershed segmentation is very effective with the use of a marker [19].
I(p) defines the gray level value of p, ne is the neighbor pixel of p, and I(ne) represents
gray level value of the next respective pixel. The array d[p] is used to store the distance
from the lowest pixels or plateau. However, the array l[p] is used to stock the labels.
DMAX and LMAX indicate the maximum distance and maximum value for the label
in the structure, respectively. DMAX determines the distance between the first pixels
of the first row to the last pixel of the last row. Image scanning can be continued
for step two if the v[p] array structure does not change. The image scanning can
continue if l[p] array structure does not change.
> Read full chapter
In our implementation, an UCM (Arbelaez et al., 2011; Arbelaez, 2006), which defines
a duality between closed, nonself-intersecting weighted contours and a hierarchy
of regions, is used to generate a pool of segmentation candidates. Because of the
nice property of UCM where the segmentation results using different thresholds are
nested into one another, we can construct a tree graph for this pool of segmentation
candidates. The final step is to solve this tree graph-based problem using dynamic
programming.
(4.15)
where denotes all possible valid configurations of x. Considering the special tree
graph structure, we can efficiently solve (4.15) via the dynamic programming
approach with a bottom-up strategy.
In order to ensure that solving (4.15) selects the desired regions, each candidate
region (node) must be assigned an appropriate muscle cell likelihood score w. In
our algorithm, each candidate region is discriminatively represented with a feature
vector that consists of a descriptor to model the convexity of the shape, and two
histograms to describe the gradient magnitudes of the pixels on the cell boundary
and inside the cell region. These morphological features are proposed based on the
following observations: (1) the shape of a muscle cell is nearly convex; (2) the cell
boundaries often exhibit higher gradient magnitudes; (3) the intensities within the
cell regions should be relatively homogeneous.
This figure starts by an example of original RGB image with all different objects,
many morphological functions and filters (edge detection, erosion, dilation, de-
termination the number of objects, watershed segmentation, etc.) are applied to
enhance the work of this image. The areas, centroids, orientations, eccentricities,
convex areas for every object can easily be determined. Moreover, the boundary
points (xij, yij) for each object are calculated individually, where i represents the
number of objects and j is the number of boundary points related to an object. These
boundary points and the objects' previous information are saved to start construction
of own proposed signature for every object based on all these information. The
relation among all these information and Euclidian distance from objects' centroids
(xtc,ytc) is plotted and saved as an individual signature for each object, that is shown
in Figure 1 by one object. These signatures for all objects are saved and waiting for
matching with any input object's signature, as in Section 5. Moreover, the contour
is drawn around all objects and tracing the exterior boundaries of them.
The third part of this proposed algorithm after segmentation and constructing
signatures is the matching process between input and saved objects' signatures as in
the above. Two different ways in matching process have used to make a comparison
between them in accuracy and activity; one is using statistical measures related to
the signatures and the second is using SURF [22].
Additionally, all shown steps in Figure 1 have applied on the input object to construct
its signature. Actually, the matching process is depending on statistical measures on
both types of objects' signatures (saved ones and input). Firstly, as shown in Figure 1,
not all objects in the example image are in the same size, orientation, or even shape,
and afterwards, their data should not be equal in length or characteristics. For that
reason and more in checking accuracy, some preprocesses of matching have carried
out; one is sorting all the data of all signatures, and then computing the variability
in the dataset by calculating the average of the absolute deviations of data points
from their mean. The equation for average deviation is
(4)
For all i's are the number of objects in the image and j's are the number of object's
signature data, xij represents the number of signature's data points, is their mean,
and ni is the number of signature data rows. Secondly, the results of Equation (4)
have applied on all input and saved objects to make a comparison between them for
get the exact matching by least error. Equation (5) introduces a method for calculate
differences between the results of Equation (4):
(5)
The components of Equation (5) are the absolute value of the difference between the
two results of Equation (4) related to saved and input object signatures. The decision
of matching based on the least value of DIF, which is given the exact matched object.
2 Related work
In recent years, research on the blood cells' detection and diagnosis of diseases using
image processing has grown rapidly. In 2010, Y M Hirimutugoda and Gamini Wija-
yarathna presented a method to detect thalassaemia and malarial parasites in blood
sample images acquired from light microscopes and investigated the possibility of
rapid and accurate automated diagnosis of red blood cell disorders. To evaluate
the accuracy of the classification in the recognition of medical image patterns,
they trained two back propagation Artificial Neural Network models (3 layers and
4 layers) together with image analysis techniques on morphological features of
RBCs. The three layers had the best performance with an error of 2.74545e-005 and
86.54% correct recognition rate. The trained three layer ANN acts as a final detection
classifier to determine diseases [7].
In December 2012, D.K. DAS, C. CHAKRABORTY, B. MITRA, A.K. MAITI, and A.K.
RAY introduced a methodology using some techniques of machine learning for
characterizing RBCs in anemia based on microscopic images of peripheral blood
smears. In the first, for reducing unevenness of background illumination and noise
they preprocessed peripheral blood smear images based on geometric mean f-
ilter and the technique of gray world assumption. Then watershed segmentation
technique applied to erythrocyte cells. The distorted RBCs, such as, sickle cells,
echinocyte, tear drop, acanthocyte, elliptocyte, and benign cells have been classified
as dependent on their morphological shape changes. They observed that when a
small subset of features used by using information gain measures, the logistic
regression classifier presented better in performance. They achieved highest pre-
diction in terms of overall accuracy by 86.87%, sensitivity to 95.3%, and specificity
was 94.13% [8].
Also in May 2013, Khan and Maruf presented an algorithm for cell segmentation and
counting via the detection of cell centroids in microscopic images. Their method was
specifically designed for counting circular cells with a high probability of occlusion.
The experimental results showed an accuracy of 92% of cell counting, even at around
60% overlap probability [10].
There is also an eikonal interpretation of the watershed. Najman and Schmitt [59]
have established that (in the continuous domain and assuming that the image is
sufficiently smooth and has isolated critical points) the continuous watershed
is equivalent to finding a skeleton by influence zones with respect to a weighted
distance function that uses points in the regional minima of the image f as sources
and || f|| as the field of indices. A similar result has been obtained by Meyer [54] for
digital images. In Maragos and Butt [47] the eikonal PDE modeling the watershed
segmentation of an image-related function f has been solved by finding a WDT
via level sets and curve evolution where the curve's normal speed is proportional
to 1/|| f||. Specifically, if is a curve representing a marker boundary, then the
propagation of the corresponding flooding wave is modeled by
(81)
Thus, the curve speed is proportional to 1/|| f||. Related ideas can be found in
[76]. Subsequently, Nguyen et al. [60] proposed a PDE segmentation approach
that combines the watershed ideas with active contours. Namely, they used curve
evolution as in (81) but with a velocity V = c0/|| f||- c1 that contains both the eikonal
speed 1/|| f|| and a term proportional to curvature that smoothens the evolving
boundaries. Further, they related this combined curve evolution to minimizing an
energy functional.
The stationary eikonal PDE || T|| = 1/V corresponding to (81), where T is the curve's
arrival time, was solved in [47] using the FMM. Further, the results of this eikonal
PDE segmentation approach were compared with the digital watershed algorithm
via flooding [83] and to the eikonal approach solved via a discrete chamfer WDT
[54, 81]. In all three approaches, robust features are extracted first as markers
of the regions, and the original image I is transformed to another function f by
smoothing via alternating opening-closings, taking the gradient magnitude of the
filtered image, and changing (via morphologic reconstruction) the homotopy of the
gradient image so that its only minima occur at the markers. The segmentation is
done on the final outcome f of the above processing.
In the standard digital watershed algorithm [8, 83], the flooding at each height
level is achieved by a planar distance propagation that uses the chessboard metric.
This kind of distance propagation is not isotropic and could give incorrect results,
particularly for images with large plateaus. Eikonal segmentation using chamfer
WDTs improves this situation a little but not entirely. In contrast, for images with
large plateaus/regions, segmentation via the eikonal PDE and curve evolution WDT
gives results close to ideal. As Fig. 11 shows, compared on a test image that is
difficult (because expanding wavefronts meet watershed lines at many angles
ranging from being perpendicular to almost parallel), the continuous segmentation
approach based on the eikonal PDE and curve evolution outperforms the discrete
segmentation results [47].
The standard watershed flooding, both in the digital algorithms [8, 83] as well as in
the eikonal curve evolution (81), corresponds to a uniform height flooding. This height
criterion corresponds to a contrast-based segmentation. Alternative criteria, such as
area (size-based segmentation) or volume (contrast and size-based segmentation)
and corresponding generalized floodings have been considered in [55]. In [75] the
uniform volume flooding was modeled via a time-dependent eikonal-type PDE and
level sets. In this new model, the evolution of a marker's boundary curve C is
described by
(82)
where A(t) denotes the instant area enclosed by the evolving curve. The correspond-
ing stationary eikonal PDE is
(83)
where T is arrival time. Whereas in the case of uniform height flooding the
FMM-based solution of the eikonal PDE was simple to implement as done in
[47], in the case of uniform volume flooding solving the stationary PDE (83) has
the peculiarity of the time-varying term A(T). This was numerically solved in [75]
using a variation of the FMM that took into consideration the time-dependent area
variations of the fronts.
Experimental results of the PDE-based height and volume flooding applied to real
image segmentation are shown in Fig. 12. The volume flooding seems to perform in
a more balanced way. This is expected since it exploits both the contrast and the (area)
size properties of the objects-regions present in an image. It is a useful segmentation
tool in cases where we want to keep a balance between the aforementioned image
properties and can give good results in cases in which contrast-based driven seg-
mentation fails. In addition, the PDE implementation has the advantage of a more
isotropic flooding.
This supervised approach was adopted by the COMPARE method (Fan et al., 2007)
that aims to perform classification of morphological patterns using adaptive regional
elements. COMPARE extracts spatially smooth clusters that can be used to train
a classifier to predict patient diagnosis by combining information stemming from
both the imaging signal and subjects’ diagnosis. The two types of information are
integrated at each image location, p, in a multiplicative fashion though the score:
(10.1)
where C(p) measures the spatial consistency of the imaging signal, while P(p)
measures discriminative power. More precisely, P is calculated as the following
leave-one-out absolute Pearson correlation:
(10.2)
where (p, i) denotes the Pearson correlation measured between the imaging signal
at p and the classification labels when excluding the ith subject/sample. The consis-
tency C(p) is the intra-class coefficient measuring the proportion of neighboring
feature variance that is explained by the inter-subject variability (McGraw and Wong,
1996; Fan et al., 2007). It takes values between 0 and 1, with higher values indicating
that the variance of the measurements across neighboring brain location is small
with respect to the inter-subject variability of the imaging signal. As a result, the
score s(p) is bounded between 0 and 1, with values close to 1 indicating that the
imaging signal around p is simultaneously highly reliable and discriminative (ie,
highly correlated or anticorrelated with patient diagnosis).
This score map is subsequently smoothed, and its gradient is used in conjunction
with a watershed segmentation algorithm (Vincent and Soille, 1991) to partition the
brain into different regions (Fig. 10.1 presents brain regions generated by watershed
from white matter tissue density maps of demented and normally aging subjects
(Shen and Davatzikos, 2002; Fan et al., 2007)). These regions are then refined by
considering only locations that optimize the classification power of the extracted
features. This is performed in a region growing fashion where, initially, only the node
of the region with the highest discriminative score is selected, and adjacent locations
are incrementally aggregated as long as the discriminative power does not decrease.
The previous steps are summarized in Algorithm??. This approach extracts a single
connected component per watershed region. Each component comprises highly
discriminative elements whose average imaging signal may be used as a feature for
training a classifier, such as an SVM (Vapnik, 2000).
Algorithm 10.1
Clustering approaches typically aim to divide the brain into spatially smooth areas
that are likely to correspond to the functional units that constitute the brain. This is
usually performed by first representing the brain in the form of a graph, where nodes
represent brain locations and edges connect nodes that correspond to spatially ad-
jacent locations. The weight of the edges represents the strength of the connectivity
between nodes and is estimated by computing the similarity between the rs-fMRI
signals that are measured at each node. The similarity is commonly measured by
the Pearson correlation or partial correlation (Smith et al., 2011). Once the node
is constructed, adjacent brain locations that are strongly connected are grouped
together in the same parcel.
Numerous methods have been proposed for this task. Among the most popular
methods, one may cite hierarchical clustering (Cordes et al., 2002), normalized
clustering (Shen et al., 2010; Craddock et al., 2012), k-means (Bellec et al., 2010),
region growing (Blumensath et al., 2013; Heller et al., 2006), and Markov random
fields (MRFs) (Ryali et al., 2013; Golland et al., 2007; Honnorat et al., 2015). Different
methods exhibit distinct advantages and disadvantages. Generally, many of the
above methods are either initialization dependent (eg, region growing (Blumensath
et al., 2013; Heller et al., 2006) and k-means (Bellec et al., 2010)), or rely on complex
models that involve a large number of parameters (Golland et al., 2007). As a result,
they are sensitive to initialization and suffer from limitations related to the employed
heuristics (eg, hierarchical clustering may lead to the creation of poorly fit parcels
at coarser scales (Cordes et al., 2002; Honnorat et al., 2015)), and the large number
of inferred parameters that may negatively impact the quality of the locally optimal
solution that is obtained (Golland et al., 2007). Moreover, not all methods produce
contiguous parcels.
Fig. 10.2. Functional parcellation of the left hemisphere of the brain, projected on
an inflated brain surface.
where p denotes a node of the brain graph, lp the parcel that should contain this
node, Vp(lp) is a cost that decreases when the node p is assigned to a parcel lp
with highly correlated rs-fMRI signal, Lp({lp}) penalizes by a positive cost K the
introduction of a parcel of functional center p, and the Sp({lp}) are the shape
priors that enforce the connectedness of each parcel p. This energy is optimized
by exploiting advanced solvers (Delong et al., 2012) that could provide a substantial
advance over existing methods. Experimental results on large datasets demonstrated
that this approach is capable of generating parcels that are all highly coherent, while
the overall parcellation is slightly more reproducible than the results produced by
hierarchical clustering and normalized cuts (Honnorat et al., 2015).
2 Segmentation
Morphological segmentation of a grayscale image is usually done by applying the
watershed algorithm to the gradient of the image. The circular centered gradient
operator allows one to segment an image containing circular data in the same
way. We present an example of the segmentation of an oriented texture. The aim
of this type of segmentation is to create regions in which the orientations are
homogeneous.
The steps in the segmentation algorithm, which are illustrated in Figure 20, are:
Figure 20. Steps in the segmentation of an oriented texture. (a) Initial image with size
420 × 1040 pixels. (b) Orientation image with size 50 × 125 pixels. (c) Morphological
circular centered gradient (with a square SE of size 2). (d) h-minima. (e) Watershed
segmentation of the circular centered gradient image (c) using the markers in (d).
The graylevel in each region encodes the mean orientation of the region.(courtesy
of Scanwood System, Pont-à-Mousson, France)
(1) The Rao and Schunck algorithm is applied to the initial image (a plank of oak,
Figure 20(a)) to calculate the orientation image (Figure 20(b)). For the example,
the parameters 1 = 1.4, 2h = 2v = 64, and Δh = Δv = 8 were used.
(2) The circular centered gradient of the orientation image is calculated (Fig-
ure 20(c)). For the example, a square SE of size 2 was used.
(3) The minima are extracted. So as to avoid finding a large number of small
minima which would result in an over-segmentation of the image, we close
the gradient image with a square SE of size 1, and then find the h-minima
(Soille, 1999) of height h = 5 (Figure 20(d)).
(4) The watershed is applied to the gradient image using the minima extract-
ed in the previous step as markers, producing the segmentation shown in
Figure 20(e). In this image, the watershed lines are shown in black, and the
graylevel of each region encodes the mean orientation of the region, calculated
using circular statistics. For visualization purposes, the segmentation obtained
is superimposed on the initial image in Figure 21(a), in which the black lines
represent the watershed lines.Figure 21. Results of segmentations of oriented
textures by the watershed algorithm for four oak images.(Images courtesy of
Scanwood System, Pont-à-Mousson, France.)
Some further results of the segmentation of oriented textures are shown in Fig-
ures 21(b)–(d). In general, this algorithm manages to segment the textures into
homogeneous regions, with the more globally homogeneous textures segmented
into the fewest regions, as in Figure 21(b) for example. Some problems are never-
theless present in the current formulation of the algorithm: the first is common to
almost all watershed segmentations, and involves the choice of markers. If we take
all the minima of the gradient as markers, an over-segmentation (segmentation
into too many regions) is produced. With the current approach, a small closing
followed by the extraction of the h-minima, the number of regions is reduced,
but a change in the value of parameter h can provoke a large difference in the
segmentation. The segmentation of an oriented texture can also be modified by
changing the scale parameter ( 1) in the calculation of the orientation image. A last
difficulty is when the orientation variations are not localized enough to be detected
by the structuring element used in the gradient calculation, which can lead to the
presence of more than one orientation in one of the regions of the segmentation.
Several possible solutions to these problems remain to be studied. For example,
starting with an over-segmentation of the image, and then fusing regions with
similar mean orientations using a graph of the partition (Meyer, 1999) so as to
eliminate over-segmented regions, or taking into account several partitions of the
same texture so as to find the most probable one (Nickels and Hutchinson, 1997).
The orientation images were calculated using the Rao and Schunck algorithm with
parameters 1 = 1.75, Δh = Δv = 2, and 2h = 2v = 16, except for images d052 and
d079 for which 2h = 2v = 32 were used. The threshold for isolating the defect
was chosen by hand for each image. The results for the five images are shown in
Figure 22. In each line, the initial image, the orientation image, the result of the
circular centered top-hat, and the borders of the thresholded regions superimposed
on the initial image are shown. In textures d052 and d053, the defects are very
subtle modifications of the structure, yet they perturb the orientations enough to
be detected. The defect in texture d077 is easily seen and easily detectable in the
orientation image. In textures d050 and d079, the defects cause perturbations in
the orientation field, but the borders of these defects are not obvious, even to the
naked eye. Among these textures, the only one made up of oriented lines is d050,
but the others are anisotropic enough to have a uniform orientation field which is
perturbed by the defects. This approach is obviously not applicable to textures which
are not anisotropic, some examples of which are given by Chetverikov and Hanbury
(2002).
Figure 22. Results of defect detection by the circular centered top-hat applied to
some Brodatz textures. The images labeled “ini” are the initial images, “ori” the
orientation images, “ath” the top-hat operator results, and “det” the regions detected
by the threshold superimposed on the initial image.
For wood, the most important structural defects are the knots, some of which do
not have a color very different to that of the wood, but which nevertheless cause
perturbations in the surrounding vein orientations. The defects identifiable only by
a change in color are obviously not detectable by these orientational methods, the
same being applicable to defects due to external influences, such as insect stings.
For the experiments, we used a database of oak boards with a very high defect
occurrence. In order to speed up the calculation, we used a large separation between
the neighborhoods in the orientation image calculation. The parameters used were
1 = 1.4, 2h = 2v = 64, and Δh = Δv = 16. Some images and their corresponding
orientation images are shown in the first two columns of Figure 23. We then
calculated the top-hat based on the labeled opening of the orientation images. The
opening was done with a sector size = 45°, and by varying from 0 = 0° to 157.5°
in steps of Δ = 22.5°. A square SE of size 3 was used. The residue of this top-hat
enlarged and superimposed on the initial image is shown in the rightmost column
of Figure 23, in which the light regions correspond to the residues.
Figure 23. Results of the detection of regions having orientation perturbations using
the top-hat based on the labeled opening. The “ini” images are the initial images
of oak boards, on which the defects found by an expert are outlined in black (the
dark horizontal lines are red chalk marks on the board, and have no bearing on the
experiment). The “ori” images are the orientation images. The light regions of the
“eye” images correspond to the residues of the orientation images detected by the
top-hat. The size of the images, in pixels, is given below each image.(courtesy of
Scanwood System, Pont-à-Mousson, France)
• For image c005, the black vein is evidently not detected as it does not perturb
the orientation. The knot at the top right is detected, but the small knots at
the bottom left do not perturb the orientation enough to be detected.
• For image c007, the large knot is detected, but the fissures to the left of the
knot, which have similar orientations to the veins, are not detected. Some false
detections near the borders of the image are also present.
• Image c034 demonstrates that this method is not very useful on boards which
contain veins having elliptical forms. Their large curvature leads to many false
detections.
If one calculates the orientation image at a finer resolution, then smaller defects can
be detected. For oak, this is useful for detecting the small light patches, some of
which are indicated in Figure 24(a). Even if these are not classified as defects, their
detection can be important if one wishes to determine the aesthetic appearance of
the wood. The detection of these light patches based only on their color is rather
difficult, as their color is very similar to that of other structures on the wood. On
looking at the orientation image, one can see that because the light patches cut the
veins, they produce perturbations in the orientation field which can be detected by
a top-hat. The orientation image of Figure 24(a), calculated using the parameters
1 = 1.4, 2h = 2v = 16, and Δh = Δv = 8 is shown in Figure 24(b). The result of a
top-hat based on a labeled opening is shown in Figure 24(c). The parameters of this
operator are = 45°, 0 = 0° and Δ = 22.5°, and a square SE of size 4 was used.
Figure 24. Detection of defects at a smaller scale. (a) An oak board with some of the
small white patches manually indicated (size 608 × 955 pixels). (b) Orientation image
(size 50 × 112 pixels). (c) Result of a top-hat based on a labeled opening. The light
pixels indicate the residue.(courtesy of Scanwood System, Pont-à-Mousson, France)
Globally, it is clear that a perturbation in the vein orientation is not always associated
with a defect, and that a defect does not always perturb the surrounding veins. This
method of defect detection can therefore not function as a total solution to the defect
detection problem. The results can nevertheless be used in a defect classification
step, which takes color and other texture variations into account along with the
orientation perturbations in the calculation of the probability of a defect being
present at a certain position on the board.
ScienceDirect is Elsevier’s leading information solution for researchers.
Copyright © 2018 Elsevier B.V. or its licensors or contributors. ScienceDirect ® is a registered trademark of Elsevier B.V. Terms and conditions apply.