Final Report
Final Report
• and monitoring. Due to the aggressive nature and diversity of gliomas, well-
organized and exact segmentation methods are used to classify tumors intra-
tumorally.
• In comparison with the current state of the art, the accuracy of brain tumor
segmentation was significantly improved using Convolutional Neural
Networks, which are frequently used in the field of biomedical image
segmentation.
INTRODUCTION
Tumor is an uncontrolled growth of cancer cells in any part of the body. Tumors
are of different types and have different characteristics and different treatments. At
present, brain tumors are classified as primary brain tumors and metastatic brain
tumors. The former begin in the brain and tend to stay in the brain, the latter begin
as a cancer elsewhere in the body and spreading to the brain. Brain tumor
segmentation is one of the crucial procedures in surgical and treatment planning.
Brain tumor segmentation using MRI has been an intense research area. Brain
tumors can have various sizes and shapes and may appear at different locations.
Varying intensity of tumors in brain magnetic resonance images (MRI) makes the
automatic segmentation of tumors extremely challenging. There are various
intensity based techniques which have been proposed to segment tumors on
magnetic resonance images. Texture is one of most popular feature for image
classification and retrieval. The multifractal texture estimation methods are more
time consuming. A texture based image segmentation using GLCM (Gray-Level
Co-occurrence Matrix) combined with AdaBoost classifier is proposed here. From
the MRI images of brain, the optimal texture features of brain tumor are extracted
by utilizing GLCM. Then using these features AdaBoost classifier algorithm
classifies the tumor and non-tumor tissues and tumor is segmented. This method
provides more efficient brain tumor segmentation compared to the segmentation
technique based on mBm and will provide more accurate result. Tumor is the
abnormal growth of the tissues. A brain tumor is a mass of unnecessary cells
growing in the brain or central spine canal. Brain cancer can be counted among the
most deadly and intractable diseases. Today, tools and methods to analyse tumors
and their behaviour are becoming more prevalent. Clearly, efforts over the past
century have yielded real advances. However, we have also come to realize that
gains in survival must be enhanced by better diagnosis tools. Although we have yet
to cure brain tumours, clear steps forward have been taken toward reaching this
ultimate goal, more and more researchers have incorporated measures into clinical
trials each advance injects hope to the team of caregivers and more importantly, to
those who live with this diagnosis. Magnetic Resonance Imaging (MRI) has
become a widely-used method of high-quality medical imaging, especially in brain
imaging where MRI’s soft tissue contrast and non-invasiveness are clear
advantages. An important use of MRI data is tracking the size of brain tumor as it
responds treatment. Therefore, an automatic and reliable method for segmenting
tumor would be a useful tool. MRI provides a digital representation of tissue
characteristics that can be obtained in any tissue plane. The images produced by an
MRI scanner are best described as slices through the brain. MRI has the added
advantage of being able to produce images which slice through the brain in both
horizontal and vertical planes. This makes the MRI-scan images an ideal source for
detecting, identifying and classifying the right infected regions of the brain. Most
of the current conventional diagnosis techniques are based on human experience in
interpreting the MRI-scan for judgment; certainly this increases the possibility to
false detection and identification of the brain tumor. On the other hand, applying
digital image processing ensures the quick and precise detection of the tumor. One
of the most effective techniques to extract information from complex medical
images that has wide application in medical field is the segmentation process. The
main objective of the image segmentation is to partition an image into mutually
exclusive and exhausted regions such that each region of interest is spatially
contiguous and the pixels within the region are homogenous with respect to a
predefined criterion. The cause of most cases is unknown. Risk factors that may
occasionally be involved include: a number of genetic syndrome such as
neurofibromatosis as well as exposure to the chemical vinyl chloride, Epstein-Barr
virus, and ionizing radiation.
DOMAIN EXPLANATION
IMAGE PROCESSING
Figure 2: Each pixel has a value from 0 (black) to 255 (white). The possible
range of the pixel values depend on the colour depth of the image, here 8 bit =
256 tones or greyscales.
Some grayscale images have more grayscales, for instance 16 bit = 65536
grayscales. In principle three grayscale images can be combined to form an image
with 281,474,976,710,656 grayscales.
There are two general groups of ‘images’: vector graphics (or line art) and bitmaps
(pixel-based or ‘images’). Some of the most common file formats are:
PSD – a dedicated Photoshop format that keeps all the information in an image
including all the layers.
Suppose we take an image, a photo, say. For the moment, lets make things
easy and suppose the photo is black and white (that is, lots of shades of grey), so
no colour. We may consider this image as being a two dimensional function, where
the function values give the brightness of the image at any given point. We may
assume that in such an image brightness values can be any real numbers in the
range (black) (white).
A digital image from a photo in that the values are all discrete. Usually they
take on only integer values. The brightness values also ranging from 0 (black) to
255 (white). A digital image can be considered as a large array of discrete dots,
each of which has a brightness associated with it. These dots are called picture
elements, or more simply pixels. The pixels surrounding a given pixel constitute its
neighborhood. A neighborhood can be characterized by its shape in the same way
as a matrix: we can speak of a neighborhood,. Except in very special
circumstances, neighborhoods have odd numbers of rows and columns; this
ensures that the current pixel is in the centre of the neighborhood.
Pixel:
Other pixel shapes and formations can be used, most notably the hexagonal grid, in
which each pixel is a small hexagon. This has some advantages in image
processing, including the fact that pixel connectivity is less ambiguously defined
than with a square grid, but hexagonal grids are not widely used. Part of the reason
is that many image capture systems (e.g. most CCD cameras and scanners)
intrinsically discretize the captured image into a rectangular grid in the first
instance.
Pixel Connectivity
The notation of pixel connectivity describes a relation between two or more pixels.
For two pixels to be connected they have to fulfill certain conditions on the pixel
brightness and spatial adjacency.
First, in order for two pixels to be considered connected, their pixel values must
both be from the same set of values V. For a grayscale image, V might be any
range of graylevels, e.g. V={22,23,...40}, for a binary image we simple have
V={1}.
Two pixels p and q, both having values from a set V are 4-connected if q is from
A set of pixels in an image which are all connected to each other is called a
connected component. Finding all connected components in an image and marking
each of them with a distinctive label is called connected component labeling.
An example of a binary image with two connected components which are based on
4-connectivity can be seen in Figure 1. If the connectivity were based on 8-
neighbors, the two connected components would merge into one.
Figure 1 Two connected components based on 4-connectivity.
Pixel Values
Each of the pixels that represents an image stored inside a computer has a pixel
value which describes how bright that pixel is, and/or what color it should be. In
the simplest case of binary images, the pixel value is a 1-bit number indicating
either foreground or background. For a gray scale images, the pixel value is a
single number that represents the brightness of the pixel. The most common pixel
format is the byte image, where this number is stored as an 8-bit integer giving a
range of possible values from 0 to 255. Typically zero is taken to be black, and 255
is taken to be white. Values in between make up the different shades of gray.
To represent colour images, separate red, green and blue components must be
specified for each pixel (assuming an RGB colour space), and so the pixel `value'
is actually a vector of three numbers. Often the three different components are
stored as three separate `grayscale' images known as color planes (one for each of
red, green and blue), which have to be recombined when displaying or processing.
Multispectral Images can contain even more than three components for each pixel,
and by extension these are stored in the same kind of way, as a vector pixel value,
or as separate color planes.
The actual grayscale or color component intensities for each pixel may not actually
be stored explicitly. Often, all that is stored for each pixel is an index into a colour
map in which the actual intensity or colors can be looked up.
Although simple 8-bit integers or vectors of 8-bit integers are the most common
sorts of pixel values used, some image formats support different types of value, for
instance 32-bit signed integers or floating point values. Such values are extremely
useful in image processing as they allow processing to be carried out on the image
where the resulting pixel values are not necessarily 8-bit integers. If this approach
is used then it is usually necessary to set up a colormap which relates particular
ranges of pixel values to particular displayed colors.
Color scale
RGB
The RGB color model is an additive color model in which red, green, and blue
light are added together in various ways to reproduce a broad array of colors. RGB
uses additive color mixing and is the basic color model used in television or any
other medium that projects color with light. It is the basic color model used in
computers and for web graphics, but it cannot be used for print production.
The secondary colors of RGB – cyan, magenta, and yellow – are formed by mixing
two of the primary colors (red, green or blue) and excluding the third color. Red
and green combine to make yellow, green and blue to make cyan, and blue and red
form magenta. The combination of red, green, and blue in full intensity makes
white.[figure4]
Figure [4]: The additive model of RGB. Red, green, and blue are the primary
stimuli for human color perception and are the primary additive colours.
-Algorithm development
-Data acquisition
Some applications:
1. Medicine
Inspection and interpretation of images obtained from X-rays, MRI or CAT
scans,
2. Agriculture
3. Industry
4. Law enforcement
Fingerprint analysis,
Example include:
highlighting edges,
Removing noise.
These classes are not disjoint; a given algorithm may be used for both image
enhancement or for image restoration. However, we should be able to decide what
it is that we are trying to do with our image: simply make it look better
(enhancement), or removing damage (restoration).
An image processing task
We will look in some detail at a particular real-world task, and see how the above
classes may be used to describe the various stages in performing this task. The job
is to obtain, by an automatic process, the postcodes from envelopes. Here is how
this may be accomplished:
Acquiring the image: First we need to produce a digital image from a paper
envelope. This can be done using either a CCD camera, or a scanner.
Preprocessing: This is the step taken before the major image processing
task. The problem here is to perform some basic tasks in order to render the
resulting image more suitable for the job to follow. In this case it may
involve enhancing the contrast, removing noise, or identifying regions likely
to contain the postcode.
Disadvantages:
This method was limited to enhancing tumors with clear enhancing edges.
This method works with two labels only - object and background.
One of the few drawbacks of the graph-based methods is that they are not
easily extended to multi-label task
The other is that they are not very flexible
The only tunable parameters are the graph weighting and cost function
coefficients.
PROPOSED SYSTEM
• The system proposed a novel semi-automatic segmentation method based on
population and individual statistical information to segment brain tumors in
magnetic resonance (MR) images.
• The probability of each pixel belonging to the foreground (tumor) and the
back ground is estimated by the k NN classifier under the learned optimal
distance metrics.
Advantages:
Flow Diagram:
TESTING OF PRODUCT
Testing of Product:
System testing is the stage of implementation, which aimed at
ensuring that system works accurately and efficiently before the live operation
commence. Testing is the process of executing a program with the intent of finding
an error. A good test case is one that has a high probability of finding an error. A
successful test is one that answers a yet undiscovered error.
UNIT TESTING:
Unit testing is the testing of each module and the integration of the overall
system is done. Unit testing becomes verification efforts on the smallest unit of
software design in the module. This is also known as ‘module testing’. The
modules of the system are tested separately. This testing is carried out during the
programming itself. In this testing step, each model is found to be working
satisfactorily as regard to the expected output from the module. There are some
validation checks for the fields. For example, the validation check is done for
verifying the data given by the user where both format and validity of the data
entered is included. It is very easy to find error and debug the system.
INTEGRATION TESTING:
Data can be lost across an interface, one module can have an adverse
effect on the other sub function, when combined, may not produce the desired
major function. Integrated testing is systematic testing that can be done with
sample data. The need for the integrated test is to find the overall system
performance. There are two types of integration testing. They are:
White Box testing is a test case design method that uses the control
structure of the procedural design to drive cases. Using the white box testing
methods, we derived test cases that guarantee that all independent paths within a
module have been exercised at least once.
VALIDATION TESTING:
User acceptance of the system is the key factor for the success of the
system. The system under consideration is tested for user acceptance by constantly
keeping in touch with prospective system at the time of developing changes
whenever required.
OUTPUT TESTING:
After performing the validation testing, the next step is output asking the
user about the format required testing of the proposed system, since no system
could be useful if it does not produce the required output in the specific format.
The output displayed or generated by the system under consideration. Here the
output format is considered in two ways. One is screen and the other is printed
format. The output format on the screen is found to be correct as the format was
designed in the system phase according to the user needs. For the hard copy also
output comes out as the specified requirements by the user. Hence the output
testing does not result in any connection in the system.
System Implementation:
The active user must be aware of the benefits of using the system
Their confidence in the software built up
Proper guidance is impaired to the user so that he is comfortable in
using the application
Before going ahead and viewing the system, the user must know that for
viewing the result, the server program should be running in the server. If the server
object is not running on the server, the actual processes will not take place.
User Training:
To achieve the objectives and benefits expected from the proposed system
it is essential for the people who will be involved to be confident of their role in the
new system. As system becomes more complex, the need for education and
training is more and more important.
Operational Documentation:
System Maintenance:
Corrective Maintenance:
Adaptive Maintenance:
Perceptive Maintenance:
The third activity that may be applied to a definition of maintenance
occurs when a software package is successful. As the software is used,
recommendations for new capabilities, modifications to existing functions, and
general enhancement are received from users. To satisfy requests in this category,
Perceptive maintenance is performed. This activity accounts for the majority of all
efforts expended on software maintenance.
Preventive Maintenance:
• Input Image
• Pre-Processing
• Segmentation
• Feature Extraction
• Classification
• Analysis
MODULE DESCRIPTION
Input Image:
• Read an image into the workspace, using the imread command. The example
reads one of the sample images included with the toolbox, an image, and
stores it in an array named I . imread infers from the file that the graphics
file format is Tagged Image File Format (TIFF).
• Display the image, using the imshow function. You can also view an image
in the Image Viewer app. The imtool function opens the Image Viewer app
which presents an integrated environment for displaying images and
performing some common image processing tasks.
• The Image Viewer app provides all the image display capabilities of imshow
but also provides access to several other tools for navigating and exploring
images, such as scroll bars, the Pixel Region tool, Image Information tool,
and the Contrast Adjustment tool.
• Preprocessing :
• Image Resize
• Preprocessing :
• In the case of decreasing the pixel number (scaling down) this usually
results in a visible quality loss. From the standpoint of digital signal
processing, the scaling of raster graphics is a two-dimensional
example of sample rate conversion, the conversion of a discrete signal
from a sampling rate (in this case the local sampling rate) to another.
Segmentation
Binarization:
• The binary image pixel values are to apply the delay register type model and
to store the pixel values.
• Then to set the minimum threshold value between the two pixels variations
limits. A mask is a filter. Concept of masking is also known as spatial
filtering. Masking is also known as filtering.
• In this concept we just deal with the filtering operation that is performed
directly on the image.
Feature Extraction:
Classification:
Performance analysis:
• Estimations:
Estimations:
• Accuracy: The accuracy of a test is its ability to differentiate the patient and
healthy cases correctly. To estimate the accuracy of a test, we should
calculate the proportion of true positive and true negative in all evaluated
cases. Mathematically, this can be stated as:
LITERATURE SURVEY
1. Title: Improved Interactive Medical Image Segmentation using Enhanced
Intelligent Scissors (EIS)
Year: 2008
Author: Akshaya Mishra, Alexander Wong, Wen Zhang, David Clausi, and
Paul Fieguth
The proposed EIS algorithm takes a fast interactive approach to the problem
of medical image segmentation, where a boundary is formed around the region of
interest based on a sequence of user-selected points. The proposed algorithm can
be described as follows. First, a phase-based representation of the image is
extracted as the external local cost using a robust iterative complex wavelet phase
moment estimation scheme. Second, the boundary extraction problem between two
user-selected points is treated as an active contour problem and formulated as a
HMM. Third, a novel approach of solving the formulated HMM using the second-
order Viterbi algorithm is performed by reformulating the second-order problem
with first-order Markovian assumptions and solving it based on the internal and
external local costs. Furthermore, a novel adaptive state pruning scheme is
performed based on the extracted external local costs to significantly reduce the
computational complexity of the proposed EIS algorithm.
In the conventional IS approach, the user starts at an initial point near the
boundary of the region of interest and moves the mouse cursor closely along the
boundary. As the mouse cursor comes close to a boundary edge, a “live-wire”
boundary snaps to the edge. Therefore, as the mouse cursor moves around the
region of interest, the live-wire wraps around the region to form a segmentation
boundary. In the proposed EIS approach, the user first selects an initial point near
the boundary of the region, as with the conventional IS approach. However, rather
than tracing the mouse cursor closely along the boundary, the user selects a
sequence of discrete points around the boundary. As the user selects points around
the boundary, the user-selected points snap to the region of interest and a boundary
is formed around the region of interest between these points. Therefore, as points
are selected, a segmentation boundary is formed. The points selected by the user in
EIS can be sparsely spaced around the region boundary and does not need to be
placed in close proximity to the region boundary. The main advantage of using the
proposed approach of user interaction over the conventional IS approach is that the
user does not need to trace around the boundary carefully. The user can simply
click around the region boundary in an imprecise manner and the EIS algorithm
will automatically create a boundary around the region of interest accordingly.
This allows for a much faster level of user interaction while still providing accurate
region boundaries based on user knowledge. The main advantage to formulating
the boundary extraction problem using a HMM is that the solution to the problem
can be found in a very efficient manner using methods such as the Viterbi
algorithm. This is as opposed to existing IS methods, where the problem
formulation does not allow such a solution. Computational efficiency is very
important for the proposed EIS algorithm since the underlying goal is to provide
fast user interaction for clinicians.
In this paper, we introduced Enhanced Intelligent Scissors (EIS), a novel fast
interactive approach to medical image segmentation. The proposed algorithm is
highly robust to contrast non-uniformities and noise through the use of an external
local cost based on complex wavelet phase coherence moments. The optimal
boundary between user-selected points is found by formulating the problem as a
HMM and solved using a novel approach to the second-order Viterbi algorithm.
Furthermore, a novel phase-adaptive state pruning scheme was proposed to
improve the computational performance of the proposed algorithm.
Advantages:
The main advantage of using the proposed approach of user interaction over
the conventional IS approach is that the user does not need to trace around
the boundary carefully.
Disadvantages:
2. Title: Watershed cuts: minimum spanning forests and the drop of water
principle
Year: 2013
Author: Jean Cousty, Gilles Bertrand, Laurent Najman and Michel Couprie
Advantages:
Disadvantages:
Year: 2011
The idea behind the present paper is to use the co-occurrence matrix and its
extracted features in face recognition. To the best of our knowledge, no one has
attempted to implement this method before. The idea is simple and straight
forward. For each face image, a feature vector is formed by converting the
generated gray-level co-occurrence matrix (GLCM) to a vector and then it is used
for classification. Additionally, Haralick features containing 14 statistical features
can be extracted from the GLCM to form a new feature vector with 14 features.
The proposed GLCM method is compared with four well-known face recognition
techniques: Principal component analysis (PCA) which is standard technique used
in statistical pattern recognition and signal processing for dimensionality reduction
and feature extraction. The proposed method is also compared with the other well-
known technique, linear discriminant analysis (LDA), which overcomes the
limitations of PCA method by applying the Fisher’s linear discriminant criterion.
This criterion tries to maximize the ratio of the determinant of the between-class
scatter matrix of the projected samples to the determinant of the within-class
scatter matrix of the projected samples. Fisher discriminant analysis method groups
images of the same class and separates images of different classes. The third
technique is local binary patterns, which is a non-parametric operator which
describes the local spatial structure of an image and high discriminative power for
classification. Finally, GLCM was also compared with Gabor wavelets transform.
The transform involves convolving an image with a group of Gabor filters with
parameterized scale and directionally.
The results in Table indicate that, the proposed direct GLCM method is
superior to the method using well-known Haralick features. Furthermore, it is
important to note that changing the number of gray levels Ng provides important
flexibility, which results scalable matrix size for the GLCM. An important notice
which can be observed in table 2 that in some cases when the co-occurrence matrix
become smaller due to the use of smaller number of gray levels, the recognition
performance of direct GLCM approach slightly increases. This can be attributed to
the quantization process that helps to suppress the noise in the input image, which
as a result, helps to increase the recognition performance.
The system shows the performance of the direct GLCM method for
changing number of training samples in the training set for different number of
gray levels. The results show that the recognition performance of the proposed
system changes slightly as we reduce the number of gray levels from 256 to 16
gray levels. Although, this observation is specific to ORL dataset and cannot be
generalized, the results confirm the scalability of the proposed direct GLCM
approach. Table compares the performance of the proposed GLCM based
recognition systems with the state of the art techniques PCA, LDA, Gabor
wavelets (GW) and LBP by using the ORL face database and nearest neighbor
classifier. The results demonstrate the superiority of the proposed direct GLCM
based face recognition system over the well know face recognition techniques PCA
and LDA. In same manner, Table shows the performance results of same
algorithms with different number of training images using subset of FRAV2D
database and nearest neighbor classifier. In both tables, the GLCM approach
outperformed the PCA and LDA approaches.
In this paper, we propose a new method, direct GLCM, which performs face
recognition by using Gray-Level Co-occurrence Matrix. The direct GLCM method
is very competitive with state of the art face recognition techniques such as PCA,
LDA, Gabor wavelets, and LBP. Using smaller number of gray levels (bins)
shrinks the size of GLCM which reduces the computational cost of the algorithm
and at the same time preserves the high recognition rates. This can be due to the
process of quantization which helps in suppressing the noise of the images at
higher grey levels.
Advantages:
It is obvious from the results that the GLCM is a robust method for face
recognition with competitive performance.
Disadvantages:
Year: 2014
Advantages:
Disadvantages:
Year: 2007
As for the proposed technique, the computation area is depends on the edge
magnitude rather than the GLCM quadrants. In the proposed technique, only the
computation area is referred to the GLCM while the threshold value computation is
totally based on the gray values of the input image (first order statistic feature).
This is varying with the entropy based technique where both the computation area
and the threshold value computation are based on the GLCM pixel pair repetition
(second order statistic feature). Thus a more precise representation of the pixel
values on the boundary of the object can be achieved by the proposed technique.
Back to the three basic images, better thresholding results as shown in Figure 6 are
obtained by the proposed technique. For the image in Figure 5(a), the proposed
technique only includes the GLCM upper right white point in its computation area.
Then, the average of the gray value contained in the computation area is computed.
This means that, average of gray value 0 and 1 is computed, which results a
threshold value equals to 0.5. Previously, the entropy based technique results the
threshold value equals to 0. For the blurred image in Figure where most of the high
values in the GLCM are located near the right and bottom limit of the GLCM, the
proposed technique also results a better representation of the boundary’s pixel
value. Thus, a better threshold value is also obtained. Based on the proposed
technique, most of the GLCM elements contained in the joint quadrant are also
included in the proposed technique computation area. By averaging all involve
gray values in the computation area, threshold value equals to 0.57 is obtained. By
applying this threshold value, a less corrupted images is obtained as shown. For the
last image, the proposed technique has overcome the problem faced by the entropy
based technique. If the entropy based technique only contained only certain GLCM
elements in the joint quadrants for any gray levels value, the proposed technique
includes the entire GLCM elements in its computation area. Thus, average value of
the gray values gives a good threshold value where the input image is segmented
evenly with bright and dark pixel.
It shows that our method provides the best result in segmenting all the test
samples compare to other methods. However, when there is an even light
concentration, all the methods gave acceptable results. For uneven light
concentration and complex background images, the results are contrary for Entropy
based method, Yun Li method and Otsu method. Entropy based method and Yun
Li method resulted in poor segmentation because false edges form from the uneven
light concentration and complex background are also computed in their algorithms.
As for the Otsu method, the poor segmentation is due to the object (defect) and the
background does not separate well in the image histogram. Otsu method assumes
that images have two normal distributions with similar variances. The threshold
value is selected by separating the image histogram into two classes such that its
inter-class variance is maximized. Unfortunately, the uneven light concentration
and complex background of the tested images cause the defects to become
invisible in the Otsu computation. For the five images, the image histogram
separation by Otsu method is shown in Figure 8. The figure also shows the
threshold value for our method, the entropy based method and Yun Li method. For
image with the uneven light concentration but without the complex background,
Otsu method recognized the darker side as object although it is not the defect
because the edges of the defects are very fuzzy where it cannot be recognize in the
image histogram.
In this paper, a thresholding technique has been proposed based on the gray
level co-occurrence matrix. The technique extracts the edge information and the
gray level transition frequency from the GLCM to compute the threshold value.
The algorithm is also designed to have the flexibility over the edge definition.
Thus, it can handle image with fuzzy boundaries between the image’s object and
background. The proposed technique was tested with starfruit defect image and
result good segmentation in order to identify the area of the defect on the starfruit
skin. The results were compared with three other techniques.
Advantages:
The proposed technique was tested with starfruit defect image and result
good segmentation in order to identify the area of the defect on the starfruit
skin.
Disadvantages:
It produce poor result when dealing with noisy, complex background and
fuzzy boundary images.
The paper depicts and elaborates a new framework for the comparison of
classifiers for medical image segmentation with transform and moment based
features. Medical images modalities such as Ultrasound (US) bladder,
Ultrasound (US) phantom, Computerized Tomography (CT) and Magnetic
Resonance (MR) images are segmented using different algorithms namely, k-
Nearest Neighbor (kNN), Grow and learn (GAL) and Incremental Supervised
Neural Networks (ISNN). Segmentation is performed by applying feature
extraction methods such as 2D Continuous Wavelet Transform (2D-CWT),
Moments of gray level histogram (MGH) and a combined version of both 2D-
CWT and MGH, called Hybrid features. With different iterations, the analysis
results indicate that kNN performs better than GAL, and the performance of GAL
is better than that of the ISNN for image segmentation. During analysis a
comparison has been drawn between the performance of kNN, GAL and ISNN on
the above three feature extraction schemes and also provides the qualitative and
quantitative analysis of three classifiers. Results indicate that the performance of
2D-CWT and that of Hybrid features is consistently better than MGH features for
all image modalities. The demonstrated frame work or the system is capable to
meet the demand for selecting best approach in order to meet the given time
constraints and accuracy standards in medical image segmentation.
The main work in this paper is to find out the best combination of classifiers
with feature extraction schemes to achieve efficient segmentation for medical
images. Recently, grow and learn (GAL) and incremental supervised neural
network (ISNN) are compared under two feature extraction methods (moment of
grey level histogram (MGH) and two dimension continuous wavelet transform
(2D-CWT)). Neural network and SVM based classifiers 22 are compared to check
which classifier has better performance. Similarly, different classifiers 23-24 are
compared for checking performance results. In this paper KNN, GAL and ISNN
under MGH, 2D-CWT and hybrid are comparatively analyzed to find out
best combination of classifier and feature extraction scheme.
In the proposed work kNN, GAL and ISNN are compared with each other
as classifiers under MGH, 2D-CWT and hybrid feature extraction method.
According to recent work ISNN performs better than GAL but according to
the proposed work GAL results are better than ISNN by comparing their
number of nodes, computational time and performance. It can also be seen from
results that kNN is better classifier than GAL and GAL is better than ISNN by
comparing their computational load and performance. The performance
evaluation is given on the basis of four modalities which are: US bladder
image, US phantom image, CT image and MRI. For accurate performance,
the results are taken on the basis of 11 images of MRI modality.
Grow and learn GAL is same as ISNN. The only difference is that when
class of winner node is equal to the class of the input vector there is no increment
in weights. But when we calculated the results of classification which is shown in
the next section, GAL perform better than ISNN. GAL performance
percentage is better than ISNN as less no of nodes generate during training and
use less training and segmentation time as compared to ISNN.
After feature extraction, the classification process takes place. GAL, ISNN
and kNN are used as classifiers. Classification process has two phases namely the
training phase and testing phase. The data is trained and weights are assigned
to that data in the training process, then labels are assigned to the whole
original image in the testing process. Hence, this is the reason to use supervised
classifiers in which expert chooses training points from each class of input
image which is to be segmented. Here kNN, GAL and ISNN are competitor
classifiers to be used for classification. In this study medical image segmentation
of four modalities as us bladder, us phantom, CT head and MR Head
images are evaluated using kNN, GAL and ISNN as classifiers under MGH,
2D-CWT and hybrid feature extraction methods. In the literature work it is claimed
that ISNN is better than GAL but on the basis of this work we can conclude that
GAL is better than ISNN because GAL uses less training time and
segmentation time and generates less no of nodes during training than ISNN.
The segmentation performance result of classifiers are evaluated on the bases of 11
MRI images and the conclusion can be drawn from the observed results that kNN
is superior and outperformed both GAL and ISNN. For segmentation, GAL is
better than ISNN under MGH, 2D-CWT and hybrid feature extraction methods.
Advantages:
The demonstrated frame work or the system is capable to meet the demand
for selecting best approach in order to meet the given time constraints and
accuracy standards in medical image segmentation.
Disadvantages:
It require too many parameters for proper performance yet these fail to
meet the desired performance level.
Year: 2011
Author: Ruogu Fang, Ramin Zabih, Ashish Raj, and Tsuhan Chen
The system propose a novel approach that applies global optimal tree-
metrics graph cuts algorithm on multi-phase contrast enhanced contrast enhanced
MRI for liver tumor segmentation. To address the difficulties caused by low
contrasted boundaries and high variability in liver tumor segmentation, we first
extract a set of features in multi-phase contrast enhanced MRI data and use color-
space mapping to reveal spatial-temporal information invisible in MRI intensity
images. Then we apply efficient tree-metrics graph cut algorithm on multi-phase
contrast enhanced MRI data to obtain global optimal labeling in an unsupervised
framework. Finally we use tree-pruning method to reduce the number of available
labels for liver tumor segmentation. Experiments on real-world clinical data show
encouraging results. This approach can be applied to various medical imaging
modalities and organs.
Advantages:
Disadvantages:
In this paper, the use of TPS to find the area with the most significant
displacement is proposed. The lateral ventricle boundary is obtained after the
lateral ventricular compartments are segmented by the modified Fuzzy-C means
algorithm in our previous work. Due to the discrepancy in age, sex, personal brain
structure, and even MR image slice number, though similar to each other, shapes
of diagnostic and template images may be varied greatly to each. Therefore, an
interpolated image should be served as a new template which morphs between the
template and diagnostic images to enhance the accuracy for the later processes. A
nonlinear morphing method based on TPS is used here to create a more accurate
interpolation result. It is found that the orientation between diagnostic and template
MR images is the key for creating an accurate result for the TPS displacement
analysis. Therefore, it is important to adjust the diagnostic MR image orientation
according to the correlation between the diagnostic and template images before the
TPS displacement analysis. Because the TPS output data is based on the
displacement between the diagnostic and template ventricles, TPS displacement
values can also serve as an error measure to be used in the process of finding the
best orientation match between the template and the diagnostic image. The
displacement information from the TPS analysis on the Skeleton has therefore used
in our first reorientation method by the Radial Basis Function Neural Network
(RBF-NN) as its error measure input. The system show that the computing speed
is at least more than 100 times faster than using TPS and RBF-NN.
Thin Plate Spline (TPS) is applied for three times in this study. It is at first
served as a morphing transformation method to retrieve an interpolated image
between standard template and the diagnostic images to reduce the effect of their
boundary shape discrepancy. As Radial Basis Neural Network (RBF-NN) is a type
of feed forward neural network that can be used to approximate any function, by
training it to function as a form of mean operator on an error measure, the
orientation that gives the least output value from the network can be used as an
indicator of the best orientation. From the results, it can be seen that the network is
able to find the best orientation using
Advantages:
Disadvantages:
Next, each of the PD images was registered to the target image, and the
registered images were averaged to form a new target image. This process was
repeated five times, when no substantive changes were seen in the tar-get after a
further iteration. Registration was performed using the root mean square difference
in intensity as the similarity measure to assess the difference between the target
and registered images, and the types of deformation of the registered image
allowed were rigid body (translation and rotation) and scaling in three orthogonal
directions. Intensity rescaling was performed as part of the registration procedure
so that all registered images contributed equally to the final average PD template
image. Two representative slices from the resulting average PD template image are
shown in in fig. 1. Using the lesion ROIs as a mask, binary images of the lesions
were produced for each patient, which were then transformed to the coordinate
space of the template image, using the same transform as was found when
registering the PD image to the tem-plate. The transformed binary images were
then summed, divided by the number of patients in the sample, and downsampled
to 256x256x256 pixels, to give the probability (relative frequency) of lesion
occurrence at each anatomical position. This downsampling provides additional
spatial smoothing so that the lesion frequency map can be constructed using a
smaller patient sample. Furthermore, any subse-quent slight errors in registration to
the template will have minimal impact because of the spatial smoothing.
Advantages:
Disadvantages:
HARDWARE REQUIREMENTS:
• Ram : 8 GB
SOFTWARE REQUIREMENTS:
• O/S : Windows 7.
• Language : Python
• Front End : Anaconda Navigator – Spyder
SOFTWARE DESCRIPTION:
Python
Python is one of those rare languages which can claim to be both simple and
powerful. You will find yourself pleasantly surprised to see how easy it is to
concentrate on the solution to the problem rather than the syntax and structure of
the language you are programming in. The official introduction to Python is
Python is an easy to learn, powerful programming language. It has efficient high-
level data structures and a simple but effective approach to object-oriented
programming. Python's elegant syntax and dynamic typing, together with its
interpreted nature, make it an ideal language for scripting and rapid application
development in many areas on most platforms. I will discuss most of these features
in more detail in the next section.
5.3.2 Features of Python
Simple
Easy to Learn
As you will see, Python is extremely easy to get started with. Python has an
extraordinarily simple syntax, as already mentioned.
High-level Language
When you write programs in Python, you never need to bother about the
low-level details such as managing the memory used by your program, etc.
Portable
Due to its open-source nature, Python has been ported to (i.e. changed to
make it work on) many platforms. All your Python programs can work on any of
these platforms without requiring any changes at all if you are careful enough to
avoid any system-dependent features.
You can even use a platform like Kivy to create games for your computer
and for iPhone, iPad, and Android.
Interpreted
Python, on the other hand, does not need compilation to binary. You just run
the program directly from the source code. Internally, Python converts the source
code into an intermediate form called bytecodes and then translates this into the
native language of your computer and then runs it. All this, actually, makes using
Python much easier since you don't have to worry about compiling the program,
making sure that the proper libraries are linked and loaded, etc. This also makes
your Python programs much more portable, since you can just copy your Python
program onto another computer and it just works!
Object Oriented
Extensible
If you need a critical piece of code to run very fast or want to have some
piece of algorithm not to be open, you can code that part of your program in C or
C++ and then use it from your Python program.
Embeddable
You can embed Python within your C/C++ programs to give scripting
capabilities for your program's users.
Extensive Libraries
The Python Standard Library is huge indeed. It can help you do various
things involving regular expressions, documentation generation, unit testing,
threading, databases, web browsers, CGI, FTP, email, XML, XML-RPC, HTML,
WAV files, cryptography, GUI (graphical user interfaces), and other system-
dependent stuff. Remember, all this is always available wherever Python is
installed. This is called the Batteries Included philosophy of Python.
Besides the standard library, there are various other high-quality libraries
which you can find at the Python Package Index.
Outside PYTHON images may be of three types i.e. black & white, grey scale and
colored. In PYTHON, however, there are four types of images. Black & White
images are called binary images, containing 1 for white and 0 for black. Grey scale
images are called intensity images, containing numbers in the range of 0 to 255 or
0 to 1. Colored images may be represented as RGB Image or Indexed Image.
In RGB Images there exist three indexed images. First image contains all the red
portion of the image, second green and third contains the blue portion. So for a 640
x 480 sized image the matrix will be 640 x 480 x 3. An alternate method of colored
image representation is Indexed Image. It actually exist of two matrices namely
image matrix and map matrix. Each color in the image is given an index number
and in image matrix each color is represented as an index number. Map matrix
contains the database of which index number belongs to which color.
Key Features
FEASIBILITY STUDY
The feasibility study is carried out to test whether the proposed
system is worth being implemented. The proposed system will be selected if it is
best enough in meeting the performance requirements.
• Economic Feasibility
• Technical Feasibility
• Behavioral Feasibility
Economic Feasibility
Technical Feasibility
This study center around the system’s department hardware, software and to
what extend it can support the proposed system department is having the required
hardware and software there is no question of increasing the cost of implementing
the proposed system. The criteria, the proposed system is technically feasible
and the proposed system can be developed with the existing facility.
Behavioral Feasibility
System Implementation
The active user must be aware of the benefits of using the system
Their confidence in the software built up
Proper guidance is impaired to the user so that he is comfortable in
using the application
Before going ahead and viewing the system, the user must know that for
viewing the result, the server program should be running in the server. If the server
object is not running on the server, the actual processes will not take place.
User Training
To achieve the objectives and benefits expected from the proposed system it
is essential for the people who will be involved to be confident of their role in the
new system. As system becomes more complex, the need for education and
training is more and more important. Education is complementary to training. It
brings life to formal training by explaining the background to the resources for
them. Education involves creating the right atmosphere and motivating user staff.
Education information can make training more interesting and more
understandable.
Training on the Application Software
Operational Documentation
Once the implementation plan is decided, it is essential that the user of the
system is made familiar and comfortable with the environment. A documentation
providing the whole operations of the system is being developed. Useful tips and
guidance is given inside the application itself to the user. The system is developed
user friendly so that the user can work the system from the tips given in the
application itself.
System Maintenance
The maintenance phase of the software cycle is the time in which software
performs useful work. After a system is successfully implemented, it should be
maintained in a proper manner. System maintenance is an important aspect in the
software development life cycle. The need for system maintenance is to make
adaptable to the changes in the system environment. There may be social, technical
and other environmental changes, which affect a system which is being
implemented. Software product enhancements may involve providing new
functional capabilities, improving user displays and mode of interaction, upgrading
the performance characteristics of the system. So only thru proper system
maintenance procedures, the system can be adapted to cope up with these changes.
Software maintenance is of course, far more than “finding mistakes”.
Corrective Maintenance
Adaptive Maintenance
Perceptive Maintenance
CONCLUSION
This paper has provided a comprehensive overview of the state of the art
MRI-based brain tumor segmentation methods. Many of the current brain tumor
segmentation methods operate MRI images due to the non-invasive and good soft
tissue contrast of MRI and employ classification and clustering methods by using
different features and taking spatial information in a local neighborhood into
account. The purpose of these methods is to provide a preliminary judgment on
diagnosis, tumor monitoring, and therapy planning for the physician. Although
most of brain tumor segmentation algorithms have relatively good results in the
field of medical image analysis, there is a certain distance in clinical applications.
Due to a lack of interaction between researchers and clinicians, clinicians still rely
on manual segmentation for brain tumor in many cases. The existence of many
tools aims to do pure research and is hardly useful for clinicians. Therefore,
embedding the developed tools into more user- friendly environments will become
inevitable in the future. Recently, some standard clinical acquisition protocols
focusing on feasibility studies are trying to formulate to improve the clinical
applications more quickly. Apart from the evaluation of accuracy and validity for
the results of brain tumor segmentation, computation time is also an important
criterion. The current standard computation time is in general a few minutes. The
real-time segmentation will be hard to achieve, but computation time over a few
minutes is unacceptable in clinical routine. Another crucial aspect for brain tumor
segmentation methods is robustness. If an automatic segmentation technique fails
in some cases, clinicians will lose their trust and not use this technique. Therefore,
the robustness is also one of the major assessment criteria for each new method
applied in clinical practice. Some current brain tumor segmentation methods
provide robust results within a reasonable computation time.
REFERENCES
[1] M. P. Gupta and M. M. Shringirishi, Implementation of brain tumor
segmentation in brain mr images using k –means clustering and fuzzy c -means
algorithm, International Journal of Computers & Technology , vol. 5, no. 1, pp. 54-
59, 2013.
[2] D. N. Louis, H. Ohgaki, O. D. Wiestler, W. K. Cavenee, P. C. Burger, A.
Jouvet, B. W. Scheithauer, and P. Kleihues, The 2007 who classification of
tumours of the central nervous system, Acta Neuropathologica, vol. 114, no. 2,
pp. 97-109, 2007.
[3] Z.-P. Liang and P. C. Lauterbur, Principles of Magnetic Resonance Imaging: A
Signal Processing Perspective . The Institute of Electrical and Electronics
Engineers Press, 2000.
[4] P. Y. Wen, D. R. Macdonald, D. A. Reardon, T. F. Cloughesy, A. G. Sorensen,
E. Galanis, J. DeGroot, W. Wick, M. R. Gilbert, A. B. Lassman, et al., Updated
response assessment criteria for high-grade gliomas:Response assessment in neuro-
oncology working group, Journal of Clinical Oncology , vol. 28, no. 11, pp. 1963-
1972, 2010.
[5] A. Drevelegas and N. Papanikolaou, Imaging modalities in brain tumors, in
Imaging of Brain Tumors with Histological Correlations . Springer, 2011, pp. 13-
33.
[6] J. J. Corso, E. Sharon, S. Dube, S. El-Saden, U. Sinha, and A. Yuille, Efficient
multilevel brain tumor segmentation with integrated bayesian model classification,
Medical Imaging, IEEE Transactions on , vol. 27, no. 5, pp. 629-640,2008.
[7] Y.-L. You, W. Xu, A. Tannenbaum, and M. Kaveh, Behavioral analysis of
anisotropic diffusion in image processing, Image Processing, IEEE Transactions
on,
vol. 5, no. 11, pp. 1539-1553, 1996.
[8] J. Weickert, Anisotropic Diffusion in Image Processing , vol. 1. Teubner
Stuttgart, 1998.
[9] T. Ogden, Essential Wavelets for Statistical Applications and Data Analysis .
Springer, 1997.
[10] R. D. Nowak, Wavelet-based rician noise removal for magnetic resonance
imaging, Image Processing, IEEETransactions on, vol. 8, no. 10, pp. 1408-1419,
1999.
[11] A. Buades, B. Coll, and J.-M. Morel, A non-local algorithm for image
denoising, in Computer Vision and Pattern Recognition, 2005. CVPR 2005. IEEE
Computer SocietyConference on , IEEE, 2005, vol. 2, pp. 60-65.
[12] J. V. Manj ́ on, P. Coup ́ e, L. Mart ́ ı-Bonmat ́ı, D. L. Collins, and M. Robles,
Adaptive non-local means denoising of mr images with spatially varying noise
levels, Journal of Magnetic Resonance Imaging , vol. 31, no. 1, pp. 192-203, 2010.
[13] S. Prima and O. Commowick, Using bilateral symmetry to improve non-local
means denoising of mr brain images, in Biomedical Imaging (ISBI), 2013 IEEE
10th International Symposium on , IEEE, 2013, pp. 1231-1234.
[14] P. Hoyer, Independent component analysis in image denoising, Master degree
dissertation, Helsinki University of Technology, 1999.
[15] K. Phatak, S. Jakhade, A. Nene, R. Kamathe, and K. Joshi, De-noising of
magnetic resonance images using independent component analysis, in Recent
Advances in Intelligent Computational Systems (RAICS), 2011 IEEE , IEEE,
2011, pp. 807-812.
[16] I. Diaz, P. Boulanger, R. Greiner, and A. Murtha, A critical review of the
effects of de-noising algorithms on mri brain tumor segmentation, in Engineering
in Medicine and Biology Society, EMBC, 2011 Annual International Conference
of the IEEE , IEEE, 2011, pp. 3934-3937.
[17] C. Fennema-Notestine, I. B. Ozyurt, C. P. Clark, S. Morris, A. Bischoff-
Grethe, M. W. Bondi, T. L. Jernigan, B. Fischl, F. Segonne, D. W. Shattuck, et al.,
Quantitative evaluation of automated skull-stripping methods applied to
contemporary and legacy images: Effects of diagnosis, bias correction, and slice
location, Human Brain Mapping , vol. 27, no. 2, pp. 99-113, 2006.
[18] A. H. Zhuang, D. J. Valentino, and A. W. Toga, Skull- stripping magnetic
resonance brain images using a model- based level set, NeuroImage , vol. 32, no.
1, pp. 79-92, 2006.
[19] R. Roslan, N. Jamil, and R. Mahmud, Skull stripping magnetic resonance
images brain images: Region growing versus mathematical morphology,
International Journal of Computer Information Systems and Industrial
Management
Applications , vol. 3, pp. 150-158, 2011.
[20] S. Bauer, L.-P. Nolte, and M. Reyes, Skull-stripping for tumor-bearing brain
images, arXiv preprint arXiv: 1204.0357, 2012.
[21] S. F. Eskildsen, P. Coup ́ e, V. Fonov, J. V. Manj ́ on, K. K. Leung, N.
Guizard, S. N. Wassef, L. R. Østergaard, and D. L. Collins, Beast: Brain extraction
based on nonlocal segmentation technique, NeuroImage , vol. 59, no. 3, pp. 2362-
2373, 2012.
SCREENSHOTS: