An Intelligent Support System For Automatic Detection of Cerebral Vascular Accidents From Brain CT Images

Download as pdf or txt
Download as pdf or txt
You are on page 1of 35

Accepted Manuscript

An Intelligent Support System for Automatic Detection of Cerebral


Vascular Accidents from Brain CT Images

Elmira Hajimani , M.G. Ruano , A.E. Ruano

PII: S0169-2607(16)30626-5
DOI: 10.1016/j.cmpb.2017.05.005
Reference: COMM 4420

To appear in: Computer Methods and Programs in Biomedicine

Received date: 16 June 2016


Revised date: 28 April 2017
Accepted date: 19 May 2017

Please cite this article as: Elmira Hajimani , M.G. Ruano , A.E. Ruano , An Intelligent Support System
for Automatic Detection of Cerebral Vascular Accidents from Brain CT Images, Computer Methods and
Programs in Biomedicine (2017), doi: 10.1016/j.cmpb.2017.05.005

This is a PDF file of an unedited manuscript that has been accepted for publication. As a service
to our customers we are providing this early version of the manuscript. The manuscript will undergo
copyediting, typesetting, and review of the resulting proof before it is published in its final form. Please
note that during the production process errors may be discovered which could affect the content, and
all legal disclaimers that apply to the journal pertain.
ACCEPTED MANUSCRIPT

Highlights

 A RBFNN based system for automatic diagnosis of CVAs from brain CT is proposed.
 The best possible RBFNN topology, inputs and parameters are identified by MOGA.
 Symmetry features along with 1st and 2nd order statistics compose the feature space.
 98.01% specificity and 98.22% sensitivity in a set of 1,867,602 pixels are achieved.
 The proposed approach compares favourably with existing approaches

T
IP
CR
US
AN
M
ED
PT
CE
AC
ACCEPTED MANUSCRIPT

An Intelligent Support System for Automatic Detection of Cerebral Vascular


Accidents from Brain CT Images
Elmira Hajimania,*, M.G. Ruanob, A. E. Ruanoc
a
Faculty of Science and Technology, University of Algarve, Faro, Portugal, (e-mail:
[email protected]).
b
Faculty of Science and Technology, University of Algarve, Faro, Portugal and CISUC,

T
University of Coimbra, Portugal (e-mail: [email protected])

IP
c
Faculty of Science and Technology, University of Algarve, Faro, Portugal and IDMEC,
Instituto Superior Técnico, University of Lisbon, Portugal, (e-mail: [email protected] )

CR
*
Corresponding author. Tel: (+351) 915 918 528. Postal address: Lab. 2.75, FCT, Universidade
do Algarve, Campus Gambelas, 8005-139 Faro, Portugal.

Abstract
US
Objective: This paper presents a Radial Basis Functions Neural Network (RBFNN) based
AN
detection system, for automatic identification of Cerebral Vascular Accidents (CVA) through
analysis of Computed Tomographic (CT) images.

Methods: For the design of a neural network classifier, a Multi Objective Genetic Algorithm
M

(MOGA) framework is used to determine the architecture of the classifier, its corresponding
parameters and input features by maximizing the classification precision, while ensuring
ED

generalization.

This approach considers a large number of input features, comprising first and second order pixel
intensity statistics, as well as symmetry/asymmetry information with respect to the ideal mid-
PT

sagittal line.

Results: Values of specificity of 98% and sensitivity of 98% were obtained, at pixel level, by an
CE

ensemble of non-dominated models generated by MOGA, in a set of 150 CT slices (1,867,602


pixels), marked by a NeuroRadiologist. This approach also compares favorably at a lesion level
with three other published solutions, in terms of specificity (86% compared with 84%), degree of
AC

coincidence of marked lesions (89% compared with 77%) and classification accuracy rate (96%
compared with 88%).

Keywords: Neural Networks; Symmetry features; Multi-Objective Genetic Algorithm;


Intelligent support systems; Cerebral Vascular Accident.

1. Introduction
ACCEPTED MANUSCRIPT

Cerebral Vascular Accident (CVA), also called stroke, is caused by the interruption of blood
supply to the brain, mainly due to a blood vessel blockage (i.e., ischemic), or by a haemorrhagic
event. The cut-off of oxygen and nutrients supplies causes brain tissue irreversible damages, if
not detected during the first 2-3 hours. Stroke accounted for approximately 1 of every 19 deaths
in the United States in 2009, according to [1]. Computed Tomography (CT) is the most
commonly used CVA imaging diagnosis modality, available at almost all emergency units.
When CVA is timely diagnosed, morbidity and mortality are minimised [2].

T
The increasing occurrence of CVAs in developed and developing countries, together with the
insufficient number of Neuroradiologists and the lack of full-time expert Radiologists in some

IP
institutions, exacerbated by the fact that each exam is constituted by several images requiring an
accurate analysis, leads to delays on the production of the clinical final report and subsequent

CR
clinical treatment, compromising the CVA’s recovery prognosis. In addition, at early stages of
CVA, changes in the tomographic images may not be perceived by the human eye due to the
inconspicuousness of the lesions.

US
For these reasoning, the existence of an automated system would be of paramount importance to
detect subtle alterations, motivating the authors to create a computational intelligent application
AN
capable of assisting the Neuroradiologist in the analysis of CT scan images. The proposed
application envisage enabling a preliminary trigger of a pathologic occurrence and a better
performance of the CVA green line.
M

In this paper, a Radial Basis Functions Neural Network (RBFNN) based system is proposed for
automatic detection of CVAs from brain CT images. The majority of the existing methods for
designing the neural network classifier do not use an approach that determines the model
ED

structure that best fits the application at their hand, while simultaneously selecting the best input
features. Moreover, their design typically does not consider multiple conflicting objectives such
as minimization of the number of False Detections (FD) in the training dataset while achieving a
PT

desired level of model generalization, or, maximising the classification precision while reducing
the model complexity. For this purpose, an hybrid of a multi-objective evolutionary technique is
used as a design framework for features and topology selection, and, state-of-the-art derivative
CE

based algorithms are used for neural network parameter estimation, taking into account multiple
objectives, as well as their corresponding restrictions and priorities [2, 3].
AC

Moreover, to the best of our knowledge, none of existing classifiers consider the asymmetry
caused by lesions in the intracranial area. In the proposed approach, a group of symmetry
features that were proposed in [4], is used along with other statistical features, as inputs to the
designed classifier.

The rest of the paper is organized as follows: Section 2 provides an overview of existing lesion
segmentation strategies. The data acquisition process is explained in section 3. Section 4
describes the features that are used in this study. Section 5 explains how the Multi Objective
ACCEPTED MANUSCRIPT

Genetic Algorithm (MOGA) is used to automatically design the RBFNN classifier. Experimental
formulations and results are given in section 6. Section 7 discusses the results obtained in
comparison with other published approaches. Conclusions are drawn in section 8.

2. Related works

Based on [5], a review of different lesion segmentation approaches, one can divide lesion
segmentation strategies into two subgroups: supervised and unsupervised strategies.

T
Supervised approaches are those that use some kind of a priori information to perform the lesion

IP
segmentation. The group of supervised strategies can be further subdivided into two sub-groups:

a) In the first subgroup, all approaches use atlas information, therefore requiring the

CR
application of a registration process to the analyzed image to perform the segmentation.

As an example, the authors in [6] used a combination of two techniques for brain lesion detection

US
from CT perfusion maps: finding asymmetries among the two hemispheres and then comparing
the captured images to a brain atlas anatomy. For generating the asymmetry map, first the
symmetry axis is approximated as the straight line that minimizes the least square error between
all centers of masses’ coordinates, and then the intensity values of the corresponding pixels on
AN
the left and right side of the image are compared. Those pixels with a significant difference are
considered as potential lesions. To perform a detailed description of lesions a second step is
required, where position image registration of the brain template is made. The goal of the
M

registration algorithm is to maximize the similarity between the template image and the newly
acquired image.
ED

The work done in [7] can also be considered in this subgroup. This study presents an automated
template-guided algorithm for the segmentation of ventricular CerebroSpinal Fluid (CSF) from
ischemic stroke CT images. In the proposed method, the authors use two ventricular templates,
PT

one extracted from a normal brain (VT1) and the other built from several pathological scans
(VT2). VT1 is used for registration and VT2 to define the region of interest. In the registration
process, they use the Fast Talairach Transformation [8], which takes care of the “tilting” angle.
CE

Automatic thresholding is applied on a slice-by-slice basis, to cater for the variability of CSF
intensity values across the slices in the same scan. The distributions of the CSF, White Matter
(WM) and Gray Matter are analyzed and only voxels in the CSF range and WM range are used
AC

in the calculation of the histogram, employed by Otsu’s automatic thresholding algorithm [9].
Finally, artifacts are removed with the help of VT2.

b) All approaches which perform an initial training step on features extracted from manually
segmented images, annotated by Neuroradiologists, can be considered as another subgroup of
supervised strategies [5]. In this subgroup, different classifiers, such as Artificial Neural
Networks, k-Nearest Neighbors, AdaBoost, Bayesian classifiers or decision trees, alone or
ACCEPTED MANUSCRIPT

combined, have been used to perform the segmentation. The work presented in this article can be
considered in this category.

The method applied in [10] is also an example of this subgroup. Primarily the method uses
morphology operations and wavelets based filtering for image denoising. Then asymmetric parts
of the brain and their neighbors are extracted being considered as the region of interest for
specifying relevant features (such as texture, contrast, homogeneity, etc). Finally, k-means
clustering and Support Vector Machines (SVM) are used for classification and provide the

T
contour of the brain tumor.

IP
The work presented in [11] uses a wavelet based statistical method for classifying brain tissues
into normal, benign and malignant tumours. The authors first obtain the second level discrete

CR
wavelet transform of each CT slice. The Gray Level Co-occurrence Matrix (GLCM) is then
calculated over the low frequency part of the transformed image. Finally, features are calculated
from the GLCM matrix. Genetic algorithms and principle component analysis are employed for
feature selection and SVM for classification.
US
In [12] a computer tomography (CT) brain image analysis system is proposed, with four phases:
enhancement, segmentation, feature extraction and classification. The enhancement phase
AN
reduces the noise using an edge-based selective median filter (ESMF); the segmentation phase
extracts the suspicious region applying a modified version of a genetic algorithm; the feature
extraction phase extracts the textural features from the segmented regions and the last phase
M

classifies the image. To diagnose and classify the image, the authors used a RBFNN classifier.

Regarding unsupervised strategies, where no prior knowledge is used, two different sub-groups
ED

can also be identified:

• A sub-group of methods that segment the brain tissue to allow lesion segmentation. These
approaches usually detect lesions as outliers on each tissue, rather than adding a new class to the
PT

classification problem. The works presented in [13] and [14] follow this strategy.

• A sub-group that uses only lesion properties for segmentation. These methods directly
CE

segment the lesions according to their properties, without providing tissue segmentation. The
works described in [15, 16] belong to this category.
AC

3. Data acquisition

A database of existing CT images was used in this prospective study. All images were obtained
from the same Siemens equipment. Each exam, where no contrast was applied, was composed by
two ranges, one to study the posterior fosse and the other covering the remaining of the brain
(5/5 mm and 10/10 mm, respectively).

In order to collect the opinion of Neuroradiologists about pathologic areas within brain CT
images in an accurate and convenient way, a web-based tool was developed [17]. Using this tool,
ACCEPTED MANUSCRIPT

the existing database of CT images was used for Neuroradiologists to analyze and mark the
images either as normal or abnormal. For the abnormal ones, the doctor was asked to designate
the lesion type and to manually trace the contours of abnormal region(s) on each CT’s slice
image. Figure 1 shows the activity diagram of the Neuroradiologist in the developed tool.

The administrator of the developed web-based tool can then download a text file in which the
coordinates of each marked pixel (i.e., a lesion) are specified. These pixels are considered as
abnormal data samples. The resolution of each CT slice is pixels and the intensity

T
value of each pixel is an integer in the range [0 255], 0 being completely black and 255
completely white. Within a CT slice, all the intracranial pixels which are not marked as lesions

IP
will be considered as normal data samples.

CR
Our collaborating Neuroradialogist registered his opinion for 7 patients (150 CT slices). 24 out
of the 150 CT slices had lesions within their intracranial area, corresponding to 64,786 abnormal
pixels. All lesions were marked as ischemic stroke. To obtain the coordinates of all normal

US
pixels, Algorithm 1 (described below) is used. It calls Algorithm 2 (described afterwards) for
artifact removal on each image. Applying Algorithm 1, we obtained 1,802,816 normal pixels. As
a result, we have a total of 1,867,602 normal and abnormal pixels to work with.
AN
M
ED
PT
CE
AC
ACCEPTED MANUSCRIPT

T
IP
CR
US
AN
M
ED
PT
CE

Figure1. Activity diagram of the Neuroradiologist in the data acquisition tool

Algorithm 1 Obtaining the coordinate of normal pixels


Input: text file, say T, where the coordinate of abnormal pixels and the path from which the
AC

image can be retrieved are saved.


1. Let be a structure that is constructed from text file T. contains the
information of each CT exam.
2. For to
2.1. Pass through Algorithm 2, to remove the skull and other artifacts.
2.2. For each image in
2.2.1. Let be two vectors containing the location of abnormal pixels
2.2.2. Let be the intensity of the pixel located in
ACCEPTED MANUSCRIPT

2.2.3. If and and


2.2.3.1. Insert and as a row in a text file O.
End if
End for
End for
Output: text file O

Algorithm 2 Artifact removal algorithm in brain CT images [18]


Input: Brain CT images of one examination

T
1. Skull detection:

IP
1.1. Remove pixels whose intensities are less than 250.
1.2. Use the Connected Component algorithm [19] to choose the largest component as the
candidate skull.

CR
1.3. Remove the small holes within the candidate skull, by inverting the matrix of candidate skull
and applying the Connected Component algorithm for the second time. Those connected
components whose area are less than 200 pixels are considered as holes and will be filled using
the bone intensity value

US
2. Removing CT slices with either unclosed skulls or skull containing too many separate regions:
Having completed step 1.3, we have already all connected components at hand. As a result, we can
count the number of big holes (e.g., areas more than 200 pixels wide). If this number is equal to 2,
AN
it will be considered as closed skull; otherwise the slice will be removed from the desired set.
3. Intracranial area detection: All CT images that successfully passed step 2 contain only two black
regions separated by the skull. To detect which black area is related to the intracranial part, the
mass centre of the skull is calculated, the region containing the mass centre being considered as
M

intracranial area.
Output: Intracranial part of a subset of input CT images.
ED

4. Feature space

Having a set of pixel coordinates whose labels (normal or abnormal) are already determined by
PT

expert, we are now able to produce our dataset by extracting the corresponding features from the
images. Each CT image is represented as a matrix with rows and columns where
stands for the intensity of pixel in row and column . Three groups of features are used to
CE

construct the feature space: first order statistics, second order statistics and symmetry features.
Table I describes the 51 features considered. First order statistics estimate properties of
individual pixel values (e.g. to and to ), ignoring the spatial interaction between
AC

the image pixels. Second order features estimate properties of two pixel values occurring at
specific locations relative to each other by constructing a Gray Level Co-occurrence Matrix. To
extract some of the first and second order statistical features, a window of size 31*31 [20, 21]
centered at the pixel is employed.

The variance of pixel intensities within a window is denoted by . Given centred at


point , is a row vector with the intensities of the 31 pixels taken from the horizontal
line centered at and is a column vector with the intensities of the 31 pixels taken from
ACCEPTED MANUSCRIPT

the vertical line centered at . For calculating features and to , gray


levels of the histogram of pixel intensities within window are calculated. Each bin of
histogram is represented by .

A GLCM matrix is a two-dimensional matrix where both the rows and the columns represent a
set of possible image values G (e.g. gray tones). The value of indicates how many times
the value co-occurs with value in some designated spatial relationship. The spatial
relationship is usually defined by a distance and a direction . In this study, to calculate the 8

T
gray level GLCM of , the displacement parameters considered were and
. As a result, 4 GLCM matrices are derived, each one belonging to one specific

IP
and then the average is computed in order to obtain a direction invariant GLCM matrix.

CR
In the formulas used in Table I, the mean value of matrix is represented by and the mean and
standard deviation for the rows and columns of are defined in (1) and (2) respectively.

∑ US ∑ (1)
AN
∑ ∑ ( ) (2)

Moreover, is the ith entry in the marginal-probability matrix obtained by summing the rows
of and is obtained by summing the columns of .
M

Given the ideal mid-sagittal line, symmetry features aim to compare one side of the brain to the
other side and discover if there are any suspicious differences. To detect the ideal midline and to
ED

rotate tilted images to make the ideal midsagittal line perpendicular to the x-axis, the method
summarized in Algorithm 3 is employed.
PT

Algorithm 3 Ideal midline detection of the brain CT [18, 22]


Input: Brain CT images of one exam
1. Use Algorithm 2 to remove artifacts.
CE

2. Since the concave shape of intracranial region will affect the accuracy of search for finding ideal
midline, CT slices with high amount of concavity are found and excluded:
2.1. For each CT slice
AC

2.1.1. Extract the contour of intracranial region.


2.1.2.
2.1.3. For to 180
2.1.3.1. Rotate contour by degree.
2.1.3.2.
2.1.3.3. For to
2.1.3.3.1. Scan the pixels of the contour in row and define the Far Left ( ) and
Far Right ( ) junctions.
2.1.3.3.2. Let be the number of pixels in row which reside between and
ACCEPTED MANUSCRIPT

and that are not located inside the intracranial region.


2.1.3.3.3.
End for
2.1.3.4.
End for
End for
2.2. Sort CT slices based on their corresponding values and select the first slices with
the least amount of concavity.

T
3. To find the line that maximizes the symmetry of the resulting halves, a rotation angle search around
the mass centre of the skull is performed:

IP
3.1. For each CT slice remaining from step 2:
3.1.1. Let be the maximum angle that a given CT image can be tilted.

CR
3.1.2. Let be the symmetry cost at angle
3.1.3. For to
3.1.3.1. Calculate ∑ | | where is the number of rows in the current CT
slice, and are the distances between the current approximate midline and the left

End for US
and right side of the skull edge in row, respectively.

3.1.4. Select rotation angle whose symmetry cost is minimum.


AN
End for
3.2. The final rotation degree for all CT slices is determined as the median value of rotation angles
obtained for each CT slice.
4. Rotate all CT slices around their corresponding skull mass centre based on the rotation degree
M

obtained in step 3.2


5. Line is considered as the ideal midline of the brain CT image .
Output: CT images are aligned and their corresponding ideal midlines determined.
ED

To extract symmetry features, a window of size centred at pixel , marked by a


PT

clinical expert as normal or abnormal, and its contralateral part with respect to the midline,
window centred at the pixel , are considered (please see Figure 2-b).
CE
AC
ACCEPTED MANUSCRIPT

T
IP
CR
(a) (b)
Figure 2. (a) Original brain CT image; (b) After skull removal and realignment, the ideal midline

US
is drawn in yellow. The green point shows the mass centre (centroid) of the skull upon which the
alignment rotation is performed. A window of size 31x31 is considered around the pixel located
at (365,279) and shown in red; its contralateral part with respect to the midline is shown in blue
AN
[4].

Having identified and , we can then specify how similar these two regions are by
M

calculating the Pearson Correlation Coefficient (PCC), as stated in (3). The norm and squared
norms are also two dissimilarity measures that can be obtained using (4) and (5), respectively.
Comparing the intensity value of the pixel that is marked by the expert, and its corresponding
ED

pixel in the contralateral part, another symmetry feature defined in (6) can be obtained.

∑ ∑ ( )( ) (3)
PT

∑ ∑ | | (4)
CE

∑ ∑ ( ) (5)
AC

(6)

In the previous equations, is the intensity value of pixel located at within the
corresponding window; , , and are the mean and standard deviation of the
intensity values within window and its contralateral part, window , respectively. In Table
I, to are symmetry features for three different window sizes ( ).
ACCEPTED MANUSCRIPT

In Table I, features to are taken from [21]; to are used in [23]; to and
to are from [24]; is a built-in MATLAB function; and are used in [25];
to are obtained from [20] and to were proposed in [4].

Table I. Features considered

Description

T
IP
CR
= ∑ ∑

US
AN

Plh = ∑ | |
M

plv = ∑ | |

cxm =
ED

Skewness= ∑ ∑
PT

Kurtosis= ∑ ∑

Energy=∑
CE

Entropy= ∑
Autocorrelation=∑

AC

Correlation=

Cluster Prominence=∑ ( )
Cluster shade=∑ ( )
Dissimilarity=∑ | |
GLCM Energy =∑
GLCM Entropy= ∑
Homogeneity= ∑ | |
ACCEPTED MANUSCRIPT

Homogeneity = ∑ ⁄
Maximum probability=
Sum of squares =∑
Sum average=∑ where ∑ ∑ |

Sum variance= ∑ where ∑ ∑ |

Sum entropy= ∑ where ∑ ∑ |

T
∑ ∑ |

IP
Difference variance= where

CR
Difference entropy= ∑ ( ) where
∑ ∑ |
Information measure of correlation1= where and are Entropies of
and and
US
Information measure of correlation2=

[
{ }
] ⁄
where
∑ { }
AN

Inverse difference normalized=∑ | |⁄


M

Inverse difference moment normalized=∑

= ∑ ∑
ED




PT


CE
AC
ACCEPTED MANUSCRIPT

5. Neural Network design using the Multi Objective Genetic Algorithm framework

The identification of Neural Network inputs, topology and parameters from data is often done
iteratively in an ad-hoc fashion, focusing mainly on parameters identification. This is because the
number of possibilities for the selection of the model structure (inputs and topology) are usually
very large. Moreover, typically the design criterion is a single measure of the error obtained by
the model, such as the mean-square error or the root-mean-square error, while typically the aim
is to obtain a satisfactory performance (determined by typically more than one criterion) with

T
small networks, i.e., the design problem should be formulated as a multiple-objective problem.

IP
The model used for pixel classification is a Radial Basis Function Neural Network (RBFNN).

CR
US
AN
M
ED

Figure 3. Topology of a RBFNN


PT

The topology of a RBFNN is presented in Figure 3. The second layer units, known as neurons,
are nonlinear functions of their vector inputs, given by,
CE

‖ ‖
(7)

where ‖ ‖ denotes the Euclidean norm, and ci and are, respectively, the location of the
AC

Gaussian function in the input space (denoted as centers) and its spread. The RBFNN output is
given by:

∑ (8)

where n is the number of neurons and the αi are the weights of the network output linear
combiner.
ACCEPTED MANUSCRIPT

As in this application the RBFNNs are used as classifiers, the output y is passed through a
threshold function, in such a way that if y>0, the pixel is classified as abnormal, and normal
otherwise.

In order to identify the best possible Radial Basis Functions neural network structure and
parameters, this work uses a Multi-Objective Genetic Algorithm (MOGA) design framework,
described in, for instance, [2, 26]. This method also allows us to handle multiple, possibly
conflicting objectives. MOGA finds a non-dominated set of individuals through number of

T
generations and then selects preferable individuals from the non-dominated set. A solution is
called non-dominated if none of the objectives can be improved in value without sacrificing

IP
some of the other objective values [27].

CR
In order to be able to use the MOGA approach for finding the best possible model structure and
its corresponding parameters, each possible structure for the NN needs to be formulated as a
chromosome. To do that, the number of neurons in hidden layer is considered as the first

US
component of the chromosome and the remaining components are the indices of features to be
selected from a feature space. Figure 4 shows the topology of the chromosome. The algorithm
starts its work by producing a pre-defined number of individuals as the first generation. The
AN
method then needs a mechanism to compare the individuals and select the best ones with respect
to pre-defined objectives. The objectives, for the case at hand, can be selected from the set as
described in (9).
M

{ | } (9)

where is the number of False Negatives (i.e., those abnormal pixels that are wrongly
ED

classified as normal); is the number of False Positives (i.e., those normal pixels that are
wrongly classified as abnormal) and stands for the Model Complexity. and represent
Test and Training sets respectively. The formula for calculating Model Complexity is given in
PT

(10).
CE

Number of neurons
in hidden layer Maximum number of feature indices that
can be in each chromosome
AC

𝑛 𝜆 𝜆 𝜆 𝑑𝑚 𝜆 𝑑𝑚 𝜆 𝑑𝑀

Minimum number of feature indices that can


be in each chromosome

Figure 4. The topology of the chromosome


ACCEPTED MANUSCRIPT

(10)

For evaluating the individuals in one generation, each NN model is trained with the provided
training dataset (i.e., using the features whose indices are depicted in chromosome). The
Levenberg-Marquardt (LM) algorithm [28, 29], with a formulation that exploits the linear-
nonlinear separability of the NN parameters [30, 31] is employed for training due to its higher
accuracy and convergence rate. Training is terminated if an user-specified number of iterations is
reached, or the performance on a test set reaches a minimum (a procedure known as early-

T
stopping [32]).

IP
Since the result of gradient-based methods, such as LM, depends on the model initial
parameters’ values, for each individual in the current generation, the training procedure is

CR
repeated α times. Within these, the best result is picked up for determining the parameters of the
individual (i.e., the centres, spreads and weights in RBFNNs). In MOGA, there are
different ways for identifying which training trial is the best one, considering as the number of

US
the objectives. The first strategy is to select the training trial which has minimized all objectives
better than the others. In other words, if we consider a dimensional space, the one whose
Euclidean distance from the origin is the smallest, will be considered as the best. The green
AN
arrow in Figure 5 indicates this situation for . In the second strategy, the average of the
objective values for all training trials is calculated and then the trial whose value is the closest to
this average will be selected as the best one (i.e., the red arrow in Figure 5).
M

The other strategies are to select the training trial which minimized the objective (
) better than the other trials. As an example, the yellow and blue arrows in Figure 5 are
the training trials which minimized objective1 and objective 2, respectively.
ED
PT
CE
AC
ACCEPTED MANUSCRIPT

Figure 5. Four different strategies for identifying the best training trial within α=10 training trials

Having trained each individual, we are now able to assign a fitness value which reflects the
individual’s quality. MOGA uses a Pareto-based fitness assignment approach, which was first
proposed by Goldberg and then modified by Fonseca and Fleming [33]. In this method, the
individuals are ranked according to the number of individuals by which they are dominated. For
example, if an individual is non-dominated, its corresponding rank is 0 and if an individual is
dominated by three other individuals, its corresponding rank will be 3. Figure 6 visualizes the

T
Pareto ranking notion.

IP
Objective 2

CR
2

0
1
US 3
AN
1
0
M

0
ED

Objective 1
PT

Figure 6. Pareto ranking [26, 34]

If there exists any preference such as assigning different priorities to each objective or defining a
CE

desired level of performance for each objective (i.e., restrictions), the ranking technique is
slightly modified to take the restrictions and priorities into account. Suppose that and are
the corresponding restrictions of objectives 1 and 2. In the case that both objectives have the
AC

same priorities, the individuals who satisfied the restrictions are assigned a rank equal to the
number of individuals by which they are dominated. The individuals which do not meet the
restrictions are penalized by assigning a higher rank. Figure 7 illustrates this situation.
ACCEPTED MANUSCRIPT

Objective 2
5
5

5
𝑐 1
3

T
0

IP
1
0

CR
5

US 𝑐 Objective 1
AN
Figure 7. Pareto ranking in the case that both objectives have equal priorities. Both objectives
should meet the defined restrictions [27, 33].
M

Figure 8 illustrates a situation in which objective 2 has a higher priority than objective 1. In this
case, individuals which do not meet restriction are the worst, independently of their
performance according to objective 1. Once is met, individuals are ranked based on how well
ED

they optimized objective 1 [27, 33].


PT
CE
AC
ACCEPTED MANUSCRIPT

Objective 2
8
7

6
𝑐 1
4

T
0

IP
3
2

CR
5

US 𝑐 Objective 1
AN
Figure 8. Pareto ranking in the case that objective 2 has higher priority than objective 1. Both
objectives should meet the defined restrictions [27, 33].

Having ranked the individuals, MOGA assigns a fitness value to each individual based on its
M

corresponding rank. To do that, the individuals are sorted based on the ranks and the fitness is
assigned by interpolating from the best individual (i.e., rank=0) to the worst according to a linear
or exponential function. Finally, a single value of fitness is calculated for the individuals with the
ED

same rank by the means of averaging. Assigning the average value to those with the same rank
will guarantee the same probability of being selected as the parent of next generation [33-35].
PT

The mating procedure uses the fitness values to generate a new population, ensuring that the
individuals with higher fitness have a larger probability of breeding. To generate a new
population, a small percentage of random immigrant from the previous generation is also
CE

introduced into the population, which makes the genetic algorithm more likely to recover
information lost through selection and thus, from genetic drift [33]. The pairs selected for mating
exchange part of their chromosome (i.e., based on the given probability crossover rate), to
AC

produce two offsprings for each pair, in the recombination phase. Parent recombination is done
in a way that the offspring respect the maximum model length. The resulting offspring may be
longer, shorter or equally sized as their parents. Once the new population is generated after
recombination, mutation is applied to randomly selected individuals. The mutation operator is
implemented by three basic operations: substitution, deletion and addition of one element. The
number of neurons is mutated, with a given probability, by adding or subtracting one neuron to
the model, verifying boundary conditions such that no neural network can have fewer or more
neurons than pre-specified values. Each model input term in the chromosome is tested and, with
ACCEPTED MANUSCRIPT

a given probability, is either replaced by a new term not in the model, or deleted. Finally, a new
term may be appended to the chromosome.

In each MOGA iteration, as shown in Figure 9, the non-dominated set is updated based on the
individuals in current generation. It is expected that, after a sufficient number of generations, the
population has evolved to achieve a non-dominated set which is not going to be altered; in this
stage, the user must select the best model, among the final non-dominated set.

T
IP
CR
US
AN

Figure 9. The update of the non-dominated set on arrival of new points. The gray area denotes all
M

generated models up to the current generation.


ED

6. Experimental results

6.1. Constructing the input dataset for MOGA


PT

Using the MOGA approach, the system must train a considerable amount of RBFNN models to
be able to construct the final non-dominated set (please recall that the training process is done
=10 times for each chromosome). As a result, in practice, some constraints should be imposed
CE

on the size of the datasets that will be provided to MOGA, otherwise the process would not be
finished in a reasonable time. As mentioned in section 3, we have 1,867,602 pixels (hereby
called as ) whose status (i.e., normal or abnormal) is already determined by the
AC

Neuroradiologist. Among these pixels 1,802,816 are normal (96.53% of the data samples) and
64,786 are abnormal (3.47% of the data samples). Hence, is an imbalanced dataset
whose size is (i.e., 51 features and 1 target column). To enable MOGA to
generate models applicable to the whole range of data where the classifier is going to be used, we
included all convex points [36] of into the training set. To obtain the convex points, the
Approxhull algorithm [37, 38] is used, resulting in 13023 samples, among which 11732 were
normal and 1291 abnormal. The convex points along with 6977 random data samples (50%
normal and 50% abnormal) constitute our training set whose size is 20,000. After excluding the
ACCEPTED MANUSCRIPT

training data samples from , 6666 random data samples were selected as a test set, and
additional 6666 random data samples as a validation set. In both test and validation sets 50% of
data samples were normal and 50% were abnormal. As a result, the input dataset for MOGA,
hereafter called has 33,332 data samples including 60% training, 20% test and 20%
validation data samples. is normalized between [-1, 1] before being passed to
MOGA, since this process reduces the chance of encountering numerical problems in the
training.

T
A flowchart, illustrating the different steps carried out for designing an RBFNN classifier for

IP
CVA detection using MOGA, is shown in Figure 10.

CR
US
AN
M
ED
PT
CE
AC
ACCEPTED MANUSCRIPT

Start

Obtain convex points of the whole available data samples using the
Approxhull approach.

Construct MOGA training set using obtained convex points, together with
some random data samples.

T
Construct MOGA test and validation sets using random data samples.

IP
Normalize train, test and validation sets within the range [-1 1].

CR
Specify MOGA objectives and restrictions, allowable range of features
and hidden neurons that can be used to construct a chromosome.

US
Specify initial center selection technique, training stopping criterion,
number of training times for each individual and strategy to select the best
AN
training trial.

Restrict trade-off surface coverage


Determine execution parameters: number of generations, population size,
M

crossover rate, selective pressure and proportion of random emigrants.

Run MOGA.
ED

Preferable set of RBFNN models


PT

𝐵𝐼𝐺 𝐷𝑆 Analyze the performance of models by calculating FP


and FN rates over 𝐵𝐼𝐺 𝐷𝑆
CE

No
Satisfied?
AC

Yes
End

Figure 10. Different steps carried out for designing an RBFNN classifier for CVA detection
using MOGA
ACCEPTED MANUSCRIPT

6.2 Experiment formulations

To identify the best possible RBFNN models two scenarios were conducted whose objectives are
shown in Table II.

Table II. Objectives for the MOGA experiments

Exp. Objectives

T
1

IP
2

CR
For both experiments, the system was allowed to choose the number of neurons in the hidden
layer and the number of input features from the ranges [2,30] and [1,30], respectively. The

US
number of generations and number of individuals in each generation were both set to 100. Early
stopping with a maximum number of 100 iterations was used as a termination criterion for the
training of each individual. The number of training trials for each individual, , was set to 10 and
AN
the nearest to the origin strategy was used to select the best training trial. The proportion of
random immigrants was 10%, the selective pressure was set to 2 and the crossover rate to 0.7.
M

The only difference between the two experiments is that restrictions were applied to the
and objectives in the second experiment, based on the results obtained from the first
experiment. To select the best model of experiment 1, we evaluated all non-dominated models on
ED

and then picked the model whose number of False Positives (FP) and False Negatives
(FN) on were minimum.

406 non-dominated models were obtained as result of experiment 1 (since there are no
PT

restrictions on the objectives of this experiment, its preferable set is the same as the non-
dominated set). Table III shows the Minimum, Average and Maximum FP and FN rates, as well
CE

as the model complexity of the non-dominated models of experiment 1. In the following Tables
and denote the training, test, validation sets, and the model complexity,
respectively. Moreover, is the number of False Detections ( ).
AC
ACCEPTED MANUSCRIPT

Table III. Min, Avg. and Max false positive and false negative rates as well as model complexity
of 406 non-dominated models obtained in experiment 1.

FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%)
Min. 0 1.86 1.08 0 1.80 2.39 0 2.28 2.91 0 2.20 2.33 6

T
Avg. 2.13 23.41 7.21 3.83 21.50 12.67 4.16 21.60 12.88 4.09 21.78 4.71 199.8

IP
Max. 8.47 100 24.16 12.27 100 50.03 13.47 100 50 12.49 100 12.74 900

CR
Table IV shows the models whose False Positive and False Negative rates are less than 3% in
. As it can be seen, both models have an equal percentage of FP (i.e., 2.96%) within
but the FN percentage of model 1371 within

US
is slightly smaller than that of
model 6009. Hence, the statistics of model 1371, shown in Table V, were used as restrictions on
experiment 2.
AN
Table IV. Models of experiment 1 whose False Positive and False Negative rates are less than
3% in
M

Model No.

FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%)
ED

1371 0.80 2.70 1.25 2.43 2.58 2.51 3.27 3.24 3.26 2.96 2.88 2.96 702

6009 0.74 2.53 1.17 2.85 2.46 2.66 3.36 3.06 3.21 2.96 2.89 2.96 870
PT

Table V. Statistics of model 1371.


CE

Model No.
FP FN FD FP FN FD FP FN FD FP FN FD
AC

1371 121 129 250 81 86 167 109 108 217 53442 1868 55310

Experiment 2 resulted in 281 non-dominated models from which 69 models are in the preferable
set. Table VI shows the Minimum, Average and Maximum FP and FN rates as well as the model
complexity over this set. Table VII shows the preferable models of experiment 2 whose FP and
FN rates are less than 2.6% in . Analyzing the results of Table VII, one can see that
ACCEPTED MANUSCRIPT

employing restrictions on and resulted in models with smaller number of false


detections in all sets, including in BI . Among the models in Table VII, model 3726 has the
minimum percentage of FP and model 3055 has the minimum percentage of FN on .

Table VI. Min, Avg. and Max false positive and false negative rates as well as model complexity
of 69 models in the preferable set of experiment 2.

T
FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%)

IP
Min. 0.49 1.38 0.77 2.19 2.06 2.36 2.24 1.75 2.05 2.40 1.98 2.40 750

CR
Avg. 0.60 1.90 0.89 2.76 2.65 2.71 2.71 2.45 2.58 2.78 2.43 2.76 862.3

Max. 0.67 2.51 1.04 3.37 3.37 3.25 3.24 2.97 2.97 3.20 2.91 3.17 900

US
Table VII. Preferable models of experiment 2 whose false positive and false negative rates are
less than 2.6% in
AN

Model
No.
FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%)
M

3726 0.60 1.79 0.87 2.27 2.90 2.58 2.58 2.37 2.48 2.40 2.34 2.40 870

4812 0.60 1.87 0.89 2.71 2.30 2.50 2.61 2.71 2.66 2.60 2.43 2.59 900
ED

3863 0.59 1.52 0.80 2.32 2.71 2.52 2.43 2.56 2.49 2.55 2.45 2.55 900

3055 0.50 1.73 0.77 2.71 2.84 2.78 2.43 2.09 2.26 2.56 2.31 2.55 900
PT
CE

6.3 Ensemble of models in the preferable set of experiment 2

Having selected a model (model 3726 shown in Table VII) with acceptable rates of specificity
AC

97.60% (i.e., 2.40 % FP rate) and sensitivity 97.66% (i.e., 2.34% FN rate) at pixel level, an
ensemble of the preferable models obtained in experiment 2 was also considered as a classifier.
Each data sample is fed to all 69 preferable models and then a majority vote determines whether
the pixel is considered normal or abnormal. Table VIII shows the results obtained on
and . Comparing the results with the ones obtained from model 3726 in Table VII, it can
be seen that 0.41% and 0.56% reductions could be obtained in the FP and FN rates over ,
ACCEPTED MANUSCRIPT

respectively. Hence, the ensemble approach achieves a specificity of 98.01% (i.e., 1.99 % FP
rate) and a sensitivity of 98.22% (i.e., 1.78% FN rate) at pixel level over .

Table VIII. Results of the ensemble of preferable models of experiment 2 on and

Ensemble

T
of
preferable
FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%)

IP
models in
experiment
2 0.44 1.18 0.61 1.90 2.03 1.96 1.93 1.74 1.83 1.99 1.78 1.99

CR
US
6.4 Visualizing abnormal regions in CT images using the ensemble of preferable models

Figure 11 shows the results of applying the ensemble of preferable models on some CT images,
where the output images of the classifier were marked with different colors, depending on the
AN
classifier output for each tested pixel. The color code is shown in Table IX.

Table IX. Colour code used for marking pixels based on the percentage of preferable models
with a positive output
M

Percentage of preferable Colour code Description


models with a positive output
ED

[66% 100%] Red Clear presence of pathology


Blue Cannot decide whether the
[50% 66%)
pixel is normal or abnormal
PT

[0% 50%) --- Clear absence of pathology


CE
AC

(a) (b) (c)


ACCEPTED MANUSCRIPT

T
IP
(d) (e) (f)

CR
US
AN

(g) (h) (i)


M

Figure 11. The result of applying the ensemble of preferable models on CT images. (a), (d) and
(g) are the original images. (b), (e) and (h) are marked by the Neuroradialogist. (c), (f) and (i) are
marked by the classifier.
ED

6.5 Features use

To understand which features are the most frequent in the preferable models of experiment 2, the
PT

relative frequency of each feature within the 69 models is shown in Figure 12. One can see
that, among the allowable 30 features within the 51 features considered, features
are the ones that have been employed in more
CE

than 80% of the models. Among this set, features belong to the set of
first order statistics, feature is a second order statistic and features are
symmetry features.
AC
ACCEPTED MANUSCRIPT

T
IP
CR
Figure 12. Relative frequency of each feature in the preferable models of experiment 2.

7. Discussion

US
As was shown before, the proposed approach achieves an excellent performance, both in terms
of sensitivity and specificity. Moreover, as it is a pixel-based classifier, lesion contours are
AN
obtained, very close to the ones marked by the Neuroradiologist. A comparison of this approach
with other works is, however, difficult as, to the best of our knowledge, there is no alternative
system that uses a pixel-based classification.
M

As Support Vector Machines (SVM) are frequently used in classification purposes, in the next
subsection we change our model, a MOGA-designed RBFNN to a SVM, and compare its use as
a pixel-based classifier. The subsequent subsection will compare the proposed approach with
ED

three existing alternatives found in the literature. As pointed out before, these systems work at a
slice level, which means that they can identify lesions but not draw their contours.
PT

7.1 Comparing MOGA RBF classifiers with Support Vector Machines

In order to compare the obtained results with a SVM [39], the MATLAB SVM tool with
CE

Gaussian RBF (Radial Basis Function) kernel was used. For determining the best penalty
parameter (C) and the spread, possible combinations obtained by selecting 2 values from the
set {0.003, 0.01, 0.03, 0.1, 0.3, 1, 3, 10, 30, 100, 300} were used for SVM training, and the
AC

combination (C=3, spread=1) whose error on the test set was minimum, was selected. In this
experiment, 69.8% of the data samples in the training set were considered as support vectors.
Table X shows the FP and FN rates when this SVM was applied.

Comparing the results with the ones obtained with the ensemble of models, shown in Table VIII,
and also with model 3726, shown in Table VII, one can see that even with a huge complexity of
the SVM model (139,600 support vectors), its FP and FN rates in are not only higher than
the ones obtained with the ensemble of preferable models, but also the ones achieved by model
ACCEPTED MANUSCRIPT

3726. Notice that a SVM model, with Gaussian Kernel can be considered a RBFNN model,
where the centers of the Gaussians are the support vectors, and with a common spread to all the
neurons. In this case, all the features (51) were considered as inputs and 13960 support vectors
were employed. This is translated into a complexity of 711,960 parameters, determined by the
SVM algorithm. This figure should be compared with a complexity of 870 (around 0.1%), for
model 3726 in Table VII.

Table X. FP and FN rates using SVM

T
IP
FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%) FP (%) FN (%) FD (%)

CR
0.16 0 0.13 2.6 2.42 2.51 2.32 2.26 2.29 2.5 2.37 2.5

US
7.2 Comparing the results obtained with other approaches

The authors in [40] presented a Computer Aided Detection (CAD) method for early detection of
AN
CVAs from CT images where, in the same way as this work, in a preprocessing phase artifacts
are removed and tilted CT images are realigned. In order to find the regions that have higher
probability of being considered lesions, a Circular Adaptive Region Of Interest (CAROI)
algorithm is applied on each CT slice, which aims to draw a circular border around areas with
M

sudden change of intensity values. Each circular region is then compared with its corresponding
region in the other side of the brain using the Pearson correlation coefficient. Those circular
areas which have the smallest PCC values are selected for further investigation. Eight second
ED

order features are calculated from the GLCM matrix of previously selected circular regions and
are passed to a 3-layer feed-forward back propagation neural network which was trained using
10 normal and 20 abnormal cases in a round robin (leave-one-out) fashion. The output of neural
PT

network identifies whether the circular region is a lesion or not.

In order to evaluate their CAD system, 31 positive cases containing 82 ischemic strokes (39
CE

acute and 43 chronic) were used as validation set. A sensitivity of 76.92% (i.e., 30/39 lesion
areas correctly detected) for acute ischemic strokes and a sensitivity of 90.70% (i.e., 39/43 lesion
areas correctly detected) for chronic strokes were reported. This gives a total sensitivity of
AC

84.14% (i.e., ).

In spite of the differences of this approach and ours, in order to be able to compare the accuracy
obtained in terms of lesions sensitivity, this measure has been calculated. A total number of 35
ischemic lesions within 150 CT images were marked by our collaborating Neuroradiologist. The
ensemble of preferable models in experiment 2 detected 30 lesions correctly, which is translated
in a sensitivity of 85.71%, slightly higher than approach [40].
ACCEPTED MANUSCRIPT

The authors in [41] developed a CAD system for detecting hemorrhagic strokes in CT images.
After removing the artifacts and realigning the tilted images, the hemorrhagic areas are
segmented by employing a threshold on the pixels’ intensity values. To detect the edema regions,
a higher contrast ratio of a given CT image is firstly obtained using a local histogram
equalization. A thresholding method is then applied to segment the edema region from the
normal tissue. The accuracy of the CAD system is evaluated by comparing the area of bleeding
region (ABR) and edema region (AER) that are detected by the CAD and the ones that are
marked by the doctor using data from 8 spontaneous hemorrhagic stroke patients. It is reported

T
that the average difference of ABRs is 8.8%, and the average of the degree of coincidence is

IP
86.4%, while the average difference of AERs is 14.1%, with an average of degree of coincidence
of 77.4%.

CR
The results obtained by this approach cannot be exactly compared with the approach presented
here, as [41] deals with hemorrhagic strokes, which typically are much easier to detect and mark
than ischemic strokes. In spite of that, the average difference of the areas as well as the average

US
degree of coincidence have been computed for the cases presented here, for the lesions both
marked by the doctor and detected by our system. The average difference is 11.4%, and the
average degree of coincidence is 88.6%. These figures are better than the values obtained for
AN
AER, in approach [41].

The authors in [42] utilize a combination of 2D and 3D Convolutional Neural Networks (CNN)
to cluster brain CT images into 3 groups: Alzheimer’s disease, lesion and normal ageing. The
M

best classification accuracy rates using the proposed CNN architecture are 85.2%, 80% and
95.3% for the classes of Alzheimer’s disease, lesion and normal, respectively, with an average of
ED

87.6%.

To be able to compare the classification accuracy rates of our work with the ones in [42], we
calculated this metric in a CT slice level, by labeling each CT slice as normal or abnormal (i.e.,
PT

having one or more lesions). Our system was able to correctly identify all 24 CT slices that had
lesions within which translates into 100% classification accuracy rate for the abnormal group.
Among the remaining 126 normal CT slices, our system identified small false lesions within 7
CE

CT slices, which means that we obtained a 94.4 % classification accuracy rate for the normal
group.
AC

8. Conclusions

In this work, a RBFNN based system for automatic identification of CVA through the analysis of
brain CT images is presented. Considering a set of 51 features, the MOGA design framework
was employed to find the best possible RBFNN structure and its corresponding parameters. Two
experiments were conducted in MOGA. The best result is obtained from an ensemble of
preferable models of experiment 2, where the and objectives were restricted based on
the results obtained by the best model from the first experiment. Values of specificity of 98.01%
ACCEPTED MANUSCRIPT

(i.e., 1.99 % FP) and sensitivity of 98.22% (i.e., 1.78% FN) were obtained at pixel level, in a set
of 150 CT slices (1,867,602 pixels).

Comparing the classification results with SVM over , despite the huge complexity of the
SVM model, the accuracy of the selected model in experiment 2, as well as the ensemble of
preferable models, are superior to that of SVM model.

The present approach compares also favorably with other similar published approaches,
achieving improved sensitivity at lesion level than [40], better average difference and degree of

T
coincidence than [41], as well as superior classification accuracy rate than [42]. It should be

IP
stressed than none of these methods are able to draw the lesion(s) contour(s), as it is achieved by
the proposed approach.

CR
As the number of abnormal pixels is much smaller than the number of normal pixels in the
existing dataset, at the present stage the classifier is able to detect the great majority of the
lesions, but sometimes will identify false lesions. Current research is tackling this problem.

US
Additionally, as the proposed classifier was designed and tested only with CT images including
ischemic CVAs, we consider enlarging the CT database to include other brain lesions, and apply
the same methodology to design an additional classifier capable of discriminating brain lesions
AN
with similar image patterns.

Conflict of interest
M

The authors declare no conflicts of interest related to this research work.

Acknowledgments
ED

The authors would like to acknowledge the support of FCT, through IDMEC, under LAETA,
project UID/EMS/50022/2013 and Dr. Luis Cerqueira, from Centro Hospitalar de Lisboa
PT

Central, Portugal, for marking the exams.

References
CE
AC

[1] A. S. Go, D. Mozaffarian, V. L. Roger, E. J. Benjamin, J. D. Berry, W. B. Borden, et al.,


"Heart disease and stroke statistics--2013 update: a report from the American Heart
Association," Circulation, vol. 127, pp. e6-e245, Jan 1 2013.
[2] P. M. Ferreira and A. E. Ruano, "Evolutionary Multiobjective Neural Network Models
Identification: Evolving Task-Optimised Models," New Advances in Intelligent Signal
Processing, vol. 372, pp. 21-53, 2011.
[3] E. Hajimani, M. G. Ruano, and A. E. Ruano, "MOGA design for neural networks based
system for automatic diagnosis of Cerebral Vascular Accidents," in 9th IEEE
International Symposium on Intelligent Signal Processing (WISP), 2015, pp. 1-6.
ACCEPTED MANUSCRIPT

[4] E. Hajimani, A. Ruano, and G. Ruano, "The Effect of Symmetry Features on Cerebral
Vascular Accident Detection Accuracy," presented at the RecPad 2015, the 21th edition
of the Portuguese Conference on Pattern Recognition, Faro, Portugal, 2015.
[5] X. Llado, A. Oliver, M. Cabezas, J. Freixenet, J. C. Vilanova, A. Quiles, et al.,
"Segmentation of multiple sclerosis lesions in brain MRI: A review of automated
approaches," Information Sciences, vol. 186, Mar 1 2012.
[6] T. Hachaj and M. R. Ogiela, "CAD system for automatic analysis of CT perfusion maps,"
Opto-Electronics Review, vol. 19, Mar 2011.
[7] L. E. Poh, V. Gupta, A. Johnson, R. Kazmierski, and W. L. Nowinski, "Automatic

T
Segmentation of Ventricular Cerebrospinal Fluid from Ischemic Stroke CT Images,"
Neuroinformatics, vol. 10, Apr 2012.

IP
[8] W. L. Nowinski, G. Qian, K. N. B. Prakash, Q. Hu, and A. Aziz, "Fast Talairach
Transformation for magnetic resonance neuroimages," Journal of Computer Assisted

CR
Tomography, vol. 30, pp. 629-641, Jul-Aug 2006.
[9] N. Otsu, "A Threshold Selection Method From Gray-level Histogram," IEEE
Transactions on Systems, Man, and Cybernetics, 1978.
[10] M. Gao and S. Chen, "Fully Automatic Segmentation of Brain Tumour in CT Images,"

[11] US
European Journal of Cancer, vol. 47, Sep 2011.
A. P. Nanthagopal and R. S. Rajamony, "Automatic classification of brain computed
tomography images using wavelet-based statistical texture features," Journal of
AN
Visualization, vol. 15, pp. 363-372, Nov 2012.
[12] T. J. Devadas and R. Ganesan, "Analysis of CT Brain images using Radial Basis
Function Neural Network," Defence Science Journal, vol. 62, Jul 2012.
[13] O. Freifeld, H. Greenspan, and J. Goldberger, "Lesion detection in noisy MR brain
M

images using constrained GMM and active contours," 4th IEEE International Symposium
on Biomedical Imaging : Macro to Nano, Vols 1-3, pp. 596-599, 2007.
[14] H. Greenspan, A. Ruf, and J. Goldberger, "Constrained Gaussian mixture model
ED

framework for automatic segmentation of MR brain images," IEEE Transactions on


Medical Imaging, vol. 25, pp. 1233-1245, Sep 2006.
[15] B. J. Bedell and P. A. Narayana, "Automatic segmentation of gadolinium-enhanced
multiple sclerosis lesions," Magnetic Resonance in Medicine, vol. 39, pp. 935-940, Jun
PT

1998.
[16] A. O. Boudraa, S. Mohammed, R. Dehak, Y. M. Zhu, C. Pachai, Y. G. Bao, et al.,
"Automated segmentation of multiple sclerosis lesions in multispectral MR imaging
CE

using fuzzy clustering," Computers in Biology and Medicine, vol. 30, pp. 23-40, Jan
2000.
[17] E. Hajimani, C. A. Ruano, M. G. Ruano, and A. E. Ruano, "A software tool for
intelligent CVA diagnosis by cerebral computerized tomography," in 8th IEEE
AC

International Symposium on Intelligent Signal Processing (WISP), 2013, pp. 103-108.


[18] X. Qi, A. Belle, S. Shandilya, W. Chen, C. Cockrell, Y. Tang, et al., "Ideal Midline
Detection Using Automated Processing of Brain CT Image," Open Journal of Medical
Imaging, vol. 3, p. 9, 2013.
[19] L. He, Y. Chao, K. Suzuki, and K. Wu, "Fast connected-component labeling," Pattern
Recognition, vol. 42, pp. 1977-1987, Sep 2009.
[20] A. Usinskas, R. A. Dobrovolskis, and B. F. Tomandl, "Ischemic stroke segmentation on
CT images using joint features," Informatica, vol. 15, pp. 283-290, 2004.
ACCEPTED MANUSCRIPT

[21] L. Ribeiro, A. E. Ruano, M. G. Ruano, and P. M. Ferreira, "Neural networks assisted


diagnosis of ischemic CVA's through CT scan," IEEE International Symposium on
Intelligent Signal Processing, Conference Proceedings Book, pp. 223-227, 2007.
[22] W. Chen, R. Smith, S.-Y. Ji, K. R. Ward, and K. Najarian, "Automated ventricular
systems segmentation in brain CT images by combining low-level segmentation and
high-level template matching," BMC Medical Informatics and Decision Making, vol. 9,
p. 14, 2009.
[23] Wei-Li Zhang and Xi-Zhao Wang, "Feature Extraction and Classification for Human
Brain CT Images," in International Conference on Machine Learning and Cybernetics,

T
2007, pp. 1155-1159.
[24] R. M. Haralick, Shanmuga.K, and I. Dinstein, "Textural Features for Image

IP
Classification," IEEE Transactions on Systems Man and Cybernetics, vol. SMC3, pp.
610-621, 1973.

CR
[25] D. A. Clausi, "An analysis of co-occurrence texture statistics as a function of grey level
quantization," Canadian Journal of Remote Sensing, vol. 28, pp. 45-62, Feb 2002.
[26] C. A. Teixeira, M. G. Ruano, A. E. Ruano, and W. C. A. Pereira, "A soft-computing
methodology for noninvasive time-spatial temperature estimation," IEEE Transactions on

[27] US
Biomedical Engineering, vol. 55, pp. 572-580, Feb 2008.
C. M. Fonseca and P. J. Fleming, "Multiobjective optimization and multiple constraint
handling with evolutionary algorithms - Part I: A unified formulation," IEEE
AN
Transactions on Systems Man and Cybernetics Part a-Systems and Humans, vol. 28, pp.
26-37, Jan 1998.
[28] D. W. Marquardt, "An Algorithm for Least-Squares Estimation of Nonlinear
Parameters," Journal of the Society for Industrial and Applied Mathematics, vol. 11, pp.
M

431-441, 1963.
[29] K. Levenberg, "A method for the solution of certain problems in least squares," Quart.
Applied Math., vol. 2, pp. 164-168, 1944.
ED

[30] P. M. Ferreira and A. E. Ruano, "Exploiting the separability of linear and nonlinear
parameters in radial basis function networks," IEEE Proceedings on Adaptive Systems for
Signal Processing, Communications, and Control Symposium, pp. 321-326, 2000.
[31] A. E. Ruano and I. o. E. Engineers, Intelligent Control Systems Using Computational
PT

Intelligence Techniques: Institution of Engineering and Technology, 2005.


[32] S. S. Haykin, Neural Networks: A Comprehensive Foundation: Prentice Hall, 1999.
[33] C. M. M. d. Fonseca, "Multiobjective genetic algorithms with application to control
CE

engineering problems," PhD, Department of Automatic Control and Systems


Engineering, University of Sheffield, 1995.
[34] C. M. Fonseca and P. J. Fleming, "Multiobjective Genetic Algorithms Made Easy:
Selection, Sharing and Mating Restriction," presented at the Genetic Algorithms in
AC

Engineering Systems: Innovations and Applications, UK, 1995.


[35] C. A. D. Teixeira, "Soft-computing techniques applied to artificial tissue temperature
estimation," PhD, FCT, University of Algarve, 2008.
[36] H. R. Khosravani, A. E. Ruano, and P. M. Ferreira, "A simple algorithm for convex hull
determination in high dimensions," presented at the 8th International Symposium
on Intelligent Signal Processing (WISP), Portugal, 2013.
ACCEPTED MANUSCRIPT

[37] A. Ruano, H. R. Khosravani, and P. M. Ferreira, "A Randomized Approximation Convex


Hull Algorithm for High Dimensions," IFAC-PapersOnLine, vol. 48, pp. 123 - 128,
2015.
[38] H. Khosravani, A. Ruano, and P. Ferreira, "A Convex Hull-based Data Selection Method
for Data Driven Models," Applied Soft Computing, 2016 (in press).
[39] C. Cortes and V. Vapnik, "Support-Vector Networks," Machine Learning, vol. 20, pp.
273-297, Sep 1995.
[40] K.-s. D. Ng, "Computer aided detection method for early detection of cerebrovascular
accident," PhD, Dept. of Health Technology and Informatics, The Hong Kong

T
Polytechnic University, The Hong Kong Polytechnic University, 2009.
[41] J.-G. Gan, Y.-W. Wang, J.-L. Su, L. Chan, and Ieee, "The Development of CAD system

IP
for Hemorrhagic Stroke in Computed Tomography Images," 2014 Ieee International
Symposium on Bioelectronics and Bioinformatics (Isbb), 2014 2014.

CR
[42] X. H. W. Gao, R. Hui, and Z. M. Tian, "Classification of CT brain images based on deep
learning networks," Computer Methods and Programs in Biomedicine, vol. 138, pp. 49-
56, Jan 2017.

US
AN
M
ED
PT
CE
AC

You might also like