An Effective and Novel Approach For Brain Tumor CL
An Effective and Novel Approach For Brain Tumor CL
An Effective and Novel Approach For Brain Tumor CL
Journal of Sensors
Volume 2023, Article ID 1224619, 19 pages
https://doi.org/10.1155/2023/1224619
Research Article
An Effective and Novel Approach for Brain Tumor Classification
Using AlexNet CNN Feature Extractor and Multiple Eminent
Machine Learning Classifiers in MRIs
Received 27 October 2022; Revised 28 January 2023; Accepted 27 February 2023; Published 8 March 2023
Copyright © 2023 Alok Sarkar et al. This is an open access article distributed under the Creative Commons Attribution License,
which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.
A brain tumor is an uncontrolled malignant cell growth in the brain, which is denoted as one of the deadliest types of cancer in
people of all ages. Early detection of brain tumors is needed to get proper and accurate treatment. Recently, deep learning
technology has attained much attraction to the physicians for the diagnosis and treatment of brain tumors. This research
presents a novel and effective brain tumor classification approach from MRIs utilizing AlexNet CNN for separating the dataset
into training and test data along with extracting the features. The extracted features are then fed to BayesNet, sequential
minimal optimization (SMO), Naïve Bayes (NB), and random forest (RF) classifiers for classifying brain tumors as no-tumor,
glioma, meningioma, and pituitary tumors. To evaluate our model’s performance, we have utilized a publicly available Kaggle
dataset. This paper demonstrates ROC, PRC, and cost curves for realizing classification performance of the models; also,
performance evaluating parameters, such as accuracy, sensitivity, specificity, false positive rate, false negative rate, precision, f-
measure, kappa statistics, MCC, ROC area, and PRC area, have been calculated for four testing options: the test data itself,
cross-validation fold (CVF) 4, CVF 10, and percentage split (PS) 34% of the test data. We have achieved 88.75%, 98.15%,
86.25% and 100% of accuracy using the AlexNet CNN+BayesNet, AlexNet CNN+SMO, AlexNet CNN+NB, and AlexNet CNN
+RF models, respectively, for the test data itself. The results imply that our approach is outstanding and very effective.
using machine learning in MRIs [12–14]. Nevertheless, second-order statistics and a least square support vector
because of high inter- and intrashape, texture, and variation machine (SVM) classifier. Sarkar et al. [38] proposed a
of contrast, this still remains a challenging task. Conven- computer-aided approach for detecting and classifying brain
tional machine learning techniques rely on artisanal features, tumors from MRIs utilizing genetic algorithm as a feature
which limit the robustness of the outcomes, while deep extractor and SVM as classifier, where they obtained 98.3%
learning techniques automatically extract important func- classification accuracy. John [25] used techniques based on
tionality that provides significantly better performance. the gray level cooccurrence matrix (GLCM) and discrete
But still, the results obtained using deep learning tech- wavelet transform (DWT) to detect and classify brain
nologies are not sufficient, and they do not always show tumors. Ullah et al. [39] applied DWT for feature extraction
good classification accuracy, uses of only AlexNet, for exam- and a feed-forward artificial neural network for brain MRI
ple [15]. The research is challenging, especially for selecting classification. Kharrat et al. [40] classified brain tumors as nor-
the region of interest (ROI), extracting deep features, using mal and abnormal classes through genetic algorithms and
clamorous images, etc. SVM classifiers. Papageorgiou et al. [33] obtained 90.26%
To minimize the aforementioned problems, in this and 93.22% accuracy for low- and high-grade glioma tumors,
study, we have proposed a novel and effective technique respectively, through fuzzy cognitive maps. Díaz-Pernas et al.
for brain tumor classification automatically employing the [26] proposed an architecture for automatic detection and seg-
AlexNet CNN feature extractor and a variety of prominent mentation of brain tumors and acquired 97.3% accuracy.
machine learning classifiers, such as BayesNet, sequential Shree and Kumar [41] used a GLCM feature extractor and
minimal optimization (SMO), Naïve Bayes (NB), and ran- probabilistic neural network classifier for classifying MRI
dom forest (RF) in MRIs. datasets and achieved 95% accuracy. Arunachalam and Savar-
We know that AlexNet CNN is a good tool to determine imuthu [42] offered an architecture that covered enhancement
ROI for extracting deep features and can easily deal with of image, transformation of image, extraction of features, and
clamorous and raw images, which are very essential for the finally, image classification. They employed shift-invariant
current research. Moreover, deploying machine learning shear-let transforms for enhancing MRIs and used numerous
classification algorithms with AlexNet CNN certainly feature extractors, such as Gabor, GLCM, and DWT to extract
enhance the performance of the proposed model. the features. Hossain et al. [30] utilized the traditional CNN
Our contributions to this work are as follows: approach to classify brain tumor images. They utilized fuzzy
c-means clustering for segmenting the tumors and achieved
(i) We have utilized AlexNet CNN for separating an accuracy of 97.87%.
human brain MRI datasets into training and test Several deep learning approaches have been extensively
data and also for extracting the features from those utilized by researchers to detect and classify brain tumors
separated datasets in the last decades. Sajid et al. [43] proposed a fully auto-
mated hybrid approach for the detection and segmentation
(ii) Those extracted features are then used for classifying
of brain tumors in MRIs using deep learning. They utilized
the brain tumors as no-tumor, glioma, meningioma,
the BRATS2013 dataset and obtained 86%, 86%, and 91%
and pituitary tumors by using eminent BayesNet,
dice scores, sensitivity, and specificity, respectively. Saxena
SMO, NB, and RF classifiers
et al. [44] utilized ResNet-50, Inception V3, VGG-16, and
This paper is organized as follows. Section 1 discusses transfer learning techniques to classify the images; the
briefly the problems, some of the existing solutions, research ResNet-50 outperformed with 95% accuracy. In [27], Çinar
gaps, objectives, contributions, and organizations of the paper. and Yildirim utilized numerous convolutional neural
Section 2 presents related works on brain tumor detection and network (CNN) models, such as Inception V3, GoogLeNet,
classification. It also discusses the research gaps at the end of ResNet-50, AlexNet, and DenseNet-201, for classifying the
this section. Section 3 describes overall working methodology, MRI datasets; whether they acquired respectable accuracies
proposed architecture, and performance measuring metrics. in every case. Khwaldeh et al. [45] customized the AlexNet
Section 4 describes the results and performance of our model architecture and employed it to classify brain MRIs, and
along with a comparison with other contemporary findings. obtained 91% accuracy. Preethi and Aishwarya [46]
Finally, Section 5 contains the conclusion of the entire study, presented an architecture that combined wavelet-based
limitations, and future works. GLCM to develop matrices of features. They utilized a deep
neural network classifier to detect and categorize brain MRIs
and obtained 92% accuracy. In [28], Hemanth et al. pro-
2. Related Works posed a modified deep convolutional neural network
(DCNN) to classify brain tumors, and they obtained excel-
Many researchers have published their works to represent lent model performance in terms of accuracy. Khan et al.
different methodologies or approaches for detection and [47] introduced a computerized multimodel categorization
classification purposes over the past few years [15–36]. technique to detect and classify brain tumor images. They
Various methodologies have been tested using different clin- utilized two presupervised CNN models, known as VGG16
ical databases, including magnetic resonance imaging (MRI) and VGG19, to extract features. They deployed the
of brain tumors. Selvaraj et al. [37] developed a binary BRATS2018 dataset to validate their model and obtained
classifier for classifying brain MRIs a utilizing first- and about 97.80% accuracy.
Journal of Sensors 3
Input MR images
Preprocessing Preprocessing
Classifier
Prediction model (BayesNet, SMO,
NB, RF)
Classified result
Figure 1: Block diagram of the proposed method for identification and classification of brain tumors.
Convolution + Convolution +
Output
ReLU Pooling ReLU Pooling
Glioma
Meningioma
No-tumor
•••••
Pituitary
{(13+2⁎1-3)/1}+1=13
4th CON 3×3,
Kemels = 384
Padding = 1,
1000 softmax Fully connected 8 13×13 13×13
Fully connected 7
6×6
Fully connected 6 {(13-3)/2}+1=6 ⁎
{(13+2 1-3)/1}+1=13
(iii) Most of the studies have used good quality MRI rated in the upcoming subsections. We have used AlexNet
images. Raw images or noisy images have been CNN in the MATLAB platform [50] for categorizing the
found well dealt by deep learning methods dataset as training and test as well as extracting the features
from it. Then, the extracted features are employed in the
(iv) Manual classification of brain tumors from MRIs is WEKA platform [51] to classify the brain tumors of MRIs
time-consuming, nonreproducible, and impractical as no-tumor, glioma, meningioma, and pituitary tumors uti-
if the data size is larger lizing prominent BayesNet, SMO, NB, and RF classifiers and
Thus, an excellent model is required for feature extraction establishing a prediction model. The performance of the
and classification of brain tumors accurately from MRIs. model is assessed by the classified results.
3. Materials and Methods 3.2. Data Description. The selection of an appropriate data-
set is the first and foremost concern for identifying and clas-
3.1. Research Implementation Block Diagram. Figure 1 shows sifying medical images. Therefore, we collected the dataset
the overall working architecture of our proposed model to from a trustworthy website, the Kaggle database [52, 53], a
accurately identify and classify brain tumors from MRIs. well-known web-based data source for brain MRIs. From
Each key component of this architecture, such as the dataset [52], we have utilized the glioma, meningioma, and pituitary
including training and test data, preprocessing, feature tumor data, and from [53], we have utilized the no tumor
extraction, prediction model, classified results, and perfor- data. In total, 3600 images have been used in this study,
mance measurement of the diagram, are sequentially nar- where each class contains the same number of images (900
Table 2: Confusion matrix for (a) test data itself, (b) CVF 4, (c) CVF 10, and (d) PS 34% of test data using BayesNet, SMO, NB, and RF classifiers.
(a)
(b)
(c)
Table 2: Continued.
(d)
1
1
0.8
0.8
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
False positive rate False positive rate
BayesNet NB BayesNet NB
SMO RF SMO RF
(a) (b)
1 1
0.8 0.8
True positive rate
True positive rate
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Figure 4: ROC curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for test data itself using BayesNet, SMO, NB,
and RF classifiers.
per class). The elaborate allocation of the data collected and clamor. Therefore, it is important to evacuate the conceiv-
deployed in this study is shown in Table 1. We split the data- able clamors through preprocessing. In addition, enhance-
set (3600 MRIs) into training and test datasets: 2520 (70% of ment of the visual nature of images offers tremendous
the total dataset) images are utilized as training images, and favor to the specialists. In this research, MRIs are prepro-
the rest 1080 (30% of the total dataset) images as test images, cessed with an anisotropic filter. The size of the images in
using AlexNet CNN. Training data are deployed for model the dataset of MRIs is not the same. These different sizes
prediction; on the other hand, test data are applied for testi- of images illustrate the input layer of the CNN network;
fying the model’s performance. therefore, these images need to be preprocessed, standard-
ized, and resized to 227 × 227 pixels.
3.3. Preprocessing. Preprocessing is a technique that aug-
ments the quality of images and makes them able for further 3.4. Feature Extraction. One presupervised CNN model,
steps. In a certain cases, the separation between normal and AlexNet [15, 31], is utilized in this investigation to include
abnormal tissue is complicated because of the high-clamor extraction, because it is a fantastically incredible model fit
level. As a result, specialists may normally commit errors for accomplishing high-level accuracy on testing datasets.
in diagnosis. Then again, minor contrasts between normal AlexNet is the most important known deep CNN structure
and abnormal tissues can likewise be concealed by the which comprises twenty-five layers, eight of which add to
8 Journal of Sensors
1 1
0.9 0.9
0.8 0.8
0.7 0.7
0.6 0.6
0.5
0.4 0.4
0.3
0.2 0.2
0.1
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
False positive rate False positive rate
BayesNet NB BayesNet NB
SMO RF SMO RF
(c) (d)
Figure 5: ROC curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for CVF 4 of the test data using BayesNet,
SMO, NB, and RF classifiers.
learning by modifying loads. Five of these are known as con- future results or conducts are predicted dependent on the
volution layers while staying three are known as fully con- former and present data. It is a measurable examination pro-
nected (FC) layers. In the AlexNet design, max-pooling cedure that empowers the assessment and count of the like-
layers are connected to convolution layers in series with each lihood of specific outcomes identified with programming or
other. The first convolution layer and then max-pooling frameworks. The prediction model works by gathering
layers utilize fluctuating kernel sizes [54]. The layer of information, making a factual model, and applying probabi-
max-pooling ensues the layer of convolution [54, 55]. FC6 listic methods to anticipate the possible results.
is the first FC layer, and FC7 is the second FC layer in acti-
vation are utilized to remove the features of vectors. There 3.6. Proposed Network Architecture
are altogether 4096 features of FC6 and FC7 vectors in the
AlexNet CNN design [56]. 3.6.1. Convolutional Neural Network. The convolutional
neural network (CNN) is a sort of multilayered feed-
3.5. Prediction Model. Consequently, we have developed pre- forward neural network and one of the significant ideas of
dictive models based on features extracted from the trained the deep learning method. Its basic structure consists of sev-
data. A prediction model is a procedure through which eral convolution layers, a rectified linear unit (ReLU) layer,
Journal of Sensors 9
1 1
0.8 0.8
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
False positive rate False positive rate
BayesNet NB BayesNet NB
SMO RF SMO RF
(a) (b)
1 1
0.8 0.8
True positive rate
True positive rate
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
False positive rate False positive rate
BayesNet NB BayesNet NB
SMO RF SMO RF
(c) (d)
Figure 6: ROC curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for the CVF 10 of test data using BayesNet,
SMO, NB and RF classifiers.
pooling layers, and FC layers, as shown in Figure 2. The (11 × 11, 5 × 5, 3 × 3, 3 × 3, 3 × 3), three max-pooling layers
basic function of the convolution layer is to extract the (3 × 3), and three FC layers. The first two max-pooling
local features of the existing layers. The ReLU layer per- layers are deployed after the first two convolutional layers
forms element-by-element activation. The pooling layers sequentially. The third, fourth, and fifth convolutional
are introduced for downsampling. Max-pooling is com- layers are connected directly. The third max-pooling layer
monly utilized to reduce the accuracy of the feature maps is inserted after the fifth convolutional layer, and its out-
by incorporating linguistically similar features. To over- put is fed into a series of three FC layers. The success of
come the fitting issues, some neurons have been removed AlexNet is usually credited to ReLU, stochastic gradient
from the CNN architecture, known as dropouts. At the descent (SGD), dropout, etc. ReLU is introduced for accel-
end, the FC layer yields the class score value ranging from erating the speed of training processes. The value of the
0 to 1 which is deployed to attain the classification deci- convolutional kernel is extracted by optimizing the total
sions. Essentially, SoftMax layers are introduced by default cost function and applying the SGD algorithm. To elimi-
in this motive. nate the overfitting problem, a dropout layer is applied
in the first two FC layers. The third FC layer, also known
3.6.2. AlexNet. Figure 3 shows the architecture of an AlexNet as the softmax layer, is used to classify various objects in a
which is mainly comprised of five convolutional layers CNN [57].
10 Journal of Sensors
1 1
0.8 0.8
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
False positive rate False positive rate
BayesNet NB BayesNet NB
SMO RF SMO RF
(a) (b)
1 1
0.9
0.8 0.8
0.7
True positive rate
True positive rate
0.6 0.6
0.5
0.4 0.4
0.3
0.2 0.2
0.1
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
False positive rate False positive rate
BayesNet NB BayesNet NB
SMO RF SMO RF
(c) (d)
Figure 7: ROC curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for PS 34% of the test data using BayesNet,
SMO, NB, and RF classifiers.
Table 3: Calculated values of ROC and PRC areas for test data itself, CVF 4, CVF 10, and PS 34% of test data using BayesNet, SMO, NB, and
RF classifiers.
1 1
Normalized expected cost
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Probability cost function Probability cost function
BayesNet NB BayesNet NB
SMO RF SMO RF
(a) (b)
1 1
Normalized expected cost
0.8 0.8
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Probability cost function Probability cost function
BayesNet NB BayesNet NB
SMO RF SMO RF
(c) (d)
Figure 8: Cost curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for the test data itself using BayesNet, SMO,
NB, and RF classifiers.
input to the model is batch size (number of experiments dure consists of building simple decision trees and finally
carried out) [61], whose value is selected as 100 for better making the decision based on the votes. It looks for the
accuracy of the model. best feature in every step (splitting nodes and growing
trees), which predominantly results in a superior model.
3.7.4. Random Forest Classifier. The random forest (RF) Batch size, number of features, and seed value are the com-
classifier is a supervised machine learning algorithm that mon inputs in the RF model [61]. For better performance,
creates a forest with a bunch of trees. Its working proce- we have considered batch size as 100 and seed value as 1.
12 Journal of Sensors
1 1
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Probability cost function Probability cost function
BayesNet NB BayesNet NB
SMO RF SMO RF
(a) (b)
1 1
Normalized expected cost
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Probability cost function Probability cost function
BayesNet NB BayesNet NB
SMO RF SMO RF
(c) (d)
Figure 9: Cost curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for CVF 4 of the test data using BayesNet,
SMO, NB, and RF classifiers.
3.8. Classified Result. In this research, it is expected that the the CVF 10 [19], and percentage split (PS) 34% [51] of
final classified results will be obtained as glioma, meningi- test data, by employing the Equations (1)–(11) to measure
oma, no-tumor, and pituitary tumors from MRI datasets. the performance of our proposed model.
Then, we wish that the qualitative and quantitative values
of the performance measurement of the proposed model’s TP + TN
will be attained to evaluate the model performance. Accuracy = , ð1Þ
TP + TN + FP + FN
1 1
0.8 0.8
Normalized expected cost
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Probability cost function Probability cost function
BayesNet NB BayesNet NB
SMO RF SMO RF
(a) (b)
1 1
0.8 0.8
Normalized expected cost
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
Probability cost function Probability cost function
BayesNet NB BayesNet NB
SMO RF SMO RF
(c) (d)
Figure 10: Cost curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for CVF 10 of the test data using BayesNet,
SMO, NB, and RF classifiers.
Here, 4 marginal, r1m is the row 1 marginal, r2m is the row 2 mar-
ginal, r3m is the row 3 marginal, r4m is the row 4 marginal,
Total Number of Correctly Classified and n is the total number of observations.
Observed Agreement = ,
n
ð5Þ FP
False Positive Rate ðFPRÞ = , ð7Þ
FP + TN
and
FN
Expected Agreement False Negative RateðFNRÞ = , ð8Þ
ÀÀ 1 Á Á ÀÀ Á Á ÀÀ Á Á ÀÀ Á Á FN + TP
cm × r1m /n + c2m × r2m /n + c3m × r3m /n + c4m × r4m /n
= ,
n TP
Precision = , ð9Þ
ð6Þ TP + FP
where c1m is the column 1 marginal, c2m is the column 2 Precision × Recall
F − measure = 2 × , ð10Þ
marginal, c3m is the column 3 marginal, c4m is the column Precision + Recall
14 Journal of Sensors
1 1
0.8 0.8
Normalized expected cost
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
BayesNet NB BayesNet NB
SMO RF SMO RF
(a) (b)
1 1
0.8 0.8
Normalized expected cost
0.6 0.6
0.4 0.4
0.2 0.2
0 0
0 0.2 0.4 0.6 0.8 1 0 0.2 0.4 0.6 0.8 1
BayesNet NB BayesNet NB
SMO RF SMO RF
(c) (d)
Figure 11: Cost curves of (a) glioma, (b) meningioma, (c) no tumor, and (d) pituitary tumors for PS 34% of the test data using BayesNet,
SMO, NB, and RF classifiers.
Table 4: Calculated values of different performance metrics for test data itself, CVF 4, CVF 10, and PS 34% of test data using BayesNet,
SMO, NB, and RF classifiers.
pituitary cases accurately. The RF classifier, our model, can rately; while we have utilized SMO classifier, our model
classify all the classes accurately. can classify 132 glioma images, 116 meningioma images,
For CVF 4 of the test data, in the table, we can see that 161 no-tumors, and 150 pituitary images accurately; again,
our model with the BayesNet classifier can classify 181 gli- while we have utilized NB classifier, our model can classify
oma images, 158 meningioma images, 233 no-tumors, and 120 glioma images, 84 meningioma images, 154 no-
205 pituitary images accurately; with the SMO classifier, it tumors, and 144 pituitary images accurately; while we have
can classify 224 glioma images, 200 meningioma images, utilized RF classifier, our model can classify 134 glioma
251 no-tumors, and 243 pituitary images accurately; with images, 107 meningioma images, 163 no-tumors, and 156
the NB classifier, it can classify 191 glioma images, 138 pituitary images accurately.
meningioma images, 235 no-tumors, and 208 pituitary
images accurately; and with the RF classifier can classify 4.2. ROC and PRC Curves of Test Data. We have plotted
207 glioma images, 187 meningioma images, 256 no-tumors, ROC (Figures 4–7) and PRC (not shown in this paper)
and 245 pituitary images accurately. curves of glioma, meningioma, no-tumor, and pituitary
As shown in the table, for CVF 10 of the test data, while tumors for the test data itself, CVF 4, CVF 10, and PS 34%
we utilizing BayesNet classifier, our model can classify 183 of test data using BayesNet, SMO, NB, and RF classifiers,
glioma images, 153 meningioma images, 232 no-tumors, respectively. And then, we have calculated the values of the
and 206 pituitary images accurately; while utilizing SMO area under each ROC and PRC curves. The average values
classifier, our model can classify 211 glioma images, 208 are summarized in Table 3. As shown in the table, the aver-
meningioma images, 254 no-tumors, and 244 pituitary age values of the ROC area (i) for the test data itself are
images accurately; again, while utilizing NB classifier, our 0.933, 0.984, 0.900, and 1.000; (ii) for CVF 4 are 0.899,
model can classify 189 glioma images, 134 meningioma 0.925, 0.890, and 0.956; (iii) for CVF 10 are 0.901, 0.924,
images, 235 no-tumors, and 205 pituitary images accurately; 0.891, and 0.956; and (iv) for PS 34% are 0.895, 0.889,
while utilizing RF classifier, our model can classify 214 gli- 0.892, and 0.932 of the test data using BayesNet, SMO, NB,
oma images, 182 meningioma images, 256 no-tumors, and and RF classifiers, respectively. These values indicate that
242 pituitary images accurately. the RF classifier performs better than the other classifiers
In the case of PS 34% of the test data, the table dem- in every case. The table also similarly reveals that the RF
onstrates that while we have utilized BayesNet classifier, classifier gives the highest PRC area values of 1.000, 0.894,
our model can classify 116 glioma images, 98 meningioma 0.897, and 0.848 for the test data itself, CVF 4, CVF 10,
images, 152 no-tumors, and 140 pituitary images accu- and PS 34%, respectively.
16 Journal of Sensors
4.3. Cost Curve Analysis of Test Data. We have also analyzed For the test data itself, Table 4 shows that the values of accuracy,
cost curves which yields model performance based on the sensitivity, specificity, FPR, FNR, precision, f-measure, kappa,
minimum misclassification cost. The minimum misclassifica- and MCC are higher for the RF classifier compared to those
tion cost can also be defined as the minimum area under the values of the other classifiers, exhibiting that the RF classifier
bottom of the envelope. Therefore, the minimum region below gives excellent performance with 100% accuracy.
the lower end of the range lowers the cost of misclassification For the case of CVF 4 of test data, Table 4 shows that the
and thus provides better classification performance. Figures 8– values of accuracy, sensitivity, specificity, precision, f-mea-
11 represent the cost curves for different testing options, such sure, kappa, and MCC are higher and the values of FPR
as the test data itself, CVF 4, CVF 10, and PS 34% of the test and FNR are lower for the SMO classifier than those values
data employing BayesNet, SMO, NB, and RF classifiers for gli- for the other classifiers, indicating that the SMO classifier
oma, meningioma, no-tumor, and pituitary classes, respec- gives better performance with 92.50% accuracy. Similar
tively. As seen in Figure 8, the RF classifier very clearly offers results are seen in the case of CVF 10 of test data, where
minimal misclassification cost; that is, better classification per- the SMO classifier gives 92.45% accuracy.
formance. However, the SMO classifier shows a slightly lower However, PS 34% of the test data in the table shows that
misclassification cost in Figures 9–11. the values of accuracy, sensitivity, specificity, FPR, FNR,
kappa, and MCC are higher for the RF classifier than the
4.4. Performance Evaluation. Finally, we have evaluated the other classifiers. The values of precision are the same for
performance of our models based on their respective confusion both SMO and RF classifiers, and the values of f-measure
matrices. We have calculated the values of accuracy, sensitivity, are very slightly lower for the RF classifier compared to the
specificity, FPR, FNR, precision, f-measure, kappa and MCC SMO classifier. From these statistical values, we can assume
and summarized them in Table 4, which represents the values that the RF classifier reveals a better performance with
for the test data itself: CVF 4, CVF 10, and PS 34%, respectively. 89.27% accuracy.
Journal of Sensors 17
vision transformers, applied to MRI: a survey,” Journal Imag- [29] Y. Zhou, Z. Li, H. Zhu et al., “Holistic brain tumor screening
ing, vol. 8, no. 8, pp. 205–240, 2022. and classification based on DenseNet and recurrent neural
[15] N. A. Ali, A. R. Syafeeza, L. J. Geok, Y. C. Wong, N. A. Hamid, network,” in Brainlesion: Glioma, Multiple Sclerosis, Stroke
and A. S. Jaafar, “Design of automated computer-aided classi- and Traumatic Brain Injuries. BrainLes 2018. Lecture Notes
fication of brain tumor using deep learning,” in Intelligent and in Computer Science, vol 11383Springer, Cham.
Interactive Computing. Lecture Notes in Networks and Systems, [30] T. Hossain, F. S. Shishir, M. Ashraf, M. A. Al Nasim, and F. M.
vol 67, V. Piuri, V. Balas, S. Borah, and S. Syed Ahmad, Eds., Shah, “Brain tumor detection using convolutional neural net-
Springer, Singapore, 2019. work,” in 2019 1st International Conference on Advances in
[16] B. S. Kumar, “Identification and classification of brain tumor Science, Engineering and Robotics Technology (ICASERT),
images using efficient classifier,” International Journal of Engi- pp. 1–6, Dhaka, Bangladesh, 2019.
neering and Advanced Technology, vol. 8, no. 6, pp. 3677–3683, [31] P. M. Krishnammal and S. S. Raja, “Convolutional neural net-
2019. work based image classification and detection of abnormalities
[17] T. Ateeq, M. N. Majeed, S. M. Anwar et al., “Ensemble-classi- in MRI brain images,” in 2019 International Conference on
fiers-assisted detection of cerebral microbleeds in brain MRI,” Communication and Signal Processing (ICCSP), pp. 0548–
Computers and Electrical Engineering, vol. 69, pp. 768–781, 0553, Chennai, India, 2019.
2018. [32] S. Deepak and P. M. Ameer, “Brain tumor classification using
[18] J. Cheng, W. Huang, S. Cao et al., “Enhanced performance of deep CNN features via transfer learning,” Computers in Biol-
brain tumor classification via tumor region augmentation ogy and Medicine, vol. 111, article 103345, pp. 1–12, 2019.
and partition,” PLoS One, vol. 10, no. 10, pp. 1–13, 2015. [33] E. I. Papageorgiou, P. P. Spyridonos, D. T. Glotsos et al., “Brain
[19] D. R. Patel, H. Thakker, M. B. Kiran, and V. Vakhariya, “Sur- tumor characterization using the soft computing technique of
face roughness prediction of machined components using gray fuzzy cognitive maps,” Applied Soft Computing, vol. 8, no. 1,
level co-occurrence matrix and bagging tree,” FME Transac- pp. 820–828, 2008.
tions, vol. 48, no. 2, pp. 468–475, 2020. [34] H. H. Sultan, N. M. Salem, and W. Al-Atabany, “Multi-classi-
[20] F. A. Zadeh, M. V. Ardalani, A. R. Salehi, R. J. Farahani, fication of brain tumor images using deep neural network,”
M. Hashemi, and A. H. Mohammed, “An analysis of new fea- IEEE Access, vol. 7, pp. 69215–69225, 2019.
ture extraction methods based on machine learning methods [35] N. Abiwinanda, M. Hanif, S. T. Hesaputra, A. Handayani, and
for classification radiological images,” Computational Intelli- T. R. Mengko, “Brain tumor classification using convolutional
gence and Neuroscience, vol. 2022, Article ID 3035426, 13 neural network,” in World Congress on Medical Physics and
pages, 2022. Biomedical Engineering 2018, pp. 183–189, Springer, 2019.
[21] G. T. Reddy, S. Bhattacharya, and S. S. Ramakrishnan, “An [36] J. Seetha and S. S. Raja, “Brain tumor classification using con-
ensemble based machine learning model for diabetic retinopa- volutional neural networks,” Biomedical & Pharmacology
thy classification,” in 2020 International Conference on Emerg- Journal, vol. 11, no. 3, pp. 1457–1461, 2018.
ing Trends in Information Technology and Engineering, pp. 1– [37] H. Selvaraj, S. T. Selvi, D. Selvathi, and L. Gewali, “Brain MRI
6, Vellore, India, 2020. slices classification using least squares support vector
[22] R. H. Jhaveri, A. Revathi, K. Ramana, R. Raut, and R. K. Dha- machine,” International Journal of Intelligent Computing in
naraj, “A review on machine learning strategies for real-world Medical Sciences & Image Processing, vol. 1, no. 1, pp. 21–33,
engineering applications,” Mobile Information Systems, 2007.
vol. 2022, Article ID 1833507, 26 pages, 2022. [38] A. Sarkar, M. Maniruzzaman, M. S. Ahsan, M. Ahmad,
[23] G. T. Reddy, M. P. K. Reddy, K. Lakshmanna, D. S. Rajput, M. I. Kadir, and S. T. Islam, “Identification and classifica-
R. Kaluri, and G. Srivastava, “Hybrid genetic algorithm and a tion of brain tumor from MRI with feature extraction by
fuzzy logic classifier for heart disease diagnosis,” Evolutionary support vector machine,” in 2020 International Conference
Intelligence, vol. 13, no. 2, pp. 185–196, 2020. for Emerging Technology (INCET), pp. 1–4, Belgaum, India,
[24] U. Ahmed, R. H. Jhaveri, G. Srivastava, and J. C. W. Lin, 2020.
Explainable deep attention active learning for sentimental ana- [39] Z. Ullah, M. U. Farooq, S. H. Lee, and D. An, “A hybrid
lytics of mental disorder, Association for Computing Machin- image enhancement based brain MRI images classification
ery, 2022. technique,” Medical Hypotheses, vol. 143, article 109922,
[25] P. John, “Brain tumor classification using wavelet and texture 2020.
based neural network,” International Journal of Scientific and [40] A. Kharrat, K. Gasmi, M. B. Messaoud, N. Benamrane, and
Engineering Research, vol. 3, no. 10, pp. 1–7, 2012. M. Abid, “A hybrid approach for automatic classification of
[26] F. J. Díaz-Pernas, M. Martínez-Zarzuela, M. Antón-Rodríguez, brain MRI using genetic algorithm and support vector
and D. González-Ortega, “A deep learning approach for brain machine,” Leonardo Journal of Sciences, vol. 17, no. 1,
tumor classification and segmentation using a multiscale convo- pp. 71–82, 2010.
lutional neural network,” Healthcare, vol. 9, no. 2, p. 153, 2021. [41] N. V. Shree and T. N. R. Kumar, “Identification and classifica-
[27] A. Çinar and M. Yildirim, “Detection of tumors on brain MRI tion of brain tumor MRI images with feature extraction using
images using the hybrid convolutional neural network archi- DWT and probabilistic neural network,” Brain Informatics,
tecture,” Medical Hypotheses, vol. 139, pp. 109684–109688, vol. 5, no. 1, pp. 23–30, 2018.
2020. [42] M. Arunachalam and S. Royappan Savarimuthu, “An efficient
[28] D. J. Hemanth, J. Anitha, A. Naaji, O. Geman, D. E. A. and automatic glioblastoma brain tumor detection using shift-
Popescu, and L. Hoang Son, “A Modified deep convolutional invariant shearlet transform and neural networks,” Interna-
neural network for abnormal brain image classification,” IEEE tional Journal of Imaging Systems and Technology, vol. 27,
Access, vol. 7, pp. 4275–4283, 2019. no. 3, pp. 216–226, 2017.
Journal of Sensors 19
[43] S. Sajid, S. Hussain, and A. Sarwar, “Brain tumor detection and national Journal of Applied Engineering Research, vol. 55,
segmentation in MR images using deep learning,” Arabian pp. 1045–1049, 2015.
Journal for Science and Engineering, vol. 44, no. 11, [61] V. Vakharia and R. Gujar, “Prediction of compressive strength
pp. 9249–9261, 2019. and portland cement composition using cross-validation and
[44] P. Saxena, A. Maheshwari, and S. Maheshwari, “Predictive feature ranking techniques,” Construction and Building Mate-
modeling of brain tumor: a deep learning approach,” in Inno- rials, vol. 225, pp. 292–301, 2019.
vations in Computational Intelligence and Computer Vision. [62] T. Saito and M. Rehmsmeier, “The precision-recall plot is
Advances in Intelligent Systems and Computing, vol more informative than the ROC plot when evaluating binary
1189Springer, Singapore. classifiers on imbalanced datasets,” PLoS One, vol. 10, no. 3,
[45] S. Khawaldeh, U. Pervaiz, A. Rafiq, and R. S. Alkhawaldeh, pp. 1–21, 2015.
“Noninvasive grading of glioma tumor using magnetic reso- [63] C. Drummond and R. C. Holte, “Cost curves: An improved
nance imaging with convolutional neural networks,” Applied method for visualizing classifier performance,” Machine
Sciences, vol. 8, no. 1, p. 27, 2018. Learning, vol. 65, no. 1, pp. 95–130, 2006.
[46] S. Preethi and P. Aishwarya, “Combining wavelet texture fea- [64] R. Chelghoum, A. Ikhlef, A. Hameurlaine, and S. Jacquir,
tures and deep neural network for tumor detection and seg- “Transfer learning using convolutional neural network archi-
mentation over MRI,” Journal of Intelligent Systems, vol. 28, tectures for brain tumor classification from MRI images,” in
no. 4, pp. 571–588, 2019. Artificial Intelligence Applications and Innovations. AIAI
[47] M. A. Khan, I. Ashraf, M. Alhaisoni et al., “Multimodal brain 2020. IFIP Advances in Information and Communication
tumor classification using deep learning and robust feature Technology, vol 583Springer, Cham.
selection: a machine learning application for radiologists,” [65] I. A. E. Kader, G. Xu, Z. Shuai et al., “Brain tumor detection
Diagnostics, vol. 10, no. 8, p. 565, 2020. and classification on MR images by a deep wavelet auto-
[48] M. Siar and M. Teshnehlab, “Brain tumor detection using deep encoder model,” Diagnostics, vol. 11, no. 9, pp. 1–19, 2021.
neural network and machine learning algorithm,” in 2019 9th [66] E. Aarthi, S. Jana, W. G. Theresa et al., “Detection and classifi-
International Conference on Computer and Knowledge Engi- cation of MRI brain tumors using S3-DRLSTM based deep
neering (ICCKE), pp. 363–368, Mashhad, Iran, 2019. learning model,” International Journal of Electrical and Elec-
[49] M. Li, L. Kuang, S. Xu, and Z. Sha, “Brain tumor detection tronics Research, vol. 10, no. 3, pp. 597–603, 2022.
based on multimodal information fusion and convolutional [67] N. A. Samee, N. F. Mahmoud, G. Atteia et al., “Classification
neural network,” IEEE Access, vol. 7, pp. 180134–180146, framework for medical diagnosis of brain tumor with an effec-
2019. tive hybrid transfer learning model,” Diagnostis, vol. 12, no. 10,
[50] MatLab, Version 9.2 (R2017a), The MathWorks Inc., Natick, p. 2541, 2022.
Massachusetts, 2017.
[51] E. Frank, M. A. Hall, and I. H. Witten, The WEKA workbench.
Online appendix for “Data mining: practical machine learning
tools and techniques”, Morgan Kaufmann, Fourth Edition edi-
tion, 2016.
[52] https://www.kaggle.com/datasets/sartajbhuvaji/brain-tumor-
classification-mri.
[53] https://www.kaggle.com/datasets/ahmedhamada0/brain-
tumor-detection?-Resource=download.
[54] M. Coşkun, Ö. Yildirim, U. Ç. A. R. Ayşegül, and Y. Demir,
“An overview of popular deep learning methods,” European
Journal of Technique, vol. 7, no. 2, pp. 165–176, 2017.
[55] https://neurohive.io/en/popular-networks/alexnet-imagenet-
classification-with-deep-convolutional-neural-networks/.
[56] K. Simonyan and A. Zisserman, “Very deep convolutional net-
works for large- scale image recognition,” 2014, http://arxiv
.org/abs/1409.1556.
[57] X. Han, Y. Zhong, L. Cao, and L. Zhang, “Pre-trained Alexnet
architecture with pyramid pooling and supervision for high
spatial resolution remote sensing image scene classification,”
Remote Sensing, vol. 9, no. 8, p. 848, 2017.
[58] A. M. Carvalho, T. T. Roos, A. L. Oliveira, and P. Myllymäki,
“Discriminative learning of Bayesian networks via factorized
conditional log-likelihood,” Journal of Machine Learning
Research, vol. 12, pp. 2181–2210, 2011.
[59] N. Friedman, D. Geiger, and M. Goldszmidt, “Bayesian net-
work classifiers,” Machine Learning, vol. 29, no. 2/3, pp. 131–
163, 1997.
[60] S. P. Rajamohana, K. Umamaheswari, and R. Karthiga, “Senti-
ment classification based on LDA using SMO classifier,” Inter-