Accurate MRI-Based Brain Tumor Diagnosis: Integrating Segmentation and Deep Learning Approaches
Accurate MRI-Based Brain Tumor Diagnosis: Integrating Segmentation and Deep Learning Approaches
sciences
Article
Accurate MRI-Based Brain Tumor Diagnosis: Integrating
Segmentation and Deep Learning Approaches
Medet Ashimgaliyev 1 , Bakhyt Matkarimov 1, * , Alibek Barlybayev 1,2 , Rita Yi Man Li 3
and Ainur Zhumadillayeva 1,4, *
Abstract: Magnetic Resonance Imaging (MRI) is vital in diagnosing brain tumours, offering crucial in-
sights into tumour morphology and precise localisation. Despite its pivotal role, accurately classifying
brain tumours from MRI scans is inherently complex due to their heterogeneous characteristics. This
study presents a novel integration of advanced segmentation methods with deep learning ensemble
algorithms to enhance the classification accuracy of MRI-based brain tumour diagnosis. We conduct
a thorough review of both traditional segmentation approaches and contemporary advancements
in region-based and machine learning-driven segmentation techniques. This paper explores the
utility of deep learning ensemble algorithms, capitalising on the diversity of model architectures
to augment tumour classification accuracy and robustness. Through the synergistic amalgamation
of sophisticated segmentation techniques and ensemble learning strategies, this research addresses
the shortcomings of traditional methodologies, thereby facilitating more precise and efficient brain
tumour classification.
Citation: Ashimgaliyev, M.;
Matkarimov, B.; Barlybayev, A.; Li,
Keywords: ensemble algorithm; image segmentation; magnetic resonance imaging; brain neoplasm;
R.Y.M.; Zhumadillayeva, A. Accurate
neural networks; categorisation
MRI-Based Brain Tumor Diagnosis:
Integrating Segmentation and Deep
Learning Approaches. Appl. Sci. 2024,
14, 7281. https://doi.org/10.3390/
app14167281 1. Introduction
The human brain, the body’s most complex organ, regulates various physiological
Academic Editors: Thomas Lindner
and Qi-Huang Zheng
functions, including sensory integration. Brain tumours (BTs), among the most common
global malignancies, disrupt these functions, leading to severe consequences, including
Received: 12 June 2024 death [1,2]. Average cellular turnover involves programmed cell death and regeneration,
Revised: 29 July 2024 but BTs cause uncontrolled cell proliferation, impairing brain functions. BTs can be malig-
Accepted: 13 August 2024 nant or benign, with symptoms like fever, headaches, and cognitive decline, often leading
Published: 19 August 2024 to fatality [3,4].
The early and accurate detection of BTs is crucial for improved patient outcomes.
Medical imaging modalities like MRI, CT, and PET scans are used. MRI, favoured for
high-resolution images, uses contrast agents like gadolinium to differentiate pathological
Copyright: © 2024 by the authors.
tissues. Computer-aided diagnosis (CAD) systems analyse these tissues for precise BT
Licensee MDPI, Basel, Switzerland.
This article is an open access article
detection [5].
distributed under the terms and
Brain tumours are categorised according to their origin, type, and malignancy, ranging
conditions of the Creative Commons from benign Grade I to aggressive Grade IV, influencing treatment approaches [6–12]. Both
Attribution (CC BY) license (https:// Machine Learning (ML) and Deep Learning (DL) improve the accuracy of brain tumour
creativecommons.org/licenses/by/ classification [1]. While ML methods such as SVM and KNN are effective, they necessitate
4.0/).
[6–12]. Both Machine Learning (ML) and Deep Learning (DL) improve the accuracy o
Appl. Sci. 2024, 14, 7281 brain tumour classification [1]. While ML methods such as SVM and KNN 2 ofare
20 effective
they necessitate manual feature extraction. In contrast, DL, mainly through Convolutiona
Neural Networks (CNNs), automates feature learning, enhancing diagnostic precisio
[13,14].
manual feature extraction. In contrast, DL, mainly through Convolutional Neural Networks
Recent
(CNNs), automates advancements
feature in AI, especially
learning, enhancing DL, have
diagnostic revolutionised
precision [13,14]. BT detection. CNN
and
Recent ensemble learning
advancements methods DL,
in AI, especially combine multiple models
have revolutionised to enhance
BT detection. CNNsclassificatio
and ensemble learning
accuracy andmethods combine
robustness multiple
[15]. This papermodels to enhance
explores classification
integrating accuracy
segmentation and ensembl
and robustness [15].techniques
learning This paperfor
explores integrating
MRI-based segmentation
BT classification, and ensemble
discussing learning clinica
advancements,
techniquesimplications,
for MRI-based BT classification, discussing advancements, clinical implications,
and future research directions to improve patient treatment and outcome
and future[16–43].
research directions to improve patient treatment and outcomes [16–43].
An overviewAn of such primary
overview of suchdeep learning-based
primary brain tumour brain
deep learning-based segmentation
tumour tech-
segmentatio
niques shows the effectiveness
techniques shows theof these methods
effectiveness (Figure
of these 1):
methods (Figure 1):
Figure 1. of
Figure 1. Overview Overview
key deepoflearning-based
key deep learning-based brain
brain tumour tumour segmentation
segmentation techniques techniques
[10,12]. [10,12].
networks offer unique advantages in processing sequential data, making them well-suited
for tasks involving temporal dynamics and spatial relationships within MRI sequences.
By modelling sequential dependencies, RNNs and LSTMs can capture temporal patterns
in tumour progression and enhance classification accuracy. DNNs encompass a broad
class of neural network architectures with multiple layers of interconnected nodes. These
networks can learn intricate representations from high-dimensional data, making them
suitable for complex tasks such as brain tumour classification. AEs leverage unsupervised
learning principles to extract latent representations of input data. By learning to reconstruct
input images from compressed representations, deep autoencoders can capture meaningful
features and reduce data dimensionality, facilitating more efficient classification [48]. GANs
offer a novel approach to data generation and representation learning by training two
neural networks simultaneously: a generator and a discriminator. GANs can generate
realistic synthetic images that resemble real MRI data, providing valuable augmentation
for training deep learning models and enhancing their generalisation capabilities [49,50].
Integrating suitable segmentation methods with deep learning-based ensemble al-
gorithms for MRI-based brain tumour classification aims to enhance the accuracy and
robustness of tumour analysis. Here is a brief overview of each component:
1. Segmentation: Suitable segmentation methods aim to accurately identify tumour
regions in MRI images, separating them from healthy tissues. This involves threshold-
ing, region growing, active contours, and machine learning-based approaches [51,52].
The choice of method depends on the tumour’s complexity and the data quality [53].
2. Ensemble algorithms: Deep learning-based algorithms combine multiple models to
create a more accurate and robust classification system. They can be used to train on
different types of data and integrate the predictions of various models to produce a
final result. This approach can help reduce errors and improve the overall accuracy of
the classification task [54].
By combining these two techniques, researchers can create a powerful tool for analysing
MRI images and detecting tumours with greater accuracy and reliability. Deep learning-
based ensemble algorithms utilise the diversity of multiple models to enhance classification
accuracy and robustness by combining their predictions. Techniques such as bagging,
boosting, and stacking combine the output of various models into a collective decision. In
the case of brain tumour classification, convolutional neural networks (CNNs) trained on
segmented regions of the tumour are utilised to extract discriminatory features that are
then fed into ensemble algorithms for final classification [52].
The process of combining segmentation methods with deep learning ensemble algo-
rithms includes the following steps:
Segmentation of tumour areas: Utilize suitable segmentation techniques to identify
tumour regions within MRI images. This involves preprocessing the data, applying appro-
priate segmentation algorithms, and creating segmented masks for the tumour areas [51].
Feature extraction: Extract features from segmented tumour regions using deep learn-
ing models such as CNNs. Train these models on the segmented regions to learn features
that can capture patterns and characteristics specific to brain tumours [52].
Ensemble learning: Use ensemble learning techniques to combine the predictions of
multiple deep learning models. This could involve training different CNN architectures
with different initialization parameters, and then combining their predictions by averaging,
voting, or other ensemble methods [55].
Classification: Use the combined predictions from the ensemble to classify tumours
into different categories. The final classification can be based on the consensus of the
individual models’ predictions, with each model’s contribution weighted according to its
performance or confidence level [53].
By integrating appropriate segmentation techniques with deep learning ensemble
algorithms, researchers seek to enhance the accuracy, dependability, and generalizability
of MRI-driven brain tumour classification models. This integrated strategy enables more
Appl. Sci. 2024, 14, 7281 4 of 20
accurate tumour analysis, facilitating improved treatment planning and patient care in
neuro-oncology [51,56].
The study employed transfer learning with five pre-trained CNN models: AlexNet,
VGG16, GoogleNet, ResNet18, and ResNet50, all originally trained on the ImageNet dataset.
It also provides a detailed discussion of each model’s architecture:
1. AlexNet:
Alex Krzyzewski introduced AlexNet in the 2012 Large-Scale Visual Recognition
Challenge (ILSVRC). It won the competition with a top-five error rate of 15.3%, surpassing
the state-of-the-art models of that time, which had error rates of 10.8% or higher [51].
Initially trained on two GPU machines, AlexNet now requires only one GPU. It is a
relatively shallow network with eight layers, comprising five convolutional layers (Conv)
followed by three fully connected (FC) layers [52].
The architecture includes layers with three different filter sizes (11 × 11, 5 × 5, and
3 × 3), data augmentation, dropout, and max pooling operations [53]. Notably, the tradi-
tional sigmoid activation function SF ( x ) was replaced with the rectified linear unit (ReLU)
activation function to address the vanishing gradient problem associated with sigmoid
functions [53]. This change enhanced the training stability of the model by preventing
the learning process from halting when the gradient approaches zero using the following
equations [52]:
1. Sigmoid function (Equation (1)):
1
SF ( x ) = (1)
1 + exp (− x )
2. Rectified linear unit (ReLU) function (Equation (2)):
Figure 3. Fine-tuned
Figure3. Fine-tuned MobileNetV2.
MobileNetV2.
Several
Several keykeyperformance
performancemetrics
metricsare commonly
are commonly used in the
used context
in the of a binary
context clas-
of a binary
sification problem, such as tumour detection. True Positive (TP) represents
classification problem, such as tumour detection. True Positive (TP) represents the the number
of correctly
number of identified
correctly positive
identifiedsamples,
positivemeaning
samples,those with tumours.
meaning False
those with PositiveFalse
tumours. (FP)
refers to the
Positive (FP)number
refers of
to incorrectly
the number identified positive
of incorrectly samples positive
identified or those without
samples tumours.
or those
True Negative (TN) denotes the number of correctly identified negative
without tumours. True Negative (TN) denotes the number of correctly identified samples negative
or those
without
samples tumours. Conversely,
or those without False
tumours. Negative (FN)
Conversely, Falseindicates
Negativethe(FN) number of the
indicates incorrectly
number
identified negative samples or those with tumours. Accuracy, as described in a relevant
of incorrectly identified negative samples or those with tumours. Accuracy, as described
equation, represents the proportion of correctly identified samples in the data. These
in a relevant equation, represents the proportion of correctly identified samples in the
metrics provide a comprehensive assessment of the model’s performance in distinguishing
data. These metrics provide a comprehensive assessment of the model’s performance in
between the presence and absence of tumours in the given dataset [1]. Equations (5)–(9)
distinguishing between the presence and absence of tumours in the given dataset [1].
describe the commonly used evaluation criteria for this study:
Equations (5)–(9) describe the commonly used evaluation criteria for this study:
( T(𝑇𝑝 − 𝑇𝑛)
p − Tn )
𝐴𝑐𝑐𝑢𝑟𝑎𝑐𝑦==( T p + Fp + Tn + Fn)
Accuracy (3)
(3)
(𝑇𝑝 + 𝐹𝑝 + 𝑇𝑛 + 𝐹𝑛)
Precision is
Precision is aa performance
performance metric
metric representing
representing the
the proportion
proportion ofof correctly
correctly identified
identified
positivesamples
positive samplesamong
amongallallthose
thoseidentified
identifiedas
aspositive
positivebybythe
themodel.
model.Recall
Recallisis the
the ratio
ratio of
of
correctlyidentified
correctly identifiedpositive
positivesamples
samplestotothe
thetotal
totalnumber
numberof ofpositive
positivesamples.
samples. The
The F1F1 score
score
is aa composite
is compositemetric
metriccombining
combiningprecision
precisionand
andrecall
recallinto
intoaasingle
singlevalue.
value.
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 T p𝑇𝑝
Precision = = (𝑇𝑝 + 𝐹𝑝) (4)
(4)
( T p + Fp)
Tp
Recall = (5)
( T p + Fn)
2 ( Precision × Recall )
F1 − score = (6)
( Precision + Recall )
The Dice coefficient is like the Intersection over Union (IoU) and ranges from 0 to 1,
with 1 indicating the highest similarity between the predicted segmentation and the ground
truth. This coefficient evaluates the effectiveness of automated probabilistic fractional
Appl. Sci. 2024, 14, 7281 8 of 20
Intersection over Union (IoU) is a performance metric used to evaluate the accuracy of
object detection models. It compares the overlap between the predicted positive regions
(including both true positives and false positives) and the actual positive regions (including
both true positives and false negatives) in the dataset. The intersection of these sets
represents the true positive (Tp) detections, while the union encompasses the true positives,
false positives (Fp), and false negatives (Fn).
Tp
oU = (8)
( T p + Fp + Fn)
The Similarity Index (SI) is a performance metric used to assess the accuracy of
tumour detection models. It measures the similarity between the ground truth annotations
representing the actual tumour regions and the model’s segmentation output, which
identifies the predicted tumour regions. A higher SI value indicates a closer match between
the predicted segmentation and the ground truth, signifying a more accurate detection of
the tumour areas.
2T p
SI = (9)
(2T p + Fp + Fn)
These evaluation metrics are essential for assessing the performance of brain tumour
detection and classification models. They measure the models’ ability to correctly identify
positive (tumour-containing) and negative (tumour-free) samples and differentiate between
different types of tumours. Metrics such as True Positive, False Positive, True Negative,
False Negative, Accuracy, Precision, Recall, F1 score, Intersection over Union, and Similarity
Index comprehensively evaluate the model’s performance in various tumour detection and
classification aspects.
3. Dataset
To evaluate the performance of our methods, we used the BRATS 2018 dataset (Menze
et al., 2015; Bakas et al., 2017 a, b, c, 2018). The training set comprised images from
285 patients, including 210 with high-grade gliomas (HGG) and 75 with low-grade gliomas
(LGG). The validation set included MRI scans from 66 patients with brain tumours of an
unknown grade, as predefined by the BRATS challenge organisers. The test set consisted
of images from 191 brain tumour patients. Of these, 77 had undergone Gross Total Resec-
tion (GTR). Their data was used to evaluate the model’s ability to predict survival.Each
patient was scanned using T1, T1Gd, T2, and FLAIR. Each image was skull-stripped and
resampled to a uniform resolution of 1 mm3 in all dimensions. The sequences for each
patient were aligned to ensure consistency. Experts manually created the ground truth
segmentation masks.
The model’s performance was assessed on the validation and test sets using the CBICA
Image Processing Portal, available at ipp.cbica.upenn.edu. Segmentation annotations
included tumour subtypes such as necrotic/non-enhancing tumour (NCR), peritumoral
oedema (ED), and gadolinium-enhancing tumour (ET).
Below is listed the step-by-step process of training deep learning models for brain
tumour classification:
Data Preparation:
1. Dataset collection: We started by collecting a comprehensive dataset of MRI scans,
such as the BRATS (Brain Tumor Segmentation Challenge) series from 2013 to 2019.
This dataset included thousands of MRI images with corresponding annotations
indicating tumour regions.
Appl. Sci. 2024, 14, 7281 9 of 20
4. Results
Figure 4 presents a sequence of MRI scans illustrating the detection and segmentation
process of a meningioma, a type of brain tumour.
and optimised for accurate and robust brain tumour classification [3,5,9].
4. Results
Figure 4 presents a sequence of MRI scans illustrating the detection and segmentation
Appl. Sci. 2024, 14, 7281 10 of 20
process of a meningioma, a type of brain tumour.
Figure
Figure 4.
4. Meningioma
Meningioma identification
identification in
in brain
brain MRI.
This diagrammatic
This diagrammaticrepresentation
representation systematically
systematicallyencapsulates
encapsulatesthe fusion
the of convolu-
fusion of
tional neural networks
convolutional with sophisticated
neural networks classification
with sophisticated algorithms,algorithms,
classification crafting a comprehen-
crafting a
sive medical imaging
comprehensive medicalsolution.
imaging This integration
solution. Thisaims to elevate
integration aimsdiagnostic
to elevateprecision and
diagnostic
contribute positively to enhanced healthcare outcomes. The first image
precision and contribute positively to enhanced healthcare outcomes. The first image on on the left displays
a transverse
the section
left displays of a brainsection
a transverse MRI scan.of aItbrain
prominently
MRI scan.shows a large, hyperintense
It prominently mass
shows a large,
in the cerebral tissue, characteristic of a meningioma. The mass
hyperintense mass in the cerebral tissue, characteristic of a meningioma. The mass appears white on the
appears white on the T1-weighted MRI image due to its contrast enhancement, indicatinga
T1-weighted MRI image due to its contrast enhancement, indicating the presence of
dense,
the extra-axial
presence of atumour.
dense,The middle image
extra-axial tumour. is a binary mask highlighting
The middle image is a the segmented
binary mask
highlighting the segmented tumour. This mask simplifies the image to show onlyfrom
tumour. This mask simplifies the image to show only the tumour region, isolated the
the surrounding
tumour brain structures.
region, isolated from the Here, the tumour
surrounding is depicted
brain structures. as Here,
a purethe
white area on
tumour is
a black background,
depicted emphasising
as a pure white area on athe exact
black shape and boundaries
background, emphasisingofthe theexact
meningioma
shape and as
identified byofimage
boundaries processing algorithms.
the meningioma as identifiedThe by final
imageimage on the right
processing reintroduces
algorithms. the
The final
context of the original MRI scan, overlaying a yellow outline that demarcates
image on the right reintroduces the context of the original MRI scan, overlaying a yellow the perimeter
of the tumour. This outline visualises the tumour’s extent and spatial relationship to nearby
outline that demarcates the perimeter of the tumour. This outline visualises the tumour’s
brain structures, crucial for surgical planning or targeted therapies. This series effectively
extent and spatial relationship to nearby brain structures, crucial for surgical planning or
demonstrates the application of image processing techniques in medical imaging, specifi-
targeted therapies. This series effectively demonstrates the application of image
cally for tumour identification, isolation, and data preparation for further clinical analysis
processing techniques in medical imaging, specifically for tumour identification, isolation,
or intervention.
and data preparation for further clinical analysis or intervention.
Integrating suitable segmentation methods with a deep learning-based ensemble
Integrating suitable segmentation methods with a deep learning-based ensemble
algorithm offered several advantages, including improved accuracy, robustness, and
algorithm offered several advantages, including improved accuracy, robustness, and
interpretability—however, challenges such as dataset bias, class imbalance, and com-
interpretability—however, challenges such as dataset bias, class imbalance, and
putational resources needed to be addressed.
computational resources needed to be addressed.
Overall, the results demonstrated that the proposed approach could enhance the
Overall, the results demonstrated that the proposed approach could enhance the
performance of MRI-based brain tumour classification, contributing to improved patient
performance
outcomes andofpersonalised
MRI-based brain tumour
treatment classification, contributing to improved patient
strategies.
outcomes and personalised treatment strategies.
Table 1 compares the specific attributes and characteristics of commonly used pre-
trained CNN architectures, including AlexNet, VGG16, GoogleNet, ResNet18, and ResNet50.
Each architecture is evaluated based on layer count, input size, model description, unique
features, top-five error rates, and the total number of parameters in millions.
Appl. Sci. 2024, 14, 7281 11 of 20
Table 1. Key attributes and characteristics of the pre-trained CNN architectures used.
The data from Table 1 underscore the varied design principles and efficiencies inherent
in each CNN architecture, offering insights into their suitability for specific image process-
ing and pattern recognition tasks. The top column indicates the total number of layers in
each model, reflecting their depth: AlexNet has 8 layers, VGG16 has 16, GoogleNet has
22, ResNet18 has 18, and ResNet50 has 50 layers. Except for AlexNet, which processes
images of 227 × 227 pixels, all other models handle images of 224 × 224 pixels with three
colour channels (RGB). The Model Description column outlines the composition of each
architecture in terms of convolutional (Conv) and fully connected (FC) layers. For instance,
AlexNet includes 5 convolutional layers and three fully connected layers, while ResNet50
features 49 convolutional layers and 1 fully connected layer. Each model has unique fea-
tures that enhance its functionality. AlexNet incorporates local response normalization
and overlapping max pooling. VGG16 is noted for its deeper layer structure. GoogleNet
integrates object localization and image classification, employs 1 × 1 convolutions, utilises
global average pooling, and includes an inception module. ResNet architectures (ResNet18
and ResNet50) are characterised by their use of skip connections, which help mitigate the
vanishing gradient problem during training. The Top-Five Error Rate column indicates the
models’ accuracy in classifying the top five predictions, with lower percentages signifying
higher accuracy. ResNet18 and ResNet50 have the lowest error rates at 3.57%, followed
by GoogleNet at 6.67%, VGG16 at 7.3%, and AlexNet at 15.3%. The number of trainable
parameters varies significantly among the models, reflecting their complexity and com-
putational demands. VGG16 has the highest number of parameters at 138 million, while
GoogleNet is the most parameter-efficient with only 4 million.
These CNN architectures are extensively used in various computer vision tasks, such
as image classification, object detection, and semantic segmentation. They are potent tools
for feature extraction and representation learning, facilitating the development of advanced
deep-learning models for various applications [57].
By incorporating these diverse deep learning architectures into ensemble frameworks,
researchers can harness the complementary strengths of individual models and improve
classification performance. Ensemble learning techniques such as bagging, boosting, and
stacking further enhance the robustness and reliability of classification systems by aggregat-
ing predictions from multiple models. Through the synergistic integration of segmentation
methods and ensemble learning, researchers aim to unlock the full potential of MRI-based
brain tumour classification, paving the way for more accurate diagnosis and personalised
treatment strategies in neuro-oncology [44].
We fine-tuned the model parameters in our machine learning process to optimise perfor-
mance and enhance results. For our study, we trained various deep learning models—AlexNet,
VGG16, ResNet18, GoogLeNet, and ResNet50—initially using the ImageNet dataset. Sub-
integration of segmentation methods and ensemble learning, researchers aim to unlock
the full potential of MRI-based brain tumour classification, paving the way for more
accurate diagnosis and personalised treatment strategies in neuro-oncology [44].
We fine-tuned the model parameters in our machine learning process to optimise
Appl. Sci. 2024, 14, 7281 performance and enhance results. For our study, we trained various deep learning 12 of 20
models—AlexNet, VGG16, ResNet18, GoogLeNet, and ResNet50—initially using the
ImageNet dataset. Subsequently, we carefully conducted additional training on
sequently, we
specialised carefully
datasets conducted
to adapt theseadditional
models fortraining on specialised
the specific datasetsimage
tasks of medical to adapt these
analysis,
models for the
particularly forspecific tasks of medical
the segmentation image analysis,
and classification ofparticularly for the segmentation and
brain tumours.
classification
Figure 5ofshows
brain tumours.
examples of segmentation outcomes contrasted with the ground
Figure
truth. 5 shows examples of segmentation outcomes contrasted with the ground truth.
Model
Dataset Preprocessing Performance
Architecture
Bias field correction,
BRATS
intensity and patch DSC 88%, SEN 89%,
2013 and Custom CNN
normalisation, PR 87%
2015
augmentation
BRATS Intensity normalisation, SEN 95%, SPE 95.5%, PR 96.5%,
HCNN + CRF-RRNN 1
2013 augmentation RE 97.8%, ACC 98.6%
Residual Network +
BRATS
Z-score normalisation dilated convolution DSC 86%
2015
RDM-Net 2
Stack Multi-connection
BRATS
Z-score normalisation Simple Reducing_Net DSC 83.42%, PR 78.96%, SEN 90.24%
2015
(SMCSRNet)
BRATS Ensemble of a 3D-CNN
- DSC 90.6%
2019 and U-Net
BRATS Bias correction, Two-PathGroup-CNN
DSC 89.2%, PR 88.22%, SEN 88.32%
2015 intensity normalisation (2PG-CNN)
BRATS Hybrid two-track U-Net
- DSC 86.5%, SEN 88.3%, SPE 99.9%
2018 (HTTU-Net)
P-Net with bounding
BRATS
- box and image-specific DSC 86.29%
2015
fine tuning (BIFSeg)
Denoising,
Multi-scale CNN
ADNI skull stripping, ACC 90.1%
(MSCNN)
sub-sampling
Intensity normalisation,
BRATS
resizing, bias field Cascaded 3D U-Nets DSC 89.4%
2017
correction
BRATS
3D centre-crop BRATS 2015: DSC 88.4%, SEN 83.8%
2015 and Downsampling
dense block BRATS 2017: DSC 88.7%, SEN 84.3%
2017
BRATS BRATS 2018: DSC 90%, SEN 90.3,
Z-score normalisation,
2018 and 3D FCN 3 SPE 99.48%; BRATS 2019: DSC 89%,
cropping
2019 SEN 88.3%, SPE 99.51%
Appl. Sci. 2024, 14, 7281 14 of 20
Table 2. Cont.
Model
Dataset Preprocessing Performance
Architecture
Intensity normalisation,
BRATS removing 1% of DCNN BRATS 2018: DSC 86.2%, SEN 84.8%,
2018 highest and lowest (Dense-MultiOCM 4) SPE 99.5%
intensity
Image cropping, DSC 84%, SEN 92%,
TCIA padding, resizing, U-Net SPE 92%,
intensity normalisation ACC 92%
BRATS
BRATS 2013 DSC 86%,
2013,
- AFPNet 5 + 3D CRF BRATS 2015 DSC 82%,
2015,
BRATS 2018 86.58%
2018
Inception-based U-Net
BRATS
+ up skip connection +
2015, Z-score normalisation DSC 89%, PR 78.5%, SEN 89.5%
cascaded training
2017
strategy
Cropping,
BRATS
z-score normalisation, Triple-intersecting BRATS 2015: DSC 85%,
2015,
min–max normalisation U-Nets (TIU-Net) BrainWeb DSC 99.5%
BrainWeb
(BrainWeb)
BRATS LSTM multi-modal DSC 73.09%, SEN 63.76%,
-
2015 U-Net PR 89.79%
1 Heterogeneous CNN combined with conditional random fields and recurrent regression-based neural algorithm.
The values reported indicate the high effectiveness of the method, with an accuracy
of 97.47%, demonstrating the model’s ability to identify both positive and negative cases
correctly. The sensitivity is recorded as 97.22%, indicating the model’s proficiency in
correctly identifying positive cases. The specificity is 97.94%, reflecting the model’s accuracy
in identifying negative cases. Finally, the Dice coefficient index, which measures the
similarity between the predicted and actual labels, is 96.71%. These metrics collectively
suggest that the proposed method exhibits robust performance in detecting brain tumours.
Table 4 provides a comparative analysis of various methodologies for brain tumour
classification using the BRATS 2018 dataset. The table outlines the classification accuracies
of different techniques, facilitating an evaluation of their efficacy in tumour detection.
The methods include combinations of CNN and the Local Binary Patterns (LBP) with
Particle Swarm Optimization (PSO) reported by Irfan et al. [44], achieving a 92.5% accuracy.
The LSTM by Amin et al. [51] shows slightly higher accuracy at 93.8%. The brain-storm
optimisation technique by Narmatha et al. [52] matches the first at 92.5%. At the same time,
a method integrating Discrete Cosine Transform (DCT), CNN, and the Extreme Learning
Machine (ELM) by Khan et al. [53] achieves a value of 93.4%. Notably, the proposed
model, which utilises a CNN, MobileNetV2, and Multi-Support Vector Machine (M-SVM),
significantly surpasses these, recording a classification accuracy of 97.5%. This indicates
Appl. Sci. 2024, 14, 7281 15 of 20
the superior performance of the proposed method in accurately classifying brain tumours,
highlighting its potential utility in clinical application.
The researchers used a multifaceted approach to improve the results of their med-
ical image analysis models. First, they fine-tuned the model architectures to match the
characteristics of medical image data, such as by adding or changing particular layers.
This architectural adaptation enhanced the models’ ability to capture medical imagery’s
unique features and patterns. Next, the researchers experimented with different training
parameters, like learning rate and batch size, to find the optimal settings for each model
and dataset. By systematically exploring the parameter space, they sought to unlock the
full potential of their models and ensure optimal performance on specific medical image
analysis tasks. Additionally, the researchers applied regularisation techniques, such as
adding regularisation layers and using early stopping, to prevent the over-training of
the models. These strategies helped to improve the models’ generalisation capabilities,
ensuring that they could reliably perform well on new, unseen medical images. Finally,
the researchers tested various loss functions to identify the one best-suited for the task, to
achieve the best possible results for their models. By carefully selecting the appropriate loss
function, they could further refine the models’ performance and enhance the accuracy and
reliability of the medical image analysis outcomes. Through this multifaceted approach,
encompassing architectural adaptations, training parameter optimisation, regularisation
strategies, and loss function selection, the researchers worked diligently to enhance the
performance of their medical image analysis models and obtain the most accurate and
reliable outcomes.
After preliminary training on a dataset, each model was further trained on a specific
dataset to improve results in a particular area, in this case medical image processing.
Additional training allowed the models to better adapt to the specifics of classifying
and segmenting medical images, including brain scans. Additionally, methods of fine-
tuning hyperparameters were applied, and the optimisation of learning algorithms was
undertaken to achieve the best results in a specific task. This approach has significantly
improved the accuracy of image classification and segmentation, which is essential for
diagnosing and treating patients with brain tumours.
As a result of these efforts, we have obtained significant improvements in the accuracy
of segmentation and the classification of brain tumours, which makes our model more
accurate and useful for medical applications.
5. Discussion
This research incorporates sophisticated deep-learning frameworks to enhance the
precision of brain tumour identification and segmentation. Our study focuses on menin-
giomas, utilising a series of MRI images to showcase our model’s capability in detecting
and outlining tumour boundaries accurately.
The visual progression presented in Figure 2, from detecting a hyperintense menin-
gioma mass to its precise segmentation and contextual visualisation, highlights the effec-
tiveness of combining CNNs with advanced classification algorithms. This methodological
fusion is intended to refine diagnostic accuracy and facilitate clinical interventions by
Appl. Sci. 2024, 14, 7281 16 of 20
providing precise, actionable imaging data. The binary mask and the subsequent overlay
on the MRI scan are pivotal in clarifying the tumour’s spatial relationship with adjacent
cerebral structures, a critical factor in planning surgical or therapeutic procedures.
The selection of CNN architectures, as detailed in Table 1, underpins our strategy to
optimise image analysis. Each architecture, from AlexNet to ResNet50, is selected based
on inherent features that support deep learning tasks specific to medical imaging, such as
object localisation and reducing vanishing gradients via skip connections. The diversity
in layer depth, error rates, and computational efficiency across these models facilitates a
tailored approach to feature extraction and image classification, ensuring robustness in
tumour segmentation outcomes.
As summarised in Table 3, our empirical evaluations demonstrate the proposed
model’s superior performance in tumour detection metrics like accuracy, sensitivity, speci-
ficity, and the Dice coefficient index. These metrics not only corroborate the high efficacy of
the Multi-Support Vector Machine classifier when applied to the BRATS 2018 dataset, but
also reflect the model’s generalizability across diverse neuroimaging datasets, including
those from ADNI, TCIA, and BrainWeb.
Despite the promising outcomes, our research recognises inherent challenges
such as dataset bias and class imbalance, which could skew model training and affect
generalizability. Addressing these challenges involves refining data preprocessing and
augmentation techniques to ensure a balanced representation of tumour types and
stages across training sets.
Moreover, the comparative performance metrics in Table 4 emphasise the advancement
of our proposed method over existing techniques. By achieving a classification accuracy of
97.5%, our model sets a new benchmark in the field, underscoring its potential for clinical
application. It surpasses traditional methods and recent innovations, which have shown
lower accuracies in similar settings.
The integration of advanced CNN architectures and ensemble learning strategies has
proven highly effective in enhancing the accuracy and reliability of brain tumour clas-
sification systems. The continuous improvement of these systems through architecture
adaptation, the optimisation of training parameters, and advanced regularisation tech-
niques points towards an exciting future where deep learning can significantly contribute
to personalised medicine and improved patient outcomes in neuro-oncology. Our future
work will further enhance these models, address existing challenges, and expand their
applicability to other medical imaging tasks.
Limitations
While the proposed method for brain tumour classification using deep learning and
ensemble algorithms shows promising results, several limitations must be acknowledged.
The model’s performance relies heavily on the training datasets used, which may not fully
represent the diversity of real-world clinical data, affecting the generalisation to unseen data.
Brain tumour datasets often exhibit class imbalances, leading to biassed model training
and suboptimal performance for less common tumour types. Additionally, the training and
deployment of deep learning models require significant computational resources, which
may not be accessible in all clinical settings. Furthermore, the interpretability of these
models remains a challenge, as clinicians require clear and understandable explanations
for their predictions to aid in decision-making, which the current models may lack. These
limitations highlight the need for continued research and development to address dataset
bias, class imbalance, computational constraints, and model interpretability, ensuring the
deployment of robust and clinically relevant brain tumour classification models.
Robustness and generalisation are critical factors for successfully translating advanced
deep-learning models for brain tumour classification from research to clinical practise.
Ensuring robustness across different imaging conditions and scanners is crucial, as vari-
ations in MRI protocols, scanner types, and patient populations can impact the model’s
performance. To ensure generalisation, the models must be validated extensively across
Appl. Sci. 2024, 14, 7281 17 of 20
diverse clinical settings. Additionally, despite applying regularisation techniques and early
stopping, the risk of overfitting persists, especially when models are trained on limited
datasets, which can lead to poor performance on new, unseen data. Furthermore, integrat-
ing these complex models into clinical workflows poses practical challenges, requiring the
development of user-friendly interfaces and seamless integration with hospital information
systems to facilitate widespread adoption. Addressing these limitations, including those
relating to robustness, generalisation, overfitting, and clinical integration, is crucial for
the reliable and effective deployment of advanced deep learning models in real-world
medical applications.
6. Conclusions
This study successfully demonstrates the capability of an integrated deep-learning
approach to enhance the accuracy and efficiency of meningioma detection and segmen-
tation using MRI scans. The research findings underscore the effectiveness of employing
convolutional neural networks combined with advanced classification algorithms, collec-
tively contributing to more precise medical diagnoses and potentially better healthcare
outcomes. The utilisation of diverse CNN architectures, as elaborated in Table 1, such as
AlexNet, VGG16, GoogleNet, ResNet18, and ResNet50, highlights their distinct advantages
in handling complex image processing tasks. Each architecture’s specific features—from
local response normalization in AlexNet to the innovative skip connections in the ResNet
series—play a crucial role in improving the model’s ability to classify accurately and seg-
ment brain tumours. These architectures are carefully chosen based on their performance
metrics and suitability for the intricate requirements of medical image analysis, as reflected
in their error rates and computational efficiency.
The outcomes of this study, detailed in Table 3, reveal high accuracy, sensitivity,
specificity, and Dice coefficient indices for the proposed method, particularly notable
when applied to the BRATS 2018 dataset. These results validate the model’s robustness
and potential for clinical applications, providing a solid foundation for surgical planning
and targeted therapies. The comparative performance analysis in Table 4 emphasises the
superiority of our proposed model over existing methods. This superiority is a testament
to the careful integration of CNNs with MobileNetV2 and Multi-Support Vector Machine
classifiers, significantly enhancing tumour classification accuracy.
However, the study also acknowledges challenges such as dataset bias, class imbal-
ance, and the required high computational resources. These challenges necessitate refining
segmentation and training techniques to ensure the models remain effective across diverse
and realistic clinical scenarios. The research confirms that the strategic application of ma-
chine learning techniques, specifically tailored CNN architectures and ensemble learning
strategies, can significantly improve the performance of MRI-based brain tumour classi-
fication systems. This advancement paves the way for more accurate and personalised
diagnostic tools and promises to enhance therapeutic strategies for patients with brain
tumours. Future work will focus on expanding these techniques to other forms of medical
imaging and continuing to refine the models to handle the variability and complexity of
real-world clinical data.
Author Contributions: Conceptualization, M.A., A.Z. and A.B.; methodology, A.B., A.Z. and B.M.;
software, M.A.; validation, M.A., A.B. and A.Z.; formal analysis, R.Y.M.L.; investigation, M.A., A.B.,
B.M., R.Y.M.L. and A.Z.; resources, M.A. and A.B.; data curation, M.A. and R.Y.M.L.; writing—
original draft preparation, M.A., A.B. and A.Z.; writing—review and editing, M.A., A.Z. and R.Y.M.L.;
visualisation, M.A., A.B. and R.Y.M.L.; supervision, B.M. and A.Z.; project administration, B.M.;
funding acquisition, B.M. and A.Z. All authors have read and agreed to the published version of the
manuscript.
Funding: The research was supported by the Science Committee of the Ministry of Science and
Higher Education of the Republic of Kazakhstan (Grant no. of the research fund: AP14869848).
Institutional Review Board Statement: Not applicable.
Appl. Sci. 2024, 14, 7281 18 of 20
References
1. Vidyarthi, A.; Agarwal, R.; Gupta, D.; Sharma, R.; Draheim, D.; Tiwari, P. Machine learning assisted methodology for multiclass
classification of malignant brain tumors. IEEE Access 2022, 10, 50624–50640. [CrossRef]
2. Steinmetz, J.D.; Seeher, K.M.; Schiess, N.; Nichols, E.; Cao, B.; Servili, C.; Cavallera, V.; Cousin, E.; Hagins, H.; Moberg, M.E.; et al.
Global, regional, and national burden of disorders affecting the nervous system, 1990–2021: A systematic analysis for the Global
Burden of Disease Study 2021. Lancet Neurol. 2024, 23, 344–381. [CrossRef]
3. Shah, H.A.; Saeed, F.; Yun, S.; Park, J.H.; Paul, A.; Kang, J.M. A robust approach for brain tumor detection in magnetic resonance
images using finetuned efficientnet. IEEE Access 2022, 10, 65426–65438. [CrossRef]
4. Musa, U.I.; Mensah, Q.E.; Falowo, R. Intracranial-tumor detection and classification system using convnet and transfer learning.
Int. Res. J. Eng. Technol. 2022, 10, 120–127.
5. Ahmad, S.; Choudhury, P.K. On the performance of deep transfer learning networks for brain tumor detection using MR images.
IEEE Access 2022, 10, 59099–59114. [CrossRef]
6. Islami, F.; Ward, E.M.; Sung, H.; Cronin, K.A.; Tangka, F.K.; Sherman, R.L.; Zhao, J.; Anderson, R.N.; Henley, S.J.; Yabroff, K.R.;
et al. Annual report to the nation on the status of cancer, part 1: National cancer statistics. JNCI J. Natl. Cancer Inst. 2021, 113,
1648–1669. [CrossRef]
7. Rizwan, M.; Shabbir, A.; Javed, A.R.; Shabbir, M.; Baker, T.; Obe, D.A.J. Brain tumor and glioma grade classification using
Gaussian convolutional neural network. IEEE Access 2022, 10, 29731–29740. [CrossRef]
8. Johnson, D.R.; O’Neill, B.P. Glioblastoma survival in the United States before and during the temozolomide era. J. Neuro-Oncol.
2012, 107, 359–364. [CrossRef] [PubMed]
9. Rehman, A.; Naz, S.; Razzak, M.I.; Akram, F.; Imran, M. A deep learning-based framework for automatic brain tumors
classification using transfer learning. Circuits Syst. Signal Process. 2020, 39, 757–775. [CrossRef]
10. Fernando, T.; Gammulle, H.; Denman, S.; Sridharan, S.; Fookes, C. Deep learning for medical anomaly detection—A survey. ACM
Comput. Surv. (CSUR) 2021, 54, 1–37. [CrossRef]
11. Somasundaram, K.; Genish, T.; Kalaiselvi, T. An atlas based approach to segment the hippocampus from MRI of human head
scans for the diagnosis of Alzheimers disease. Int. J. Comput. Intell. Inform. 2015, 5.
12. Xie, Y.; Zaccagna, F.; Rundo, L.; Testa, C.; Agati, R.; Lodi, R.; Tonon, C. Convolutional neural network techniques for brain tumor
classification (from 2015 to 2022): Review, challenges, and future perspectives. Diagnostics 2022, 12, 1850. [CrossRef]
13. Louis, D.N.; Perry, A.; Reifenberger, G.; Von Deimling, A.; Figarella-Branger, D.; Cavenee, W.K.; Ohgaki, H.; Wiestler, O.D.;
Kleihues, P.; Ellison, D.W. The 2016 World Health Organization classification of tumors of the central nervous system: A summary.
Acta Neuropathol. 2016, 131, 803–820. [CrossRef]
14. Montaha, S.; Azam, S.; Rafid, A.R.H.; Hasan, M.Z.; Karim, A.; Islam, A. Timedistributed-cnn-lstm: A hybrid approach combining
cnn and lstm to classify brain tumor on 3d mri scans performing ablation study. IEEE Access 2022, 10, 60039–60059. [CrossRef]
15. Lundervold, A.S.; Lundervold, A. An overview of deep learning in medical imaging focusing on MRI. Z. Für Med. Phys. 2019, 29,
102–127. [CrossRef]
16. Rundo, L.; Militello, C.; Vitabile, S.; Russo, G.; Pisciotta, P.; Marletta, F.; Ippolito, M.; D’Arrigo, C.; Midiri, M.; Gilardi, M.C.
Semi-automatic brain lesion segmentation in Gamma Knife treatments using an unsupervised fuzzy c-means clustering technique.
In Proceedings of the Advances in Neural Networks: Computational Intelligence for ICT. International Workshop on Neural
Networks, WIRN 2015, Vietri sul Mare, Italy, 20–22 May 2015; pp. 15–26.
17. Liang, N.Y.; Huang, G.B.; Saratchandran, P.; Sundararajan, N. A fast and accurate online sequential learning algorithm for
feedforward networks. IEEE Trans. Neural Netw. 2006, 17, 1411–1423. [CrossRef]
18. Nie, D.; Wang, L.; Gao, Y.; Lian, J.; Shen, D. STRAINet: Spatially varying sTochastic residual AdversarIal networks for MRI pelvic
organ segmentation. IEEE Trans. Neural Netw. Learn. Syst. 2018, 30, 1552–1564. [CrossRef] [PubMed]
19. Hollon, T.C.; Pandian, B.; Adapa, A.R.; Urias, E.; Save, A.V.; Khalsa, S.S.S.; Eichberg, D.G.; D’Amico, R.S.; Farooq, Z.U.; Lewis, S.
Near real-time intraoperative brain tumor diagnosis using stimulated Raman histology and deep neural networks. Nat. Med.
2020, 26, 52–58. [CrossRef]
20. Natarajan, A.; Kumarasamy, S. Efficient segmentation of brain tumor using FL-SNM with a metaheuristic approach to optimization.
J. Med. Syst. 2019, 43, 25. [CrossRef] [PubMed]
21. Muhammad, K.; Khan, S.; Del Ser, J.; De Albuquerque, V.H.C. Deep learning for multigrade brain tumor classification in smart
healthcare systems: A prospective survey. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 507–522. [CrossRef] [PubMed]
22. Mohan, G.; Subashini, M.M. MRI based medical image analysis: Survey on brain tumor grade classification. Biomed. Signal
Process. Control 2018, 39, 139–161. [CrossRef]
Appl. Sci. 2024, 14, 7281 19 of 20
23. Hussain, L.; Saeed, S.; Awan, I.A.; Idris, A.; Nadeem, M.S.A.; Chaudhry, Q.U.A. Detecting brain tumor using machines learning
techniques based on different features extracting strategies. Curr. Med. Imaging 2019, 15, 595–606. [CrossRef] [PubMed]
24. Gurbină, M.; Lascu, M.; Lascu, D. Tumor detection and classification of MRI brain image using different wavelet transforms
and support vector machines. In Proceedings of the 2019 42nd International Conference on Telecommunications and Signal
Processing (TSP), Budapest, Hungary, 1–3 July 2019; pp. 505–508.
25. Sarkar, A.; Maniruzzaman, M.; Ahsan, M.S.; Ahmad, M.; Kadir, M.I.; Islam, S.T. Identification and classification of brain tumor
from MRI with feature extraction by support vector machine. In Proceedings of the 2020 International Conference for Emerging
Technology (INCET), Belgaum, India, 5–7 June 2020; pp. 1–4.
26. Sekhar, A.; Biswas, S.; Hazra, R.; Sunaniya, A.K.; Mukherjee, A.; Yang, L. Brain tumor classification using fine-tuned GoogLeNet
features and machine learning algorithms: IoMT enabled CAD system. IEEE J. Biomed. Health Inform. 2021, 26, 983–991. [CrossRef]
[PubMed]
27. Ramdlon, R.H.; Kusumaningtyas, E.M.; Karlita, T. Brain tumor classification using MRI images with K-nearest neighbor method.
In Proceedings of the 2019 International Electronics Symposium (IES), Surabaya, Indonesia, 27–28 September 2019; pp. 660–667.
28. Sathi, K.A.; Islam, M.S. Hybrid feature extraction based brain tumor classification using an artificial neural network. In
Proceedings of the 2020 IEEE 5th International Conference on Computing Communication and Automation (ICCCA), Greater
Noida, India, 30–31 October 2020; pp. 155–160.
29. Charan, K.S.; Chokkalingam, S.P.; Sundari, K.S. Efficiency of Decision tree algorithm for Brain Tumor MRI Images comparing
with SVM Algorithm. In Proceedings of the 2022 14th International Conference on Mathematics, Actuarial Science, Computer
Science and Statistics (MACS), Karachi, Pakistan, 12–13 November 2022; pp. 1–4.
30. Amran, G.A.; Alsharam, M.S.; Blajam, A.O.A.; Hasan, A.A.; Alfaifi, M.Y.; Amran, M.H.; Gumaei, A.; Eldin, S.M. Brain tumor
classification and detection using hybrid deep tumor network. Electronics 2022, 11, 3457. [CrossRef]
31. Sajjad, M.; Khan, S.; Muhammad, K.; Wu, W.; Ullah, A.; Baik, S.W. Multi-grade brain tumor classification using deep CNN with
extensive data augmentation. J. Comput. Sci. 2019, 30, 174–182. [CrossRef]
32. Ottom, M.A.; Rahman, H.A.; Dinov, I.D. Znet: Deep learning approach for 2D MRI brain tumor segmentation. IEEE J. Transl. Eng.
Health Med. 2022, 10, 1–8. [CrossRef]
33. Asif, S.; Yi, W.; Ain, Q.U.; Hou, J.; Yi, T.; Si, J. Improving effectiveness of different deep transfer learning-based models for
detecting brain tumors from MR images. IEEE Access 2022, 10, 34716–34730. [CrossRef]
34. Park, H.; Sjösund, L.L.; Yoo, Y.; Bang, J.; Kwak, N. Extremec3net: Extreme lightweight portrait segmentation networks using
advanced c3-modules. arXiv 2019, arXiv:1908.03093.
35. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017,
arXiv:1706.05587.
36. Chen, X.; Qi, D.; Shen, J. Boundary-aware network for fast and high-accuracy portrait segmentation. arXiv 2019, arXiv:1901.03814.
37. Toufiq, D.M.; Sagheer, A.M.; Veisi, H. A review on brain tumor classification in mri images. Turk. J. Comput. Math. Educ.
(TURCOMAT) 2021, 12, 1958–1969.
38. Hossain, M.F.; Islam, M.A.; Hussain, S.N.; Das, D.; Amin, R.; Alam, M.S. Brain Tumor Classification from MRI Images Using
Convolutional Neural Network. In Proceedings of the 2021 IEEE International Conference on Artificial Intelligence in Engineering
and Technology (IICAIET), Kota Kinabalu, Malaysia, 13–15 September 2021; pp. 1–6.
39. Saranya, N.; Renuka, D.K. Brain tumor classification using convolution neural network. In Journal of Physics: Conference Series;
IOP Publishing: Bristol, UK, 2021; Volume 1916, p. 012206.
40. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.-M.; Larochelle, H. Brain tumor
segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [CrossRef]
41. Bayoumi, E.; Khalaf, A.A.; Gharieb, R.R. Brain tumor automatic detection from MRI images using transfer learning model with
deep convolutional neural network. J. Adv. Eng. Trends 2021, 41, 19–30. [CrossRef]
42. Minarno, A.E.; Mandiri, M.H.C.; Munarko, Y.; Hariyady, H. Convolutional neural network with hyperparameter tuning for brain
tumor classification. Kinet. Game Technol. Inf. Syst. Comput. Netw. Comput. Electron. Control 2021, 4, 127–132. [CrossRef]
43. Wolpert, D.H. Stacked generalization. Neural Netw. 1992, 5, 241–259. [CrossRef]
44. Nigam, I.; Huang, C.; Ramanan, D. Ensemble knowledge transfer for semantic segmentation. In Proceedings of the 2018 IEEE
Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1499–1508.
45. Zhang, S.H.; Dong, X.; Li, H.; Li, R.; Yang, Y.L. PortraitNet: Real-time portrait segmentation network for mobile device. Comput.
Graph. 2019, 80, 104–113. [CrossRef]
46. Kim, Y.W.; Rose, J.I.; Krishna, A.V. Accuracy enhancement of portrait segmentation by ensembling deep learning models. In
Proceedings of the 2020 Fifth International Conference on Research in Computational Intelligence and Communication Networks
(ICRCICN), Bangalore, India, 26–27 November 2020; pp. 59–64.
47. Rohlfing, T.; Maurer, C.R., Jr. Shape-based averaging for combination of multiple segmentations. In Proceedings of the
International Conference on Medical Image Computing and Computer-Assisted Intervention, Palm Springs, CA, USA, 26–29
October 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 838–845.
48. Hansen, L.K.; Salamon, P. Neural network ensembles. IEEE Trans. Pattern Anal. Mach. Intell. 1990, 12, 993–1001. [CrossRef]
49. Singh, V.; Mukherjee, L.; Peng, J.; Xu, J. Ensemble clustering using semidefinite programming with applications. Mach. Learn.
2010, 79, 177–200. [CrossRef]
Appl. Sci. 2024, 14, 7281 20 of 20
50. Zhao, B.; Feng, J.; Wu, X.; Yan, S. A survey on deep learning-based fine-grained object classification and semantic segmentation.
Int. J. Autom. Comput. 2017, 14, 119–135. [CrossRef]
51. Narmatha, C.; Eljack, S.M.; Tuka, A.A.R.M.; Manimurugan, S.; Mustafa, M. A hybrid fuzzy brain-storm optimization algorithm
for the classification of brain tumor MRI images. J. Ambient. Intell. Humaniz. Comput. 2020, 1–9. [CrossRef]
52. Khan, M.A.; Ashraf, I.; Alhaisoni, M.; Damaševičius, R.; Scherer, R.; Rehman, A.; Bukhari, S.A.C. Multimodal brain tumor
classification using deep learning and robust feature selection: A machine learning application for radiologists. Diagnostics 2020,
10, 565. [CrossRef] [PubMed]
53. Cheng, J.; Huang, W.; Cao, S.; Yang, R.; Yang, W.; Yun, Z.; Wang, Z.; Feng, Q. Enhanced performance of brain tumor classification
via tumor region augmentation and partition. PLoS ONE 2015, 10, e0140381. [CrossRef] [PubMed]
54. Rokach, L. Ensemble-based classifiers. Artif. Intell. Rev. 2010, 33, 1–39. [CrossRef]
55. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016.
56. Amin, J.; Sharif, M.; Raza, M.; Saba, T.; Sial, R.; Shad, S.A. Brain tumor detection: A long short-term memory (LSTM)-based
learning model. Neural Comput. Appl. 2020, 32, 15965–15973. [CrossRef]
57. Muthukrishnan, R.; Radha, M. Edge detection techniques for image segmentation. Int. J. Comput. Sci. Inf. Technol. 2011, 3, 259.
[CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual
author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to
people or property resulting from any ideas, methods, instructions or products referred to in the content.