Brain Tumor Diagnosis Using Machine Learning, Convolutional Neural Networks, Capsule Neural Networks and Vision Transformers, Applied To MRI: A Survey
Brain Tumor Diagnosis Using Machine Learning, Convolutional Neural Networks, Capsule Neural Networks and Vision Transformers, Applied To MRI: A Survey
Brain Tumor Diagnosis Using Machine Learning, Convolutional Neural Networks, Capsule Neural Networks and Vision Transformers, Applied To MRI: A Survey
Imaging
Review
Brain Tumor Diagnosis Using Machine Learning,
Convolutional Neural Networks, Capsule Neural Networks and
Vision Transformers, Applied to MRI: A Survey
Andronicus A. Akinyelu 1,2, *, Fulvio Zaccagna 3,4 , James T. Grist 5,6,7,8 , Mauro Castelli 1,†
and Leonardo Rundo 9, *,†
often of lower grades [3]. Malignant brain tumors infiltrate the surrounding parenchyma
with variable aggressiveness. Glioblastoma (GBM) is the most common and aggressive
type of malignant brain tumor, commonly classified as a grade 4 CNS tumor with dismal
survival [3,4]. Brain tumors can also be categorized according to their origin into primary
and secondary brain tumors, with the former originating in the brain and the latter usually
in a distant site [3].
At present, diagnosis and management are based upon clinical and radiological
information. Magnetic Resonance Imaging (MRI) is the mainstay for the assessment of
patients with brain tumors [5], although conventional imaging has significant limitations
in evaluating tumor extent, predicting grade, and assessing treatment response [6]. Novel
acquisition techniques are under development to improve lesion characterization, therapy
assessment and management [7]; however, novel approaches to image analysis have been
gaining traction thanks to the wealth of information possessed by radiological images [8].
In this context, brain tumor classification and segmentation have become pivotal
for image analysis. Brain tumor classification can be performed with different methods,
including manual classification and computer-aided classification. Manual classification
of brain tumors is very time-consuming [9], and prone to error [10]. However, manual
classification cannot be ignored, as it is still the reference standard both for clinical care and
used as a comparison for other techniques [11].
Computer-aided diagnosis has proven to be useful in supporting medical practi-
tioners [12]. It can be performed using different techniques including classical machine
learning (ML) [13,14] (such as clustering [15]), Convolutional Neural Network (CNN)
approaches [16,17], recently considering also capsule neural networks (CapsNets) [18,19],
and Vision Transformers (ViTs) [20].
This paper presents a survey on brain tumor segmentation and classification tech-
niques, with a particular emphasis on ML-based, CNN-based, CapsNet-based, and ViT-
based techniques.
The primary objective of this survey is to highlight the current state-of-the-art of
ML-based, CNN-based, CapsNet-based, and ViT-based brain tumor segmentation and
classification methods. This paper also highlights recent achievements, relevant research
challenges, and future research directions. This survey is different from the existing
surveys [9,21–23] in the following ways:
1. Most review papers considered one or two types of ML-based techniques, while
we included four types of brain tumor segmentation and classification techniques:
classical ML algorithms, CNN-based techniques, CapsNet-based techniques, and ViT-
based techniques. This survey summarizes the current state-of-the-art in brain tumor
segmentation, grade estimation, and classification. It outlines the advances made
in ML-based, CNN-based, CapsNet-based, and ViT-based brain tumor diagnosis
between 2019 and 2022.
2. Most of the previous studies [9,21–23] presented an overview of either CNN-based
techniques, ML-based techniques, or both ML- and CNN-based approaches. In
addition to CNN-based and ML-based techniques, we present a summary of CapsNet-
based, and ViT-based brain tumor segmentation and classification techniques. Cap-
sNet is one of the state-of-the-art techniques for brain tumor segmentation and classi-
fication. The findings of this survey show that CapsNet outperformed CNN-based
brain tumor diagnosis techniques, as they require significantly less training data
compared to CNNs. Moreover, CapsNets are very robust to different rotations and
image transformations. Furthermore, unlike CNN, ViT-based models can effectively
model local and global feature information, which is very critical for accurate brain
tumor segmentation and classification.
3. Some review papers did not provide a comprehensive discussion and pertinent
findings. This survey presents significant findings and a comprehensive discussion
of approaches for segmenting and classifying brain tumors. We also identify open
J. Imaging 2022, 8, 205 3 of 40
problems and significant areas for future research, which should be beneficial to
interested scholars and medical practitioners.
The remainder of this manuscript is structured as follows: Section 2 introduces
some fundamental concepts necessary for an accurate understanding of this manuscript.
Section 3 introduces the material and methods used for this survey. Section 4 provides an
in-depth review of recently published ML, CNN, CapsNet, and ViT-based brain tumor
classification and segmentation techniques. Section 4 also highlights the core contributions
and performance of various studies. Section 5 discusses several open research problems
and critical issues related to brain tumor segmentation and classification. Section 5 also
presents a performance analysis of brain tumor diagnosis techniques, as well as details
on several popular datasets. Section 6 concludes the paper with some final thoughts and
potential research directions.
2. Background
2.1. Brain Tumors and Magnetic Resonance Imaging
Brain tumors can be intra-axial (e.g., gliomas) or extra-axial (e.g., meningiomas or
pituitary adenomas). Intra-axial brain tumors are particularly difficult to treat, especially at
advanced stages, when they are usually discovered due to the symptoms caused by the
mass effect on the surrounding brain [5]. Treatment failure can be due to several factors,
including the limited capacity of current imaging modalities to identify the boundaries
of the lesion within the normal appearing brain parenchyma [24]. Hence, more advanced
imaging techniques for assessment of brain tumors and surrounding structures is critical
to improve overall management [5,24,25]. Extra-axial brain cancers also require special
attention, as these tumors (such as pituitary adenoma and meningiomas) can result in
complications and long-term impairment [26–28].
MRI is the workhorse for brain tumor imaging in clinical practice providing structural,
microstructural, functional, and metabolic information [29]. Moreover, novel advanced
imaging techniques are continuously developed to improve identification, characterization,
and response assessment of brain tumors [6]. Hence, many artificial intelligence (AI)
applications in brain tumor imaging have been based on MRI. For more information on
brain tumors, kindly refer to [5].
capsule layer is typically the first one, followed by an undetermined number of capsule
layers. The capsule layer is followed by the class capsule layer. The convolutional layer
is used to extract features, which are then transmitted to the primary capsule layer. The
primary capsule performs a series of operations and transmits the resulting feature map to
the digit capsule. Typically, the digit capsule is composed of a n × m weight matrix, where
n denotes the number of classes and m the size of each digit capsule. The digit capsule is
used to classify the input image before it is fed into the decoder. The decoder consists of
three fully connected layers that are used to reconstruct or decode the selected digit capsule
into an image.
Figure 2. Overview of a Vision Transformers (ViT) model. The image is partitioned into N small
patches (e.g., 9 patches). Each of the image patches contains n × n pixels (e.g., 16 × 16 pixels). After
partitioning, each image patch is flattened: each of the flattened image patches is fed into a linear
projection layer to obtain a lower-dimensional linear embedding. Moreover, positional embeddings
are added to the sequence of image patches to ensure that each image keeps its positional information.
The input sequences and position embedded sequence are fed into a standard transformer encoder for
training. The training can be conducted by an MLP or CNN head stacked on top of the transformer.
The “*” symbol refers to an additional learnable (class) embedding that is appended to the sequence
based on the position of the image patch. This class embedding is used to predict the class of an input
image after self-attention updates it.
Figure 3. General scheme of a capsule neural network (CapsNet). A CapsNet is a three-layer network
composed of convolutional, primary capsule, and class capsule layers. The primary capsule layer is
typically the first one, followed by an undetermined number of capsule layers. The capsule layer is
followed by the class capsule layer. The convolutional layer is used to extract features, which are
then transmitted to the primary capsule layer. The primary capsule performs a series of operations
and transmits the resulting feature map to the digit capsule.
J. Imaging 2022, 8, 205 6 of 40
CapsNet can recognize spatial and hierarchical relationships among objects in im-
ages [40]. They are resistant to rotation and image transformations [38]. Additionally, as
shown in [41], CapsNet requires substantially less training data than CNN [18]. Moreover,
results reported in the literature [41] show that CapsNet has the potential to improve the
accuracy of CNN-based brain tumor diagnosis using a very small number of network
parameters [42].
It is worth noting that the pooling operation in CNNs makes it robust to small input
transformation. However, for CNN to perform well, it must be trained on augmented
data in terms of scale, rotation, and varying perspective. Despite this, results reported in
the literature [39,43] indicate that, in some cases, CapsNet performs comparably to CNN
models trained on augmented datasets. CapsNet does not need to be trained on large-scale
or augmented data to produce very good results. This makes it a suitable model for medical
image datasets, which are typically small. For more information on CapsNet, please refer
to [38].
Figure 4. Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) diagram
of the proposed review on AI applications to brain tumor MRI.
J. Imaging 2022, 8, 205 7 of 40
Figure 5. (a) percentage of reviewed articles; (b) number of reviewed articles and their
publication year.
3.1. Datasets
The Medical Image Computing and Computer Assisted Intervention (MICCAI) Society
has funded numerous events and open challenges over the years to stimulate the develop-
ment of DL tools and medical devices for computer-aided diagnosis. Most studies used
the datasets provided by the MICCAI Society to evaluate the efficiency of their techniques.
Details of the other datasets are also shown in Table 1. As shown in the table, most of the
benchmark datasets are small, making it challenging to build DL models from end-to-end.
Table 1. Cont.
4. Literature Survey
Brain tumors can be located anywhere in the human brain and assume virtually
any shape, size, or contrast (dissimilarity) [10]. This shows that ML-based solutions that
can effectively and automatically classify and segment brain tumors are needed. The
J. Imaging 2022, 8, 205 9 of 40
introduction of powerful computing devices and lower hardware prices have prompted
the scientific community to develop numerous techniques for brain tumor segmentation
and classification. Some of the techniques were designed with classical ML algorithms,
while other techniques were designed with CNN algorithms and CapsNet. This section
presents a review of ML-based, CNN-based, CapsNet-based, and ViT-based techniques for
brain tumor segmentation and classification. It is expected that these techniques will assist
medical practitioners in improving the accuracy and consistency of diagnosis.
4.1.1. Brain Tumor Classification and Segmentation Using Hybrid Texture-Based Features
The texture of an image is an important feature that can be used to identify different
regions of interest. The texture of a region in an image is determined by the distribution
of Gray levels across the image pixels [15]. Jena et al. [15] proposed a brain tumor classi-
fication and segmentation technique using texture features and multiple ML algorithms.
The technique is divided into two stages: tumor classification and tumor segmentation.
In the tumor classification stage, the MRI scans are pre-processed, and texture features
are extracted from the images using different texture extraction techniques. The following
texture-based features were explored in the study: first-order statistical feature [76], Gray-
level co-occurrence matrix (GLCM) feature [77], Gray-level run length matrix (GLRLM)
feature [78], Histogram-oriented gradient (HOG) feature [79], Local binary patterns (LBP)
feature [80], Cross-diagonal texture matrix (CDTM) feature [81], and simplified texture
spectrum feature [82]. All the features were extracted from 100 tumor and 100 non-tumor
images. The extracted features were combined to form a feature vector matrix of size
200 × 471. Subsequently, the feature vector matrix was used to train five ML algorithms:
SVM, k-NN, binary decision trees, RF, and ensemble methods. The ensemble methods
consist of seven algorithms, namely: Adaboost, Gentleboost, Logitboost, LPboost, Robust-
boost, RUSboost, and Totalboost. After training, the tumorous images were identified
and used as input to a hybrid tumor segmentation technique designed in the study. The
hybrid technique consists of k-NN and fuzzy C-means clustering algorithms. The hybrid
technique was used to segment the tumor regions in the images, and it was evaluated on
two datasets based on the following performance metrics: average Dice similarity coeffi-
cient (DSC), average Jaccard similarity coefficient, and average accuracy. The dataset used
to evaluate the model include: BraTS2017, BraTS2019, and the Cancer Imaging Archive
(TCIA). Experiments show that the ensemble methods produced the best result, achieving a
classification accuracy of 96.98% and 97.01% for BraTS2017 + TCIA and BraTS2019 + TCIA,
respectively. RF produced the second-best result, achieving an accuracy of 96.5% and
96.99% for BraTS2017 + TCIA and BraTS2019 + TCIA, respectively. The results also show
that the segmentation technique produced a Dice similarity score and accuracy of 90.16%
and 98.4%, respectively for BraTS2017.
recall of 96.02% and 97.00% for glioma, respectively, using softmax activation function. The
performance of the model was improved by over 2.5% when the SVM classifier was used.
It achieved precision and specificity of 98.76% and 98.93% for glioma, respectively. The
performance of GoogleNet was also improved by over 2.3% when the k-NN classifier was
used. It produced precision and specificity of 98.41% and 98.63% for glioma, respectively.
This shows that features extracted from pre-trained CNN models can be used to build
effective ML-based classifiers.
4.1.3. Brain Tumor Classification Using Ensemble of Deep Features and ML Classifiers
Kang et al. [84] proposed a method for brain tumor classification using an ensemble
of deep features. The technique consists of three stages. In the first stage, input images
are pre-processed, and more images are generated using data augmentation. The pre-
processed images are then used as input to 13 pre-trained CNN models, namely: ResNet-
50 [85], ResNet101 [85], DenseNet-121 [86], DenseNet-169 [86], VGG-16 [87], VGG-19 [87],
AlexNet [88], Inception-v3 [89], ResNeXt-50 [90], ResNeXt-101 [90], ShuffleNet-v2 [91],
MobileNet-v2 [92], and MnasNet [93]. The pre-trained CNN models are used to extract
features from the images. In particular, the features are extracted from the fully connected
layers of the pre-trained models. The extracted features are used to train nine ML classifiers,
namely: Gaussian Naïve Bayes, Extreme Learning Machine (ELM), Adaptive Boosting
(AdaBoost), k-NN, RF, SVM and neural networks with a fully connected layer. Moreover,
the pre-trained models that produced the three best results are identified, and the extracted
features from the three best pre-trained models are combined into one sequence. Finally, the
combined features are used to train the nine ML classifiers. The technique was evaluated
on three brain MRI datasets downloaded from Kaggle websites. Results showed that
DenseNet-169, Inception-v3, and ResNeXt-50 produced the best features, achieving an
accuracy of 96.08%, 92.16%, and 94.12%, respectively.
study. The technique was tested on the FigShare dataset. The dataset consists of 3064 T1 w
post GBCA MR images (from 233 patients) belonging to three classes of brain tumors:
glioma, meningioma, and pituitary tumor. The proposed technique achieved an accuracy
of 95.82% when trained on the 10 feature vectors extracted from the fully connected layer.
The accuracy reduced to 93.83% when it was trained on the 3136-feature set extracted from
the fifth convolution layer. This shows that SVM models trained on small-scale feature sets
have the potential to produce better results than those trained on large-scale feature sets.
CNN produced an accuracy of 94.26% when it was trained as a standalone classifier.
The summary of all the ML-based algorithms is presented in Table 2. Furthermore, the
building blocks of a typical ML-based brain tumor segmentation and classification model
are shown in Figure 6.
Classes
Ref. Year Method Main Highlight Dataset Performance
Considered
DenseNet-169,
Inception-v3, and
13 pre-trained Concatenated three deep
Three brain ResNeXt-50 produced
CNN models and Normal and features from pre-trained
[84] 2021 MRI datasets the best deep features,
nine ML tumor images CNN models and trained
[56–58] achieving an accuracy of
classifiers nine ML classifiers.
96.08%, 92.16% and
94.12%, respectively.
SVM, k-NN, Extracted features from
SVM and k-NN
and modified Glioma, modified GoogleNet
produced a specificity of
[14] 2021 GoogleNet meningioma, pre-trained architecture CE-MRI dataset
98.93% and 98.63%,
pre-trained and pituitary. and used it to train SVM
respectively.
architecture and k-NN
Developed multiple ML
Classification accuracy of
models on six
SVM, k-NN, 96.98% and 97.01% for
FLAIR, T1 w, T1 w texture-based features.
binary decision BraTS 2017 and BraTS 2017 and 2019,
[15] 2022 post GBCA, and Applied a hybrid of
tree, RF, ensemble BraTS 2019 respectively. DSC and
T2 w k-NN and C-means
methods. accuracy of 90.16% and
clustering for tumor
98.4%, respectively.
segmentation.
NB, RNN, bat Designed hybrid
and lion techniques model using
Tumor and Classification accuracy
[94] 2022 optimization, the combination of TCIA
normal images of 98.61%.
PCA, ICA, and metaheuristics and ML
cuckoo search. algorithms.
Proposed a hybrid
Glioma,
technique using FigShare
[13] 2021 SVM and CNN meningioma, and 95.82% accuracy
CNN-based features dataset
pituitary tumor
and SVM.
Figure 6. Building blocks of a typical ML-based brain tumor classification and segmentation model.
J. Imaging 2022, 8, 205 12 of 40
4.2.3. Brain Tumor Classification for Multi-Class Brain Tumor Image Using Block-Wise
Fine-Tuning
Swati et al. [97] introduced a block-wise fine-tuning technique for multi-class brain
tumor MR images. They considered three classes of abnormal brain tumors: (i) glioma,
(ii) meningioma, and (iii) pituitary tumors. In the study, the VGG-19 pre-trained network
was used to initialize the weights of the CNN model designed in the study. Additionally,
rather than fine-tuning the VGG-19 pre-trained network all at once, the previous layers of
the CNN were fine-tuned using a block-wise approach. In this approach, the 19 layers of
the VGG-19 architecture were divided into six blocks. The first and second blocks consist of
two convolutional layers and one pooling layer. The third, fourth, and fifth blocks consist
of four convolutional layers and one pooling layer. The last block consists of two fully
J. Imaging 2022, 8, 205 13 of 40
connected layers and one output layer. The block-wise fine-tuning is initiated by fine-
tuning the layers in the last block and freezing the weights of the layers in the other blocks.
A similar approach is taken until the whole six blocks were fine-tuned. The block-wise
fine-tuning technique was evaluated on the CE-MRI dataset [56]. Results showed that the
technique achieved a classification accuracy of 94.82% using the 5-fold cross-validation.
study that produced the best result achieved an average classification accuracy, sensitivity,
and specificity of 88.82%, 81.81%, and 92.17%, respectively.
The flow for most of the proposed CNN-based brain tumor classification technique is
shown in Figure 7. The summary of all the CNN-based brain classification algorithms is
presented in Table 3.
Classes
Ref. Year Method Main Highlight Dataset Performance
Considered
Classification accuracy
of 95.5%, 92.66%,
Adopted four data 87.77%, and 86.71%.
augmentation techniques. for Grades I–IV,
InputCascadeCNN
Also used Radiopedia respectively on
+ data
[16] 2019 Grades I–IV InputCascadeNN and Brain radiopedia dataset.
augmentation +
architecture for data tumor dataset Sensitivity and
VGG-19
augmentation and VGG-19 specificity 88.41% and
for fine-tuning. 96.12%, respectively,
on the brain
tumor dataset.
Applied user-defined
hyperactive values and a Classification accuracy,
Differential deep differential operator to TUCMD sensitivity, and
T1 w, T2 w and
[95] 2021 CNN + Data generate feature maps for (17,600 MR specificity of 99.25%,
FLAIR images
augmentation CNN. Proposed several brain images) 95.89%, and 93.75%,
data augmentation respectively
techniques.
Introduced a block-wise CE-MRI [56]
Glioma,
fine-tuning technique for 3064 images Classification accuracy:
[97] 2019 meningioma, and VGG-19
multi-class brain tumor from 233 94.82%
pituitary tumor
MRI image. patients.
Proposed a multi-scale 3D
CNN architecture for
grade classification BraTS2018:
capable of learning both Training-209
local and global brain HGG and 75
tumor features. Applied LGG from Classification
[62] 2020 LGG and HGG 3D CNN
two image pre-processing 284 patients. accuracy: 96.49%
techniques for reducing Validation:
thermal noise and 67 mixed
scanner-related artifacts in grades.
brain MRI. Used data
augmentation.
Average classification
GANs was used to TCGA-GBM
T1 w, T1 w post accuracy, sensitivity,
U-Net architecture, generate synthetic images [99] and
[75] 2020 GBCA, T2 w, and specificity of
GANs for four modalities of MRI: TCGA-LGG
FLAIR 88.82%, 81.81%, and
T1 , T1e, T2 , FLAIR. [99].
92.17%, respectively
Used BAT algorithm to
optimize the loss function Accuracy, recall (or
Complete, core, Custom CNN
of CNN. In addition, used sensitivity), and
[17] 2019 and enhancing architecture, Bat BraTS2015
skull stripping and image precision of 92%, 87%,
tumors algorithm
enhancement techniques and 90%, respectively
for image pre-processing.
J. Imaging 2022, 8, 205 15 of 40
4.3.1. Enhanced CNN and Bat Algorithm for Brain Tumor Classification
Thaha et al. [17] introduced a strategy for segmenting brain tumors using an enhanced
CNN coupled with the Bat algorithm. The Bat algorithm was used to optimize the loss
function (cross-entropy). The loss function is adjusted with the goal of improving the
segmentation accuracy. In the experiments, MR images were preprocessed using a skull
stripping and image enhancement technique. The authors added a central-point-enhanced
layer in the study to address misclassification concerns. The approach was evaluated, and
the results indicated that the improved CNN achieved a precision, recall (or sensitivity),
and accuracy of 87%, 90%, and 92%, respectively, compared to the standard CNN, which
achieved 82%, 85%, and 89%, respectively.
near the tumor, and other healthy patches. Based on patient data, the authors further
categorized the tumor images into six classes: 0 (normal), 1 (ground truth region based on
a combination of classes 2–5), 2 (enhancing region), 3 (necrotic region), 4 (T1-abnormality),
and 5 (T2-abnormality). The models were evaluated using DSC and the Hausdorff distance.
The performance metric was utilized to compare the ground truth with the automatic
segmentation. The performance of several techniques was compared using 10-fold cross-
validation method. The results demonstrated that CNN models achieved the following
Dice coefficients when it was evaluated on different institutional data: 0.68 ± 0.19 and
0.59 ± 0.19, respectively. The performance declined when CNN was trained on the same
institutional data. It produced a Dice coefficient of 0.72 ± 0.17 and 0.76 ± 0.12, respec-
tively. The authors stated that further research is necessary to determine the cause of this
performance decrease.
output layer takes care of calibrating the output probabilities (thanks to the re-training that
was performed on the output layer).
The un-cascaded and cascaded models were subjected to a variety of experiments.
Results from the un-cascaded models indicated that the double-pathway models generated
the best results. The results also indicated that adding a second training phase considerably
improved the model’s performance. Additionally, training the local and global routes con-
currently produced better outcomes than training each pathway separately and averaging
the results. The results from the cascaded models showed that the first cascade architecture
produced the best Dice similarity of 0.88. The models increased the Dice similarity scores
for all tumor regions.
98%, 99%, and 99% for BraTS2018, BraTS2019, and BraTS2020, respectively. The results also
show that the YOLOv2-inception-v3 model produced a mean average precision (mAP) of
0.98, 0.99, and 1.00 for BraTS2018, BraTS2019, and BraTS2020, respectively.
The general workflow for CNN-based brain tumor segmentation is shown in Figure 8.
Moreover, Table 4 shows the summary of all the CNN-based brain tumor segmentation
techniques surveyed in this study.
Classes
Ref. Year Method Main Highlight Dataset Performance
Considered
Designed a patching-based DSC-Same institution:
Metastasis, technique for brain tumor 0.72 ± 0.17 and
[73] 2018 meningiomas CNN segmentation. Evaluated TCIA 0.76 ± 0.12. Different
gliomas the impact of Institution: 0.68 ± 0.19
inter-institutional dataset. and 0.59 ± 0.19
Designed a two-pathway
architecture for capturing
Necrosis, edema,
[10] 2017 CNN global and local features. BraTS2013 DSC: 0.88
non-ET, ET.
Also designed three
cascade architectures.
Developed a triple
Triple CNN network architecture to
DSC: 0.90, 0.82, and
architecture for simplify the multiclass BraT2018
[74] 2020 WT, TC, and ET 0.79 for WT, TC, and
multi-class segmentation problem to a dataset
ET, respectively
segmentation. single binary
segmentation problem.
Modified U-Net
architecture, Data Developed an DSC, sensitivity, and
T1 w, T1 w post
augmentation, batch encoder-decoder specificity of 0.814,
[41] 2020 GBCA, T2 w, and BraTS 2019
normalization using architecture for brain 0.783, 0.999,
FLAIR
the N3 bias tumor segmentation. respectively.
correction tool [100].
Inception-v3, NSGA, Designed a CNN-based Classification accuracy
LDA, SVM, k-NN, hybrid framework for of 98%, 99%, and 99%
BraTS 2018,
softmax, CART, tumor enhancement, for BraTS2018,
[101] 2021 HGG and LGG BraTS2019,
YOLOv2, and feature extraction and BraTS2019, and
and BraTS2020
McCulloch’s Kapur selection, localization, and BraTS2020,
entropy tumor segmentation respectively.
Used BAT algorithm to
optimize the loss function Accuracy, recall (or
Complete, core, Custom CNN
of CNN. In addition, used sensitivity), and
[17] 2019 and enhancing architecture, Bat BraTS2015
skull stripping and image precision of 92%, 87%,
tumors algorithm
enhancement techniques and 90%, respectively
for image pre-processing.
J. Imaging 2022, 8, 205 19 of 40
4.4.3. Brain Tumor Segmentation Using Transformers Encoders and CNN Decoder
Some of the ViT-based techniques used CNN for encoding. Ali et al. [105] intro-
duced a ViT-based technique for 3D image segmentation that uses ViT for encoding. The
transformer encoder directly accepts a one-dimensional sequence of 3D image patches as
inputs. The image sequence was created by dividing the input images into non-overlapping
patches of uniform size. For each layer in the transformer model, the extracted patches are
projected into an N-dimensional embedding space. Moreover, a one-dimensional positional
embedding was added to the embedding space to preserve the spatial information of the ex-
tracted patches. Additionally, the output was routed through a stack of transformer blocks
for encoding. During the encoding, sequence representations at different resolutions are
extracted from the transformer and reshaped into a tensor of a specific size. Convolutional
and normalization layers are used to project the reshaped tensors from the embedding space
into the input space at each image resolution. The output of the transformer’s final layer is
fed into a CNN decoder. The CNN decoder employs deconvolutional layers to increase
the resolution of the feature maps by a user-defined factor. All of the transformer blocks’
feature maps are concatenated and fed into another set of convolutional layers. The feature
map from the convolutional layers is fed into a deconvolutional layer for upsampling. This
procedure is repeated for all the layers. The final output is fed into a 3D convolutional
layer with a softmax activation function to construct the predicted segmented image. On
the Medical Segmentation Decathlon (MSD) dataset [59], the technique produced a DSC of
0.789, 0.585, and 0.761 for WT, ET, and TC, respectively.
transformer encoder blocks. The 3D patch partitioning layer is used to divide the three-
dimensional input volume into non-overlapping three-dimensional patches. The authors
proposed two techniques for partitioning volumetric data into non-overlapping patches.
The partitioned 3D tokens are fed into the linear embedding layer, which converts each
token to a vector space of N dimensions. Additionally, the 3D patch merging layer is used
to generate feature representation hierarchies. The encoder module’s output is passed to the
decoder module. The module consists of a transformer encoder block, a 3D patch expand-
ing layer, and a classifier. The decoder block is used to compute cross and self-attention
independently. Self-attention is used to exploit global interactions between encoder fea-
tures, whereas cross-attention is used to filter out local or non-semantic features. The
3D patch expanding layer is used to transform the encoded feature maps into a format
compatible with the input image’s spatial resolution. The classifier layers are composed
of a single 3D convolutional layer that is used to map the feature representations to the
specified segmentation classes. The technique was evaluated on the BraTS2021 dataset,
and it achieved a DSC of 85.59%, 87.41%, and 91.20% for ET, TC, and WT, respectively.
The workflow for a ViT-based brain tumor segmentation technique is shown in
Figure 9. Furthermore, the summary for the ViT-based techniques reviewed in this study is
shown in Table 5.
Classes
Ref. Year Method Dataset Main Highlight Performance
Considered
BraTS2020-DSC of
90.09%, 78.73%, and
81.73% for WT, ET, and
TC, respectively.
Developed a
BraTS2019–DSC of
Transformers and BraTS2019 and transformer-based
[35] 2021 ET, WT, EC 90%, 78.93%, and
3D CNN BraTS2020 network for 3D brain
81.84% for WT, ET, and
tumor segmentation.
TC, respectively DSC
of 90%, 78.93%, and
81.84% for WT, ET, and
TC, respectively
Developed a technique for
DSC of 0.891, 0.933,
Swin transformers multi-modal brain tumor
[32] 2022 ET, WT, and TC BraTS2021 and 0.917 for ET, WT,
and CNN images using Swin
and TC, respectively.
transformers and CNN.
Developed a segmentation
DSC of 0.789, 0.585„
Transformers and technique for multi-modal
[105] 2022 WT, ET, and TC MSD dataset and 0.761 for WT, ET,
CNN brain tumor image using
and TC, respectively.
transformers and CNN.
Designed a
CNN-transformer DSC of 0.823, 0.908,
Transformers and
[104] 2021 WT, ET, and TC BraTS2021 technique for multi-modal and 0.839 for ET, WT,
3D CNN
brain MRI scan and TC, respectively.
segmentation.
Developed a U-Net shaped
encoder-decoder technique
using only transformers.
The transformer encoder DSC of 85.59%, 87.41%,
Transformers and
[107] 2021 WT, ET, and TC BraTS2021 can capture local and and 91.20% for ET, TC,
3D CNN
global information. The and WT, respectively
decoder block allows
parallel computation of
cross- and self-attention.
J. Imaging 2022, 8, 205 22 of 40
4.5.2. Brain Tumor Classification Using Bayesian Theory and Capsule Neural Network
CapsNet, like other standard DL networks, suffers from model uncertainty [108].
Model uncertainty refers to the degree to which a model is uncertain about its weights and
predictions. Developing a network that models uncertainty is important because it serves
as a medium for returning the uncertain predictions to experts for further verification. Most
CNN networks use the softmax activation function. However, the output of the softmax
does not reflect the degree of uncertainty of a model’s prediction [108]. Afshar et al. [40]
proposed a brain tumor classification network (called BayesCap) using Bayesian theory
and CapsNet. Bayesian theory is used to model the uncertainty associated with predictions
of the model. The technique was evaluated on the CR-MRI dataset [56] consisting of
3604 brain MR images from 233 patients. The results show that the technique produced an
accuracy of 68.3%. The same authors [40] in another study [18] evaluated CapsNet on the
J. Imaging 2022, 8, 205 23 of 40
same dataset, and the results show that CapsNet (without Bayesian theory) outperforms
the Bayesian variant, producing a classification accuracy of 78%. The reduced accuracy
was expected because the Bayesian variant was developed with the objective of modeling
the uncertainty and learning the posterior of the model’s weights rather than improving
the classification accuracy [40]. Overall, the results show that Bayesian theory improves
the interpretability of a network, which is very important for medical applications.
connected layer, and an output layer with three neurons, one for each of the three output
classes. The CapsNet was evaluated on the CE-MRI dataset. The dataset was used to train
the CapsNet before and after image pre-processing. Results reported in the study show
that the proposed technique produced an accuracy of 87% and 92.6%, before and after
pre-processing, respectively. This shows that the pre-processing improved the accuracy of
CapsNet by 5.6%.
4.5.6. Brain Tumor Segmentation Using a Modified Version of Capsule Neural Network
Aziz et al. [37] introduced a CapsNet-based technique for segmenting brain tumors.
They used the modified version of the original CapsNet architecture called SegCaps [42].
SegCaps was first proposed by Lalonde and Bagci [42] to address CapsNet’s [37] runtime
and memory constraint issues. The routing method is one of the primary distinctions
between CapsNet and SegCaps. In the SegCaps design, routing between lower layer
capsules and upper layer capsules takes place entirely within a single spatial window, and
transformation matrices are also shared between capsules of each layer. The proposed
technique was evaluated on the BraTS2020 dataset. Twenty percent of the tumor slices
were randomly selected and utilized to train SegCaps in the studies. Additionally, 85% of
tumor slices were randomly chosen and used to train the U-Net architecture. Results show
that SegCaps surpassed the U-Net architecture by 3%, obtaining a DSC of 87.96%. SegCaps
accomplished this improvement by using 95.4% fewer parameters than U-Net.
The workflow for a typical CapsNet-based brain tumor segmentation technique is
shown in Figure 10. Moreover, the summary of all the CapsNet-based techniques reviewed
in this study is presented in Table 6.
Figure 10. Workflow for a Capsule Network (CapsNet)-based brain tumor segmentation technique.
Classes
Ref. Year Method Dataset Main Highlight Performance
Considered
Designed a DL technique
Meningioma,
CapsNet and that can model uncertainty Classification
[40] 2020 glioma, and Cancer dataset [56]
Bayesian theory. associated with predictions accuracy: 68.3%
pituitary.
of CapsNet models.
Brain tumor dataset.
Meningioma
Introduced a new
Meningioma, (937 images),
activation function for Classification
[19] 2021 Glioma, Pituitary, CapsNet Glioma (926 images),
CapsNet, called PSTanh accuracy of 96.70%.
normal Pituitary
activation function.
(901 images),
normal (500)
J. Imaging 2022, 8, 205 25 of 40
Table 6. Cont.
Classes
Ref. Year Method Dataset Main Highlight Performance
Considered
Brain tumor dataset
Developed a
[56]. 3064 images
CapsNet-based technique
from 233 patients.
Meningioma, using dilation convolution
CapsNet, dilation Meningioma Classification
[109] 2019 Glioma, Pituitary, with the objective of
convolution (708 slices), Glioma accuracy: 95.54%.
normal maintaining the high
(1426 slices),
resolution of the images
Pituitary (930 slices),
for accurate classification.
normal
Presented a performance
CapsNet; analysis on the effect of
Meningioma, Brain tumor dataset: Classification
[110] 2019 classification; Data image pre-processing on
Glioma, Pituitary 3064 [56]. accuracy: 92.6%
pre-processing CapsNet for brain tumor
segmentation.
SegCaps–Capsule
T1 w, T2 w, T1 w Designed a modified
network; brain
[37] 2021 post GBCA and BraTS 2020 version of CapsNet using DSC of 87.96%.
tumor
FLAIR SegCaps network.
segmentation
Developed different
CapsNet for brain tumor
segmentation. Investigated 86.5% for
Meningioma,
Brain tumor dataset the performance of input segmented tumor,
[18] 2018 Pituitary, and Capsule network
proposed by [56] data on capsule network. and 78% for whole
Glioma
Developed a visualization brain image
paradigm for the output of
capsule network.
5. Discussion
Different ML-based, CNN-based, CapsNet-based, and ViT-based techniques have
been developed so far. Generally, most of the techniques are divided into two stages:
segmentation and classification stage. The segmentation stage is used to identify the
specific region where the tumor is located and the subregions within the tumor, while the
classification stage is used to identify the specific type or grade of the segmented tumor.
Most of the techniques were developed to distinguish the different classes of CNS tumors
including meningioma, pituitary tumors, glioma, and metastases. SVM, DT, NB, Bayesian
algorithm, k-NN, ANNs, and CNN are the most widely used ML and DL algorithms for
brain tumor classification and segmentation.
Brain tumors can be diagnosed using different types of medical images, such as CT
and MRI. However, MRI is the reference standard in the clinical routine. This is due to
the multiparametric nature of MRI. This section presents a discussion on the design and
development of various brain tumor classification and segmentation techniques.
data imbalance in many datasets. Results obtained in the study show that training the
model on local and global features concurrently produced better outcomes than training
separately and averaging the results.
A few authors [41,60] investigated the usage of small kernels with the goal of con-
structing deeper networks that do not overfit. Results achieved by the studies show that
small kernels produce effective brain segmentation results. Furthermore, some studies
designed effective multi-purpose frameworks for brain tumor diagnosis. For example,
Sharif et al. [101] designed a multi-purpose framework that can perform different functions
including tumor enhancement, feature extraction and selection, localization, and tumor seg-
mentation. In the study, homomorphic wavelet filter, inception-v3 model, non-dominated
sorted genetic algorithm (NGSA), YOLOv2, and ML algorithms were used for tumor en-
hancement, feature extraction, feature selection, localization, and tumor segmentation,
respectively. Experiments performed in the study show that the multi-purpose framework
produced good results. This type of framework will be very useful to medical practitioners,
as it can help them to handle multiple tasks effectively.
Most of the CNN-based and CapsNet-based architectures developed in the literature
used the ReLU activation function. The performance produced by ReLU-based networks
can be improved [19]. Adu et al. [19] introduced a new activation function for improved
tumor classification and segmentation, called PSTanh. Results reported in the study show
that PSTanh activation function outperformed ReLU by 7.11%.
5.4. Capsule Neural Network-Based Brain Tumor Classification and Segmentation Techniques
Many CNN-based image classification techniques have been introduced in the litera-
ture, and most of them produced state-of-the-art results. CNNs can automatically learn
high-level and low-level features from images without the need for feature engineering.
However, CNNs require large datasets for training. Additionally, CNNs are incapable of
J. Imaging 2022, 8, 205 28 of 40
then combined to produce improved results. Jena et al. [15] designed a tumor classifica-
tion technique using a combination of seven algorithms, namely: Adaboost, Gentleboost,
Logitboost, LPboost, Robustboost, RUSboost, and Totalboost. Results reported in the study
show that the ensemble method produced the best classification accuracy compared to
stand-alone classification algorithms, such as RF.
In addition to the U-Net architecture, the following network architectures are widely
used in the literature: ResNet, VGG-19, and InputCascadeCNN. The VGG-19 architecture
is pre-trained on the ImageNet dataset [118] consisting of 1.2 million high-quality images.
It consists of 16 convolutional layers and three fully connected layers. The InputCas-
cadeCNN architecture [10] is designed to extract local and global contextual features. It
consists of 7 × 7 feature maps for extracting local features (small details of the tumors) and
11 × 11 feature maps for extracting global features.
Some researchers introduced novel fine-tuning techniques for brain tumor classifi-
cation and segmentation. Swati et al. [97] introduced a block-wise fine-tuning technique
using the VGG-19 network architecture. They divided the layers of the VGG-19 network
into different blocks and fine-tuned each block sequentially. While the layers in one block
are fine-tuned, the layers in the other blocks are frozen. Experiments show that this tech-
nique produced promising results. Thus, more fine-tuning approaches can be developed to
improve the performance of brain tumor diagnosis systems.
Most studies also applied data augmentation techniques to increase the dataset size
and consequently improve the performance of CNN models. Some of the data augmenta-
tion techniques used in the literature include flipping, rotation, skewness, shears, sharpen-
ing, Gaussian blur, emboss, and edge detection. Some studies [16,95] used a combination of
these techniques. Results reported in the literature show that data augmentation techniques
improved the performance of DL-based brain tumor segmentation and classification mod-
els. In addition to data augmentation, some studies [75] used data enrichment techniques,
such as GANs, to generate synthetic MR images. Many studies did not explore the use of
advanced data enrichment techniques, and it is still an active research area.
presented in this study. The authors introduced a two-pathway architecture that can learn
to segment both localized and global-scale features of the brain. This shows that building
a model that can learn both local and global features concurrently will produce better
outcomes compared to building separate models and averaging their results. Figure 13 also
shows that the patching-based technique developed by AlBadawy et al. [73] produced the
worst result, by achieving a DSC of 0.68.
Figure 14 shows the performance overview of the ViT-based techniques reviewed in
this study. As shown in Figure 14, the ViT-based technique developed by
Hatamizadeh et al. [32] produced the best result, achieving a WT, ET, and TC of 93.3%,
89.1%, and 91.7%, respectively. In the study [32], the authors developed a U-shaped ar-
chitecture consisting of CNN and Swin. Moreover, the ViT-based technique developed
by Jia et al. [104] produced the second-best result. The authors designed a UNet-based
architecture for multi-modal MRI scans using CNN and ViT.
Figure 15 shows the results of the CapsNet-based brain tumor segmentation and
classification techniques surveyed in this paper. As shown in this plot, the technique
proposed by Thaha et al. [17] yielded the best result for CapsNet-based brain tumor
segmentation and classification. The technique achieved a classification accuracy of 96.7%.
In the study, the authors [17] developed a loss function optimization technique using the
BAT algorithm. This demonstrates how bio-inspired algorithms can be used to improve
CapsNet-based models. Figure 15 also shows that the method developed by Adu et al. [19]
produced the second-best result. The authors [19] used a new activation function with
CapsNet, and it achieved a classification accuracy of 95.54%.
Figure 11. Summarized results for classical ML-based brain tumor classification techniques. For
studies that used the same algorithm, we selected the algorithm that produced the best performance.
The SVM model designed by Sekhar et al. [14] produced the best performance for ML-based tumor
classification. The RNN constructed by Kaur et al. [94] generated the second-best result for ML-
based brain tumor classification. The remaining competitors include the DenseNet-169 proposed in
Kang et al. [84], the ensemble approach exploited by Jena et al. [15], and the work of Deepak and
Ameer [13] that combined SVM with CNN.
J. Imaging 2022, 8, 205 32 of 40
Figure 12. Performance overview of CNN-based brain tumor classification techniques. The tech-
nique developed by Isselmou et al. [95] achieved the best performance for CNN-based brain tumor
classification. The techniques proposed by Mzoughi et al. [62] yielded the second-best classification
accuracy. The remaining competitors include the work of Sajjad et al. [16], the VGG19-based approach
of Swati et al. [97], the work of Ge et al. [75] based on UNet and GAN, and, finally, the combination
of CNN and BAT proposed by Thaha et al. [17].
Figure 13. Performance overview of CNN-based brain tumor segmentation techniques. The tech-
nique developed by Havaei et al. [10] outperformed all the CNN-based brain tumor segmentation
techniques presented in this study. The remaining competitors include the patching-based technique
developed by AlBadawy et al. [73], the encoder-decoder architecture proposed by Zeineldin et al. [41],
and the triple network architecture developed by Yogananda et al. [74].
J. Imaging 2022, 8, 205 33 of 40
Figure 14. Performance overview of ViT-based techniques. The ViT-based technique developed
by Hatamizadeh et al. [32] produced the best result, achieving a WT, ET, and TC of 93.3%, 89.1%,
and 91.7%, respectively. The remaining competitors include the ViT-based technique developed by
Jia et al. [104], the work of Peiris et al. [107], the study of Wang et al. [35], and the method proposed
by Ali et al. [105].
Figure 15. Performance overview of CapsNet-based brain tumor techniques. The technique proposed
by Thaha et al. [17] in 2020 yielded the best result for CapsNet-based brain tumor segmentation
and classification. The method developed by Adu et al. [19] produced the second-best result. The
remaining competitors include the work of Afshar et al. [40], the studies of Kurup et al. [110] and
Aziz et al. [37], and the method proposed by Afshar et al. [18] in 2018.
outperformed the ReLU activation function. Future research could explore more acti-
vation functions for improved brain tumor classification and segmentation methods.
8. The max-pooling operation used in network architectures normally affects the image
quality and resolution of an image, and consequently the accuracy of a network.
Very few studies developed techniques that maintain high-resolution feature maps.
Further work is therefore required in this area. Future research can develop effective
techniques that can preserve the quality and spatial recognizability of an image.
9. Most of the currently available CNN-based approaches were developed for a specific
form of cancer. A general-purpose DL-based framework that can diagnose different
types of cancer will be very useful to medical practitioners
10. Most datasets cited in the literature suffer from data imbalance problems. For instance,
98% of samples in one of the benchmark datasets–Brain Tumor Segmentation (BraTS)
dataset–belong to a single class, whereas the remaining 2% belong to another class.
Clearly, building a model on such an imbalanced dataset will result in overfitting. Fur-
thermore, most studies did not explore the use of advanced data enrichment methods,
such as GANs, for improving the performance of brain tumor diagnosis. In addition,
most studies did not investigate the performance of different data augmentation
techniques for brain cancer diagnosis. Moreover, most studies did not investigate
the use of different state-of-the-art pre-trained networks for brain cancer diagnosis.
The problem of data imbalance and small-scale dataset in brain tumor diagnosis may
be addressed by developing techniques that combine advanced data augmentation
techniques and state-of-the-art pre-trained network architectures.
11. AlBadawy et al. [73] reported that there was a significant decrease in the performance
of CNN models when they were trained for patients from the same institution com-
pared to when they are trained for patients from different institutions. The reason
behind the reduced performance requires systematic investigations.
12. The fusion of multi-modal data can improve the performance of brain tumor diagnosis
models. Utilizing complementary information from multiple imaging modalities has
sparked a rise in recent research interest in cross-modality MR image synthesis. Recent
studies [119,120] have developed multi-modal brain tumor segmentation systems
that can learn high-level features from multi-modal data. Future research can design
enhanced multi-modal diagnostic frameworks for brain tumors.
13. Correct classification of multi-modal images of brain tumors is a vital step towards
accurate diagnosis and successful treatment of brain tumors. However, resolving
incomplete multi-modal issues is a challenging task in brain tumor diagnosis. Some
techniques [121,122] have been proposed to address this difficulty, but more research
is still required.
14. An important aspect that deserves further investigation is the development of inte-
grative approaches, by considering clinical and multiomics data along with imag-
ing [123,124], such as autoencoders [125] and variational autoencoders [126].
Author Contributions: Conceptualization, A.A.A., M.C. and L.R.; Methodology, A.A.A.; writing—
original draft preparation, A.A.A.; writing—review and editing, A.A.A., F.Z., J.T.G., M.C. and
L.R.; supervision, M.C. and L.R. All authors have read and agreed to the published version of
the manuscript.
Funding: We gratefully acknowledge financial support from FCT Fundação para a Ciência e a
Tecnologia (Portugal), national funding through research grant Information Management Research
Center—MagIC/NOVA IMS (UIDB/04152/2020).
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Not applicable.
Data Availability Statement: Not applicable.
Conflicts of Interest: The authors declare no conflict of interest.
J. Imaging 2022, 8, 205 36 of 40
References
1. Louis, D.N.; Perry, A.; Wesseling, P.; Brat, D.J.; Cree, I.A.; Figarella-Branger, D.; Hawkins, C.; Ng, H.K.; Pfister, S.M.;
Reifenberger, G. The 2021 WHO classification of tumors of the central nervous system: A summary. Neuro-Oncology 2021, 23,
1231–1251. [CrossRef] [PubMed]
2. The Brain Tumor Charity. Brain Tumor Basics. Available online: https://www.thebraintumourcharity.org/ (accessed on
1 June 2022).
3. Tandel, G.S.; Biswas, M.; Kakde, O.G.; Tiwari, A.; Suri, H.S.; Turk, M.; Laird, J.R.; Asare, C.K.; Ankrah, A.A.; Khanna, N.N. A
review on a deep learning perspective in brain cancer classification. Cancers 2019, 11, 111. [CrossRef] [PubMed]
4. Tamimi, A.F.; Juweid, M. Epidemiology and Outcome of Glioblastoma; Exon Publications: Brisbane, Australia, 2017; pp. 143–153.
[CrossRef]
5. Villanueva-Meyer, J.E.; Mabray, M.C.; Cha, S. Current clinical brain tumor imaging. Neurosurgery 2017, 81, 397–415. [CrossRef]
[PubMed]
6. Overcast, W.B.; Davis, K.M.; Ho, C.Y.; Hutchins, G.D.; Green, M.A.; Graner, B.D.; Veronesi, M.C. Advanced imaging techniques
for neuro-oncologic tumor diagnosis, with an emphasis on PET-MRI imaging of malignant brain tumors. Curr. Oncol. Rep. 2021,
23, 34. [CrossRef]
7. Zaccagna, F.; Grist, J.T.; Quartuccio, N.; Riemer, F.; Fraioli, F.; Caracò, C.; Halsey, R.; Aldalilah, Y.; Cunningham, C.H.; Massoud,
T.F. Imaging and treatment of brain tumors through molecular targeting: Recent clinical advances. Eur. J. Radiol. 2021, 142, 109842.
[CrossRef]
8. Zhang, Z.; Sejdić, E. Radiological images and machine learning: Trends, perspectives, and prospects. Comput. Biol. Med. 2019, 108,
354–370. [CrossRef]
9. Biratu, E.S.; Schwenker, F.; Ayano, Y.M.; Debelee, T.G. A survey of brain tumor segmentation and classification algorithms.
J. Imaging 2021, 7, 179. [CrossRef]
10. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain tumor
segmentation with deep neural networks. Med. Image Anal. 2017, 35, 18–31. [CrossRef]
11. Işın, A.; Direkoğlu, C.; Şah, M. Review of MRI-based brain tumor image segmentation using deep learning methods. Procedia
Comput. Sci. 2016, 102, 317–324. [CrossRef]
12. Chen, A.; Zhu, L.; Zang, H.; Ding, Z.; Zhan, S. Computer-aided diagnosis and decision-making system for medical data analysis:
A case study on prostate MR images. J. Manag. Sci. Eng. 2019, 4, 266–278. [CrossRef]
13. Deepak, S.; Ameer, P.M. Automated categorization of brain tumor from mri using cnn features and svm. J. Ambient Intell. Humaniz.
Comput. 2021, 12, 8357–8369. [CrossRef]
14. Sekhar, A.; Biswas, S.; Hazra, R.; Sunaniya, A.K.; Mukherjee, A.; Yang, L. Brain tumor classification using fine-tuned GoogLeNet
features and machine learning algorithms: IoMT enabled CAD system. IEEE J. Biomed. Health Inform. 2021, 26, 983–991. [CrossRef]
[PubMed]
15. Jena, B.; Nayak, G.K.; Saxena, S. An empirical study of different machine learning techniques for brain tumor classification and
subsequent segmentation using hybrid texture feature. Mach. Vis. Appl. 2022, 33, 6. [CrossRef]
16. Sajjad, M.; Khan, S.; Muhammad, K.; Wu, W.; Ullah, A.; Baik, S.W. Multi-grade brain tumor classification using deep CNN with
extensive data augmentation. J. Comput. Sci. 2019, 30, 174–182. [CrossRef]
17. Thaha, M.M.; Kumar, K.P.M.; Murugan, B.S.; Dhanasekeran, S.; Vijayakarthick, P.; Selvi, A.S. Brain tumor segmentation using
convolutional neural networks in MRI images. J. Med. Syst. 2019, 43, 1240–1251. [CrossRef]
18. Afshar, P.; Mohammadi, A.; Plataniotis, K.N. Brain tumor type classification via capsule networks. In Proceedings of the 2018
25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 3129–3133.
19. Adu, K.; Yu, Y.; Cai, J.; Asare, I.; Quahin, J. The influence of the activation function in a capsule network for brain tumor type
classification. Int. J. Imaging Syst. Technol. 2021, 32, 123–143. [CrossRef]
20. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.;
Gelly, S. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929.
21. Muhammad, K.; Khan, S.; Del Ser, J.; De Albuquerque, V.H.C. Deep learning for multigrade brain tumor classification in smart
healthcare systems: A prospective survey. IEEE Trans. Neural Netw. Learn. Syst. 2020, 32, 507–522. [CrossRef]
22. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in Medical Imaging: A Survey. arXiv
2022, arXiv:2201.09873.
23. Magadza, T.; Viriri, S. Deep learning for brain tumor segmentation: A survey of state-of-the-art. J. Imaging 2021, 7, 19. [CrossRef]
24. Zaccagna, F.; Riemer, F.; Priest, A.N.; McLean, M.A.; Allinson, K.; Grist, J.T.; Dragos, C.; Matys, T.; Gillard, J.H.; Watts, C.
Non-invasive assessment of glioma microstructure using VERDICT MRI: Correlation with histology. Eur. Radiol. 2019, 29,
5559–5566. [CrossRef] [PubMed]
25. Xiao, H.-F.; Chen, Z.-Y.; Lou, X.; Wang, Y.-L.; Gui, Q.-P.; Wang, Y.; Shi, K.-N.; Zhou, Z.-Y.; Zheng, D.-D.; Wang, D.J.J. Astrocytic
tumour grading: A comparative study of three-dimensional pseudocontinuous arterial spin labelling, dynamic susceptibility
contrast-enhanced perfusion-weighted imaging, and diffusion-weighted imaging. Eur. Radiol. 2015, 25, 3423–3430. [CrossRef]
[PubMed]
J. Imaging 2022, 8, 205 37 of 40
26. Dijkstra, M.; van Nieuwenhuizen, D.; Stalpers, L.J.A.; Wumkes, M.; Waagemans, M.; Vandertop, W.P.; Heimans, J.J.; Leenstra, S.;
Dirven, C.M.; Reijneveld, J.C. Late neurocognitive sequelae in patients with WHO grade I meningioma. J. Neurol. Neurosurg.
Psychiatry 2009, 80, 910–915. [CrossRef]
27. Waagemans, M.L.; van Nieuwenhuizen, D.; Dijkstra, M.; Wumkes, M.; Dirven, C.M.F.; Leenstra, S.; Reijneveld, J.C.; Klein, M.;
Stalpers, L.J.A. Long-term impact of cognitive deficits and epilepsy on quality of life in patients with low-grade meningiomas.
Neurosurgery 2011, 69, 72–79. [CrossRef]
28. van Alkemade, H.; de Leau, M.; Dieleman, E.M.T.; Kardaun, J.W.P.F.; van Os, R.; Vandertop, W.P.; van Furth, W.R.; Stalpers, L.J.A.
Impaired survival and long-term neurological problems in benign meningioma. Neuro. Oncol. 2012, 14, 658–666. [CrossRef]
29. Grist, J.T.; Miller, J.J.; Zaccagna, F.; McLean, M.A.; Riemer, F.; Matys, T.; Tyler, D.J.; Laustsen, C.; Coles, A.J.; Gallagher, F.A.
Hyperpolarized 13C MRI: A novel approach for probing cerebral metabolism in health and neurological disease. J. Cereb. Blood
Flow Metab. 2020, 40, 1137–1147. [CrossRef]
30. Fan, J.; Fang, L.; Wu, J.; Guo, Y.; Dai, Q. From brain science to artificial intelligence. Engineering 2020, 6, 248–252. [CrossRef]
31. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; MIT Press: Cambridge, MA, USA, 2016; ISBN 0262337371.
32. Hatamizadeh, A.; Nath, V.; Tang, Y.; Yang, D.; Roth, H.; Xu, D. Swin UNETR: Swin Transformers for Semantic Segmentation of
Brain Tumors in MRI Images. arXiv 2022, arXiv:2201.01266.
33. Dai, Y.; Gao, Y.; Liu, F. Transmed: Transformers advance multi-modal medical image classification. Diagnostics 2021, 11, 1384.
[CrossRef]
34. Raghu, M.; Unterthiner, T.; Kornblith, S.; Zhang, C.; Dosovitskiy, A. Do vision transformers see like convolutional neural
networks? Adv. Neural Inf. Process. Syst. 2021, 34, 12116–12128.
35. Wang, W.; Chen, C.; Ding, M.; Yu, H.; Zha, S.; Li, J. Transbts: Multimodal brain tumor segmentation using transformer. In
Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Virtual Event,
27 September–1 October 2021; Springer: Berlin/Heidelberg, Germany, 2021; pp. 109–119.
36. Rodriguez, R.; Dokladalova, E.; Dokládal, P. Rotation invariant CNN using scattering transform for image classification. In
Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019;
pp. 654–658.
37. Aziz, M.J.; Zade, A.A.T.; Farnia, P.; Alimohamadi, M.; Makkiabadi, B.; Ahmadian, A.; Alirezaie, J. Accurate Automatic Glioma
Segmentation in Brain MRI images Based on CapsNet. In Proceedings of the 2021 43rd Annual International Conference of the
IEEE Engineering in Medicine & Biology Society (EMBC), Virtual Conference, 31 October–4 November 2021; pp. 3882–3885.
38. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic routing between capsules. arXiv 2017, arXiv:1710.09829.
39. Patrick, M.K.; Adekoya, A.F.; Mighty, A.A.; Edward, B.Y. Capsule networks–a survey. J. King Saud Univ. Inf. Sci. 2022, 34,
1295–1310.
40. Afshar, P.; Mohammadi, A.; Plataniotis, K.N. BayesCap: A Bayesian Approach to Brain Tumor Classification Using Capsule
Networks. IEEE Signal Process. Lett. 2020, 27, 2024–2028. [CrossRef]
41. Zeineldin, R.A.; Karar, M.E.; Coburger, J.; Wirtz, C.R.; Burgert, O. DeepSeg: Deep neural network framework for automatic brain
tumor segmentation using magnetic resonance FLAIR images. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 909–920. [CrossRef]
42. LaLonde, R.; Bagci, U. Capsules for object segmentation. arXiv 2018, arXiv:1804.04241.
43. Nguyen, H.P.; Ribeiro, B. Advanced capsule networks via context awareness. In Proceedings of the International Conference on
Artificial Neural Networks, Munich, Germany, 17–19 September 2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 166–177.
44. BraTS. Multimodal Brain Tumor Segmentation. 2012. Available online: https://www2.imm.dtu.dk/projects/BRATS2012/
(accessed on 8 November 2021).
45. QTIM. The Quantitative Translational Imaging in Medicine Lab at the Martinos Center. 2013. Available online: https:%0A//qtim-
lab.github.io/ (accessed on 8 December 2021).
46. BraTS. MICCAI-BRATS 2014. 2014. Available online: https://sites.google.com/site/miccaibrats2014/ (accessed on
8 December 2021).
47. BraTS. BraTS 2015. 2015. Available online: https://sites.google.com/site/braintumorsegmentation/home/brats2015 (accessed
on 8 December 2021).
48. BraTS. BraTS 2016. 2016. Available online: https://sites.google.com/site/braintumorsegmentation/home/brats_2016 (accessed
on 1 December 2021).
49. MICCAI 2017. 20th International Conference on Medical Image Computing and Computer Assisted Intervention 2017. Available
online: https://www.miccai2017.org/ (accessed on 8 December 2021).
50. BraTS. Multimodal Brain Tumor Segmentation Challenge 2018. 2018. Available online: https://www.med.upenn.edu/sbia/
brats2018.html (accessed on 8 December 2021).
51. BraTS. Multimodal Brain Tumor Segmentation Challenge 2019: Data. 2019. Available online: https://www.med.upenn.edu/
cbica/brats2019/data.html (accessed on 8 December 2021).
52. BraTS. Multimodal Brain Tumor Segmentation Challenge 2020: Data. 2021. Available online: https://www.med.upenn.edu/
cbica/brats2020/data.html (accessed on 8 December 2021).
53. BraTS. RSNA-ASNR-MICCAI Brain Tumor Segmentation (BraTS) Challenge 2021. 2021. Available online: http://
braintumorsegmentation.org/ (accessed on 8 December 2021).
54. TCIA. Cancer Imaging Archive. Available online: https://www.cancerimagingarchive.net/ (accessed on 8 December 2021).
J. Imaging 2022, 8, 205 38 of 40
55. Radiopaedia. Brain Tumor Dataset. Available online: https://radiopaedia.org/ (accessed on 30 May 2022).
56. Cheng, J. Brain Tumor Dataset. Figshare. Dataset 2017. Available online: https://scholar.google.com/scholar_lookup?title=
Braintumordataset&author=J.Cheng&publication_year=2017 (accessed on 30 May 2022).
57. Chakrabarty, N. Brain MRI Images for Brain Tumor Detection Dataset, 2019. Available online: https://www.kaggle.com/
navoneel/brain-mri-images-for-brain-tumor-detection (accessed on 28 January 2022).
58. Hamada, A. Br35H Brain Tumor Detection 2020 Dataset. 2020. Available online: https://www.kaggle.com/ahmedhamada0
/brain-tumor-detection (accessed on 28 January 2022).
59. Simpson, A.L.; Antonelli, M.; Bakas, S.; Bilello, M.; Farahani, K.; Van Ginneken, B.; Kopp-Schneider, A.; Landman, B.A.; Litjens, G.;
Menze, B. A large annotated medical image dataset for the development and evaluation of segmentation algorithms. arXiv 2019,
arXiv:1902.09063.
60. Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Brain tumor segmentation using convolutional neural networks in MRI images. IEEE
Trans. Med. Imaging 2016, 35, 1240–1251. [CrossRef]
61. Brooks, J.C.W.; Faull, O.K.; Pattinson, K.T.S.; Jenkinson, M. Physiological noise in brainstem FMRI. Front. Hum. Neurosci. 2013,
7, 623. [CrossRef] [PubMed]
62. Mzoughi, H.; Njeh, I.; Wali, A.; Slima, M.B.; BenHamida, A.; Mhiri, C.; Mahfoudhe, K.B. Deep multi-scale 3D convolutional
neural network (CNN) for MRI gliomas brain tumor classification. J. Digit. Imaging 2020, 33, 903–915. [CrossRef]
63. Mohan, J.; Krishnaveni, V.; Guo, Y. A survey on the magnetic resonance image denoising methods. Biomed. Signal Process. Control
2014, 9, 56–69. [CrossRef]
64. Chen, H.; Dou, Q.; Yu, L.; Qin, J.; Heng, P.-A. VoxResNet: Deep voxelwise residual networks for brain segmentation from 3D MR
images. Neuroimage 2018, 170, 446–455. [CrossRef] [PubMed]
65. Chen, W.; Liu, B.; Peng, S.; Sun, J.; Qiao, X. Computer-aided grading of gliomas combining automatic segmentation and radiomics.
Int. J. Biomed. Imaging 2018, 2018, 2512037. [CrossRef] [PubMed]
66. Mzoughi, H.; Njeh, I.; Slima, M.B.; Hamida, A.B. Histogram equalization-based techniques for contrast enhancement of MRI brain
Glioma tumor images: Comparative study. In Proceedings of the 2018 4th International Conference on Advanced Technologies
for Signal and Image Processing (ATSIP), Sousse, Tunisia, 21–24 March 2018; pp. 1–6.
67. Pizer, S.M.; Johnston, E.; Ericksen, J.P.; Yankaskas, B.C.; Muller, K.E.; Medical Image Display Research Group. Contrast-limited
adaptive histogram equalization: Speed and effectiveness. In Proceedings of the First Conference on Visualization in Biomedical
Computing, Atlanta, GA, USA, 22–25 May 1990; Volume 337.
68. Ramesh, S.; Sasikala, S.; Paramanandham, N. Segmentation and classification of brain tumors using modified median noise filter
and deep learning approaches. Multimed. Tools Appl. 2021, 80, 11789–11813. [CrossRef]
69. Zeng, Y.; Zhang, B.; Zhao, W.; Xiao, S.; Zhang, G.; Ren, H.; Zhao, W.; Peng, Y.; Xiao, Y.; Lu, Y. Magnetic resonance image denoising
algorithm based on cartoon, texture, and residual parts. Comput. Math. Methods Med. 2020, 2020, 1405647. [CrossRef] [PubMed]
70. Heo, Y.-C.; Kim, K.; Lee, Y. Image Denoising Using Non-Local Means (NLM) Approach in Magnetic Resonance (MR) Imaging: A
Systematic Review. Appl. Sci. 2020, 10, 7028. [CrossRef]
71. Kidoh, M.; Shinoda, K.; Kitajima, M.; Isogawa, K.; Nambu, M.; Uetani, H.; Morita, K.; Nakaura, T.; Tateishi, M.; Yamashita, Y.
Deep learning based noise reduction for brain MR imaging: Tests on phantoms and healthy volunteers. Magn. Reson. Med. Sci.
2020, 19, 195. [CrossRef]
72. Moreno López, M.; Frederick, J.M.; Ventura, J. Evaluation of MRI Denoising Methods Using Unsupervised Learning. Front. Artif.
Intell. 2021, 4, 75. [CrossRef] [PubMed]
73. AlBadawy, E.A.; Saha, A.; Mazurowski, M.A. Deep learning for segmentation of brain tumors: Impact of cross-institutional
training and testing. Med. Phys. 2018, 45, 1150–1158. [CrossRef]
74. Yogananda, C.G.B.; Shah, B.R.; Vejdani-Jahromi, M.; Nalawade, S.S.; Murugesan, G.K.; Yu, F.F.; Pinho, M.C.; Wagner, B.C.;
Emblem, K.E.; Bjørnerud, A. A Fully automated deep learning network for brain tumor segmentation. Tomography 2020, 6,
186–193. [CrossRef]
75. Ge, C.; Gu, I.Y.-H.; Jakola, A.S.; Yang, J. Enlarged training dataset by pairwise gans for molecular-based brain tumor classification.
IEEE Access 2020, 8, 22560–22570. [CrossRef]
76. Haralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man. Cybern. 1973, 6,
610–621. [CrossRef]
77. Aouat, S.; Ait-hammi, I.; Hamouchene, I. A new approach for texture segmentation based on the Gray Level Co-occurrence
Matrix. Multimed. Tools Appl. 2021, 80, 24027–24052. [CrossRef]
78. Chu, A.; Sehgal, C.M.; Greenleaf, J.F. Use of gray value distribution of run lengths for texture analysis. Pattern Recognit. Lett. 1990,
11, 415–419. [CrossRef]
79. Tian, S.; Bhattacharya, U.; Lu, S.; Su, B.; Wang, Q.; Wei, X.; Lu, Y.; Tan, C.L. Multilingual scene character recognition with
co-occurrence of histogram of oriented gradients. Pattern Recognit. 2016, 51, 125–134. [CrossRef]
80. Prakasa, E. Texture feature extraction by using local binary pattern. INKOM J. 2016, 9, 45–48. [CrossRef]
81. Al-Janobi, A. Performance evaluation of cross-diagonal texture matrix method of texture analysis. Pattern Recognit. 2001, 34,
171–180. [CrossRef]
82. He, D.-C.; Wang, L. Simplified texture spectrum for texture analysis. J. Commun. Comput. 2010, 7, 44–53.
J. Imaging 2022, 8, 205 39 of 40
83. Khan, R.U.; Zhang, X.; Kumar, R. Analysis of ResNet and GoogleNet models for malware detection. J. Comput. Virol. Hacking Tech.
2019, 15, 29–37. [CrossRef]
84. Kang, J.; Ullah, Z.; Gwak, J. MRI-Based Brain Tumor Classification Using Ensemble of Deep Features and Machine Learning
Classifiers. Sensors 2021, 21, 2222. [CrossRef]
85. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on
Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778.
86. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE
Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708.
87. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556.
88. Krizhevsky, A. One weird trick for parallelizing convolutional neural networks. arXiv 2014, arXiv:1404.5997.
89. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings
of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826.
90. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the
IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500.
91. Ma, N.; Zhang, X.; Zheng, H.-T.; Sun, J. Shufflenet v2: Practical guidelines for efficient cnn architecture design. In Proceedings of
the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 116–131.
92. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. Mobilenetv2: Inverted residuals and linear bottlenecks. In
Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018;
pp. 4510–4520.
93. Tan, M.; Chen, B.; Pang, R.; Vasudevan, V.; Sandler, M.; Howard, A.; Le, Q. V Mnasnet: Platform-aware neural architecture search
for mobile. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA,
15–20 June 2019; pp. 2820–2828.
94. Kaur, D.; Singh, S.; Mansoor, W.; Kumar, Y.; Verma, S.; Dash, S.; Koul, A. Computational Intelligence and Metaheuristic Techniques
for Brain Tumor Detection through IoMT-Enabled MRI Devices. Wirel. Commun. Mob. Comput. 2022, 2022, 1519198. [CrossRef]
95. Abd El Kader, I.; Xu, G.; Shuai, Z.; Saminu, S.; Javaid, I.; Salim Ahmad, I. Differential deep convolutional neural network model
for brain tumor classification. Brain Sci. 2021, 11, 352. [CrossRef] [PubMed]
96. Lei, B.; Huang, S.; Li, R.; Bian, C.; Li, H.; Chou, Y.-H.; Cheng, J.-Z. Segmentation of breast anatomy for automated whole breast
ultrasound images with boundary regularized convolutional encoder–decoder network. Neurocomputing 2018, 321, 178–186.
[CrossRef]
97. Swati, Z.N.K.; Zhao, Q.; Kabir, M.; Ali, F.; Ali, Z.; Ahmed, S.; Lu, J. Brain tumor classification for MR images using transfer
learning and fine-tuning. Comput. Med. Imaging Graph. 2019, 75, 34–46. [CrossRef] [PubMed]
98. Mzoughi, H.; Njeh, I.; Slima, M.B.; Hamida, A.B.; Mhiri, C.; Mahfoudh, K.B. Denoising and contrast-enhancement approach of
magnetic resonance imaging glioblastoma brain tumors. J. Med. Imaging 2019, 6, 44002. [CrossRef]
99. Bakas, S.; Akbari, H.; Sotiras, A.; Bilello, M.; Rozycki, M.; Kirby, J.; Freymann, J.; Farahani, K.; Davatzikos, C. Segmentation labels
and radiomic features for the pre-operative scans of the TCGA-GBM collection. The Cancer Imaging Archive. Nat. Sci. Data 2017,
4, 170117. [CrossRef]
100. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 bias correction. IEEE
Trans. Med. Imaging 2010, 29, 1310–1320. [CrossRef]
101. Sharif, M.I.; Li, J.P.; Amin, J.; Sharif, A. An improved framework for brain tumor analysis using MRI based on YOLOv2 and
convolutional neural network. Complex Intell. Syst. 2021, 7, 2023–2036. [CrossRef]
102. Conn, A.R.; Gould, N.I.M.; Toint, P. A globally convergent augmented Lagrangian algorithm for optimization with general
constraints and simple bounds. SIAM J. Numer. Anal. 1991, 28, 545–572. [CrossRef]
103. Kapur, J.N.; Sahoo, P.K.; Wong, A.K.C. A new method for gray-level picture thresholding using the entropy of the histogram.
Comput. Vis. Graph. Image Process. 1985, 29, 273–285. [CrossRef]
104. Jia, Q.; Shu, H. BiTr-Unet: A CNN-Transformer Combined Network for MRI Brain Tumor Segmentation. arXiv 2021,
arXiv:2109.12271.
105. Hatamizadeh, A.; Tang, Y.; Nath, V.; Yang, D.; Myronenko, A.; Landman, B.; Roth, H.R.; Xu, D. Unetr: Transformers for 3d
medical image segmentation. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa,
HI, USA, 4–8 January 2022; pp. 574–584.
106. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted
windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11–17 October
2021; pp. 10012–10022.
107. Peiris, H.; Hayat, M.; Chen, Z.; Egan, G.; Harandi, M. A Volumetric Transformer for Accurate 3D Tumor Segmentation. arXiv
2021, arXiv:2111.13300.
108. Gal, Y.; Ghahramani, Z. Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In Proceedings
of the International Conference on Machine Learning, New York City, NY, USA, 19–24 June 2016; pp. 1050–1059.
109. Adu, K.; Yu, Y.; Cai, J.; Tashi, N. Dilated Capsule Network for Brain Tumor Type Classification Via MRI Segmented Tumor Region.
In Proceedings of the 2019 IEEE International Conference on Robotics and Biomimetics (ROBIO), Dali, China, 6–8 December 2019;
pp. 942–947.
J. Imaging 2022, 8, 205 40 of 40
110. Kurup, R.V.; Sowmya, V.; Soman, K.P. Effect of data pre-processing on brain tumor classification using capsulenet. In Proceedings
of the International Conference on Intelligent Computing and Communication Technologies, Chongqing, China, 6–8 December
2019; Springer: Berlin/Heidelberg, Germany, 2019; pp. 110–119.
111. Verma, R.; Zacharaki, E.I.; Ou, Y.; Cai, H.; Chawla, S.; Lee, S.-K.; Melhem, E.R.; Wolf, R.; Davatzikos, C. Multiparametric tissue
characterization of brain neoplasms and their recurrence using pattern classification of MR images. Acad. Radiol. 2008, 15, 966–977.
[CrossRef] [PubMed]
112. Zacharaki, E.I.; Wang, S.; Chawla, S.; Yoo, D.S.; Wolf, R.; Melhem, E.R.; Davatzikos, C. MRI-based classification of brain tumor
type and grade using SVM-RFE. In Proceedings of the 2009 IEEE International Symposium on Biomedical Imaging: From Nano
to Macro, Boston, MA, USA, 28 June–1 July 2009; pp. 1035–1038.
113. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M. Imagenet
large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [CrossRef]
114. Sasikala, M.; Kumaravel, N. A wavelet-based optimal texture feature set for classification of brain tumours. J. Med. Eng. Technol.
2008, 32, 198–205. [CrossRef]
115. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the
International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October
2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241.
116. Wang, X.; Yang, L.T.; Feng, J.; Chen, X.; Deen, M.J. A Tensor-Based Big Service Framework for Enhanced Living Environments.
IEEE Cloud Comput. 2016, 3, 36–43. [CrossRef]
117. Naser, M.A.; Deen, M.J. Brain tumor segmentation and grading of lower-grade glioma using deep learning in MRI images.
Comput. Biol. Med. 2020, 121, 103758. [CrossRef]
118. Deng, J.; Dong, W.; Socher, R.; Li, L.-J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of
the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 248–255.
119. Zhou, T.; Fu, H.; Zhang, Y.; Zhang, C.; Lu, X.; Shen, J.; Shao, L. M2 Net: Multi-modal Multi-channel Network for Overall Survival
Time Prediction of Brain Tumor Patients. In Proceedings of the International Conference on Medical Image Computing and
Computer-Assisted Intervention, Strasbourg, France, 27 September–1 October 2020; Springer: Berlin/Heidelberg, Germany, 2020;
pp. 221–231.
120. Fang, F.; Yao, Y.; Zhou, T.; Xie, G.; Lu, J. Self-supervised Multi-modal Hybrid Fusion Network for Brain Tumor Segmentation.
IEEE J. Biomed. Health Inform. 2021, 1. [CrossRef]
121. Yurt, M.; Dar, S.U.H.; Erdem, A.; Erdem, E.; Oguz, K.K.; Çukur, T. mustGAN: Multi-stream generative adversarial networks for
MR image synthesis. Med. Image Anal. 2021, 70, 101944. [CrossRef]
122. Yu, B.; Zhou, L.; Wang, L.; Shi, Y.; Fripp, J.; Bourgeat, P. Ea-GANs: Edge-aware generative adversarial networks for cross-modality
MR image synthesis. IEEE Trans. Med. Imaging 2019, 38, 1750–1762. [CrossRef] [PubMed]
123. Rundo, L.; Militello, C.; Vitabile, S.; Russo, G.; Sala, E.; Gilardi, M.C. A survey on nature-inspired medical image analysis: A step
further in biomedical data integration. Fundam. Inform. 2020, 171, 345–365. [CrossRef]
124. Ressler, K.J.; Williams, L.M. Big data in psychiatry: Multiomics, neuroimaging, computational modeling, and digital phenotyping.
Neuropsychopharmacology 2021, 46, 1–2. [CrossRef] [PubMed]
125. Biswas, N.; Chakrabarti, S. Artificial intelligence (AI)-based systems biology approaches in multi-omics data analysis of cancer.
Front. Oncol. 2020, 10, 588221. [CrossRef]
126. Simidjievski, N.; Bodnar, C.; Tariq, I.; Scherer, P.; Andres Terre, H.; Shams, Z.; Jamnik, M.; Liò, P. Variational autoencoders for
cancer data integration: Design principles and computational practice. Front. Genet. 2019, 10, 1205. [CrossRef] [PubMed]