Jimaging 07 00019
Jimaging 07 00019
Imaging
Review
Deep Learning for Brain Tumor Segmentation: A Survey of
State-of-the-Art
Tirivangani Magadza and Serestina Viriri *
Abstract: Quantitative analysis of the brain tumors provides valuable information for understanding
the tumor characteristics and treatment planning better. The accurate segmentation of lesions requires
more than one image modalities with varying contrasts. As a result, manual segmentation, which is
arguably the most accurate segmentation method, would be impractical for more extensive studies.
Deep learning has recently emerged as a solution for quantitative analysis due to its record-shattering
performance. However, medical image analysis has its unique challenges. This paper presents a
review of state-of-the-art deep learning methods for brain tumor segmentation, clearly highlighting
their building blocks and various strategies. We end with a critical discussion of open challenges in
medical image analysis.
Keywords: brain tumor segmentation; deep learning; magnetic resonance imaging; survey
1. Introduction
Brain tumors are an abnormal growth of cells in the brain. Their exact causes are
not yet known, but there are factors that can increase the risk of brain tumor, such as
Citation: Magadza, T.; Viriri, S. Deep
exposure to radiation and a family history of brain cancer. There has been an increase in
Learning for Brain Tumor
Segmentation: A Survey of
incidences of brain tumors in all ages globally over the past few years [1]. In the United
State-of-the-Art. J. Imaging 2021, 7, 19.
States alone, an estimate of 78,980 new cases of primary malignant and non-malignant
https://doi.org/10.3390/ tumors were expected to be diagonized in 2018. Despite considerable efforts in brain tumor
jimaging7020019 segmentation research, patient diagnosis remains poor [2]. The most common types of
tumors in adults are meningiomas (low grade tumors) and gliomas and glioblastomas
Academic Editor: Leonardo Rundo (high grade tumors). Low grade tumors are less aggressive and they come with a life
Received: 23 November 2020 expectancy of several years. High grade tumors are much more aggressive and they have a
Accepted: 11 January 2021 median survival rate of less than two years.
Published: 29 January 2021 Medical imaging techniques, such as Magnetic Resonance Imaging (MRI), CT scans,
Positron emission tomography (PET), among others, play a crucial role in the diagnosis
Publisher’s Note: MDPI stays neu- of the tumors. These techniques are used to locate and assess the progression of the
tral with regard to jurisdictional clai- tumor before and after treatment. MRI is usually the modality of choice for diagnosis
ms in published maps and institutio- and treatment planning for brain tumors [2] due to its high resolution, soft tissue contrast,
nal affiliations. and non-invasive characteristics. Surgery is the most common form of treatment for
brain tumors, but radiation and chemotherapy can also be used to slow the growth of the
tumor [1]. More than one MRI slice is required to view different regions of the brain, e.g.,
Copyright: © 2021 by the authors. Li-
T1, T2, T1 contrast and FLAIR images.
censee MDPI, Basel, Switzerland. Again, in clinical practice, delineation of the tumor is usually done manually. An
This article is an open access article experienced radiologist will carefully study the scanned medical images of the patient
distributed under the terms and con- segmenting all of the affected regions. Apart from being time consuming, manual seg-
ditions of the Creative Commons At- mentation is dependent on the radiologist and it is subject to large intra and inter rater
tribution (CC BY) license (https:// variability [3]. Consequently, manual segmentation is limited to qualitative assessment or
creativecommons.org/licenses/by/ visual inspection only.
4.0/).
Figure 1. Labeled example of a brain tumor illustrating the importance of the different modalities
(adapted from [13]).
This is done by identifying abnormal areas when compared to normal tissues [1].
Some tumors, like glioblastomas, are hard to distinguish from normal tissues, because they
infiltrate surrounding tissues causing unclear boundaries. As a solution, more than one im-
age modalities with varying contrasts are often employed. In Figure 1, two MRI modalities
(T1 with contrast and T2) were used in order to accurately delineate tumor regions.
methods are less time consuming than manual segmentation, their results are still depen-
dent on the operator.
3. Deep Learning
Deep learning is a class of machine learning algorithms that uses multiple layers
to learn a hierarchy of increasingly complex presentations directly from the raw input.
Machine learning models are all about finding appropriate representations for their input
data. In this section, we will describe the building blocks, and recent techniques and
architectures of deep learning algorithms for brain tumor segmentation that we found in
papers surveyed in this work, as summarized in Figure 2.
Figure 2. Building blocks, architectures and techniques for deep learning algorithms for brain
tumor segmentation.
J. Imaging 2021, 7, 19 5 of 22
f ( x : θ ) = σ(w · x + b) (2)
where σ (·) is element-wise non-linearity activation function.
Figure 3. Typical feed-forward neural network composed of three layers. (adapted from [18]).
In a typical neural network, neurons are organized in layers. The input of each neuron
in a layer is connected to all or some of the output of neurons in the up-stream layer.
Likewise, the output of each neuron is connected to all or some of the input of neurons in
the downstream layer. The first layer in the network is the input layer, and the final layer is
the output layer. Layers in the middle are referred to as hidden layers. When each neuron
in a layer is connected to all of the neurons in the next layer, the network is called fully
connected network. A deep neural network is formed when there are many hidden layers,
hence the term deep learning.
analysis task. Furthermore, units in convolutional layers share weights, thereby reducing
the number of parameter to learn and improve the efficiency of the network.
3.3.5. Optimization
The performance of the deep CNN can be improved (or optimized) by training the
network on a large dataset. Training involves finding the parameters θ of the model that
significantly reduce a cost function J (θ ). Gradient descent is the widely used method for
updating network parameters through a back-propagation algorithm. Optimization can
be done per single sample, subset, or full set of the training samples. Thus, stochastic,
mini-batch, or batch gradient descent, respectively. Today, many optimization algorithms
for deep learning use mini-batches and it is now common to just call them stochastic
methods [16].
Stochastic gradient descent (SDG) comes with few notable challenges. Choosing an
appropriate learning rate can be difficult. A learning rate that is too small leads to very
slow convergence (tiny updates to the model parameters) and, at the same time, too large
will result in undesired divergence behavior in the loss function. All of the parameter
updates are based on the same learning rate, disregarding the fact that some of the features
might have higher frequency than other. Another key challenge is that optimization
can be trapped in sub-optimal local minima or saddle points, especially for non-convex
optimization [28].
Various variants of SDG have been proposed in the literature that address the afore-
mented challenges. Memontum-based SDG methods [29] can help in accelerating SDG
in relevant direction, dampening undesirable oscillations in local optima. Adagrad [30]
addressed the issue of manually turning the learning by adapting the learning rate to the
parameters, performing larger updates for infrequent parameters as compared to frequent
ones. However, Adagrad suffers from monotonically decreasing learning rate to a point
at which the algorithm stops learning. Adadelta [31], RMSprop [32], and Adam [33] ad-
dressed the shortcomings of Adagrad by dividing the learning rate by an exponentially
decaying average of past gradients.
The choice of the cost function depends on the problem area, whether it is a classifica-
tion or regression problem and the choice of the output unit [16]. The majority of the image
classification algorithms use softmax loss, withhs a combination of softmax and CE loss or
log-loss [28]. The softmax function produces a probability distribution over a number of
given output classes, while the CE loss takes the probability of predictions and penalizes
predictions that are confident but wrong. Class imbalance is one major issue in medical
image analysis, where one class will have fewer instances than the other. For example,
a brain tumor occupies a small portion when compared to healthy tissues. As a result,
the classifier will tend to be biased to the majority class. One way of addressing such a
problem is to adapt loss functions for class imbalance. Some works [34–36] proposed a loss
function that is based on the Dice coefficient. Ronneberger et al. [37] proposed a weighted
CE loss, which gives more importance to some pixels in the training data.
3.3.9. Regularization
Regularization is a technique for improving the performance of a machine learning
algorithm on unseen data. It is a way of reducing over-fitting on training set. Over-
fitting occurs when the gap between the training error and test error is too large [16].
When that happens, the model performs well on training data, but poorly on previously
unseen data. There are various techniques that can be employed in order to reduce the
generalization error, such as reducing the model capacity, which is, reducing the number
of learnable parameters in the model; adding L2 or L1 weight decay regularization term
to the cost function to force the model to only take small weight values; introducing
early stopping whenever the model performance stops improving on validation dataset;
randomly dropping out (skipping) the output of some units during training [41]. The last
approach is one of the most effective and most commonly used technique [17], mainly
because it is computationally inexpensive and prevents interdependent learning amongst
units. Batch Normalization [42] can also be used as a regularizer by ensuring that the
distribution of non-linearity inputs remains more stable as the model trains, thereby
improving the training of the model.
J. Imaging 2021, 7, 19 9 of 22
Training a machine learning model with more data is the best way to reduce the
generalization error. However, in the medical domain, acquiring a training dataset is time-
consuming, more expensive, and requires highly trained personnel to annotate ground
truth labels. Data augmentation can increase the dataset and reduce over-fitting by flipping,
applying small rotations, warping, and using the non-rigid deformation transformation
of images. However, great care must be taken when performing transformations of the
medical image dataset since the patch’s label is determined by the center of pixel [43].
Some recent works used generative models that include variational autoencoders [44] and
generative adversarial networks [45] to act as additional regularization that deals with
data scarcity.
the first stage is used as an input to perform either a multi-class intra-tumoral segmentation,
as in [49], or perform successive binary segmentation of the remain tumor sub-regions [34].
Wang et al. [34] observed an increase in the training and inference time of a multi-stage
binary segmentation as compared to a single multi-class network approach.
3.4.4. UNET
The UNET architecture [37] is an improvement of FCN [27], which resembles an
encoder and decoder network designed specifically for biomedical image segmentation.
The network consists of a contracting path (encoder) and an expansive path (decoder),
which gives it the u-shaped architecture. The contracting path consists of the repeated
application of two convolutional layers, followed by a rectified linear unit (ReLU) and max
pooling layer. Along the path, the spacial information is reduced, while feature information
is increased. The expansive path consists of a series of up-sampling operations combined
with high-resolution features from the contracting path through skip connections.
3.5.2. Post-Processing
The post-processing step is performed to further refine the segmentation results. It
helps in reducing the number of misclassifications or false positives in the segmentation
results while using algorithms, like conditional random fields (CRF) [4,34,53], markov
random fields (MRF) [54], connected component analysis [1,53,55], and morphological
operators [48,56]. CRF and MRF based techniques effectively remove false positives by com-
bining model predictions with low-level image information, like local interations of pixels
and edges when making finer adjustments. However, these techniques are computationaly
expensive [14]. Connected compents analysis involves finding and extracting connected
components and then applying a simple thresholding technique to remove unwanted blobs.
Another technique of removing false positive around edges of the segmentation image is
to apply morphological operations, erosion, and dilation in succession.
3.6. Datasets
Over the past few years, there have been considerable research interests in automatic
brain tumor segmentation. As research output continued to grow, the objective evaluation
of different algorithms became a challenge because researchers used private datasets with
varying attributes. As a result, benchmarking challenges, such as Multi-modal Brain Tumor
Image Segmentation (BRATS), emerged to standardize performance evaluation while using
publicly accessible datasets. Table 1 show a summary of the mostly used datasets for brain
tumor segmentation.
Since 2012, the BRATS Challenge [2], in conjunction with the International Conference
on Medical Image Computing and Computer-Assisted Interventions (MICCAI), has been
J. Imaging 2021, 7, 19 12 of 22
the primary bench-marking resource for brain tumor segmentation. It offers the medical re-
search community publicly accessible datasets for training and validation and standardized
metrics in order to objectively evaluate model performance against an online evaluation
platform. The dataset initially contained as small as 30 clinically acquired scans of glioma
patience, and the number has continued to grow over the subsequent years.
Table 1. Summary of commonly used public datasets for brain tumor segmentation.
Medical Segmentation Decathlon Challenge offers a relatively large dataset that sup-
ports a wide range of segmentation task. The Challenge aims to facilitate research in
general-purpose segmentation algorithms that solve various functions without any human
intervention. For brain tumor segmentation, the dataset comprises a subset of the 2016 and
2017 BRATS Challenge data.
Caffe [73] is a C++ deep learning framework that was initially developed for computer
vision applications and later spread to other domains like robotics, neuroscience, and
astronomy. It offers a complete toolkit for a deep learning pipeline, from training to
production deployment. Each processing stage is supplemented with well-documented
examples. Moreover, the framework is shipped with implementations of popular deep
learning building block and reference models allowing for quick experimentation with state-
of-the-art deep learning methods. The definition of models is done in config files, rather
than being hard-coded, ensuring the separation of representation from implementation.
Pytorch [74] is yet another fully-fledged open-source deep learning framework. Its de-
sign philosophy moved away from the define and execute style, as in many frameworks
that create a static computational graph before running the model. While this approach is
powerful, it sacrifices usability, the ease of debugging, and flexibility. Instead, Pytorch took
an imperative approach by dynamically constructing the computational graph, allowing
for the models to be idiomatically defined following the python programming model.
The framework also offers a seamless transition from research to production, distributed
training, and the seamless execution of models on edge devices.
Tensorflow [75] is an end-to-end distributed deep learning platform for large scale
machine learning applications. The platform supports the execution of dataflow graphs
across a span of heterogeneous devices, such as mobile devices and large-scale distributed
systems, with little or no change. Its design philosophy has been used to simplify model
parallelism within a single machine and across thousands of distributed systems. It has a
complete toolbox for quick experimentation with state-of-the-art deep learning models,
seamless transition from research to heterogeneous deployments, and the visualization
and debugging of large-scale models.
Keras [76] is a fast-growing high-level API for deep learning applications. Although it
initially supported multiple data-flow graph back-ends, like Theano, it is now deeply
woven into the Tensorflow 2 ecosystem. It provides consistent and simple APIs to quickly
experiment with new models and leverage Tensorflow in order to export the models to run
in browsers and mobile devices. Moreover, it comes bundled with building blocks and pre-
trained state-of-the-art models for various machine learning domains. The industry and
the research community have adopted the platform, because of its ease of use, user-centric
approach, and extensive documentation.
4. Discussion
Deep learning methods to medical image analysis have received tremendous attention
over the past few years. This is evident in the considerable increase in the number of
published works each year [2]. Deep learning techniques are able to learn a hierarchy of
increasingly complex features directly from data, as stated earlier. For example, in brain
tumor segmentation, deep learning algorithms can learn to segment MRI images by being
trained on a sufficiently large dataset. For this reason, CNN based models have been widely
adopted in medical image analysis, following their success in solving many problems in
computer vision, speech recognition, and natural language processing. Table 2 shows a
summary of deep learning methods that were reviewed in this work. Many techniques
differ considerably in terms of architectural design, with recent works following the
Unet [37] architecture and ensemble methods as shown in Table 3. Moreover, several
techniques have been developed in order to address inherent problems in automated brain
MRI analysis.
J. Imaging 2021, 7, 19 14 of 22
Table 2. Overview of Deep learning methods for brain tumor segmentation. BN = Batch normalization, GN = Group
normalization, outliers = remove top 1%, hist-norms = Histogram normalization,RN = Range normalization, HS = Histogram
standardization, slice-norm = Slice-based normalization, PLN = Piece-wise linear normalization, IN = Instant normalization,
CE = Cross entropy,BS = Bootstrapping, SS = Sensitivity-specification, NM = Negative Mining, WCE = Weighted cross-
entropy, neg-mining = Hard negative mining.
Unet Architecture
Table 2. Cont.
Ensemble Architecture
Table 3. A summary of top performing methods on BraTS 2017, 2018, and 2019 validation data as reported by the online evaluation platform. ET—Enhancing tumor, WT—Whole tumor,
and TC—Tumor core.
BraTS 2017
1 [79] Ensemble 0.738 0.901 0.797 0.783 0.895 0.762 0.998 0.995 0.998 4.499 4.229 6.562
2 [34] Cascaded 0.786 0.905 0.838 0.771 0.915 0.822 0.999 0.995 0.998 3.282 3.890 6.479
3 [83] Unet 0.776 0.903 0.819 0.803 0.902 0.786 0.998 0.996 0.999 3.163 6.767 8.642
3 [101] SegNet 0.706 0.857 0.716 0.687 0.811 0.660 0.999 0.997 0.999 6.835 5.872 10.925
BraTS 2018
1 [44] Ensemble 0.825 0.912 0.870 0.845 0.923 0.864 0.998 0.995 0.998 3.997 4.537 6.761
2 [88] Unet 0.809 0.913 0.863 0.831 0.919 0.844 0.998 0.995 0.999 2.413 4.268 6.518
3 [102] Ensemble 0.792 0.901 0.847 0.829 0.911 0.836 0.998 0.994 0.998 3.603 4.063 4.988
3 [103] Ensemble 0.814 0.909 0.865 0.813 0.914 0.868 0.998 0.995 0.997 2.716 4.172 6.545
BraTS 2019
1 [91] two-stage Unet 0.802 0.909 0.865 0.804 0.924 0.862 0.998 0.994 0.997 3.146 4.264 5.439
2 [92] Unet 0.746 0.904 0.840 0.780 0.901 0.811 0.990 0.987 0.990 27.403 7.485 9.029
3 [104] Ensemble 0.634 0.790 0.661 0.604 0.727 0.587 0.983 0.980 0.983 47.059 14.256 26.504
J. Imaging 2021, 7, 19 17 of 22
Deep learning algorithms require a relatively large amount of training data to gen-
eralize well on unseen data. However, this poses many challenge in the medical domain.
Firstly, it takes a well trained radiologist a considerable amount of time to annotate even
a single MRI volume. Moreover, the work is subject to an intra-rater and inter-rater
variability. Therefore, all of the annotations are approved by one to many experienced
neuro-radiologists [105], before they can be used in supervised training, which makes the
process of creating training and testing datasets not only time consuming, but expensive.
Secondly, medical data is protected by data protection laws that restrict the usage and
sharing of this kind of data to other parties. Consequently, a lot of time is spent seeking
approvals and removing personal identifiable information from medical data. Fortunately,
Table 1 shows a consistent increase of training and testing data for the BraTS Challenge.
Hopefully, this trend will continue in the coming years. Thus, facilitating training relative
deep networks and reducing over-fitting.
Because the lack of large-scale datasets restricts deep learning models’ full potential, re-
searchers have adopted data augmentation as an immediate solution to the data challenges
that are mentioned above. Other works have recently explored weakly-supervised learn-
ing [106–108] as a promising solution to address the need for fully annotated pixel-wise
labels. Instead of performing pixel-level annotations, known to be tedious and time-
consuming, weakly-supervised annotation uses bounding box or image-level annotations in
order to signify the presence or absence of lesions in images. This approach has the benefit
of being cheap, contains less labeling noise [107], far larger volumes of data can be generated
than pixel-level annotation, and training of deep learning models can leverage both kinds
of datasets.
Moreover, deep learning techniques require a huge amount of computational and
memory resources [28]. Very deep networks, which are becoming a widespread, have
millions of parameters that result in many costly mathematical computations that are
restrictive on the kind of computational hardware that can be used by researchers. Fur-
thermore, the use of 3D deep learning models increases the computational and memory
requirements by large margins. All of the reviewed literature use deep learning software
libraries to provide an infrastructure to define and train deep neural networks in parallel
or distributed manner while leveraging multi-core or multi-GPU environments. Currently,
researchers are being limited by the amount of GPU memory at their disposal (typically 12
gigabytes). For this reason, batch sizes and model complexities are being limited to what
can fit into the available memory.
The performance of brain tumor segmentation algorithms have continued to increase
over the past few years due to the availability of more training data and use of more sophis-
ticated CNN architectures and training schemes. However, their robustness is still lagging
behind expert performance [105]. Recently, researchers have used the ensemble methods
to achieve state-of-the-art performance (see Table 3). Precisely, the ensemble methods
fuse the segmentation results of several models to improve the robustness of individual
approach, resulting in superior performance as compared to inter-rater agreements [105].
Interestingly, single Unet [37] based models [91] continue to produce exceptional per-
formance, supporting the argument that: “a well trained Unet is hard to beat” [88]. The
reviewed literature have shown that careful initialization of hyper-parameters, a selection
of pre-processing techniques, employing advanced training schemes, as well as dealing
with the class imbalance problem will immensely improve the accuracy and robustness of
segmentation algorithms.
5. Summary
This paper has discussed several building blocks, state-of-the-art techniques, and tools
for implementing automatic brain tumor segmentation algorithms. Despite the tremendous
advance in the field, the robustness of deep learning methods are still inferior to expert
performance. Some notable architectures, including ensemble methods and UNet based
models, have shown great potential for improving the state-of-the-art with careful pre-
J. Imaging 2021, 7, 19 18 of 22
Author Contributions: Conceptualization, T.M. and S.V.; methodology, T.M. and S.V; formal analysis,
S.V.; investigation, T.M.; resources, S.V.; writing original draft preparation, T.M.; writing review and
editing, S.V.; supervision, S.V. Both authors have read and agreed to the published version of the
manuscript.
Funding: This research received no external funding.
Institutional Review Board Statement: Not applicable.
Informed Consent Statement: Informed consent was obtained from all subjects involved in the
study.
Data Availability Statement: Data available in publicly accessible repositories.
Conflicts of Interest: The authors declare no conflict of interest.
References
1. Havaei, M.; Davy, A.; Warde-Farley, D.; Biard, A.; Courville, A.; Bengio, Y.; Pal, C.; Jodoin, P.M.; Larochelle, H. Brain tumor
segmentation with Deep Neural Networks. Med. Image Anal. 2017, 35, 18–31. [CrossRef] [PubMed]
2. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The
Multimodal Brain Tumor Image Segmentation Benchmark (BRATS). IEEE Trans. Med. Imaging 2015, 34, 1993–2024. [CrossRef]
[PubMed]
3. Işın, A.; Direkoğlu, C.; Şah, M. Review of MRI-Based Brain Tumor Image Segmentation Using Deep Learning Methods. Procedia
Comput. Sci. 2016, 102, 317–324. [CrossRef]
4. Kamnitsas, K.; Ledig, C.; Newcombe, V.F.J.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient multi-scale
3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78. [CrossRef] [PubMed]
5. Razzak, M.I.; Imran, M.; Xu, G. Efficient Brain Tumor Segmentation With Multiscale Two-Pathway-Group Conventional Neural
Networks. IEEE J. Biomed. Health Inform. 2019, 23, 1911–1919. [CrossRef]
6. Muhammad, K.; Khan, S.; Ser, J.D.; de Albuquerque, V.H.C. Deep Learning for Multigrade Brain Tumor Classification in Smart
Healthcare Systems: A Prospective Survey. IEEE Trans. Neural Netw. Learn. Syst. 2020, 1–16. [CrossRef]
7. Wadhwa, A.; Bhardwaj, A.; Singh Verma, V. A review on brain tumor segmentation of MRI images. Magn. Reson. Imaging 2019,
61, 247–259. [CrossRef]
8. Alom, M.Z.; Taha, T.M.; Yakopcic, C.; Westberg, S.; Sidike, P.; Nasrin, M.S.; Hasan, M.; Van Essen, B.C.; Awwal, A.A.S.; Asari, V.K.
A State-of-the-Art Survey on Deep Learning Theory andArchitectures. Electronics 2019, 8, 292. [CrossRef]
9. Zikic, D.; Ioannou, Y.; Brown, M.; Criminisi, A. Segmentation of Brain Tumor Tissues with Convolutional Neural Networks. In
Proceedings of the BRATS-MICCAI, Boston, MA, USA, 14 September 2014; pp. 36–39.
10. Urban, G.; Bendszus, M.; Hamprecht, F.A.; Kleesiek, J. Multi-Modal Brain Tumor Segmentation Using Deep Convolutional
Neural Networks. In Proceedings of the BRATS-MICCAI, Boston, MA, USA, 14 September 2014; pp. 31–35.
11. Shen, D.; Wu, G.; Suk, H.I. Deep learning in medical image analysis. Annu. Rev. Biomed. 2017, 19, 221–248. [CrossRef]
12. Havaei, M.; Guizard, N.; Larochelle, H.; Jodoin, P.M. Deep Learning Trends for Focal Brain Pathology Segmentation in MRI.
In Machine Learning for Health Informatics; Holzinger, A., Ed.; Springer International Publishing: Berlin/Heidelberg, Germany,
2016; Volume 9605, pp. 125–148. [CrossRef]
13. Corso, J.J.; Sharon, E.; Dube, S.; El-Saden, S.; Sinha, U.; Yuille, A. Efficient Multilevel Brain Tumor Segmentation With Integrated
Bayesian Model Classification. IEEE Trans. Med. Imaging 2008, 27, 629–640. [CrossRef]
14. Gordillo, N.; Montseny, E.; Sobrevilla, P. State of the Art Survey on MRI Brain Tumor Segmentation. Magn. Reson. Imaging 2013,
31, 1426–1438. [CrossRef] [PubMed]
15. Chen, L.; Bentley, P.; Mori, K.; Misawa, K.; Fujiwara, M.; Rueckert, D. DRINet for Medical Image Segmentation. IEEE Trans. Med.
Imaging 2018, 37, 2453–2462. [CrossRef] [PubMed]
16. Goodfellow, I.; Bengio, Y.; Courville, A. Deep Learning; Adaptive Computation and Machine Learning; The MIT Press: Cambridge,
MA, USA, 2016.
17. Chollet, F. Deep Learning with Python; Manning Publications Co.: Shelter Island, NY, USA, 2018.
18. Svozil, D.; Kvasnicka, V.; Pospichal, J. Introduction to Multi-Layer Feed-Forward Neural Networks. Chemom. Intell. Lab. Syst.
1997, 39, 43–62. [CrossRef]
19. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with
Convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA,
7–12 June 2015.
J. Imaging 2021, 7, 19 19 of 22
20. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings
of the Advances in Neural Information Processing Systems; Pereira, F., Burges, C.J.C., Bottou, L., Weinberger, K.Q., Eds.; Curran
Associates, Inc.: Red Hook, NY, USA, 2012; Volume 25, pp. 1097–1105.
21. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556.
22. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385.
23. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. arXiv 2019, arXiv:1709.01507.
24. Glorot, X.; Bordes, A.; Bengio, Y. Deep Sparse Rectifier Neural Networks. In Proceedings of the Fourteenth International
Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Ft. Lauderdale, FL, USA,
11–13 April 2011; pp. 315–323.
25. Maas, A.L.; Hannun, A.Y.; Ng, A.Y. Rectifier Nonlinearities Improve Neural Network Acoustic Models. In Proceedings of the
ICML Workshop on Deep Learning for Audio, Speech and Language Processing, Atlanta, GA, USA, 16 June 2013.
26. He, K.; Zhang, X.; Ren, S.; Sun, J. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification.
arXiv 2015, arXiv:1502.01852.
27. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. arXiv 2015, arXiv:1411.4038.
28. Bernal, J.; Kushibar, K.; Asfaw, D.S.; Valverde, S.; Oliver, A.; Martí, R.; Lladó, X. Deep Convolutional Neural Networks for Brain
Image Analysis on Magnetic Resonance Imaging: A Review. Artif. Intell. Med. 2019, 95, 64–81. [CrossRef]
29. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the Importance of Initialization and Momentum in Deep Learning. In Proceedings
of the 30th International Conference on Machine Learning; Dasgupta, S., McAllester, D., Eds.; PMLR: Atlanta, GA, USA, 2013;
Volume 28, pp. 1139–1147.
30. Duchi, J.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res.
2011, 12, 2121–2159.
31. Zeiler, M.D. ADADELTA: An Adaptive Learning Rate Method. arXiv 2012, arXiv:1212.5701.
32. Tieleman, T.; Hinton, G. Lecture 6.5-rmsprop: Divide the gradient by a running average of its recent magnitude. COURSERA
Neural Netw. Mach. Learn. 2012, 4, 26–31.
33. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980.
34. Wang, G.; Li, W.; Ourselin, S.; Vercauteren, T. Automatic Brain Tumor Segmentation Based on Cascaded Convolutional Neural
Networks With Uncertainty Estimation. Front. Comput. Neurosci. 2019, 13, 56. [CrossRef] [PubMed]
35. Li, H.; Li, A.; Wang, M. A novel end-to-end brain tumor segmentation method using improved fully convolutional networks.
Comput. Biol. Med. 2019, 108, 150–160. [CrossRef]
36. Cahall, D.E.; Rasool, G.; Bouaynaya, N.C.; Fathallah-Shaykh, H.M. Inception Modules Enhance Brain Tumor Segmentation. Front.
Comput. Neurosci. 2019, 13, 44. [CrossRef]
37. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015,
arXiv:1505.04597.
38. Glorot, X.; Bengio, Y. Understanding the Difficulty of Training Deep Feedforward Neural Networks. In Proceedings of the
Thirteenth International Conference on Artificial Intelligence and Statistics, Chia Laguna Resort, Sardinia, Italy, 13–15 May 2010;
Volume 9, pp. 249–256.
39. Bengio, Y.; Lamblin, P.; Popovici, D.; Larochelle, H. Greedy Layer-Wise Training of Deep Networks. In Advances in Neural
Information Processing Systems 19; Schölkopf, B., Platt, J.C., Hoffman, T., Eds.; MIT Press: Cambridge, MA, USA, 2007; pp. 153–160.
40. Claesen, M.; De Moor, B. Hyperparameter Search in Machine Learning. arXiv 2015, arXiv:1502.02127.
41. Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A Simple Way to Prevent Neural Networks
from Overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958.
42. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015,
arXiv:1502.03167.
43. Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Brain Tumor Segmentation Using Convolutional Neural Networks in MRI Images.
IEEE Trans. Med. Imaging 2016, 35, 1240–1251. [CrossRef] [PubMed]
44. Myronenko, A. 3D MRI Brain Tumor Segmentation Using Autoencoder Regularization. arXiv 2018, arXiv:1810.11654.
45. Rezaei, M.; Harmuth, K.; Gierke, W.; Kellermeier, T.; Fischer, M.; Yang, H.; Meinel, C. Conditional Adversarial Network for
Semantic Segmentation of Brain Tumor. arXiv 2017, arXiv:1708.05227.
46. Rao, V.; Sarabi, M.S.; Jaiswal, A. Brain tumor segmentation with deep learning. In Proceedings of the MICCAI Multimodal
Brain Tumor Segmentation Challenge (BraTS), 2015; pp. 56–59. Available online: https://www.researchgate.net/profile/Mona_
Sharifi2/publication/309456897_Brain_tumor_segmentation_with_deep_learning/links/5b444445458515f71cb8a65d/Brain-
tumor-segmentation-with-deep-learning.pdf (accessed on 1 June 2020).
47. Casamitjana, A.; Puch, S.; Aduriz, A.; Sayrol, E.; Vilaplana, V. 3D Convolutional Networks for Brain Tumor Segmentation.
In Proceedings of the MICCAI Challenge on Multimodal Brain Tumor Image Segmentation (BRATS), 2016; pp. 65–68. Available
online: https://imatge.upc.edu/web/sites/default/files/pub/cCasamitjana16.pdf (accessed on 1 June 2020).
48. Hussain, S.; Anwar, S.M.; Majid, M. Brain Tumor Segmentation Using Cascaded Deep Convolutional Neural Network. In
Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC),
Seogwipo, Korea, 11–15 July 2017; pp. 1998–2001. [CrossRef]
J. Imaging 2021, 7, 19 20 of 22
49. Pereira, S.; Oliveira, A.; Alves, V.; Silva, C.A. On hierarchical brain tumor segmentation in MRI using fully convolutional neural
networks: A preliminary study. In Proceedings of the 2017 IEEE 5th Portuguese Meeting on Bioengineering (ENBENG), Coimbra,
Portugal, 16–18 February 2017; pp. 1–4. [CrossRef]
50. Sled, J.; Zijdenbos, A.; Evans, A. A nonparametric method for automatic correction of intensity nonuniformity in MRI data. IEEE
Trans. Med. Imaging 1998, 17, 87–97. [CrossRef]
51. Tustison, N.J.; Avants, B.B.; Cook, P.A.; Zheng, Y.; Egan, A.; Yushkevich, P.A.; Gee, J.C. N4ITK: Improved N3 Bias Correction.
IEEE Trans. Med. Imaging 2010, 29, 1310–1320. [CrossRef]
52. Nyul, L.; Udupa, J.; Zhang, X. New variants of a method of MRI scale standardization. IEEE Trans. Med. Imaging 2000, 19, 143–150.
[CrossRef]
53. Zhao, X.; Wu, Y.; Song, G.; Li, Z.; Zhang, Y.; Fan, Y. A deep learning model integrating FCNNs and CRFs for brain tumor
segmentation. Med. Image Anal. 2018, 43, 98–111. [CrossRef]
54. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation.
arXiv 2016, arXiv:1606.04797.
55. Vaidhya, K.; Thirunavukkarasu, S.; Alex, V.; Krishnamurthi, G. Multi-Modal Brain Tumor Segmentation Using Stacked Denoising
Autoencoders. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Menze, B.,
Maier, O., Reyes, M., Handels, H., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg,
Germany, 2016; pp. 181–194. [CrossRef]
56. Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Deep Convolutional Neural Networks for the Segmentation of Gliomas in Multi-
Sequence MRI. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Menze, B.,
Maier, O., Reyes, M., Handels, H., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg,
Germany, 2016; pp. 131–143. [CrossRef]
57. Hussain, S.; Anwar, S.M.; Majid, M. Segmentation of glioma tumors in brain using deep convolutional neural network.
Neurocomputing 2018, 282, 248–261. [CrossRef]
58. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. arXiv 2018, arXiv:1708.02002.
59. Sun, L.; Zhang, S.; Chen, H.; Luo, L. Brain Tumor Segmentation and Survival Prediction Using Multimodal MRI Scans with Deep
Learning. Front. Neurosci. 2019, 13, 810. [CrossRef] [PubMed]
60. Mlynarski, P.; Delingette, H.; Criminisi, A.; Ayache, N. Deep learning with mixed supervision for brain tumor segmentation.
J. Med. Imaging 2019, 6, 034002. [CrossRef] [PubMed]
61. Iqbal, S.; Ghani Khan, M.U.; Saba, T.; Mehmood, Z.; Javaid, N.; Rehman, A.; Abbasi, R. Deep learning model integrating features
and novel classifiers fusion for brain tumor segmentation. Microsc. Res. Tech. 2019, 82, 1302–1315. [CrossRef] [PubMed]
62. Kuzina, A.; Egorov, E.; Burnaev, E. Bayesian Generative Models for Knowledge Transfer in MRI Semantic Segmentation Problems.
Front. Neurosci. 2019, 13, 844. [CrossRef]
63. Kao, P.Y.; Ngo, T.; Zhang, A.; Chen, J.W.; Manjunath, B.S., Brain Tumor Segmentation and Tractographic Feature Extraction from
Structural MR Images for Overall Survival Prediction. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries;
Crimi, A., Bakas, S., Kuijf, H., Keyvan, F., Reyes, M., van Walsum, T., Eds.; Springer International Publishing: Berlin/Heidelberg,
Germany, 2019; Volume 11384, pp. 128–141. [CrossRef]
64. Kao, P.Y.; Shailja, F.; Jiang, J.; Zhang, A.; Khan, A.; Chen, J.W.; Manjunath, B.S. Improving Patch-Based Convolutional Neural
Networks for MRI Brain Tumor Segmentation by Leveraging Location Information. Front. Neurosci. 2020, 13. [CrossRef]
65. Nalepa, J.; Marcinkiewicz, M.; Kawulok, M. Data Augmentation for Brain-Tumor Segmentation: A Review. Front. Comput.
Neurosci. 2019, 13, 83. [CrossRef]
66. Shin, H.C.; Tenenholtz, N.A.; Rogers, J.K.; Schwarz, C.G.; Senjem, M.L.; Gunter, J.L.; Andriole, K.; Michalski, M. Medical Image
Synthesis for Data Augmentation and Anonymization Using Generative Adversarial Networks. arXiv 2018, arXiv:1807.10225.
67. Han, C.; Rundo, L.; Araki, R.; Nagano, Y.; Furukawa, Y.; Mauri, G.; Nakayama, H.; Hayashi, H. Combining Noise-to-Image and
Image-to-Image GANs: Brain MR Image Augmentation for Tumor Detection. IEEE Access 2019, 7, 156966–156977. [CrossRef]
68. Han, C.; Murao, K.; Noguchi, T.; Kawata, Y.; Uchiyama, F.; Rundo, L.; Nakayama, H.; Satoh, S. Learning More with Less:
Conditional PGGAN-Based Data Augmentation for Brain Metastases Detection Using Highly-Rough Annotation on MR Images.
In Proceedings of the 28th ACM International Conference on Information and Knowledge Management, Beijing China, 3–7
November 2019; pp. 119–127. [CrossRef]
69. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial
Networks. arXiv 2014, arXiv:1406.2661.
70. Simpson, A.L.; Antonelli, M.; Bakas, S.; Bilello, M.; Farahani, K.; van Ginneken, B.; Kopp-Schneider, A.; Landman, B.A.; Litjens, G.;
Menze, B.; et al. A Large Annotated Medical Image Dataset for the Development and Evaluation of Segmentation Algorithms.
arXiv 2019, arXiv:1902.09063.
71. Team, T.T.D.; Al-Rfou, R.; Alain, G.; Almahairi, A.; Angermueller, C.; Bahdanau, D.; Bastien, F.; Bayer, J.; Belikov, A.; Belopolsky,
A.; et al. Theano: A Python Framework for Fast Computation of Mathematical Expressions. arXiv 2016, arXiv:1605.02688.
72. Goodfellow, I.J.; Warde-Farley, D.; Lamblin, P.; Dumoulin, V.; Mirza, M.; Pascanu, R.; Bergstra, J.; Bastien, F.; Bengio, Y. Pylearn2:
A Machine Learning Research Library. arXiv 2013, arXiv:1308.4214.
73. Jia, Y.; Shelhamer, E.; Donahue, J.; Karayev, S.; Long, J.; Girshick, R.; Guadarrama, S.; Darrell, T. Caffe: Convolutional Architecture
for Fast Feature Embedding. arXiv 2014, arXiv:1408.5093.
J. Imaging 2021, 7, 19 21 of 22
74. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch:
An Imperative Style, High-Performance Deep Learning Library. arXiv 2019, arXiv:1912.01703.
75. Abadi, M.; Agarwal, A.; Barham, P.; Brevdo, E.; Chen, Z.; Citro, C.; Corrado, G.S.; Davis, A.; Dean, J.; Devin, M.; et al. Tensorflow:
Large-Scale Machine Learning on Heterogeneous Distributed Systems. arXiv 2016, arXiv:1603.04467.
76. Chollet, F. Keras: The Python Deep Learning API. 2020. Available online: https://keras.io/ (accessed on 1 June 2020).
77. Zhang, J.; Shen, X.; Zhuo, T.; Zhou, H. Brain tumor segmentation based on refined fully convolutional neural networks with a
hierarchical dice loss. arXiv 2017, arXiv:1712.09093.
78. Kayalibay, B.; Jensen, G.; Smagt, P.V.D. CNN-based segmentation of medical imaging data. arXiv 2017, arXiv:1701.03056.
79. Kamnitsas, K.; Bai, W.; Ferrante, E.; McDonagh, S.; Sinclair, M.; Pawlowski, N.; Rajchl, M.; Lee, M.; Kainz, B.; Rueckert, D.; et al.
Ensembles of Multiple Models and Architectures for Robust Brain Tumour Segmentation. arXiv 2017, arXiv:1711.01468.
80. Dong, H.; Yang, G.; Liu, F.; Mo, Y.; Guo, Y. Automatic Brain Tumor Detection and Segmentation Using U-Net Based Fully
Convolutional Networks. In Proceedings of the Medical Image Understanding and Analysis; Valdés Hernández, M., González-Castro,
V., Eds.; Communications in Computer and Information Science; Springer International Publishing: Berlin/Heidelberg, Germany,
2017; pp. 506–517. [CrossRef]
81. Alex, V.; Safwan, M.; Krishnamurthi, G. Automatic Segmentation and Overall Survival Prediction in Gliomas Using Fully
Convolutional Neural Network and Texture Analysis. arXiv 2017, arXiv:1712.02066.
82. Erden, B.; Gamboa, N.; Wood, S. 3D Convolutional Neural Network for Brain Tumor Segmentation; Technical Report; Computer
Science, Stanford University: Stanford, CA, USA, 2017.
83. Isensee, F.; Kickingereder, P.; Wick, W.; Bendszus, M.; Maier-Hein, K.H. Brain Tumor Segmentation and Radiomics Survival
Prediction: Contribution to the BRATS 2017 Challenge. arXiv 2018, arXiv:1802.10508.
84. Meng, Z.; Fan, Z.; Zhao, Z.; Su, F. ENS-Unet: End-to-End Noise Suppression U-Net for Brain Tumor Segmentation. In Proceedings
of the 2018 40th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Honolulu,
HI, USA, 18–21 July 2018; pp. 5886–5889. [CrossRef]
85. Liu, J.; Chen, F.; Pan, C.; Zhu, M.; Zhang, X.; Zhang, L.; Liao, H. A Cascaded Deep Convolutional Neural Network for Joint
Segmentation and Genotype Prediction of Brainstem Gliomas. IEEE Trans. Bio-Med. Eng. 2018, 65, 1943–1952. [CrossRef]
[PubMed]
86. Pereira, S.; Pinto, A.; Amorim, J.; Ribeiro, A.; Alves, V.; Silva, C.A. Adaptive feature recombination and recalibration for semantic
segmentation with Fully Convolutional Networks. IEEE Trans. Med. Imaging 2019. [CrossRef] [PubMed]
87. Kermi, A.; Mahmoudi, I.; Khadir, M.T. Deep Convolutional Neural Networks Using U-Net for Automatic Brain Tumor
Segmentation in Multimodal MRI Volumes. In Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi,
A., Bakas, S., Kuijf, H., Keyvan, F., Reyes, M., van Walsum, T., Eds.; Lecture Notes in Computer Science; Springer International
Publishing: Berlin/Heidelberg, Germany, 2019; pp. 37–48. [CrossRef]
88. Isensee, F.; Kickingereder, P.; Wick, W.; Bendszus, M.; Maier-Hein, K.H. No New-Net. arXiv 2019, arXiv:1809.10483.
89. Wang, L.; Wang, S.; Chen, R.; Qu, X.; Chen, Y.; Huang, S.; Liu, C. Nested Dilation Networks for Brain Tumor Segmentation Based
on Magnetic Resonance Imaging. Front. Neurosci. 2019, 13, 285. [CrossRef]
90. Ribalta Lorenzo, P.; Nalepa, J.; Bobek-Billewicz, B.; Wawrzyniak, P.; Mrukwa, G.; Kawulok, M.; Ulrych, P.; Hayball, M.P.
Segmenting brain tumors from FLAIR MRI using fully convolutional neural networks. Comput. Methods Programs Biomed. 2019,
176, 135–148. [CrossRef]
91. Jiang, Z.; Ding, C.; Liu, M.; Tao, D. Two-Stage Cascaded U-Net: 1st Place Solution to BraTS Challenge 2019 Segmentation Task.
In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Bakas, S., Eds.; Lecture
Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 231–241. [CrossRef]
92. Zhao, Y.X.; Zhang, Y.M.; Liu, C.L. Bag of Tricks for 3D MRI Brain Tumor Segmentation. In Brainlesion: Glioma, Multiple Sclerosis,
Stroke and Traumatic Brain Injuries; Lecture Notes in Computer Science; Crimi, A., Bakas, S., Eds.; Springer International Publishing:
Berlin/Heidelberg, Germany, 2020; pp. 210–220. [CrossRef]
93. Zhuge, Y.; Krauze, A.V.; Ning, H.; Cheng, J.Y.; Arora, B.C.; Camphausen, K.; Miller, R.W. Brain tumor segmentation using
holistically nested neural networks in MRI images. Med. Phys. 2017, 44, 5234–5243. [CrossRef]
94. Liu, Y.; Stojadinovic, S.; Hrycushko, B.; Wardak, Z.; Lau, S.; Lu, W.; Yan, Y.; Jiang, S.B.; Zhen, X.; Timmerman, R.; et al. A deep
convolutional neural network-based automatic delineation strategy for multiple brain metastases stereotactic radiosurgery.
PLoS ONE 2017, 12, e0185844. [CrossRef]
95. Li, Z.; Wang, Y.; Yu, J.; Guo, Y.; Cao, W. Deep Learning based Radiomics (DLR) and its usage in noninvasive IDH1 prediction for
low grade glioma. Sci. Rep. 2017, 7, 5467. [CrossRef]
96. Kamnitsas, K.; Chen, L.; Ledig, C.; Rueckert, D.; Glocker, B. Multi-Scale 3D Convolutional Neural Networks for Lesion
Segmentation in Brain MRI. Ischemic Stroke Lesion Segm. 2015, 13, 46.
97. Hoseini, F.; Shahbahrami, A.; Bayat, P. AdaptAhead Optimization Algorithm for Learning Deep CNN Applied to MRI
Segmentation. J. Digit. Imaging 2019, 32, 105–115. [CrossRef]
98. Naceur, M.B.; Saouli, R.; Akil, M.; Kachouri, R. Fully Automatic Brain Tumor Segmentation using End-To-End Incremental Deep
Neural Networks in MRI images. Comput. Methods Programs Biomed. 2018, 166, 39–49. [CrossRef] [PubMed]
99. Yi, D.; Zhou, M.; Chen, Z.; Gevaert, O. 3-D convolutional neural networks for glioblastoma segmentation. arXiv 2016,
arXiv:1611.04534.
J. Imaging 2021, 7, 19 22 of 22
100. Cui, S.; Mao, L.; Jiang, J.; Liu, C.; Xiong, S. Automatic Semantic Segmentation of Brain Gliomas from MRI Images Using a Deep
Cascaded Neural Network. J. Healthc. Eng. 2018, 2018, 4940593. [CrossRef] [PubMed]
101. Yang, T.; Ou, Y.; Huang, T. Automatic Segmentation of Brain Tumor from MR Images Using SegNet: Selection of Training Data
Sets. In Proceedings of the 6th MICCAI BraTS Challenge, Quebec City, QC, Canada, 14 September 2017; pp. 309–312.
102. McKinley, R.; Meier, R.; Wiest, R. Ensembles of Densely-Connected CNNs with Label-Uncertainty for Brain Tumor Segmentation.
In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Bakas, S., Kuijf, H., Keyvan,
F., Reyes, M., van Walsum, T., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg,
Germany, 2019; pp. 456–465. [CrossRef]
103. Zhou, C.; Chen, S.; Ding, C.; Tao, D. Learning Contextual and Attentive Information for Brain Tumor Segmentation. In Proceedings
of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Bakas, S., Kuijf, H., Keyvan, F., Reyes, M.,
van Walsum, T., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2019;
pp. 497–507. [CrossRef]
104. McKinley, R.; Rebsamen, M.; Meier, R.; Wiest, R. Triplanar Ensemble of 3D-to-2D CNNs with Label-Uncertainty for Brain Tumor
Segmentation. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries; Crimi, A., Bakas, S.,
Eds.; Lecture Notes in Computer Science; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 379–387.
[CrossRef]
105. Bakas, S.; Reyes, M.; Jakab, A.; Bauer, S.; Rempfler, M.; Crimi, A.; Shinohara, R.T.; Berger, C.; Rozycki, M.; Prastawa, M.; et al.
Identifying the Best Machine Learning Algorithms for Brain Tumor Segmentation, Progression Assessment, and Overall Survival
Prediction in the BRATS Challenge. arXiv 2019, arXiv:1811.02629.
106. Ji, Z.; Shen, Y.; Ma, C.; Gao, M. Scribble-Based Hierarchical Weakly Supervised Learning for Brain Tumor Segmentation. arXiv
2019, arXiv:1911.02014.
107. Pavlov, S.; Artemov, A.; Sharaev, M.; Bernstein, A.; Burnaev, E. Weakly Supervised Fine Tuning Approach for Brain Tumor
Segmentation Problem. arXiv 2019, arXiv:1911.01738.
108. Wu, K.; Du, B.; Luo, M.; Wen, H.; Shen, Y.; Feng, J. Weakly Supervised Brain Lesion Segmentation via Attentional Representation
Learning. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2019; Shen, D., Liu, T.,
Peters, T.M., Staib, L.H., Essert, C., Zhou, S., Yap, P.T., Khan, A., Eds.; Lecture Notes in Computer Science; Springer International
Publishing: Berlin/Heidelberg, Germany, 2019; pp. 211–219. [CrossRef]