Paper 1-Deep Neural Network Based Methods For Brain Image
Paper 1-Deep Neural Network Based Methods For Brain Image
Abstract—Various types of noise may affect the visual quality Although the results were satisfactory, the researchers did
of images during capturing and transmitting procedures. Finding not want to stop at this stage, and they wanted to go beyond
a proper technique to remove the possible noise and improve the conventional gaussian denoisers. In this regard,
both quantitative and qualitative results is always considered as convolutional neural networks are widely used in image
one of the most important and challenging pre -processing tasks processing due to their excellent performance for obtaining
in image and signal processing. In this paper, we made a short high quality output images. Jian and Seung developed a
comparison between two well-known approaches called combined CNN with unsupervised learning for natural image
thresholding neural network (TNN) and deep neural network de-noising [7]. Vincent et al. developed a new training
(DNN) based methods for image de-noising. De-noising results of
TNNs, Dn-CNNs, Flashlight CNN (FLCNN) and Diamond de - principle for unsupervised learning and it became one of the
noising networks (DmDN) have been compared with each other. basic deep learning techniques for noise removal aspects [8].
In this regard, several experiments have been performed in terms While using deep convolutional neural networks there is an
of Peak S ignal to Noise Ratio (PS NR) to validate the performance issue in which we cannot train deeper networks easily. To
analysis of various de-noising methods. The analysis indicates address this problem, Mao et al. proposed symmetric skip
that DmDNs perform better than other learning-based connection combined with auto-encoders [9]. Zhang in [10]
algorithms for de-noising brain MR images. DmDN achieved a proposed a Dn-CNN method consisting of two main stages,
PS NR value of 29.85 dB, 30.74 dB, 29.15 dB, and 29.45 dB for de- residual learning and batch normalization. Deeper networks
noising MR image 1, MR image 2, MR image 3 and MR Image 4, also cause gradient dispersion in which residual learning has
respectively for a standard deviation of 15. been utilized in Dn-CNNs to tackle this issue [11]. There are
also some other issues which deep neural network-based
Keywords—CNN; Deep neural network; de-noising; MR methods are suffering from. One is diminishing feature reuse,
image; PSNR and the other is that increasing the number of parameters and
layers does not have any advantage for them [12]. To address
I. INT RODUCT ION these issues Bin et al. developed a flashlight CNN method
Noise is considered as unwanted signals causing based on deep residual and inception networks that is able to
imperfections and low resolution in image and signal hold many parameters [12]. Additionally, J. Zhang in [11]
processing, and may happen during the receiving and developed a diamond denoiser to deal with the issue of losing
transmitting processes. Thus, further image analysis and network’s gradient caused by deeper networks.
processing may not be possible until we discard or reduce the
noise in the images. In image de-noising, the main goal is A self-supervised based method for fluorescence image
enhancing the visual quality. Various methods are available in denoising has been proposed by Huang et al., [16]. In this
literature for removing the possible noise from images. approach, the authors utilized Wiener filtering and wavelet
transformation, as two classic denoising techniques as well as
Donoho and Johnstone proposed adapting to unknown DeepCAD to perform comparative experiments [16]. In
smoothness [1] and ideal spatial adaptation [2] using wavelet another study conducted by Yang et al. [17], an efficient auto-
shrinkage for de-noising in 1994 and 1995, respectively. encoder technique using convolutional neural networks to
These techniques became the foundation for further gradient perform both classification and de-noising has been
descent learning based methods. Zhang took one step forward developed.
in de-noising by proposing a learning-based method for
improving the conventional approaches [3]. He developed a Content-noise complementary learning has been presented
thresholding neural network using an improved and non -linear in [18] to denoise medical images. In this study to validate the
hard-soft threshold function. Sahraeian et al., proposed an performance of various de-noising methods, MR, CT, and
improved TNN and cycle spinning for image de-noising [4]. PET images have been utilized. Structural priors based deep
Nasri and Nezamabadipour tried to improve Zhang’s results MRI super resolution has been developed in a study conducted
by proposing another data driven function with three shape by Cherukuri et al., [19]. Low rank structure and sharpness
tuning parameters [5]. To enhance the results of TNN based priors have been utilized in this study to enhance the visual
methods, instead of using gradient descent algorithm, the quality of images. Convolutional de-noising autoencoders to
authors in [6] proposed an optimized based technique. discard noise from MR images has also been proposed in [20].
This technique provided better accuracy with less computation
and data for de-noising the medical images.
1|Pag e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 2, 2024
In this paper we have a brief survey on several state-of- One of the methods we discuss however uses a deep feed -
the-art de-noising approaches. We analyzed the results for forward network, which can not only learn with overall
MRI brain image de-noising. Thresholding neural networks, smaller sample sizes but uses residual learning. It trains on
Dn-CNNs, Flashlight CNNs, and Diamond de-noising images that already have noise and learns from it, working
networks have been taken into account. The results indicate along with batch normalization in order to increase its
that deep neural network based methods have superior results accuracy [23]. In the case of the flashlight CNN, it uses a very
compared to TNN based techniques. Among these deep neural similar strategy, while also using inception layers that help the
network based approaches, Diamond de-noising networks network better handle Gaussian white noise. Fig. 1 shows the
(DmDN) perform well, followed closely by FLCNN, and main procedure of de-noising using learning based
DnCNN. approaches. Images have been obtained from [21].
The rest of the paper is organized as follows: Section II is III. TNN BASED M ET HODS
about CNN based image de-noising. A brief discussion about
CNNs and how to perform CNN based de-noising has been Standard hard and soft thresholding functions were first
provided. In Section III, we discuss image de-noising using proposed in [3]. In this case, these functions became the basis
thresholding neural network. In Section IV, we discuss several and foundation of further thresholding based de-noising. Since
deep neural network methods. Section V is results and the obtained results using these functions were not
discussion. Finally, Section VI concludes the paper. satisfactory, the researchers in the fields of image and signal
processing attempted to enhance these methods and add more
II. DE-NOISING USING CNN parameters to make them non-linear and differentiable to be
used in a network called, “thresholding neural network’’.
Sitting as a contrast from more traditional methods, These functions which are the enhanced version of standard
convolutional neural networks can be used to great effect on
thresholds are called “improved thresholding functions’’
de-noising images. CNNs have been the neural network of which were first introduced by Zhang [3]. The equations
choice in the field of image processing due to their high
below indicate these improved soft and improved threshold
effectiveness and can also be used when de-noising. These
functions:
networks use their convolutional layers. There are multiple
different methods regarding deep learning, but the ones that 1
we discuss in this paper are feed-forward convolutional neural L (u, ) u ( (u - )2 l - (u )2 - l )
soft 2 (1)
networks (DnCNN) and flashlight CNNs (FLCNN).
In order to de-noise an image, CNNs traditionally require a where, Lsoft (u, ) denotes the non-linear soft threshold,
large training sample size, and learns by training with input- u is the WT components, is the threshold value and l 0
output pairs, images of noisy scans, followed by its clean is a function parameter (user defined) [3].
variation. The network learns kernels through its
1 1
convolutional layers, small weights that can detect patterns Lhard (u, ) ( - 1)u
over the input image. The convolutional layers create a -u -u -
1 exp 1 exp
hierarchical representation of the input and can use this (2)
separation to learn to differentiate between the noise and the
LS (u, )
( u he ).sgn(u ), u
-n u
(3)
2|Pag e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 2, 2024
capability. Thereafter, Nasri and Nezamabadi-pour [5] in the proposed network with depth D, there are three types of
presented other nonlinear functions with three shape tuning layers [10]:
parameters which are formulated.
Conv+ReLU is used for the first layer with 64 filters
i g with the size 3×3×c. Note that c is the channels’
u - 0.5 u i -1 ( g -1) , u number. Also, ReLU has been utilized to give
nonlinearity.
g u j
(u, , i, j , g ) 0.5 j -1 sgn(u ) , u Conv+BN+ReLU is used from layer 2-D-1 with 64
filters of size 3×3×64. Batch normalization (BN) has
(- )i g also been used in these layers.
u 0.5 - ( g -1) , u -
u i -1
(4) Conv is utilized in the very last layer with c filter of
where, is the threshold value, u denotes the WT
size 3×3×64 for reconstructing the output image.
Conv+BN+ReLU
Conv+BN+RelU
Residual Image
Conv+NB+Relu
Conv+ReLU
noising, please refer to [3].
Conv
…. ….
IV. DEEP LEARNING BASED M ET HODS
A. DnCNN
Nowadays, due to the availability of large-scale datasets
and progress in deep learning algorithms, CNN approaches
attract lots of attention in imaging technologies [10]. The Fig. 2. T he structure of DnCNN [10].
construction of feed-forward convolutional neural networks
(DnCNNs) for de-noising has become the basis for de-noising B. Flashlight CNN (FLCNN)
using deep learning [10]. In this structure, to improve the Flashlight CNNs are another type of convolutional neural
computational time and also to enhance the quality of the de- network implementing deep NN for noise removal processes.
noised image, batch normalization and residual learning have The main structure of this method is based on the combination
been utilized, leading to this approach becoming one of the of deep residual and inception networks [12]. Utilizing
more efficient and effective gaussian denoisers. Conventional inception layers provides us with overcoming and addressing
deep NNs can estimate a clean image directly, but DnCNNs the reuse of diminishing features while tackling additive white
can remove and discard the clean image by adapting it to the gaussian noise. As shown in Fig. 3, this network consists of
residual learning strategy [10]. Training a single DnCNN as a two main phases [12]:
blind gaussian denoiser gives better results compared to
alternative methods. As mentioned earlier, residual learning Warmup phase which utilizes convolutional layers
and batch normalization are used in this structure. Residual (typical or conventional CNN). There are two main
learning has been utilized for solving performance degradation stages in this phase. The first one employs 3×3 kernels
issues [14]. with 64 features and the second one employs 5×5
kernels with 64 features.
The developed DnCNN utilizes only one residual unit for
predicting the residual image [10]. If we compare residual Boost phase utilizes wider inception layers (residual)
mapping with the original unreferenced mapping in terms of leading to growth and increment in the number of
learning, residual mapping is easier, so deep CNN models can networks’ parameters while overcoming the reducing
be trained easily [14] [10]. On the other hand, although feature reuse.
training based on stochastic gradient descent (SGD) is
Warm-up Phase Boost Phase
effective and simple, internal covariance shifts can largely
reduce the training efficiency [15] [10]. So, alleviating the Conv Conv Conv Conv estimate x
y Incept -
covariance shift is also a challenging task in deep CNN 3× 3 3× 3 3× 3
A
3× 3 +
models and is the reason that batch normalization is used in 64 64 64 1
+
these networks [15] [10]. The combination of residual learning 1 layers m layers n layers
and batch normalization provides us with stable training, fast
training procedure (because of using batch normalization),
better qualitative and quantitative results [10]. The main Fig. 3. T he architecture of FLCNN with noisy input of y and estimate x [12].
structure of the DnCNN model is depicted in Fig. 2.
C. Diamond De-noising Network (DmDN)
As can be seen, the network’s input is a noisy image
corrupted by gaussian noise. Here, instead of learning a Images’ detail and important characteristics and
mapping function, we can proceed by adapting residual information may be diminished by doing excessive scaling
learning for training the residual mapping [10]. Additionally, [11]. Although the convolutional network is deeper, it may be
easy to lose the gradient of the network. To address these
issues, Diamond Shaped (DS) multi-scale feature extraction
3|Pag e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 2, 2024
4|Pag e
www.ijacsa.thesai.org
(IJACSA) International Journal of Advanced Computer Science and Applications,
Vol. 15, No. 2, 2024
5|Pag e
www.ijacsa.thesai.org