VCEH B Tech Project Report
VCEH B Tech Project Report
BACHELOR OF TECHNOLOGY
IN
Submitted by
SUPERVISOR
K.ASHWINI
Ass.professor
March, 2023
Department of Electronics and Communication Engineering
CERTIFICATE
Examiner
We avail this opportunity to express our deep sense of gratitude and heart-
ful thanks to Dr. Teegala Vijender Reddy, Chairman and Sri Teegala
Upender Reddy, Secretary of VCE, for providing a congenial atmosphere to
complete this project successfully.
T.SAI TEJA
J.RAVALI
A.NIKHIL
ii
Abstract
iv
Table of Contents
vi
3.11.1 Test strategy . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.11.2 Test data . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.11.3 Test plans . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.11.4 Test scenarios . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.11.5 Test cases . . . . . . . . . . . . . . . . . . . . . . . . . . . . 28
3.11.6 Traceability Matrix . . . . . . . . . . . . . . . . . . . . . . . 28
3.12 Unit Testing . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
CHAPTER 4 Network Architecture . . . . . . . . . . . . . . . . . . 30
4.1 Proposed System . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
4.1.1 Advantages of Proposed System . . . . . . . . . . . . . . . 31
4.2 Existing System . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
4.3 Drawbacks of Existing System . . . . . . . . . . . . . . . . . . . . 33
4.4 System Block Diagram . . . . . . . . . . . . . . . . . . . . . . . . . 34
4.4.1 Labeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 34
4.5 Image Segmentation using Unet . . . . . . . . . . . . . . . . . . . 34
4.5.1 Process of Image segmentation using Unet . . . . . . . . . 35
4.6 Feature Extraction using ResNext50 . . . . . . . . . . . . . . . . . 36
4.7 Image Retrieving using Convolution Neural Network . . . . . . . 37
4.8 Image Enhancement . . . . . . . . . . . . . . . . . . . . . . . . . . 38
4.9 Thresholding . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 39
4.10 Tumor classification using CNN . . . . . . . . . . . . . . . . . . . 40
4.11 Pooling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 41
4.12 Advantages of Implemented Algorithm . . . . . . . . . . . . . . . . 41
CHAPTER 5 Result . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
5.1 Experiment results . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
5.1.1 Bar Graph of the Data set . . . . . . . . . . . . . . . . . . 45
CHAPTER 6 Conclusions and Future Scope . . . . . . . . . . . . . 49
6.1 Conclusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 49
6.2 Future scope . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 50
REFERENCES . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 51
List of Tables
viii
List of Figures
ix
Abbreviations
Abbreviation Description
Introduction
1
desktop computer is difficult. Transfer learning is used to overcome this issue
by implementing popular CNN architectures with pre-trained weights and minor
model changes to fit the dataset. Image denoising, data augmentation, and
segmentation methods are used for both traditional machine learning classifiers
and sophisticated CNN models to improve prediction accuracy without human
involvement.Interpreting medical images such as CT and MRI requires extensive
training and ability because organ and lesion segmentation must be done layer
by layer. Manual segmentation places a significant burden on doctors, which
can introduce bias if subjective views are involved. To analyze complicated
images, doctors must typically make a joint diagnosis, which takes time.
Furthermore, automatic segmentation remains a difficult and unsolved issue.
1.2 Motivation
Creating exact outlines of the tumor areas is the goal of brain tumor
segmentation. In tackling a variety of computer vision tasks, such as image
classification, object detection, and semantic segmentation, deep learning meth-
ods have shown remarkable results. These methods can now be used to more
accurately and quickly identify brain lesions at an early stage. Additionally,
this technique automates image processing and analysis, makes it easier for
scientists to identify different types of brain structures, and enhances diagnosis.
1.5 Objective
The goal of this research is to create a semantic similarity method for
retrieving images of brain tumors. The suggested method combines a number
of deep learning methods, including Unet for segmenting images, ResNeXt50 for
extracting features, and CNN for retrieving images. Based on their semantic
content, the system seeks to offer a quick and accurate method of retrieving
images of brain tumors that are comparable. The ability to access a sizable
database of comparable images and compare them to the patient’s MRI scans
could possibly help medical professionals diagnose and treat brain tumors
more successfully. The overall objective is to support the creation of more
sophisticated and intelligent medical imaging systems for improved patient
treatment.
Chapter 5: Result - This chapter gives the details about the simulation results
using Python Language.
Literature Survey
There are many researchers which discuss MRI image detection methods
with the help of deep learning.Neelum Noreen[1] et al. A combination method
was used to develop a deep learning model for detecting tumors in the brain.
To enhance diagnostic precision, the model was developed to integrate both
MRI and CT images. The authors’ findings were encouraging, showing that
the model they developed outperformed conventional machine learning models
in terms of accuracy, sensitivity, and specificity. The suggested model has the
ability to transform brain tumor evaluation, allowing for early detection and
more efficient therapy. More study is required, however, to verify the model’s
efficacy on bigger data sets and in clinical situations.
Abdu Gumaei[2] et al. For tumors in the brain categorization, a combi-
nation of extraction of features technique coupled with a regularized extreme
learning machine was suggested. To extract useful information from images
generated by MRI, the technique used a mix of wavelet-based features and
gray-level co-occurrence matrix features. The study’s authors reported encour-
aging findings, showing that their suggested approach outperformed several
traditional machine learning models in terms of accuracy. The suggested ap-
proach has the ability to enhance brain tumor categorization accuracy and
speed, resulting in early discovery and more effective therapy. However, ad-
ditional validation of the technique is required to guarantee its efficacy in
clinical situations.
Hossam H. Sultan [3]et al. A deep neural network model for multi-
classification of brain tumor pictures was suggested. The model extracted
pertinent characteristics from magnetic resonance imaging (MRI) pictures using
convolutional neural networks and obtained high accuracy in categorizing brain
tumor photos into various kinds. The authors’ findings were encouraging,
showing that their suggested model outperformed several traditional machine
6
learning models. The suggested model has the ability to speed up and increase
the precision of brain tumor categorization, resulting in early discovery and
improved therapy. However, additional validation of the model is required to
guarantee its efficacy in clinical situations.
Yun-Qian Li [4]et al.PS-OCT (polarization-sensitive optical coherent scan-
ning) has been suggested for identifying brain tumors. The technique dis-
tinguishes between normal and abnormal tissues by utilizing the birefringent
characteristics of tumor tissues. The authors published encouraging findings,
showing that their proposed PS-OCT method can detect cancers of the brain
a particularly prevalent form of brain tumor. The suggested technique has the
ability to increase brain tumor diagnostic precision, allowing for early detection
and more efficient treatment. However, additional validation of the technique
is required to guarantee its efficacy in clinical situations.
Stefan T. Lang [5]et al. A mathematical modeling research was carried
out to examine the effect of peritumoral edema during tumor treatment field
therapy. A mathematical model was used in the research to mimic the electric
field dispersion and tumor reaction to treatment. The authors presented
encouraging findings, showing that the existence of peritumoral edema can
have a substantial impact on the efficacy of tumor treatment field therapy.
The suggested computational model has the ability to enhance cancer field
therapeutic design and optimizing, resulting in more successful results from
treatment. However, additional validation of the model is required to guarantee
its efficacy in clinical situations.
Jianxin Zhang [6]et al.An Attention Gate ResU-Net model for automated
MRI neural tumor separation was suggested. To correctly separate brain
tumor areas in MRI pictures, the algorithm used a mix of residual links,
U-Net design, and focus gates. The authors revealed encouraging findings,
showing the fact that the model they suggested outperformed several state-of-
the-art segmentation algorithms in terms of accuracy. The suggested model has
the ability to speed up and increase the precision of brain tumor segmentation,
resulting in early discovery and more effective therapy. However, additional
validation of the model is required to guarantee its efficacy in clinical situations.
Methodology
17
of the project depends on both the quantity and quality of the data used
to train and evaluate the models. To obtain the necessary data, the Cancer
Imaging Archive (TCIA), a publicly available resource of medical imaging data,
including brain MRI scans, is the primary source. Additional sources of data
include hospital and study center archives, where patient data is gathered
and stored for medical diagnosis and research. Prior to using the images
for analysis, pre-processing techniques are applied to ensure accuracy and
consistency. To enhance the quality and clarity of the images, pre-processing
methods such as noise reduction, normalization, and contrast enhancement are
used.”
Data collection is a crucial first step in any research project, and this applies
to retrieving brain tumor images using semantic-based similarity methods as
well. The quality and quantity of data used for training and evaluating the
models are critical to the success of the project. The Cancer Imaging Archive
(TCIA) provides a large collection of publicly available medical imaging data,
including brain MRI scans, and serves as the primary source of data for
this project. Additional sources of data include hospital and study center
archives where patient data is collected and saved for medical diagnosis and
research. Pre-processing methods such as noise reduction, normalization, and
contrast enhancement are utilized to ensure accuracy and uniformity in the
project’s images. Medical professionals then mark the pre-processed images
to identify the regions of interest (ROIs) containing the tumor. Annotations
3.3 Pre-processing
Pre-processing refers to a set of operations that are performed on raw data
to prepare it for further analysis. In the context of medical image analysis,
preprocessing typically involves a series of steps to improve the quality of the
images, remove noise and artifacts, and enhance the contrast between different
tissues or structures.
The preprocessing steps for brain tumor images typically include the fol-
lowing:
3.8 Evaluation
Evaluation is a critical component of any machine learning project to assess
the accuracy and effectiveness of the proposed model. In this project, we will
evaluate the performance of a semantic-based similarity technique for retrieving
brain tumor images using various metrics. The evaluation metrics used in this
project will include precision, recall, and F1 score. Precision measures the
number of relevant images among the retrieved images, while recall measures
the proportion of relevant images that were successfully retrieved. The F1
score is the harmonic mean of precision and recall. Furthermore, we will
plot the receiver operating characteristic (ROC) curve to evaluate the model’s
performance in terms of true positive rate (TPR) and false positive rate
(FPR). We will also calculate the area under the ROC curve (AUC), which
provides a comprehensive measure of the model’s performance.
To assess the effectiveness of the proposed technique, we will compare it
to other state-of-the-art image retrieval techniques using the same dataset.
The comparison will be based on the metrics mentioned above. Overall, the
evaluation process will provide valuable insights into the effectiveness of the
proposed technique and its potential for enhancing the accuracy of brain tumor
image retrieval.
3.10 Deployment
The deployment stage involves integrating the developed model into an
application or platform where it can be used to serve its intended purpose.
In the case of this project, the deployment stage would involve integrating
the developed image retrieval model into a software application or platform
that can be used by medical professionals to retrieve relevant brain tumor
images based on semantic similarity. The deployment process may involve
several steps such as converting the model to an appropriate format that can
be easily integrated into the application, setting up a suitable server or cloud
environment to host the model, and creating an interface that allows users to
interact with the application and retrieve the desired images.
The deployment stage also involves testing the application to ensure that
it is functioning properly and meeting the required performance criteria. This
may involve testing the application on a small scale initially and gradually
Network Architecture
30
technique to train the Convolutional Neural Network.Trained over millions of
images which not only makes the learning process faster but also makes it
more accurate while predicting while in UNet, learning new tasks is based on
previously taught tasks.Call backs can help with faster error resolution and the
creation of better models. They can help you see how your model’s training
is progressing and avoid overfitting by introducing early stops or changing the
learning rate for each iteration. In this case, we’ll use the Model Checkpoint
and Reduce LR On Plateau callback methods. We will use Global Average
Pooling D instead of MaxPooling to create a second pooling layer while using
transfer learning.
4.4.1 Labeling
The act of giving a class or group to a data sample is known as labeling.
Labeling in the context of image analysis entails identifying and giving a name
to each pixel or region in an image based on the class or group to which
it belongs.This process is typically done manually by human annotators or
using automated techniques, such as computer vision algorithms. Labeling is
an important step in supervised machine learning, where labeled data is used
to train a model to make accurate predictions on new, unseen data.Labeling
is critical in the context of medical image analysis for jobs such as tumor
segmentation, where precise labeling is required to guarantee effective diagnosis
and therapy planning.Proper labeling also helps in building robust and accurate
machine learning models for medical imaging applications.
expansive path to access low-level features from the contracting path. This
helps the model to better preserve spatial information and capture fine details
in the segmented regions.During training, the U-Net model is optimized using
a loss function that calculates the disparity between the expected and true
segmentation maps. The Dice coefficient loss function is the most frequently
used loss function for picture segmentation tasks.In conclusion, the U-Net design
is a strong deep learning model for image segmentation tasks, especially in
medical image analysis. Its capacity to catch fine features while preserving
spatial information makes it ideal for jobs like tumor segmentation in brain
MRI pictures.
Figure 4.4: How does the model Faster R-CNN ResNet 50 work
4.9 Thresholding
Thresholding is a fundamental method for image segmentation, which is
a nonlinear process that converts a gray scale image into a binary image by
assigning two levels to pixels based on a predetermined threshold value. In
Open CV, the cv2.threshold() method is used for thresholding, which takes in
a single-channel matrix and performs fixed-level thresholding. This function is
commonly used to create a binary image from a gray scale image to remove
noise, such as filtering out pixels with excessively small or large values.
The ”maxval” parameter is the predefined threshold value used to compare
input values. When the input value is greater than the predefined threshold
value, the output is set to the predefined maxval value. When the intensity
levels of the incoming pixels are below the threshold, the result is dark. The
tool supports several thresholding methods, and it returns both the calculated
threshold value and the thresholded image.
4.11 Pooling
The Pooling layer is an important component in Convolutional Neural
Networks (CNNs) as it reduces the spatial size of the convolved features
while retaining their important information. This process is also known as
dimensionality reduction, and it helps to reduce the computational complexity
of the model. Additionally, the pooling layer can help to identify important
features that are invariant to rotation and position, thereby enhancing the
model’s ability to learn effectively. During the pooling layer, the feature map
is sub sampled to a smaller size, typically using a 2x2 pool size. This reduces
the spatial size of the feature map while preserving its important features.
By reducing the size of the feature map, the computational complexity of the
model is reduced, allowing for faster processing and training. Overall, the
pooling layer is an important step in CNNs that helps to extract important
features while reducing the size and complexity of the data.
classifier.add (Max Pooling2D (pool size = (2,2)))
Result
The above MRI images show the generated output of semantic based image
retrieval and we can see that it shows the predictions of brain tumor im-
ages using CNN in which it shows the prediction results and the ground truth.
43
It detects the location of brain tumor accurately with 98.87 percentage.
where we used python language for coding .
We are able to successfully locate the location of the brain tumor with
semantic based similarity technique for retrieving brain tumor images.
Sr Performance criteria
No Type of clusters recall execution Time(s)
1 Fuzzy c-means cluster 94.83 1.875
2 K-means cluster 88.45 55.65
Sr Performance criteria
No Type of clusters recall execution Time(s)
1 Fuzzy c-means cluster 95.83 6.875
2 K-means cluster 84.65 17.75
It detects the brain tumor images using MRI by feature extraction method.
Sr Performance criteria
No Technique recall F1-source Precision MAP
1 Latent Semantic Analysis (LSA) 0.89 0.85 0.82 0.75
2 Convolutional Neural Networks (CNNs) 0.94 0.93 0.92 0.83
3 Transfer Learning 0.92 0.90 0.88 0.80
4 Auto encoders 0.88 0.86 0.85 0.77
5 Siamese Neural Networks 0.93 0.92 0.91 0.88
Figure 5.9: Coding part used for Feature Extraction of brain tumor images
using ResNext50
6.1 Conclusion
The creation of a semantic-based image retrieval system for brain tumor
images has the potential to revolutionize medical image analysis. It could
result in a quicker and more accurate diagnosis, better therapy planning,
and advanced medical research, despite difficulties like dataset selection and
labeling, robust algorithm development, and optimization. To completely reap
its benefits, more study is required.This system uses deep learning and natural
language processing to understand user queries, extract useful features, and
retrieve pertinent medical images from huge databases. However, it is essential
to guarantee user friendliness, adherence to moral principles, and compliance
with data protection laws. Overall, the development of a semantic-based image
retrieval system for pictures of brain tumors has the potential to significantly
advance the field of medical image analysis and patient
49
6.2 Future scope
In a semi-supervised and favored semi-paired learning environment, we will
continue to investigate the segmentation of brain tumors. Future research can
also concentrate on the intrinsic value that uncertainty estimation provides for
automated diagnosis when differentiating between the two sources of uncer-
tainty. The predictive uncertainty can be divided into epistemic and aleatoric
uncertainty. The third path involves testing this framework in additional di-
agnostic contexts, perhaps favoring the fusion of more multi modal data sources.
51
[14] Zar Nawab Khan Swati, Qinghua Zhao, Muhammad Kabir, Farman Ali,
Zakir Ali, Saeed Ahmed, and Jianfeng Lu. “Content-based brain tumor
retrieval for MR images using transfer learning”. In: IEEE Access 7
(2019), pp. 17809–17822.
[15] Palwinder Kaur and Rajesh Kumar Singh. “A Panoramic View of
Content-based Medical Image Retrieval system”. In: 2020 International
Conference on Intelligent Engineering and Management (ICIEM). 2020,
pp. 187–192.
[16] S Maheswaran, S Sathesh, E D Bhaarathei, and D Kavin. “‘Design
and development of chemical free green embedded weeder for row based
crops,” in: J. Green Eng 10.5 (2020), pp. 2103–2120.
[17] D Patil, S Krishnan, and S Gharge. “‘Medical image retrieval by region
based shape feature for CT images,” in: Proc. Int. Conf. Mach. Learn.,
Big Data. 2019, pp. 155–159.
[18] Senthil Kumar Sundararajan, B Sankaragomathi, and D Saravana Priya.
“Deep Belief CNN Feature Representation based content based image
retrieval for medical images”. In: J. Med. Syst. 43.6 (May 2019), p. 174.
[19] Shengcong Chen, Changxing Ding, and Minfeng Liu. “Dual-force convo-
lutional neural networks for accurate brain tumor segmentation”. en. In:
Pattern Recognit. 88 (Apr. 2019), pp. 90–100.
[20] Neelum Noreen, Sellappan Palaniappan, Abdul Qayyum, Iftikhar Ahmad,
Muhammad Imran, and Muhammad Shoaib. “A deep learning model
based on concatenation approach for the diagnosis of brain tumor”. In:
IEEE Access 8 (2020), pp. 55135–55144.
[21] Ahmed H Abdel-Gawad, Lobna A Said, and Ahmed G Radwan. “Opti-
mized edge detection technique for brain tumor detection in MR images”.
In: IEEE Access 8 (2020), pp. 136243–136259.
[22] B Deepa, M G Sumithra, R M Kumar, and M Suriya. “‘Weiner fil-
ter based Hough transform and wavelet feature extraction with neural
network for classifying brain tumor,” in: Proc. 6th Int. Conf. Inventive
Comput. Technol. (ICICT). 2021, pp. 637–641.
[23] Linwei Fan, Fan Zhang, Hui Fan, and Caiming Zhang. “Brief review
of image denoising techniques”. In: Vis. Comput. Ind. Biomed. Art 2.1
(Dec. 2019).
[24] M Gurbina, M Lascu, and D Lascu. “‘Tumor detection and classification
of MRI brain image using different wavelet transforms and support vector
machines,” in: Proc. 42nd Int. Conf. Telecommun. Signal Process. (TSP).
2019, pp. 505–508.
[25] Ming Li, Lishan Kuang, Shuhua Xu, and Zhanguo Sha. “Brain tumor de-
tection based on multimodal information fusion and convolutional neural
network”. In: IEEE Access 7 (2019), pp. 180134–180146.
[26] Amr M Abdelaty, Merna Roshdy, Lobna A Said, and Ahmed G Radwan.
“Numerical simulations and FPGA implementations of fractional-order
systems based on product integration rules”. In: IEEE Access 8 (2020),
pp. 102093–102105.