Alzheimers and Brain Tumor Detection Using Deep Learning
Alzheimers and Brain Tumor Detection Using Deep Learning
Abstract: In several industries, such as manufacturing, construction, and the Accurate detection of brain tumors and
Alzheimer’s disease is essential for effective treatment and disease management. With the rapid progress in deep learning
technologies, the field of medical imaging—particularly the interpretation of brain scans—has seen remarkable
improvements. This research focuses on utilizing two well-established convolutional neural network (CNN) architectures,
VGG-19 and ResNet-50, for brain tumor classification, while employing a standard CNN model for detecting Alzheimer’s
disease.
VGG-19, characterized by its consistent and deep structure comprising 19 layers, is particularly effective in extracting
complex features due to its sequential convolutional layers. This makes it well-suited for identifying subtle patterns in MRI
images of the brain. In contrast, ResNet-50 incorporates residual connections within its 50-layer design, allowing the model
to mitigate issues like vanishing gradients and improving learning efficiency by enabling the network to focus on residual
mappings. This study compares both models to evaluate their accuracy, resilience, and computational efficiency in detecting
brain abnormalities.
Moreover, the research examines each model's ability to generalize across various datasets and tumor types, aiming to
provide insights into their clinical applicability. The results may contribute to refining current diagnostic techniques,
promoting earlier detection, and assisting in the development of advanced tools for accurate brain tumor diagnosis and
treatment planning. Integrating these models into healthcare systems could improve diagnostic accuracy and enhance
patient care outcomes. Alzheimer’s disease, the most prevalent form of dementia, leads to progressive memory loss, impaired
thinking, and behavioral changes. Its symptoms typically worsen over time, eventually hindering the ability to perform
everyday tasks. Dementia is a broad term describing a range of symptoms caused by cognitive decline, with Alzheimer’s
accounting for 60% to 80% of all cases. Vascular dementia, often following a stroke, is the second most common type, though
several other reversible conditions—like thyroid imbalances and vitamin deficiencies—can produce similar symptoms.In
this study, we use publicly available datasets for Alzheimer’s detection. The system employs deep learning models,
particularly CNN and ResNet, to analyze the data. The outcomes demonstrate the model's ability to accurately classify the
disease into categories such as mild, moderate, very moderate, and dementia, based on performance metrics like prediction
accuracy.
Keywords: Deep Learning, Convolutional Neural Networks (CNN), VGG-19, ResNet-50, Brain Tumor Detection, Alzheimer’s Disease, Medical
Image Analysis, MRI, Classification, Transfer Learning.
How to Cite: Prof Bandu Meshram ; Shubham Sudarshan More; Ajinkya Padmakar Sagane; Datta Meghe; Riddhesh Santosh
Sarode; Arpit Suryakant Lende (2025), Alzheimers and Brain Tumor Detection Using Deep Learning. International Journal of
Innovative Science and Research Technology, 10(4), 2073-2081. https://doi.org/10.38124/ijisrt/25apr1270
Fig 1 Workflow Diagram A custom CNN model was also designed and optimized
specifically for Alzheimer’s detection. This model included
Feature Extraction multiple convolutional and pooling layers followed by dense
Both statistical and textural features were extracted layers, using ReLU activations and softmax for final
from the preprocessed images. Techniques such as mean, classification. While less deep than the VGG-19 and ResNet-
median, variance, and Gray-Level Co-occurrence Matrix 50 models, it offered flexibility and computational efficiency
(GLCM) were employed to derive meaningful insights. for the Alzheimer's-specific dataset.
GLCM features like contrast, correlation, energy, entropy,
and homogeneity helped capture spatial texture relationships, All models were trained using transfer learning, where
essential for distinguishing healthy and abnormal brain pretrained weights from large datasets like ImageNet were
regions. adapted to the medical imaging task. Fine-tuning of upper
In addition to GLCM, histogram-based features were layers was performed to tailor the models to domain-specific
analyzed to understand pixel intensity distribution, which features, ensuring improved accuracy even with limited
assists in identifying abnormalities in grayscale values. Edge labeled data.
detection filters were also utilized to highlight boundaries of
tumor-affected areas. These extracted features were then The dataset was divided into training (70%) and testing
flattened into a numerical vector format suitable for feeding (30%) subsets using a stratified approach to preserve class
into the neural network models. The combination of both distribution. The models were compiled using categorical
global statistical and local texture-based features ensured cross-entropy as the loss function and optimized using the
robust input representation for accurate classification and Adam optimizer with learning rate scheduling and dropout
prediction. for regularization. Training was conducted over multiple
epochs with batch normalization to stabilize and accelerate
Model Development convergence.
To achieve accurate classification of brain tumors and
Alzheimer’s disease, this study implemented three deep Performance metrics such as accuracy, precision, recall,
learning models—VGG-19, ResNet-50, and a custom- and F1-score were tracked during training, and early stopping
designed Convolutional Neural Network (CNN). Each model was used to prevent overfitting. This multi-model strategy
was carefully chosen or developed to exploit its strengths in allowed for comparative analysis and identification of the
medical image analysis. most suitable architecture for medical diagnosis.
VGG-19 is a deep convolutional neural network known for its ability to handle large-scale image
consisting of 19 weight layers, primarily using small 3x3 classification tasks. VGG-19 is effective in recognizing
convolution filters and 2x2 max-pooling layers. Its subtle variations in MRI scans due to its depth and fine-
architecture follows a very uniform and simple structure, grained feature extraction. Despite its high accuracy, it
making it easier to implement and tune. The model extracts requires significant computational resources due to the large
hierarchical features from images, starting from edges in the number of parameters.
initial layers to complex patterns in deeper layers. It is widely
CNN is a class of deep learning models designed classification. CNNs automatically learn to extract relevant
specifically for analyzing visual data. It processes input spatial information from raw images without manual feature
images through multiple layers of convolution, activation engineering. They are widely used in medical image analysis
(usually ReLU), and pooling, which help in detecting features due to their high accuracy and adaptability. By stacking
such as edges, textures, and shapes. These features are then layers, CNNs can learn increasingly complex patterns to
passed through fully connected layers to perform distinguish between healthy and diseased tissues.
ResNet-50 is a 50-layer deep neural network that During the classification process, the input image
incorporates residual blocks, allowing it to train much deeper passed through the network’s layers where important features
architectures without performance degradation. The key were automatically extracted. Convolution and pooling
innovation in ResNet is the use of shortcut or skip operations helped in reducing the image's dimensionality
connections that pass the input of a layer directly to a deeper while preserving crucial information needed for accurate
layer, helping to maintain gradient flow during recognition. After feature extraction, the information was
backpropagation. This architecture addresses the vanishing aggregated through fully connected layers, preparing the data
gradient problem commonly seen in very deep networks. for final classification. The softmax layer then generated the
ResNet-50 is highly effective in complex image recognition probabilities for each possible class, and the model selected
tasks, offering improved accuracy and faster convergence. In the class with the maximum probability as the final output.
medical imaging, it is particularly useful for identifying
detailed patterns in high-resolution scans and handling In the case of Alzheimer’s detection, the system not
diverse datasets. only identified the presence of the disease but also assessed
the severity of its progression. By analyzing structural brain
V. CLASSIFICATION AND PREDICTION changes, such as hippocampal shrinkage and cortical atrophy,
the models could classify patients into different stages of the
After the deep learning models (VGG-19, ResNet-50, disease. This detailed classification was critical in helping
and a custom CNN) were effectively trained using the healthcare professionals provide early diagnosis and plan
preprocessed and labeled medical images, the next important effective treatment strategies tailored to each patient’s
phase involved classifying new and unseen MRI scans. The condition.
main objective was to determine whether a brain MRI
indicated the presence of a tumor or Alzheimer’s disease and, After obtaining predictions, the model outputs were
if applicable, accurately classify the type or stage of the compared with the ground truth labels from the test dataset.
condition. To evaluate the performance and reliability of the
For brain tumor detection, the models were trained to classification results, metrics such as accuracy, precision,
distinguish between normal brain scans and those exhibiting recall, and F1-score were calculated. Accuracy measured the
tumor-affected areas. In more complex cases, the models overall percentage of correct predictions, while precision
were capable of further classifying different tumor types, assessed how many of the positive predictions were actually
such as glioma, meningioma, and pituitary tumors, depending correct. Recall determined the model’s ability to detect all
on the dataset that was used. Similarly, for Alzheimer’s actual positive cases, and the F1-score provided a balanced
disease detection, the models categorized brain images into evaluation by combining both precision and recall. Together,
various stages, including mild cognitive impairment, these performance indicators offered a comprehensive
moderate Alzheimer’s, severe Alzheimer’s, and full understanding of how well the models could classify brain
dementia, allowing for more precise clinical insights. tumors and Alzheimer's disease, highlighting their potential
to support clinical decision-making and early intervention.
To achieve this, each model’s final output layer was
constructed to perform multi-class classification through the
use of the softmax activation function. Softmax transforms
the model’s raw output scores into a probability distribution
across all possible categories. This approach allowed the
model to predict the most likely class by selecting the label
with the highest probability, offering a clear and confident
final decision for each input image.
REFERENCES