Automatic Severity Classification of Diabetic Retinopathy Based On Densenet and Convolutional Block Attention Module
Automatic Severity Classification of Diabetic Retinopathy Based On Densenet and Convolutional Block Attention Module
Received March 24, 2022, accepted April 1, 2022, date of publication April 6, 2022, date of current version April 14, 2022.
Digital Object Identifier 10.1109/ACCESS.2022.3165193
ABSTRACT Diabetic Retinopathy (DR) - a complication developed due to heightened blood glucose levels-
is deemed one of the most sight-threatening diseases. Unfortunately, DR screening is manually acquired by
an ophthalmologist, a process that can be considered erroneous and time-consuming. Accordingly, automated
DR diagnostics have become a focus of research in recent years due to the tremendous increase in diabetic
patients. Moreover, the recent accomplishments demonstrated by Convolutional Neural Networks (CNN)
settle them as state-of-the-art for DR stage identification. This paper proposes a new automatic deep-
learning-based approach for severity detection by utilizing a single Color Fundus photograph (CFP).
The proposed technique employs DenseNet169’s encoder to construct a visual embedding. Furthermore,
Convolutional Block Attention Module (CBAM) is introduced on top of the encoder to reinforce its
discriminative power. Finally, the model is trained using cross-entropy loss on the Kaggle Asia Pacific
Tele-Ophthalmology Society’s (APTOS) dataset. On the binary classification task, we accomplished (97%
accuracy - 97% sensitivity - 98.3% specificity - 0.9455, Quadratic Weighted Kappa score (QWK)) compared
to the state-of-the-art. Moreover, Our network showed high competency (82% accuracy - 0.888 (QWK))
for severity grading. The significant contribution of the proposed framework is that it efficiently grades
the severity level of diabetic retinopathy while reducing the time and space complexity required, which
demonstrates it as a promising candidate for autonomous diagnosis.
INDEX TERMS Diabetic retinopathy, convolutional neural networks (CNN), attention mechanism, deep
learning.
This work is licensed under a Creative Commons Attribution 4.0 License. For more information, see https://creativecommons.org/licenses/by/4.0/
VOLUME 10, 2022 38299
M. M. Farag et al.: Automatic Severity Classification of DR Based on DenseNet and CBAM
Artificial Flora Algorithm (AFA) [5] for feature selection in devices to get high-quality images. (ii) Implicit stochasticity.
addition to using Gradient Boosted Trees (GBT) [6] as a clas- Retinal fundus images experience large variations caused
sification model. Furthermore, exploited by Gharaibeh et al. by different devices and environmental conditions regarding
in [7] and [8] by employing feature engineering process, then color, contrast, illumination, and size. As a result, the model’s
applying Support Vector Machines (SVM) as a classifier for decision may be distorted. (iii) Fading classes’ disparity. The
DR detection [9]. ML algorithms need personalized expe- threshold chosen for image classification between two closely
rience and domain knowledge to find the most informative distributed classes (e.g., mild and moderate in the APTOS
representation despite its effectiveness. dataset) is blurry, as will be shown in Section III.C, due to
Deep Learning (DL) has gained a foothold in various the dependence on microscale ocular pathologies. To solve
fields by representing the world as a nested hierarchy of the problem of fading disparity, large CNN architectures
concepts, with each concept defined through its relation were employed in the literature to extract more informative
to simpler concepts [10]. Convolutional Neural Networks features, data augmentation and preprocessing were used to
was the standout DL architecture in the late nineties. Since enhance CNNs’ generalizability. Finally, transfer learning
then, it has been used extensively for processing data such was exploited to overcome data shortage.
as images and time series. Moreover, it has demonstrated In this paper, we investigate the efficacy of light-weight
outstanding performance in practical applications such as deep learning architecture for fast and robust severity grad-
Natural Language Processing (NLP) [11], [12] and Computer ing of diabetic retinopathy. Our framework is based on
Vision (CV) problems [13]–[15]. a modified version of DenseNet [31] with integrating an
Exploiting convolutional neural networks’ power for a attention mechanism with the former architecture for more
medical domain has developed more robust solutions, specif- feature refinement. Furthermore, we observe the effect of
ically in the DR domain. [16] and [17] demonstrated data imbalance on the model performance and mitigate
the effectiveness of such a technique for retinal vessel such an effect by using an imbalanced learning technique.
segmentation. Similarly, by leveraging Generative Adver- As shown in Fig.1, we first pass and preprocess the reti-
sarial Networks (GANs), Zhao et al. [18] could synthesize nal image for quality enhancement, afterward, the images
fundus images. Dai et al. [19] utilized multi-sieving con- were passed to the DenseNet encoder C for feature extrac-
volutional neural network and image to text mapping for tion, then the features are sent to the attention module A
Micro-aneurysms (MA) early detection. [20] evaluated the for more improved representation. We train our model by
performance of three recognized CNN architectures: VGG16, freezing Densnet’s encoder, trained on the ImageNet [32]
VGG19, and InceptionV3 [21], [22] by employing transfer dataset for the model’s convergence acceleration by using the
learning and fine-tuning for binary and multi-class classi- pre-trained weights θC and training only the attention module
fication. Zeng et al. [23] introduced Siamese-like architec- and the classification head using APTOS data in a supervised
ture [24] trained with transfer learning to classify fundus approach to update θA & θM . Our main contributions are as
images into two grades. Kassani et al. [25] used a Multi- follows:
Layer Perceptron (MLP) as a classification head on top of
1) We developed a modified architecture to reduce the
the modified Xception network [26] by concatenating differ-
time needed for training and inference while enhancing
ent feature maps from different convolutional layers. Four
DR severity grading by using a relatively small model
Inception models were utilized [27] for multi-class classi-
with 8.5 million parameters compared to 10.8 million
fication, each fundus image was sliced into four quadrants,
in the previous work.
and each quadrant will be classified by one of the four
2) We exploited the effect of using an attention mecha-
models. [28] exploited blended models to enhance data rep-
nism as a supplementary module for feature refinement
resentation, Gangwar et al. [29] investigated a new hybrid
which led to an increase in accuracy while preserving
model inherited from Inception and ResNet architectures.
low model complexity.
Al Antary et al. [30] designed ResNet architecture integrated
3) We tested the effect of using an imbalanced learning
with a Multi-Scale Attention mechanism (MSA) to enhance
approach to alleviate the impact of data imbalance on
the representational power of the encoder. Moreover, they
the model’s performance and proved its efficiency in
employed a multi-level approach for feature reuse for more
enhancing the overall metrics.
improvements. Since our focus in this paper is to enhance the
4) We utilized transfer learning only by freezing
grading system both on binary and multi-class classification
the convolutional encoder without extra fine-tuning
tasks, we observed drawbacks related to the aforementioned
which led to relatively low number of learnable
algorithms despite their success ranging from high time and
parameters (150K).
space complexity to drop out mitigating the severe data
imbalance inherited. The paper is divided as follows. The related work
DR severity grading remains a challenging task due to three is presented in Section II. In Section III, the methodol-
factors: (i) Data rarity. Acquiring massive labeled data is ogy is presented. In Section IV, the results and discus-
a crucial issue for DL and more significant in the medical sions are demonstrated. Finally, conclusions are provided in
domain due to the data privacy issues or/and having costly Section V.
FIGURE 1. In the scheme of our proposed approach, In the network training step (upper), we pass a batch of labeled preprocessed images X to
our convolutional encoder C for feature extraction, then an attention mechanism A for feature refinement. Finally, in the testing phase (lower),
we directly pass the data to the network to predict the image class.
TABLE 2. DR severity grading results on APTOS dataset. The best, second best, and third best are marked by italics, boldface, and underline, respectively.
M:million.
TABLE 3. Binary classification results on APTOS dataset. The best, second best, and third best are marked by italics, boldface, and underline, respectively.
TABLE 4. Statistics for training and validation datasets. when we have a different data domain compared to ImageNet
data, and we took our decision based on the interesting results
provided by [44], where ImageNet weights demonstrated its
robustness as a feature extractor for retinal disease detec-
tion. A reduction ratio (r = 32) and kernel size (K 7×7 ) at
channel and spatial modules, respectively for CBAM. Due
to its performance, our fourth trial was compared to other
state-of-the-art techniques. Detailed information regarding
our architecture is demonstrated in Table.1.
Four trials were investigated to show the gradual increase in E. TRAINING SETTINGS
performance: Our splitting policy was 90% to 10% of our dataset to
• Baseline DensNet169. form a training and validation set. A stratified data splitting
• DenseNet169 + INS. technique was exploited to preserve the same distribution
• DenseNet169 + CBAM. to ensure the classes’ distribution consistency between the
• DenseNet169 + CBAM + INS. aforementioned subsets and the original set. Table.4 demon-
Where our baseline has only DenseNet169’s modified strates the training and validation data statistics. Furthermore,
encoder without attaching CBAM as a supplementary mod- K-fold validation was implemented to have more robust
ule, moreover as well as not deal with the class imbalance results, and due to the size of the dataset, we used 5-folds
inherited in APTOS data. For the second trial, we demon- to train on 80% and test using 20% of the original dataset
strated the effectiveness of using cost-sensitive learning to at each trial. Furthermore, the maximum number of epochs
penalize our model when dealing with minor classes and was limited to 400 while using an early stopping callback to
vice-versa. CBAM was added to DenseNet without using avoid overfitting by saving the best weights corresponding
INS to investigate its effectiveness in the third trial. Finally, to the minimum validation loss. Finally, we used the exact
we investigated the enhancements added by CBAM and INS stratified data splitting mechanism to ensure the same class
together. The four experiments had followed the same settings distribution at each fold.
by freezing DenseNet’s encoder and using transfer learning Our algorithm was implemented using TensorFlow [45]
to accelerate the training of CBAM and Softmax layers. Fine- and trained on Tesla V100 GPU provided by Google Co-lab.
tuning was not used in contrast to the conventional framework We trained four networks for 1000 epochs, and with a small
F. EVALUATION METRICS
Five common metrics were used to evaluate the model’s
performance.
1) ACCURACY (ACC)
The percentage of correct predictions that a model can
achieve. Accuracy is defined as
TP + TN
Acc = (10)
TP + TN + FP + FN
2) SENSITIVITY (SENS)
is the percentage of positive cases that is classified as actual
positive. Identified as follows
TP
Sens = (11)
TP + FN
3) SPECIFICITY (SPEC)
is the percentage of negative cases that are detected as actual
negative. Identified as follows
TN
Spec = (12)
TN + FP
4) F1-SCORE (F1)
is the harmonic mean of precision and recall and is identified
as
TP
F1 = 1
(13)
TP + 2 (FN + FP)
5) KAPPA-SCORE
to assess the agreement between our model and the original
rater. Identified as follows
P
i,j wi,j Oi,j
k =1− P (14)
i,j wi,j Ei,j
FIGURE 6. Normalized confusion matrices for (a) Baseline DenseNet169
where true positives (TP) are the classes classified correctly (b) DenseNet + CBAM (c) DenseNet169 + INS (d) DenseNET169 + INS +
by the algorithm, true negatives (TN) are samples predicted CBAM.
correctly as negative, false positives (FP) are samples that are
miss-classified as a positive class, and false negatives (FN) IV. RESULTS AND DISCUSSIONS
are samples miss-classified as negative class. Oi,j is the Fig.6 illustrates the performance of our four algorithms.
observed matrices, and Ei,j is the expected one. In Fig.6.a, we observe that without the weighted loss function,
it is easier for our model to be distorted and have robust specificity [48]. Finally, our model achieved low training
behavior only in detecting major classes (0 and 2) and vice- time (9 seconds/epoch) and relatively high inference speed
versa. As can be shown in Fig.6.b, attaching CBAM to our (1.166 seconds/32 images) compared to the MSA network
encoder enhanced the detection of classes 1 and 3 by 63.3% that achieved 5 seconds exploiting the same batch size.
and 90.9%, while reducing class 2 only by 4.6%. Class imbal-
ance mitigation allowed better performance, as can be seen in V. CONCLUSION
Fig.6.c, class 1,3 detection is enhanced by 43.3% and 236.4% In this study, we exploited a new CNN model based on
respectively, with respect to the baseline algorithm. Finally, DenseNet169 architecture integrated with CBAM as an addi-
using CBAM with DenseNet169 while adding weighted loss tional component to be added for representational power
has demonstrated thriving performance across all classes. enhancement. The proposed method demonstrated robust per-
Regardless of the reduction in class 2 by 14.63%, classes formance and comparable quality metrics while reducing
(1,3 and 4) exhibit significant improvements by 44.2%, the burden of space and time complexity. Furthermore, a
43.24%, and 235%. An average QWK and accuracy values 2-D Gaussian filter enhances fundus images’ quality. Finally,
of 0.8072 and 72.3% were achieved, respectively, using the we used INS to form our weighted loss function to tackle
5-fold k-validation technique. As shown in Section III.E, the class imbalance to improve the model’s prediction across
we trained our algorithm only for 400 epochs to reduce the all classes. For future research direction, we evaluate the
computational cost of training five different models, further performance of different CBAM configurations. Moreover,
training will provide more intact results. experimenting with different imbalanced learning techniques
As shown in Table.2, the proposed method outperformed and increasing the dataset size will lead to better performance.
the literature work on the severity grading task and showed
comparable results. Our model enhanced accuracy and QWK REFERENCES
by 0.4% and 24.9% while decreasing inference time by [1] Global Report on Diabetes, World Health Organization (WHO), Geneva,
cutting down the number of parameters by 83% compared Switzerland, 2016.
[2] Diabetes Atlas, International Diabetes Federation (IDF), Brussels,
to [28]. We achieved almost the same accuracy as [29] while Belgium, 2019.
reducing the model size. Our best trial had an increase in [3] Diabetes Eye Health a Guide for Health Professionals, International Dia-
accuracy of about 7% compared to the AM-InceptionV3 [37] betes Federation (IDF), Brussels, Belgium, 2017.
[4] P. Nagaraj, P. Deepalakshmi, R. F. Mansour, and A. Almazroa, ‘‘Arti-
method. SFTL model achieved high accuracy at the severity ficial flora algorithm-based feature selection with gradient boosted
grading task. However, they did not tackle the problem of tree model for diabetes classification,’’ Diabetes, Metabolic Syndrome
data imbalance. EfficientNet-B3 [36] achieved higher accu- Obesity: Targets Therapy, vol. 14, pp. 2789–2806, Jun. 2021, doi:
10.2147/DMSO.S312787.
racy but only for major classes, while we achieved compa- [5] L. Cheng, X.-H. Wu, and Y. Wang, ‘‘Artificial flora (AF) optimization
rable accuracy in minor classes, and finally, We compared algorithm,’’ Appl. Sci., vol. 8, no. 3, p. 329, Feb. 2018. [Online]. Available:
our best trial with the MSA network without multi-level https://www.mdpi.com/2076-3417/8/3/329, doi: 10.3390/app8030329.
[6] T. Chen and C. Guestrin, ‘‘XGBoost: A scalable tree boosting system,’’
feature reuse [30]. We had almost the same accuracy with 2016, arXiv:1603.02754.
an increase in QWK by 3.6%. Furthermore, we achieved [7] N. Gharaibeh, O. M. Al-hazaimeh, A. Abu-Ein, and K. M. O. Nahar,
a better confusion matrix across all classes than the litera- ‘‘A hybrid SVM NAÏVE-bayes classifier for bright lesions recognition
in eye fundus images,’’ Int. J. Electr. Eng. Informat., vol. 13, no. 3,
ture while reducing time and space complexity by a 45% pp. 530–545, Sep. 2021, doi: 10.15676/ijeei.2021.13.3.2.
reduction in parameters. Severity grading f1-score was not [8] O. M. Al Hazaimeh, K. M. O. Nahar, B. Al Naami, and N. Gharaibeh,
mentioned in the literature. However, by using CBAM and ‘‘An effective image processing method for detection of diabetic retinopa-
thy diseases from retinal fundus images,’’ Int. J. Signal Imag. Syst. Eng.,
INS, an enhancement was established by 21.4% with respect vol. 11, no. 4, p. 206, 2018, doi: 10.1504/IJSISE.2018.093825.
to the baseline DenseNet169. [9] T. Evgeniou and M. Pontil, ‘‘Support vector machines: Theory and appli-
Our algorithm demonstrated robustness against other deep cations,’’ in Advanced Course on Artificial Intelligence, vol. 2049. Berlin,
Germany: Springer, 2001, pp. 249–257, doi: 10.1007/3-540-44673-7_12.
learning architectures for the binary classification task,
[10] I. Goodfellow, Y. Bengio, and A. Courville, Deep Learning. Cambridge,
as shown in Table.3. Above all, the literature did not deal MA, USA: MIT Press, 2016.
with the class imbalance problem. Most of the algorithms [11] O. Abdel-Hamid, A.-R. Mohamed, H. Jiang, L. Deng, G. Penn, and
implemented did not consider its effect on quality metrics D. Yu, ‘‘Convolutional neural networks for speech recognition,’’
IEEE/ACM Trans. Audio, Speech, Language Process., vol. 22, no. 10,
which provided overestimated outcomes, as most of them pp. 1533–1545, Jul. 2014.
were predicting perfectly only for major classes due to ignor- [12] A. van den Oord, S. Dieleman, H. Zen, K. Simonyan, O. Vinyals,
ing data inherited imbalance. Furthermore, as mentioned A. Graves, N. Kalchbrenner, A. Senior, and K. Kavukcuoglu, ‘‘WaveNet:
A generative model for raw audio,’’ 2016, arXiv:1609.03499.
in Section III.C, binary grading did not require complex [13] K. Xu, J. Ba, R. Kiros, K. Cho, A. Courville, R. Salakhutdinov, R. Zemel,
architectures to solve it, our algorithm with lower param- and Y. Bengio, ‘‘Show, attend and tell: Neural image caption generation
eters achieved almost the same metrics compared to other with visual attention,’’ 2015, arXiv:1502.03044.
[14] J. Redmon and A. Farhadi, ‘‘YOLOv3: An incremental improvement,’’
algorithms, plus when we artificially formed two clusters 2018, arXiv:1804.02767.
(infected and normal), the classes were balanced which [15] L. A. Gatys, A. S. Ecker, and M. Bethge, ‘‘A neural algorithm of artistic
helped literature algorithms to excel in such a task. More- style,’’ 2015, arXiv:1508.06576.
[16] Z. Yan, X. Yang, and K.-T. Cheng, ‘‘Joint segment-level and pixel-wise
over, our algorithm exceeds the minimum limits provided losses for deep learning based retinal vessel segmentation,’’ IEEE Trans.
by English National Screening Program for sensitivity, and Biomed. Eng., vol. 65, no. 9, pp. 1912–1923, Sep. 2018.
[17] Y. Wu, Y. Xia, Y. Song, Y. Zhang, and W. Cai, ‘‘NFN+: A novel network [41] S. Ioffe and C. Szegedy, ‘‘Batch normalization: Accelerating deep network
followed network for retinal vessel segmentation,’’ Neural Netw., vol. 126, training by reducing internal covariate shift,’’ 2015, arXiv:1502.03167.
pp. 153–162, Jun. 2020. [42] X. Glorot, A. Bordes, and Y. Bengio, ‘‘Deep sparse rectifier neural net-
[18] H. Zhao, H. Li, S. Maurer-Stroh, and L. Cheng, ‘‘Synthesizing retinal works,’’ in Proc. 14th Int. Conf. Artif. Intell. Statist., vol 15, Fort Laud-
and neuronal images with generative adversarial nets,’’ Med. Image Anal., erdale, FL, USA, 2011.
vol. 49, pp. 14–26, Oct. 2018. [43] S. Woo, J. Park, J.-Y. Lee, and I. So Kweon, ‘‘CBAM: Convolutional block
[19] L. Dai, R. Fang, H. Li, X. Hou, B. Sheng, Q. Wu, and W. Jia, ‘‘Clin- attention module,’’ 2018, arXiv:1807.06521.
ical report guided retinal microaneurysm detection with multi-sieving [44] M. Raghu, C. Zhang, J. Kleinberg, and S. Bengio, ‘‘Transfusion: Under-
deep learning,’’ IEEE Trans. Med. Imag., vol. 37, no. 5, pp. 1149–1161, standing transfer learning for medical imaging,’’ 2019, arXiv:1902.07208.
May 2018. [45] M. Abadi et al., ‘‘TensorFlow: A system for large-scale machine learning,’’
[20] A. Jain, A. Jalui, J. Jasani, Y. Lahoti, and R. Karani, ‘‘Deep learning for 2016, arXiv:1605.08695.
detection and severity classification of diabetic retinopathy,’’ in Proc. 1st [46] K. He, X. Zhang, S. Ren, and J. Sun, ‘‘Delving deep into rectifiers:
Int. Conf. Innov. Inf. Commun. Technol. (ICIICT), Apr. 2019, pp. 1–6. Surpassing human-level performance on ImageNet classification,’’ in Proc.
[21] K. Simonyan and A. Zisserman, ‘‘Very deep convolutional networks for IEEE Int. Conf. Comput. Vis. (ICCV), Dec. 2015, pp. 1026–1034.
large-scale image recognition,’’ 2014, arXiv:1409.1556. [47] N. Srivastava, G. Hinton, A. Krizhevsky, I. Sutskever, and
[22] C. Szegedy, V. Vanhoucke, S. Ioffe, J. Shlens, and Z. Wojna, ‘‘Rethinking R. Salakhutdinov, ‘‘Dropout: A simple way to prevent neural networks
the inception architecture for computer vision,’’ 2015, arXiv:1512.00567. from overfitting,’’ J. Mach. Learn. Res., vol. 15, no. 56, pp. 1929–1958,
[23] X. Zeng, H. Chen, Y. Luo, and W. Ye, ‘‘Automated diabetic retinopathy 2014.
detection based on binocular siamese-like convolutional neural network,’’ [48] R. Taylor and D. Batey, The Need to Screen. Hoboken, NJ, USA: Wiley,
IEEE Access, vol. 7, pp. 30744–30753, 2019. 2012, ch. 4, pp. 29–41, doi: 10.1002/9781119968573.ch4.
[24] G. Koch, R. Zemel, and R. Salakhutdinov, ‘‘Siamese neural networks for
one-shot image recognition,’’ in Proc. ICML Deep Learn. Workshop, vol. 2,
Lille, France, 2015.
[25] S. H. Kassani, P. H. Kassani, R. Khazaeinezhad, M. J. Wesolowski,
K. A. Schneider, and R. Deters, ‘‘Diabetic retinopathy classification using MOHAMED M. FARAG received the B.Sc.
a modified xception architecture,’’ in Proc. IEEE Int. Symp. Signal Process. degree in renewable energy engineering from The
Inf. Technol. (ISSPIT), Dec. 2019, pp. 1–6. University of Ain-Shams, Cairo, Egypt, in 2018.
[26] F. Chollet, ‘‘Xception: Deep learning with depthwise separable convolu- He is currently pursuing the M.Sc. degree in elec-
tions,’’ 2016, arXiv:1610.02357. tronics engineering with the German University in
[27] Z. Gao, J. Li, J. Guo, Y. Chen, Z. Yi, and J. Zhong, ‘‘Diagnosis of Cairo (GUC), Cairo. He is currently working as
diabetic retinopathy using deep neural networks,’’ IEEE Access, vol. 7, a Teaching Assistant with the Electronics Depart-
pp. 3360–3370, 2019. ment, German University in Cairo. His research
[28] J. D. Bodapati, V. Naralasetti, S. N. Shareef, S. Hakak, M. Bilal,
and professional interests include data science,
P. K. R. Maddikunta, and O. Jo, ‘‘Blended multi-modal deep ConvNet
medical image processing, machine learning, and
features for diabetic retinopathy severity prediction,’’ Electronics, vol. 9,
no. 6, p. 914, May 2020. deep learning.
[29] A. K. Gangwar and V. Ravi, ‘‘Diabetic retinopathy detection using transfer
learning and deep learning,’’ in Evolution in Computational Intelligence,
V. Bhateja, S.-L. Peng, S. C. Satapathy, and Y.-D. Zhang, Eds. Singapore:
Springer, 2021, pp. 679–689.
[30] M. T. Al-Antary and Y. Arafa, ‘‘Multi-scale attention network for diabetic MARIAM FOUAD was born in Cairo, Egypt,
retinopathy classification,’’ IEEE Access, vol. 9, pp. 54190–54200, 2021. in 1993. She received the bachelor’s degree in
[31] G. Huang, Z. Liu, L. van der Maaten, and K. Q. Weinberger, ‘‘Densely electronics engineering from the German Univer-
connected convolutional networks,’’ 2016, arXiv:1608.06993. sity, Cairo, in 2015, and the master’s degree in
[32] O. Russakovsky, J. Deng, H. Su, J. Krause, S. Satheesh, S. Ma, Z. Huang, 2017 with the thesis titled ‘‘Joint Near-Infrared
A. Karpathy, A. Khosla, M. Bernstein, A. C. Berg, and L. Fei-Fei, ‘‘Ima- and Bio-Impedance Spectroscopy Non-Invasive
geNet large scale visual recognition challenge,’’ 2014, arXiv:1409.0575. Glucose Monitoring’’. She is currently pursuing
[33] Y. Wang, M. Yu, B. Hu, X. Jin, and Y. Li, ‘‘Deep learning-based detec- the Ph.D. degree with the Department of Medical
tion and stage grading for optimising diagnosis of diabetic retinopa-
Engineering, Ruhr University Bochum in collab-
thy,’’ Diabetes/Metabolism Res. Rev., vol. 37, no. 4, p. e3445, 2021, doi:
oration with the German University in Cairo. Her
10.1002/dmrr.3445.
[34] S. Qummar, F. G. Khan, S. Shah, A. Khan, S. Shamshirband, Z. U. Rehman, current research interests include the utilization of deep learning concepts in
I. Ahmed Khan, and W. Jadoon, ‘‘A deep learning ensemble approach for ultrasound special applications such as harmonic imaging and synthetic data
diabetic retinopathy detection,’’ IEEE Access, vol. 7, pp. 150530–150539, generation.
2019, doi: 10.1109/ACCESS.2019.2947484.
[35] S. Toledo-Cortés, M. De La Pava, O. Perdómo, and F. A. González,
‘‘Hybrid deep learning Gaussian process for diabetic retinopathy diagnosis
and uncertainty quantification,’’ 2020, arXiv:2007.14994.
[36] A. Sugeno, Y. Ishikawa, T. Ohshima, and R. Muramatsu, ‘‘Simple methods AMR T. ABDEL-HAMID was born in Cairo,
for the lesion detection and severity grading of diabetic retinopathy by Egypt, in 1974. He received the B.S. degree in
image processing and transfer learning,’’ Comput. Biol. Med., vol. 137, electronics and communications engineering from
Oct. 2021, Art. no. 104795, doi: 10.1016/j.compbiomed.2021.104795. Cairo University, Cairo, in 1997, and the M.S.
[37] V. Vives-Boix and D. Ruiz-Fernández, ‘‘Diabetic retinopathy detection and Ph.D. degrees in electrical and computer engi-
through convolutional neural networks with synaptic metaplasticity,’’
neering from Concordia University, Canada, in
Comput. Methods Programs Biomed., vol. 206, Jul. 2021, Art. no. 106094,
2001 and 2005, respectively. Currently, he is an
doi: 10.1016/j.cmpb.2021.106094.
[38] C. Zhang, T. Lei, and P. Chen, ‘‘Diabetic retinopathy grading by a source-
Assistant Professor with the Department of Elec-
free transfer learning approach,’’ Biomed. Signal Process. Control, vol. 73, tronics Engineering and the Vice Dean of Stu-
Mar. 2022, Art. no. 103423, doi: 10.1016/j.bspc.2021.103423. dent Affairs with the German University in Cairo
[39] B. Graham. (2015). Kaggle Diabetic Retinopathy Detection Compe- (GUC), Egypt. His main research interests include the Internet of Things
tition Report. [Online]. Available: https://www.kaggle.com/c/diabetic- applications and security, system-on-a-chip design and verification, func-
retinopathy-detection/discussion/15801 tional verification techniques, tools, and languages, IP watermarking, and
[40] G. Hinton and S. Roweis, ‘‘Stochastic neighbor embedding,’’ in Proc. Adv. security protocols verification.
Neural Inf. Process. Syst., 2002, pp. 833–840.