Issamlagoiti Transfer Learningarticle
Issamlagoiti Transfer Learningarticle
net/publication/374134473
CITATIONS READS
0 190
2 authors, including:
Issam Lagoiti
Faculté des Sciences Ain Chock - Casablanca
1 PUBLICATION 0 CITATIONS
SEE PROFILE
All content following this page was uploaded by Issam Lagoiti on 23 September 2023.
ARTICLEINFO ABSTRACT
Keywords: In the realm of satellite imagery analysis, automated and accurate image classification remains pivotal for
various applications, spanning from environmental monitoring to urban development. This study delves into
Satellite Imagery, the efficacy of transfer learning, a potent deep learning technique, in enhancing the classification performance
Transfer Learning, of satellite images. We juxtaposed a custom Convolutional Neural Network (CNN) against renowned pre-
CNN, trained architectures, namely ResNet, MobileNetV2, and InceptionV3, fine-tuning them for our specific task.
ResNet, Our findings elucidate that while the custom CNN yielded commendable results, the pre-trained models,
MobileNetV2, benefiting from the generic features learned on vast datasets like ImageNet, showcased superior performance.
InceptionV3, Among the pre-trained contenders, MobileNet stood out, achieving an accuracy of approximately 98.93%. This
Comparative Study, research underscores the potential of transfer learning in satellite image classification, offering a pathway for
ImageNet, future endeavors seeking optimized performance with limited labeled data.
Deep Learning.
Figure 4: Comparative Analysis of Training and Validation Accuracy Across Epochs for Various Deep
Learning Models.
Figure 5: Comparative Analysis of Training and Validation Loss Across Epochs for Various Deep Learning
Models.
5
Confusion Matrix for CNN Confusion Matrix for MobileNet
models models
Confusion Matrix for ResNet models Confusion Matrix for inceptionv3 models
By analyzing these matrices, we were able to pinpoint specific areas where Notably, when we applied this function to our custom CNN model, we
each model excels and where they struggle, providing a roadmap for future observed a few instances of incorrect predictions, a testament to the
optimizations and adjustments. Furthermore, the confusion matrices serve as challenges associated with perfecting the classification process. These
a transparent tool for readers to grasp the real-world implications of the visualizations serve as a valuable tool for intuitively gauging the model's
models' performances, fostering a deeper understanding and facilitating performance, allowing us to identify potential areas where the model may
subsequent research and development in this domain. benefit from further optimization and refinement.
5.4. Visual Analysis of Model Predictions: As we continue to refine our models, this visual tool will remain a
In our endeavor to offer a more tangible perspective of each model's cornerstone in our analytical arsenal, assisting us in continually assessing and
predictive prowess, we developed to visually illustrate the models' predictions enhancing our models' predictive accuracy and reliability. Moreover, it offers
on a random set of images from the dataset. The essence of this function is to readers a tangible insight into the complex process of image classification
select a random batch of images and portray them alongside both the through deep learning architectures.
predicted and true labels, thereby offering a vivid illustration of the model's
classification capabilities.
As depicted in Figure 7, the function operates by selecting a random batch
of images from the dataset and predicting their respective labels using the
trained model. Following this, it plots the images in a grid format, wherein each
image is accompanied by its predicted label (along with the associated
confidence percentage) and its true label. The color coordination in the labels
serves as an intuitive indicator of the prediction's accuracy, with green
denoting correct predictions and red highlighting incorrect ones.
6
CNN
MobileNet
Inceptionv3
ResNet
8
8. Acknowledgments [12]. Hughes, G. “On the Mean Accuracy of Statistical Pattern
Recognizers.” IEEE Transactions on Information Theory, vol. 14, no.
The author would like to extend their sincere gratitude to the creators and
1, Jan. 1968, pp. 55–63, https://doi.org/10.1109/tit.1968.1054102.
contributors of the Satellite Image Classification Dataset-RSI-CB256. This
dataset, comprised of diverse classes sourced from Sensors and Google map [13]. Castelluccio, Marco, et al. “Land Use Classification in Remote
snapshots, has been instrumental in facilitating our research endeavors. Sensing Images by Convolutional Neural Networks.”
ArXiv:1508.00092 [Cs], 1 Aug. 2015,
Additionally, our appreciation extends to our peers and reviewers who
https://arxiv.org/abs/1508.00092.
have provided invaluable feedback and insights throughout the course of this
research. [14]. Lopez, Jorge. “Combining Satellite Imagery and Machine Learning
to Predict Poverty.” Medium, 23 Sept. 2022,
Lastly, we acknowledge the institutions and organizations that have
https://towardsdatascience.com/combining-satellite-imagery-and-
supported our work. Their belief in the potential of our research has been a
machine-learning-to-predict-poverty-884e0e200969. Accessed 10
driving force behind our endeavors.
Sept. 2023.
9. References [15]. Xie, Michael E, et al. “Transfer Learning from Deep Features for
Remote Sensing and Poverty Mapping.” ArXiv (Cornell University),
[1]. Abburu, Sunitha, and Suresh Babu Golla. “Satellite Image 1 Oct. 2015, https://doi.org/10.48550/arxiv.1510.00098 . Accessed
Classification Methods and Techniques: A Review.” International 21 Apr. 2023.
Journal of Computer Applications, vol. 119, no. 8, 18 June 2015, pp.
20–25, [16]. Sameera, Nerella, and M. Shashi. “Deep Transductive Transfer
pdfs.semanticscholar.org/6961/3390ca76bf103791ef251e1568de Learning Framework for Zero-Day Attack Detection.” ICT Express,
b5fe91dd.pdf, https://doi.org/10.5120/21088-3779. Mar. 2020, https://doi.org/10.1016/j.icte.2020.03.003.
[2]. Carlos Favis Ezequiel, et al. “UAV Aerial Imaging Applications for [17]. Joseph, V. Roshan, and Akhil Vakayil. “SPlit: An Optimal Method for
Post-Disaster Assessment, Environmental Management and Data Splitting.” Technometrics, 28 Apr. 2021, pp. 1–23,
Infrastructure Development.” International Conference on https://doi.org/10.1080/00401706.2021.1921037. Accessed 16
Unmanned Aircraft Systems, 27 May 2014, Dec. 2021.
https://doi.org/10.1109/icuas.2014.6842266.Accessed 30 Apr.
2023. [18]. Shorten, Connor, and Taghi M. Khoshgoftaar. “A Survey on Image
Data Augmentation for Deep Learning.” Journal of Big Data, vol. 6,
[3]. Tempa, Karma, and Komal Raj Aryal. “Semi-Automatic Classification no. 1, 6 July 2019, https://doi.org/10.1186/s40537-019-0197-0.
for Rapid Delineation of the Geohazard-Prone Areas Using Sentinel-
2 Satellite Imagery.” SN Applied Sciences, vol. 4, no. 5, 9 Apr. 2022, [19]. Bengio, Y., et al. “Representation Learning: A Review and New
https://doi.org/10.1007/s42452-022-05028-6. Accessed 4 May Perspectives.” IEEE Transactions on Pattern Analysis and Machine
2022. Intelligence, vol. 35, no. 8, Aug. 2013, pp. 1798–1828,
https://doi.org/10.1109/tpami.2013.50.
[4]. Hu, Yunfeng, et al. “A Deep Convolution Neural Network Method
for Land Cover Mapping: A Case Study of Qinhuangdao, China.” [20]. Howard, Andrew, et al. “Searching for MobileNetV3.”
Remote Sensing, vol. 10, no. 12, 17 Dec. 2018, p. 2053, Openaccess.thecvf.com,2019,
https://doi.org/10.3390/rs10122053. https://openaccess.thecvf.com/content_ICCV_2019/html/Howard
_Searching_for_MobileNetV3_ICCV_2019_paper.html.
[5]. Shallu, and Rajesh Mehra. “Breast Cancer Histology Images
Classification: Training from Scratch or Transfer Learning?” ICT [21]. Russakovsky, Olga, et al. “ImageNet Large Scale Visual Recognition
Express, vol. 4, no. 4, Dec. 2018, pp. 247–254, Challenge.” ArXiv.org, 2014, https://arxiv.org/abs/1409.0575 .
https://doi.org/10.1016/j.icte.2018.10.007.Accessed 31 Mar. 2020. [22]. LeCun, Yann, et al. “Deep Learning.” Nature, vol. 521, no. 7553, May
[6]. Cheng, Gong, et al. “Remote Sensing Image Scene Classification: 2015, pp. 436–444, www.nature.com/articles/nature14539,
Benchmark and State of the Art.” Proceedings of the IEEE, vol. 105, https://doi.org/10.1038/nature14539.
no. 10, Oct. 2017, pp. 1865–1883, [23]. Dahl, G. E., et al. “Context-Dependent Pre-Trained Deep Neural
https://doi.org/10.1109/jproc.2017.2675998. Networks for Large-Vocabulary Speech Recognition.” IEEE
[7]. Fauvel, Mathieu, et al. “Spectral and Spatial Classification of Transactions on Audio, Speech, and Language Processing, vol. 20,
Hyperspectral Data Using SVMs and Morphological Profiles.” IEEE no.1,Jan.2012,pp.30–42,
Transactions on Geoscience and Remote Sensing, vol. 46, no. 11, https://doi.org/10.1109/tasl.2011.2134090.Accessed 10 May
Nov. 2008, pp. 3804–3814, 2019.
https://doi.org/10.1109/tgrs.2008.922034. [24]. Wilson, Ashia C., et al. “The Marginal Value of Adaptive Gradient
[8]. Laleh Armi, and Shervan Fekri-Ershad. “Texture Image Analysis and Methods in Machine Learning.” ArXiv:1705.08292 [Cs, Stat], 21 May
Texture Classification Methods - a Review.” ArXiv (Cornell 2018, https://arxiv.org/abs/1705.08292 .
University), 13 Apr. 2019. Accessed 10 Sept. 2023. [25]. Sokolova, Marina, and Guy Lapalme. “A Systematic Analysis of
[9]. Bruzzone, L., and D.F. Prieto. “Automatic Analysis of the Difference Performance Measures for Classification Tasks.” Information
Image for Unsupervised Change Detection.” IEEE Transactions on Processing & Management, vol. 45, no. 4, July 2009, pp. 427–437,
Geoscience and Remote Sensing, vol. 38, no. 3, May 2000, pp. www.sciencedirect.com/science/article/pii/S0306457309000259,
1171–1182, https://doi.org/10.1109/36.843009. Accessed 24 May https://doi.org/10.1016/j.ipm.2009.03.002. Accessed 29 Aug.
2020. 2019.
[10]. Simonyan, Karen, and Andrew Zisserman. “Very Deep [26]. Chawla, Nitesh V., et al. “Editorial.” ACM SIGKDD Explorations
Convolutional Networks for Large-Scale Image Recognition.” Arxiv- Newsletter, vol. 6, no. 1, 1 June 2004, p. 1,
Vanity.com, 2014, www.arxiv-vanity.com/papers/1409.1556/. https://doi.org/10.1145/1007730.1007733.
[11]. Yosinski, Jason, et al. “How Transferable Are Features in Deep [27]. Smith, Linda, and Michael Gasser. “The Development of Embodied
Neural Networks?” ArXiv.org, 2014, Cognition: Six Lessons from Babies.” Artificial Life, vol. 11, no. 1-2,
https://arxiv.org/abs/1411.1792. Jan.2005, pp. 13–29, https://doi.org/10.1162/1064546053278973.
9
[28]. Hohman, Fred, et al. “Visual Analytics in Deep Learning: An
Interrogative Survey for the next Frontiers.” IEEE Transactions on
Visualization and Computer Graphics, vol. 25, no. 8, 1 Aug. 2019,
pp.2674–2693,arxiv.org/pdf/1801.06889.pdf,
https://doi.org/10.1109/tvcg.2018.2843369.
[29]. Davis, Jesse, and Mark Goadrich. “The Relationship between
Precision-Recall and ROC Curves.” Proceedings of the 23rd
International Conference on Machine Learning - ICML ’06, 2006,
pages.cs.wisc.edu/~jdavis/davisgoadrichcamera2.pdf,
https://doi.org/10.1145/1143844.1143874.
[30]. Chawla, N. V., et al. “SMOTE: Synthetic Minority Over-Sampling
Technique.” Journal of Artificial Intelligence Research, vol. 16, no.
16,1June2002,pp.321–357,
www.jair.org/index.php/jair/article/view/10302,
https://doi.org/10.1613/jair.953.
10