Development of Hybrid Convolutional Neural Network and Autoregressive Integrated Moving Average On Computed Tomography Image Classification
Development of Hybrid Convolutional Neural Network and Autoregressive Integrated Moving Average On Computed Tomography Image Classification
Corresponding Author:
Chee Ka Chin
Department of Electrical and Electronic, Faculty of Engineering, Universiti Malaysia Sarawak
94300 Kota Samarahan, Sarawak, Malaysia
Email: [email protected]
1. INTRODUCTION
The most common cause of mortality worldwide is lung cancer. In comparison to other cancers, lung
cancer is responsible for roughly 1.8 million fatalities (18.0% of all cancer deaths) in 2020 [1]. If lung cancer
is detected earlier, it can be treated effectively and even cured. The pulmonary nodules are the abnormal
development of lung tissue and are generally the first sign of lung cancer. As a result, it is critical to detect and
diagnose lung nodules earlier to reduce lung cancer mortality [2]. Radiologists utilize a variety of diagnostic
techniques to detect lung cancer, including magnetic resonance imaging (MRI), computed tomography (CT),
chest X-ray, and other methods. However, a CT scan is the most reliable screening approach due to its low
noise and robustness in determining tumor size [3]. With the widespread use of CT scanning technology, the
need for medical image analysis has risen and causes an increased workload for radiologists [4]. Traditionally,
achieving an accurate radiologic diagnosis requires the expertise of the radiologist throughout the analysing
process. The emergence of computer-aided tools has relieved the burden on radiologists and offered a second
opinion for doctors to diagnose lung cancer [5]. In traditional methods, there exist various techniques for feature
extraction and classification in the diagnosis of lung cancer. Among feature extraction techniques are the
threshold method, the region growth method, the morphology approach, and so on [6]. Machine learning
technologies, such as support vector machines (SVM) and k-nearest neighbor, are extensively employed to
distinguish the false-positive nodules [6]. However, the process of these traditional methods is complex, which
leads to poor generalization. This issue is due to the pulmonary nodule features needing to extract manually,
which causes the automation degree to be limited [7]. Motivated by the lung nodule analysis 16 (LUNA16)
challenge [8], numerous research has sought to automate the diagnosis of pulmonary nodules using deep
learning, particularly convolutional neural networks (CNN), to aid radiologists in the lung screening process
[9]. These techniques have substantially enhanced the treatment’s quality and dependability, particularly the
early detection of lung cancer. While lung cancer is slow and unreliable in survival rates compared to other
cancers, CNN techniques show promise performance. The computer-aided diagnosis (CAD) systems that
focused on CNN strategies can reduce lung cancer death rates by a factor of 25% in the last five years [10].
Despite the advancements in using CNN for lung cancer detection, there are still several issues that need to be
addressed. These include a high false-positive rate that limits its clinical application and the challenge of
surpassing the performance of conventional CAD systems in detecting lung cancer from CT images [11].
Moreover, the approach of adding more training data or fine-tuning hyperparameters to improve CNN’s
performance faces three main challenges. Firstly, there is observer variability among radiologists, which can
significantly affect the interpretation of the images. Secondly, detection networks may miss nodules that are
under-represented in the training set. Finally, neural networks are susceptible to unanticipated image distortions
[9].
To address the challenges of lung nodule classification, we propose a hybrid approach that combines
CNN and auto-regressive integrated moving average (ARIMA) for lung nodule classification using CT scan
images. This approach was motivated by previous research that used the hybrid CNN-ARIMA model for skin
cancer classification, which demonstrated promising results [12], [13]. In this approach, the CNN algorithm
serves as a feature extractor, while the ARIMA model functions as a classifier. The aim of this study is to
assess the performance of the hybrid CNN-ARIMA model for lung cancer classification.
2. RELATED WORKS
Several deep learning approaches had effectively applied in medical domains to address classification
problems [12]–[15]. A recent study showed that deep learning methods outperform conventional CAD systems
in detecting lung cancer with CT scans [11]. As a result, deep learning algorithms become beneficial for nodule
detection and classification, with very intriguing results.
Haiying et al. [16] proposed a new deformable convolutional neural network (DCNN) architecture to
address the problem of the similarity between true and false-positive nodules in early morphology in China.
The efficacy of the DCNN model was evaluated on the Lung Nodule Analysis 2016 dataset, achieving an
average competitive performance metric score of 0.835 and a high sensitivity of 0.941 and 0.958 to 4 and 8
false positives per scan, respectively. Sheng et al. proposed a method for detecting and diagnosing lung nodules
using two networks: Prediction network (PredNet) and diagnosis network (DiagNet). The PredNet uses
convolutional networks to predict the growth of nodules between consecutive CT scans through spatial
transformation. The DiagNet categorizes nodules based on their growth and previous diagnosis. The model
was tested on a dataset of 615 low-dose computed tomography (LDCT) images of 153 early-stage lung nodules
from 125 individuals. The proposed method demonstrated promising performance with a classification
accuracy of 90%. Many researchers have also shown interest in hybridization techniques due to their
outstanding performance in addressing cancer problems by enhancing algorithm performance in recent times
[10], [12], [13], [17]–[19]. Agarwal et al. [10] proposed a method to classify lung cancers as either malignant
or benign by combining a CNN with the AlexNet network model whereas Ardimento et al. [17] suggested a
novel ensemble-based method for classifying lung cancer more accurately using CT scan images.
Saleh et al. [20] developed an approach for classifying CT images of the lung using a hybrid of CNN and SVM
algorithms. The goal was to effectively detect the presence of cancer cells. The results showed that the proposed
CNN-SVM approach achieved an accuracy of 97.91% and had the ability to accurately classify lung cancer in
CT scans.
3. METHODOLOGY
This section discusses the study's methodology. The proposed approach in this study does not require
any prior techniques because the CNN can classify lesions based on high-level features without the need
for low-level nodule visual information and segmentation steps as required in traditional methods [10], [19].
Figure 1 depicts the image classification algorithm used in this study to detect lung cancer.
Development of hybrid convolutional neural network and autoregressive …(Abdulrazak Yahya Saleh)
1866 ISSN: 2252-8938
3.1. Dataset
In this study, the doctor from Sarawak General Hospital provided the dataset for the experiment. These
datasets contain 1,286 digital imaging and communications in medicine (DICOM) images divided between
normal lung and lung cancer images. There were 890 images of lung cancer and 396 images of normal lung.
These DICOM-format datasets need to convert to JPEG format for deep learning applications. Furthermore,
all images have a resolution of 512×512 pixels. Figure 2(a) shows the examples of normal lung. Meanwhile,
Figure 2(b) illustrates the lung cancer images and the red circle indicates the nodules.
(a) (b)
Figure 2. Type of lungs: (a) Normal lung and (b) Lung cancer – red circle indicates the nodules
3.4. CNN-ARIMA
The aim of this study was to combine the CNN and ARIMA algorithms to extract image features and
classify them. The CNN algorithm was used as a feature extractor, while the ARIMA algorithm acted as a
classifier by being integrated into a fully connected layer, which transformed the extracted features non-
linearly. To output two classes (normal lung and lung cancer), the traditional CNN algorithm was modified by
placing the ARIMA model between the flatten and dense layers of the fully connected layer. This hybridization
also evaluated the similarity between images of different categories by calculating the performance evaluation
for all classes when the proposed CNN-ARIMA model predicted a new image. Based on the approach
presented in [13], the decision-making process of the CNN-ARIMA model proposed in this study is achieved
through the implementation of a modified ARIMA algorithm in the fully connected layer.
The value 9 was used as the average out-shape value in the fully connected layer, which showed the
best performance in [13]. In addition, 𝑝, 𝑑, and 𝑞 are the auto-regressive, differencing, and moving average
values, respectively. However, the performance in this experiment was not affected by the specific values
chosen for 𝑝, 𝑑, and 𝑞. Therefore, these 𝑝, 𝑑, and 𝑞 set to a value of 1. Figure 4 illustrates the architecture of
CNN-ARIMA.
The datasets are resized to a 32×32 shape and converted into feature maps, which are arrays of every
pixel in the images, before being pre-processed and augmented for use in the model. The CT image, represented
as a 2D array, is then input into three components, each consisting of two convolutional layers with 64, 32, and
16 filters, respectively. After each convolutional layer in the feature extractor, a rectified linear unit (ReLU) is
applied as an activation function, and dropout with a ratio of 0.5 is utilised.
In the classification stage, the output obtained from the last dropout layer in the feature extraction stage
is flattened into a single array. Then, the ARIMA layer is used to improve the classification accuracy. A dense
layer with 512 units and a dropout ratio of 0.5 is added, followed by a final layer with two nodes for classifying
the CT images into two categories: lung cancer and normal lung. The SoftMax activation function is used in the
last three layers for classification.
Development of hybrid convolutional neural network and autoregressive …(Abdulrazak Yahya Saleh)
1868 ISSN: 2252-8938
𝑇𝑃
𝑃𝑟𝑒𝑐𝑖𝑠𝑖𝑜𝑛 = (3)
(𝑇𝑃+𝐹𝑃)
𝑇𝑁
𝑆𝑝𝑒𝑐𝑖𝑓𝑖𝑐𝑖𝑡𝑦 = (4)
(𝑇𝑁+𝐹𝑃)
after that, the ROC curve will show the performance of the binary classification model at all classification
thresholds due to their decision-making ability in the classification issues. The prediction performance will
show the output of an algorithm after being trained on the dataset.
(a) (b)
Figure 5. CLAHE implementation: (a) Sample of lung cancer image and (b) Clean image after CLAHE
This study achieved a test accuracy performance of 99.61% for CT image classification using the
CNN-ARIMA model. The loss performance decreased exponentially while the accuracy performance
increased exponentially for training and validation, as shown in Figure 6(a) and Figure 6(b), respectively,
leading to high accuracy. The results demonstrated that the training and validation performance of the CNN-
ARIMA on CT image classification were almost identical at every epoch, as illustrated in Figure 6. This
condition can deliver high classification accuracy in the proposed method.
(a) (b)
Figure 6. Performance during training and validation: (a) Loss and (b) Accuracy
Development of hybrid convolutional neural network and autoregressive …(Abdulrazak Yahya Saleh)
1870 ISSN: 2252-8938
According to Figure 7(a), the test datasets are labelled as “True label” in the confusion matrix. A total
of 258 test images (20% of the overall dataset) are distributed randomly, of which 172 images were distributed
in lung cancer and 86 images distributed in normal lung. Moreover, Figure 7(a) also shows the outstanding
confusion matrix performance in CT images classification, in which 171 images categorize as lung cancer and
86 images categorize as normal lung. According to the results presented in Table 2, the performance metrics
for CT image classification using the CNN-ARIMA model achieved an average sensitivity of 99.71%, an
average precision of 99.43%, an average specificity of 99.71%, and an average F1-score of 99.57%.
Furthermore, the ROC curve, as shown in Figure 7(b), demonstrated excellent performance with an area under
the curve (AUC) of 1.000.
(a) (b)
The prediction performance of the proposed work on the testing dataset shows remarkable
performance with accurate prediction. Figure 8(a) and Figure 8(b) show the examples of predicted lung cancer
and normal lung images from the testing dataset respectively. The comparative analysis in Table 3 demonstrates
that the performance of the proposed CNN-ARIMA model surpasses that of the existing techniques. Only work
from Saleh et al. [20] obtained a similar result with the proposed work in terms of AUC.
(a) (b)
Figure 8. Detection Performance: (a) Predict: Lung cancer, Actual: Lung cancer and
(b) Predict: Normal lung, Actual: Normal lung
5. CONCLUSION
This study presented the results of using a hybrid approach of CNN and ARIMA for CT image
classification. The hybrid of CNN and ARIMA shows better performance than the state-of-art deep learning
method. This study can further establish a CAD system, which assists radiologists and doctors in automatically
detecting and predicting lung conditions effectively. Nevertheless, this proposed work did not localize and
classify the pulmonary nodules in lung cancer image in depth. To further enhance the performance, the
segmentation step will consider as a future work because the efficiency of segmentation performance will
increase the diagnosis performance.
ACKNOWLEDGEMENTS
The authors express their gratitude towards the Ministry of Higher Education Malaysia for providing
financial assistance through the Fundamental Research Grant Scheme (FRGS)
FRGS/1/2021/SS0/UNIMAS/02/4. Additionally, the authors would like to extend their appreciation to the
Faculty of Cognitive Sciences and Human Development at Universiti Malaysia Sarawak (UNIMAS) for their
support and funding of the publication.
REFERENCES
[1] H. Sung et al., “Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185
countries,” CA: A Cancer Journal for Clinicians, vol. 71, no. 3, pp. 209–249, 2021, doi: 10.3322/caac.21660.
[2] Y. Chen, C. Liu, W. Peng, and S. Xia, “Thyroid nodule detection using attenuation value based on non-enhancement CT images,”
IET Conference Publications, vol. 2017, no. CP718, 2017, doi: 10.1049/cp.2017.0079.
[3] A. P. A. Thabsheera, T. M. Thasleema, and R. Rajesh, “Lung cancer detection using CT scan images: A review on various image
processing techniques,” Lecture Notes in Networks and Systems, vol. 43, pp. 413–419, 2019, doi: 10.1007/978-981-13-2514-4_34.
[4] R. J. McDonald et al., “The effects of changes in utilization and technological advancements of cross-sectional imaging on
radiologist workload,” Academic Radiology, vol. 22, no. 9, pp. 1191–1198, 2015, doi: 10.1016/j.acra.2015.05.007.
[5] A. Hosny, C. Parmar, J. Quackenbush, L. H. Schwartz, and H. J. W. L. Aerts, “Artificial intelligence in radiology,” Nature Reviews
Cancer, vol. 18, no. 8, pp. 500–510, 2018, doi: 10.1038/s41568-018-0016-5.
[6] L. Haibo, T. Shanli, S. Shuang, and L. Haoran, “An improved yolov3 algorithm for pulmonary nodule detection,” IMCEC 2021 -
IEEE 4th Advanced Information Management, Communicates, Electronic and Automation Control Conference, pp. 1068–1072,
2021, doi: 10.1109/IMCEC51613.2021.9482291.
[7] J. Liu, J. Li, F. Xue, and C. Wu, “Dense attention module for accurate pulmonary nodule detection,” ICASSP, IEEE International
Conference on Acoustics, Speech and Signal Processing - Proceedings, vol. 2021-June, pp. 1220–1224, 2021, doi:
10.1109/ICASSP39728.2021.9413936.
[8] A. A. A. Setio et al., “Validation, comparison, and combination of algorithms for automatic detection of pulmonary nodules in
computed tomography images: The LUNA16 challenge,” Medical Image Analysis, vol. 42, pp. 1–13, 2017, doi:
10.1016/j.media.2017.06.015.
[9] S. Liu et al., “No surprises: Training robust lung nodule detection for low-dose CT scans by augmenting with adversarial attacks,”
IEEE Transactions on Medical Imaging, vol. 40, no. 1, pp. 335–345, 2021, doi: 10.1109/TMI.2020.3026261.
[10] A. Agarwal, K. Patni, and D. Rajeswari, “Lung cancer detection and classification based on AlexNet CNN,” Proceedings of the 6th
International Conference on Communication and Electronics Systems, ICCES 2021, pp. 1390–1397, 2021, doi:
10.1109/ICCES51350.2021.9489033.
[11] S. Das and S. Majumder, “Lung cancer detection using deep learning network: A comparative analysis,” Proceedings - 2020 5th
International Conference on Research in Computational Intelligence and Communication Networks, ICRCICN 2020, pp. 30–35,
2020, doi: 10.1109/ICRCICN50933.2020.9296197.
[12] C. K. Chin, D. A. B. A. Mat, and A. Y. Saleh, “Hybrid of convolutional neural network algorithm and autoregressive integrated
moving average model for skin cancer classification among Malaysian,” IAES International Journal of Artificial Intelligence, vol.
10, no. 3, pp. 707–716, 2021, doi: 10.11591/ijai.v10.i3.pp707-716.
[13] C. K. Chin, D. A. Binti Awang Mat, and A. Y. Saleh, “Skin cancer classification using convolutional neural network with
autoregressive integrated moving average,” ACM International Conference Proceeding Series, pp. 18–22, 2021, doi:
10.1145/3467691.3467693.
[14] A. Tripathi, A. Arora, and A. Bhan, “Classification of cervical cancer using deep learning algorithm,” Proceedings - 5th
International Conference on Intelligent Computing and Control Systems, ICICCS 2021, pp. 1210–1218, 2021, doi:
10.1109/ICICCS51141.2021.9432382.
Development of hybrid convolutional neural network and autoregressive …(Abdulrazak Yahya Saleh)
1872 ISSN: 2252-8938
[15] A. Naeem, M. S. Farooq, A. Khelifi, and A. Abid, “Malignant melanoma classification using deep learning: Datasets, performance
measurements, challenges and opportunities,” IEEE Access, vol. 8, pp. 110575–110597, 2020, doi:
10.1109/ACCESS.2020.3001507.
[16] Y. Haiying, F. Zhongwei, D. Ding, and S. Zengyang, “False-positive reduction of pulmonary nodule detection based on deformable
convolutional neural networks,” 2021 IEEE 9th International Conference on Bioinformatics and Computational Biology, ICBCB
2021, pp. 130–134, 2021, doi: 10.1109/ICBCB52223.2021.9459209.
[17] P. Ardimento, L. Aversano, M. L. Bernardi, and M. Cimitile, “Deep neural networks ensemble for lung nodule detection on chest
CT scans,” Proceedings of the International Joint Conference on Neural Networks, vol. 2021-July, 2021, doi:
10.1109/IJCNN52387.2021.9534176.
[18] J. Sheng, Y. Li, G. Cao, and K. Hou, “Modeling nodule growth via spatial transformation for follow-up prediction and diagnosis,”
Proceedings of the International Joint Conference on Neural Networks, vol. 2021-July, 2021, doi:
10.1109/IJCNN52387.2021.9534163.
[19] P. Musa, F. Al Rafi, and M. Lamsani, “A review: Contrast-limited adaptive histogram equalization (CLAHE) methods to help the
application of face recognition,” Proceedings of the 3rd International Conference on Informatics and Computing, ICIC 2018, 2018,
doi: 10.1109/IAC.2018.8780492.
[20] A. Y. Saleh, C. K. Chin, V. Penshie, and H. R. H. Al-Absi, “Lung cancer medical images classification using hybrid CNN-SVM,”
International Journal of Advances in Intelligent Informatics, vol. 7, no. 2, pp. 151–162, 2021, doi: 10.26555/ijain.v7i2.317.
[21] G. Yadav, S. Maheshwari, and A. Agarwal, “Contrast limited adaptive histogram equalization based enhancement for real time
video system,” Proceedings of the 2014 International Conference on Advances in Computing, Communications and Informatics,
ICACCI 2014, pp. 2392–2397, 2014, doi: 10.1109/ICACCI.2014.6968381.
[22] A. Mikołajczyk and M. Grochowski, “Data augmentation for improving deep learning in image classification problem,” 2018
International Interdisciplinary PhD Workshop, IIPhDW 2018, pp. 117–122, 2018, doi: 10.1109/IIPHDW.2018.8388338.
[23] T. Tsao and D. Y. Tsao, “A topological solution to object segmentation and tracking,” Proceedings of the National Academy of
Sciences of the United States of America, vol. 119, no. 41, 2022, doi: 10.1073/pnas.2204248119.
[24] P. M. Radiuk, “Impact of training set batch size on the performance of convolutional neural networks for diverse datasets,”
Information Technology and Management Science, vol. 20, no. 1, 2018, doi: 10.1515/itms-2017-0003.
[25] A. M. Saleh and T. Hamoud, “Analysis and best parameters selection for person recognition based on gait model using CNN
algorithm and image augmentation,” Journal of Big Data, vol. 8, no. 1, 2021, doi: 10.1186/s40537-020-00387-6.
[26] O. G. Ajayi and J. Ashi, “Effect of varying training epochs of a faster region-based convolutional neural network on the accuracy
of an automatic weed classification scheme,” Smart Agricultural Technology, vol. 3, 2023, doi: 10.1016/j.atech.2022.100128.
[27] D. P. Kingma and J. L. Ba, “Adam: A method for stochastic optimization,” 3rd International Conference on Learning
Representations, ICLR 2015 - Conference Track Proceedings, 2015.
[28] H. Benbrahim, H. Hachimi, and A. Amine, “Deep convolutional neural network with TensorFlow and Keras to classify skin cancer
images,” Scalable Computing: Practice and Experience, vol. 21, no. 3, pp. 379–390, 2020, doi: 10.12694/scpe.v21i3.1725.
[29] M. Abadi et al., “TensorFlow: A system for large-scale machine learning,” Proceedings of the 12th USENIX Symposium on
Operating Systems Design and Implementation, OSDI 2016, pp. 265–283, 2016.
[30] M. Stojanović et al., “Understanding sensitivity, specificity and predictive values,” Vojnosanitetski Pregled, vol. 71, no. 11, pp.
1062–1065, 2014, doi: 10.2298/VSP1411062S.
BIOGRAPHIES OF AUTHORS
Mr Chee Ka Chin received his MSc and BSc from Faculty of Engineering, University
Malaysia Sarawak (UNIMAS). His research is mainly on the artificial intelligence and neural
network. The current focus of his research is on deep learning applications, data science and
computer vision. He can be contacted at email: [email protected].