TSP CMC 44065
TSP CMC 44065
DOI: 10.32604/cmc.2024.044065
ARTICLE
Mutasem K. Alsmadi*
Department of Management Information Systems, College of Applied Studies and Community Service, Imam Abdulrahman Bin
Faisal University, Dammam, Saudi Arabia
*Corresponding Author: Mutasem K. Alsmadi. Email: [email protected]
Received: 20 July 2023 Accepted: 24 October 2023 Published: 20 June 2024
ABSTRACT
Lung cancer is among the most frequent cancers in the world, with over one million deaths per year. Classification is
required for lung cancer diagnosis and therapy to be effective, accurate, and reliable. Gene expression microarrays
have made it possible to find genetic biomarkers for cancer diagnosis and prediction in a high-throughput manner.
Machine Learning (ML) has been widely used to diagnose and classify lung cancer where the performance of ML
methods is evaluated to identify the appropriate technique. Identifying and selecting the gene expression patterns
can help in lung cancer diagnoses and classification. Normally, microarrays include several genes and may cause
confusion or false prediction. Therefore, the Arithmetic Optimization Algorithm (AOA) is used to identify the
optimal gene subset to reduce the number of selected genes. Which can allow the classifiers to yield the best
performance for lung cancer classification. In addition, we proposed a modified version of AOA which can work
effectively on the high dimensional dataset. In the modified AOA, the features are ranked by their weights and
are used to initialize the AOA population. The exploitation process of AOA is then enhanced by developing a
local search algorithm based on two neighborhood strategies. Finally, the efficiency of the proposed methods
was evaluated on gene expression datasets related to Lung cancer using stratified 4-fold cross-validation. The
method’s efficacy in selecting the optimal gene subset is underscored by its ability to maintain feature proportions
between 10% to 25%. Moreover, the approach significantly enhances lung cancer prediction accuracy. For instance,
Lung_Harvard1 achieved an accuracy of 97.5%, Lung_Harvard2 and Lung_Michigan datasets both achieved 100%,
Lung_Adenocarcinoma obtained an accuracy of 88.2%, and Lung_Ontario achieved an accuracy of 87.5%. In
conclusion, the results indicate the potential promise of the proposed modified AOA approach in classifying
microarray cancer data.
KEYWORDS
Lung cancer; gene selection; improved arithmetic optimization algorithm and machine learning
1 Introduction
Classification is one of the most active areas of ML [1]. It entails training a predictive model
using a collection of samples containing a target output, such that, after training, the model could
predict the output for data that has not yet been seen. The field of biomedicine has recently become
This work is licensed under a Creative Commons Attribution 4.0 International License,
which permits unrestricted use, distribution, and reproduction in any medium, provided the
original work is properly cited.
5176 CMC, 2024, vol.79, no.3
quite prominent in ML due to the vast quantity of data that can be derived from genetic tissue [2–4].
Specifically, the rise of microarray datasets has led to the establishment of an active research area in
the domains of bioinformatics and ML.
Lung cancer detection has seen significant advancements in recent years, thanks to the integration
of machine learning and deep learning techniques. Recent meta-heuristic algorithms and method-
ologies have played a pivotal role in improving the accuracy and efficiency of image segmentation
including the CT scans of lung cancer detection models. These algorithms include Hunger Games
Search (HGS) [5], Slime Mould Algorithm (SMA) [6], Moth Search Algorithm (MSA) [7], Grey
Wolf Optimization (GWO) [8], Runge Kutta Method (RUN) [9], Monarch Butterfly Optimization
(MBO) [10], Weighted Mean of Vectors (INFO) [11], and Harris Hawks Optimization (HHO) [12]. By
incorporating these innovative approaches, researchers have significantly enhanced the performance
of machine learning and deep learning models in classification of lung cancer, ultimately contributing
to more accurate diagnoses and better patient outcomes. This integration of cutting-edge techniques
has brought about a paradigm shift in the field of lung cancer detection [13–18].
Lung cancer is among the most prevalent cancers in the world, causing over one million deaths
each year. Cancer is the greatest cause of death globally, with 10 million deaths in 2020, the most
prevalent in 2020 (new cancer cases) being the lung (2.21 million cases). Lung cancer (1.80 million
deaths) was the leading cause of death for cancer patients in 2020 [19–21].
Successful cancer diagnosis and therapy depend on accurate and reliable classification. Therefore,
attention has increased due to advancements in cancer classification [22]. Lung cancer classification
depends on clinicopathological characteristics. Gene expression microarrays have provided a high-
throughput platform for identifying genomic biomarkers for cancer diagnosis and prognosis [23,24].
The microarray data utilized by the classification algorithms may be shown as a matrix composed
of samples (rows) and features (columns) that describe the microarray data. Classification algorithms
need a sufficient number of samples since a limited number of samples might result in overfitting
and decreased generalizability. For lung cancer classification, microarray data is considered structured
data, described by the fact that it contains very few samples (often less than a hundred) yet comprises
a large number of features (often numbering in the thousands). Dealing with such many features with
few samples presents a machine learning research challenge, primarily due to the chances of “false
positives” being elevated in feature-rich microarray data (during the construction of the predictive
model or the gene selection process) [25]. Feature selection is an important and essential technique
used in data filtering to improve machine learning models in large databases and for dimensionality
reduction, where important features are chosen and unnecessary and redundant features are rejected
[26]. A decrease in input dimensionality may boost performance by reducing either the learning
speed and complexity of the model or increasing the generalization capacity and precision of the
classification. Selecting suitable characteristics may also lower measurement costs and enhance
comprehension of the problem.
The work presented by [27] showed that only a part of this significant number of genes or features
present on the microarray is important for solving certain classification problems. Feature selection is
required in this situation for two reasons. It removes extraneous and redundant functions and helps
experts identify key associations between specific diseases and gene expression. Fig. 1 is an illustration
of a common feature selection method used on microarray data.
Different feature selection models were developed in the literature, each using a feature score
measure to extract feature subsets from the same databases [27]. Training classification models with
all their high-dimensional features can lead to serious performance and scalability issues [28–30].
CMC, 2024, vol.79, no.3 5177
Selecting features that form a subset of the original features is useful where interpretability and
knowledge retrieval are important, like in medicine, although this sometimes results in some loss of
precision. Our approach focuses on feature selection since it is focused on medical issues. Feature
selection techniques may use individual scoring (also recognized as feature ranking) or feature subset
assessment, depending on the outcome. In the first case, the attributes are evaluated individually with
a weight that exhibits their relevance. In the latter case, the subsets of candidate traits are assessed
using some measures to choose the best trait.
Figure 1: A typical feature selection process applied to microarray data. Reprinted with permission
from [27]. Copyright © 2024 Elsevier Ltd.
Feature selection approaches are categorized into filters, embedded approaches, and wrappers
based on their correlation to the learning approach [26,31]. Filters are not dependent on learning
methods since they concentrate on the fundamental properties of the data. They are not compu-
tationally expensive and have strong generalization capability due to their independence from the
induction technique. To choose features, wrappers and embedded methods need a learning technique.
An induction approach assesses feature candidate subsets for wrappers. Although interacting with the
classifier makes it more computationally complex than filters, this strategy frequently produces better
results. Because the search for the ideal feature subset is done during classifier training, embedded
approaches, which fall between filters and wrappers, are computationally more effective than wrappers
[27]. There is also a trend to blend algorithms, as with hybrid approaches, which often integrate two
or more selection algorithms of characteristics with a distinct conceptual background. A common
example is to first use a less calculated filter to delete certain functions, then use a mathematically
expensive wrapper for fine nailing.
5178 CMC, 2024, vol.79, no.3
In hybrid methods based on filters and optimization methods, the dataset is filtered, and the
optimization algorithm iteratively chooses a subset of gene features from the chosen subset until opti-
mum classification accuracy is achieved. Such different algorithms are the brainstorming optimization
algorithm [31], Particle Swarm Optimization (PSO) [32], Whale Optimization [33], Genetic Algorithm
(GA) [34], and Moth Flam Optimization [35]. AOA was recently suggested by [30] and uses arithmetic
operators, including multiplication and division, for exploring the search space and addition and
subtraction operators for exploiting the search space, to discover the best solution for a given problem.
In summary, the gene selection method based on AOA and the feature importance ranking method is
proposed for the gene selection problem to work effectively on the high-dimensional dataset. Our goal
is to decrease the number of genes and improve classification accuracy based on microarray datasets
related to lung cancer. The contributions of this paper can be summarized as follows:
• A novel gene selection method combining AOA and feature importance ranking for high-
dimensional datasets is introduced.
• A modified AOA (MAOA) is presented to effectively navigate high-dimensional feature spaces,
addressing the exponential complexity associated with dimensionality.
• MAOA incorporates an innovative initialization process and a local search strategy, improving
the exploitation process of AOA and enhancing classification accuracy.
• Novel neighborhood strategies are proposed to enhance the effectiveness of feature subset
selection, further extending the adaptability and efficiency of the proposed approach
The structure of the paper is as follows: Section 2 offers a detailed literature review. Section 3
covers the methods and materials, including a discussion of AOA and MAO. Section 4 delves into the
results and facilitates a comprehensive discussion. In Section 5, conclusion and future works are given.
2 Literature Review
Lung cancer is a type of cancer that begins in the cells of the lungs. It is one of the most common
and deadliest forms of cancer worldwide [36]. The incidence and mortality rates vary by region, with
higher rates typically seen in developed countries and regions with a higher prevalence of smoking.
Besides smoking, other risk factors include exposure to secondhand smoke, occupational exposure to
carcinogens like asbestos and radon gas, air pollution, and a family history of lung cancer. In addition,
genetic and molecular studies have revealed various mutations and alterations associated with the
development of lung cancer.
According to Planchard et al. [37], symptoms of lung cancer can vary and may include persistent
coughing up blood, shortness of breath, chest pain, hoarseness, weight loss, and recurrent respiratory
infections. Diagnosing typically involves a combination of imaging tests such as X-rays, CT scans, and
PET scans, along with a tissue biopsy to determine the type and stage of the cancer. Staging helps
determine the extent of the cancer’s spread and guides treatment decisions. Treatment approaches
depend on the type, stage, and overall health of the patient, including surgery, radiation therapy,
chemotherapy, targeted therapy, immunotherapy, and palliative care. Newer treatment options, such as
immunotherapy and targeted therapy, have shown promise for improving outcomes for some patients
[38]. The prognosis varies widely depending on the stage of diagnosis, the type of lung cancer, and the
patient’s overall health. Early-stage lung cancers that are localized have a better prognosis compared
to advanced-stage cancers that have spread to other parts of the body.
There are several job selection methods described in the literature [31,39]. Community optimiza-
tion problems, feature selection, and intelligent optimization algorithms were suggested as efficient
CMC, 2024, vol.79, no.3 5179
solutions to the feature selection problem [40]. The authors in [40] proposed a GA-assisted feature
selection approach, including the ability to define subsets of features for recognition, to solve problems
in computer-aided diagnosis of learned tumors (in dermatology). The process improves classification
performance.
A multitude of optimization methods has been developed over the years. These methods can
be broadly categorized into different classes, including mathematical optimization, heuristic and
metaheuristic methods, evolutionary algorithms, and more. Some widely recognized optimization
techniques include linear programming, genetic algorithms, simulated annealing, particle swarm
optimization, and many others. Each of these methods possesses its unique characteristics, advan-
tages, and applications. For example, the study of Yu et al. [41] presented a hybrid model called
bERIME_FKNN, which combines the enhanced rime algorithm (ERIME) and fuzzy K-Nearest
Neighbor (FKNN) technique. The model uses a triangular game search strategy and a random
follower search strategy to enhance global exploration. Compared to 11 state-of-the-art algorithms,
the bERIME_FKNN model achieved an accuracy and specificity of over 99% in PH classification.
This makes it a valuable computer-aided tool for diagnosing and evaluating pulmonary hypertension,
preventing disease progression and reducing complications.
A GA was also used in [23] where a GA is integrated with the closet neighbor technique
to distinguish seeds through computer vision. The ‘invalid nearest neighbor’ approach determined
that the number be increased as the correct classification percent. The best findings were found
at generation 400 with an initialization probability of 0.1, resulting in a misclassification of 3.00%
between four types of seeds. Zhang et al. [42] represented a binary “feature selection algorithm based
on barebones particle swarm optimization” that was effectively used to address feature selection issues.
The quantitative findings exhibit that the suggested approach attains the best mean classification
accuracy for seven of the eight data sets utilized in the assessment. Yang et al. [43] suggested a method
to characterize selection with MA (“Memetic Algorithms”).
Yang et al. [43] stated that their suggested strategy can produce high grading accuracy with a
limited number of characteristics and is superior to PSO and GA approaches in accuracy, especially
for large problems. The given experimental findings demonstrate that this approach allows for greater
search efficiency and can produce excellent classification accuracy while minimizing the number of
features. Kabir et al. [44] represented a novel hybrid ant colony optimization algorithm for trait
selection (ACOFS). It is argued that this method not only strikes an efficient balance between the
exploitation and exploration of ants in research but also enhances the global research capabilities
of optimizing ant colonies in realizing high-quality solutions to trait selection problems. The findings
noted in the experimental tests demonstrate that ACOFS offers the notable capability to create subsets
of smaller sizes of relevant features while providing substantial classification accuracy. Huang et al. [45]
also addressed the feature selection problem with a nature-inspired solution. The suggested method
uses a novel hybrid method based on PSO-SVM (“Particle Swarm Optimization and Support Vector
Machines”) with feature selection as well as parameter optimization to address feature “subset
selection” with kernel parameter tuning.
To reduce the computation time, a data mining system is used, which is implemented through
a distributed design employing web services technology. The experimental findings demonstrate the
suggested method’s ability to choose discriminating input characteristics accurately while obtaining
high classification accuracy. Wang et al. [46] presented a new ACO (“Ant Colony Optimization”)
approach to enable feature selection based on coarse sentences and PSO to categorize hand motion
SEMG (“Surface Electromyography”) signals. The experimental findings imply that the recommended
5180 CMC, 2024, vol.79, no.3
strategy may attain high classification levels in the SEMG motion classification job when compared
to PCA (“Principal Component Analysis”).
An enhanced approach to IBPSO (“Binary Particle Swarm Optimization”) is suggested in [47]
to execute trait selection, a KNN (“K-Nearest Neighbor”) that acts as an IBPSO evaluator for
classification problems with data of gene expression. The experimental findings demonstrate that
the approach efficiently simplified the selection of genes (traits) and decreased the total number of
traits (genes) needed. Relevant research has shown that nature-inspired systems provide an efficient
foundation upon which feature selection can be attained. In this study, we used our new extended
SFLA [ISFLA] to select high-dimensional data of biomedical characteristics using a method that was
inspired by nature.
Although optimization algorithms have been utilized in the feature selection process as filter
or wrapper methods in the cancer classification task, it is still being investigated in recent studies.
Furthermore, there is still a need to conduct more research to investigate different hybrid approaches
and combinations of filter methods with optimization algorithms on different cancer datasets.
Several feature selection and machine learning methods were recently used on genetic datasets.
For instance, Ali et al. [48] presented a hybrid filter-genetic feature selection approach to address high-
dimensional microarray datasets. Elemam et al. [49] presented a highly discriminative hybrid feature
selection algorithm for cancer diagnosis. In addition, Almazrua et al. [50] and Vahmiyan et al. [51]
provided comprehensive reviews of feature selection and machine learning methods applied to cancer
datasets, including lung cancer.
Accordingly, presented challenges or limitations in hybrid gene selection methods for high-
dimensional lung cancer are the computational cost and complexity of hybrid methods; data imbalance
and standardization; and of course, parameter tuning and overfitting control. However, related works
discussed the concerns about some challenges of hybrid approaches when dealing with lung cancer,
mainly handling (1) heterogeneity; (2) small sample size; and (3) noisy imbalanced data. To overcome
these challenges, it is required to deeply understand lung cancer through gene selection methods.
Hence, researchers need to customize hybrid approaches to specific characteristics of lung cancer data,
and a comprehensive experimentation and clinical validation to validate the obtained results. Hybrid
approaches must address the three main challenges.
Therefore, this study proposes a hybrid gene selection method with an arithmetic optimization
algorithm (AOA). The AOA helps identify the optimal gene subset to reduce the number of selected
genes. Which can allow the classifiers to yield the best performance for lung cancer classification. In
this study, the AOA is chosen due to several advantages, which are presented in different related state-
of-the-art applications and implementations, such as: (1) it can efficiently handle high-dimensional
data using an exploration mechanism; (2) using an exploitation mechanism, it can identify the most
relevant set of genes to class labels; and (3) it has few parameters to tune with a simple structure. Thus,
the use of AOA will perform an effective feature selection that reduces the model’s overfitting and
speeds up training.
Algorithm 1 (continued)
if rand3 > 0.5:
Subtraction operator: Use Eq. (4) to update the solution
else:
Addition operator: Use Eq. (4) to update the solution
end if
end if
end for
end for
Crnt_Iter ← Crnt_Iter + 1
end while
Return the best solution (x).
AOA exhibits exploratory activity throughout the exploration phase. As per arithmetic operators,
the mathematical computations employing either the D (division) or even the M (multiplication)
operator produced highly scattered decisions or values (refer to several reigns), which commit to
the exploratory search process. Nevertheless, owing to their significant dispersion compared to other
operators (A and S), these operators (M and D) find it difficult to approach the objective. The impact
of the distribution values of the various operators is shown using a function built on the foundation
of four mathematical operations. Therefore, the scout research discovers a near-optimal solution that
could be derived after multiple iterations. Moreover, the exploration operators (M and D) were used in
this optimization phase to help the other phase (exploitation) in the research process via better mutual
communication.
AOA’s exploration operators randomly determine the search area in different areas and get closer
to discovering a better solution based on two primary search strategies (Multiplication and Division
search strategy), which are expressed in Eq. (2). This search phase (exploratory search by performing
M or D) is determined by the “accelerated function” (AF) presented in Eq. (1) for the condition that
any number is greater than AF. The first operator (D) is determined by a random number (r1) less
than 0.5 and 2nd operator (M) is neglected until this operator ends its present task. Otherwise, the
2nd operator (M) is activated to conduct the present task rather than D. A stochastic scale coefficient
for the element is considered to create additional diversification paths and investigate different areas
of the space of research. AOA envisions the following position update equations being offered for
exploration games:
best (xj ) ÷ (PF + ) × ((UBj − LBj ) × μ + LBj ), r2 < 0.5
xi,j (CrntIter + 1) = (2)
best (xj ) × PF × ((UBj − LBj ) × μ + LBj ), otherwise
where the solution within the next iteration is denoted by xi,j (Crnt_Iter + 1) which signifies the ith
solution’s j th position at the present iteration, and best(xj ) indicates the j th position in the best-obtained
solution so far. indicates a small integer number, UBj & LBj signify the upper and lower bound value
of the j th position, respectively. μ indicates the control parameter to modify the search process.
1
CrntIter α
PF (CrntIter ) = 1 − 1
(3)
MaxIter α
where probability function (PF) represents a coefficient, PF (Crnt_Iter) indicates the function value at
the tth iteration, Crnt_Iter signifies the current iteration, and (Max_Iter) signifies the highest number
CMC, 2024, vol.79, no.3 5185
of iterations. α represents a sensitive parameter and describes the accuracy of the exploitation across
iterations.
In the exploitation phase, mathematical computations that either used addition (A) or subtraction
(S) obtained high-density findings related to the exploitation search process. Nevertheless, these
traders (A and S) may quickly reach the target owing to their low spread, unlike other traders.
Therefore, despite the low spread of these traders (A and S), they are adept at rapidly approaching
the optimal solution, which may be derived after multiple iterations [30]. Moreover, the exploitation
operators (A & S) were operated in this phase of research (exploitation research by performing S or
A) conditioned by the value of the function AF for the condition that r1 is not higher than the current
AF Crnt_Iter value, as shown in Eq. (1). In AOA, the exploitation operators (addition (A) and subtraction
(S)) of AOA examine the search area in different dense areas and converge to discover the best solution
based on two primary search strategies (search by addition (A) and subtraction (S) strategy) modeled
in Eq. (4).
best xj − (PF + ) × ((UBj − LBj ) × μ + LBj ), r3 < 0.5
xi,j (CrntIter + 1) = (4)
best xj + PF × ((UBj − LBj ) × μ + LBj ), otherwise
This phase utilizes the search space by executing a depth-first search. In this phase, modeled in
Eq. (4), a random number (r3) conditions the 1st operator (S), less than 0.5, and the other operator
(A) is neglected until this operator ends its current task. Otherwise, the 2nd operator (A) oversees
performing the present task rather than S. These methods in this phase are like the partitions in the
previous phase. Nevertheless, exploitative search operators (A and S) frequently try to avoid being
stuck inside the local search perimeter. This approach supports exploration search methods to discover
the best solution and keep the variety of candidate solutions. The parameter μ is designed to create a
“stochastic value” at every iteration to maintain the exploration not only in the 1st iterations but also
during the last iterations.
S-shaped transfer function [55] to resolve the issues into binary search space as in Eq. (5).
1
TF(xij (CrntIter )) = (5)
1 + exp(−xij (CrntIter ))
where xij (CrntIter ) is the ith element in the dimension j and CrntIter indicates the current iteration number.
Eq. (6) is used in AOA to update a solution for S-shaped transfer functions to acquire the subset of
features.
0, if rnd > TF(xij (CrntIter ))
xij (CrntIter + 1) = (6)
1, if rnd ≤ TF(xij (CrntIter ))
The minimization of the number of chosen features while reducing the classification error rate
may be thought of as a “multi-objective optimization” problem for feature selection. Each solution is
assessed in accordance with the suggested fitness function, which relies on three classifiers to determine
the solution’s classification error rate and several features that were specifically chosen for the solution.
Both the AOA and modified AOA algorithms assess search performance using the objective function
in Eq. (7).
SF
OF = γ e + (1 − γ ) (7)
TF
where OF is the objective function to be minimized, γ ∈ [0, 1] represents the importance of
classification quality (1 − γ ) the importance of the number of features chosen and e indicates the
error rate of a particular classier. Also, SF indicates the number of “selected features” and TF denotes
the total number of features within the dataset [56].
feature in Sol, which was chosen at random and has a value of 1, will now be changed to 0 by the not
operator. Therefore, the updated solution would be as follows: x∗ = [1, 0, 0, 0, 1, 0],
NBMove neighborhood method is used instead of the addition operator, which chooses a random
feature and moves its location to a new position, For instance, assuming that the randomly selected
feature is positioned at the first position in vector x (which equals 1), then it is relocated to a new
random position” (let’s suppose the 4th position) as illustrated in Fig. 3.
Table 2: Average cross-validation Accuracy, Recall, Precision, and F1_Score using different classifica-
tion algorithms
Dataset name Classifier Mean accuracy Precision Recall F1_Score
KNN 92.59 88.38 82.76 83.91
SVM 68.48 13.70 20.00 16.26
Lung_Harvard1 DT 86.17 71.86 74.09 71.37
RF 92.12 79.80 71.25 73.78
NB 89.62 77.62 79.22 77.50
KNN 100 100 100 100
SVM 82.88 41.44 50.00 45.32
Lung_Harvard2 DT 92.25 88.52 83.71 84.78
RF 98.90 99.68 98.57 99.07
NB 98.90 99.35 96.90 97.99
KNN 68.63 53.24 51.56 48.91
SVM 72.09 36.05 50.00 41.88
Lung_Adenocarcinoma DT 58.17 55.67 56.18 55.13
RF 72.09 43.36 52.40 46.68
NB 70.92 68.63 60.92 61.33
KNN 98.00 89.00 90.00 89.47
SVM 89.58 44.79 50.00 47.25
Lung_Michigan DT 95.89 96.67 99.41 97.70
RF 96.95 89.00 90.00 89.47
NB 98.95 99.44 95.00 96.38
KNN 66.79 67.83 65.00 64.24
SVM 61.43 30.71 50.00 38.04
Lung_Ontario DT 54.64 62.14 58.17 57.11
RF 69.64 72.98 68.83 66.75
NB 64.29 70.45 65.67 62.31
CMC, 2024, vol.79, no.3 5189
Figure 4: Average cross-validation Accuracy, Precision, Recall, and F1_Score using different classifi-
cation algorithms
(“Grey Wolf Optimizer”) [8], MFO (“Moth-Flame Optimization”) [62], FA (“Firefly Algorithm”) [63]
and WOA (“Whale optimization algorithm”) [64].
The parameters settings of AOA and MOA used are:
• α (sensitivity of exploitation, Eq. (3)) = 5.
• μ (Adjustment of the search process, Eqs. (2) and (4)) = 0.5.
• min (range min of AF, Eq. (1)) = 0.2.
• max (range max of AF, Eq. (1)) = 0.9.
• RSN (the number of solutions to be considered for the update used for MAOA) = 10
To obtain a fair comparison, the algorithms under consideration have been applied with the same
number of iterations along with population sizes of 100, and 20, respectively, with Accuracy where
Max, Average Accuracy, and standard deviation are presented the Table 3 where the best-obtained
results are bolded.
Table 3: Min, Max, Mean, and Std classification accuracies MAOA with AOA and other compared
algorithms
Algorithm PSO GWO MFO WOA FFA AOA MAOA
Min 90.24 90.24 90.06 90.24 90.24 90.24 92.68
Max 95.12 92.68 95.12 92.68 95.12 97.12 97.56
Lung_Harvard1
Mean 93.17 91.92 91.97 91.52 92.50 93.41 95.61
Std 0.02 0.01 0.02 0.01 0.02 0.02 0.02
Min 100 100 97.22 97.22 100 97.22 97.22
Max 100 100 100 100 100 100 100
Lung_Harvard2
Mean 100 100 99.07 99.07 100 99.07 99.07
Std 0.00 0.00 0.02 0.02 0.00 0.02 0.02
Min 70.59 76.47 70.59 70.59 58.82 70.59 76.47
Max 88.24 88.24 88.24 88.24 82.35 88.24 88.24
Lung_Adenocarcinoma
Mean 77.65 80.59 82.94 79.41 75.88 78.24 84.12
Std 0.05 0.04 0.06 0.06 0.08 0.06 0.05
Min 100 100 100 100 94.74 100 100
Max 100 100 100 100 100 100 100
Lung_Michigan
Mean 100 100 100 100 98.25 100 100
Std 0.00 0.00 0.00 0.00 0.03 0.00 0.00
Min 75.00 75.00 75.00 75.00 75.00 75.00 75.00
Max 75.00 87.50 87.50 87.50 87.50 87.50 87.50
Lung_Ontario
Mean 75.00 81.25 76.25 85.00 81.25 83.92 81.25
Std 0.00 0.07 0.04 0.05 0.07 0.04 0.07
search using the neighborhood strategy and improvements in initial solution diversity using the AOA
solution initialization strategy.
To demonstrate the consistency of the outcomes produced by the suggested MAOA in terms of
feature reduction, the boxplot is presented in Fig. 5. The datasets that show differences are plotted,
including the Lung_Harvard2, Lung_Adenocarcinoma, and Lung_Ontario datasets. Notably, the
Lung_Harvard2 and Lung_Michigan datasets are not included due to their similar results observed
across various runs.
As shown in Fig. 5, MAOA is superior to the other algorithms, as can be seen in the line in
the middle of the box, along with the upper and lower bounds for the three data sets used in the
comparisons. The Wilcoxon statistical test was utilized to assess the importance of the reported
findings. This test makes it possible to evaluate the importance and robustness of the MAOA compared
to the original AOA and other optimization methods. The average classification accuracy across 10
runs was used to define the significance threshold for this test, which was set at 5%.
Table 4 shows the values of p achieved from the Wilcoxon trial of all competing methods. If the
p-value in this test is below 5%, it means that MAOA shows significant enhancement over competing
algorithms. Otherwise, there is no significant enhancement. The p-values for Lung_Harvard2 and
Lung_Michigan are indicated as ‘-’ due to the uniform identity of outcomes across all comparators,
where accuracy reached 100% (Table 3). This uniformity makes a t-test impractical. The findings in
Table 5 show that MAOA shows an important enhancement over other algorithms in terms of average
5192 CMC, 2024, vol.79, no.3
classification accuracy in most datasets. Table 5 shows the minimum, maximum, and mean number of
features chosen.
Table 4: The p-values of the Wilcoxon test for MAOA and other competing algorithms in terms of the
Mean accuracy over 10 independent runs
Algorithm PSO GWO MFO WOA FFA AOA
Lung_Harvard1 0.0264 0.0020 0.0264 0.0020 0.0098 0.0473
Lung_Harvard2 – – – – – –
Lung_Adenocarcinoma 0.0173 0.1552 0.7981 0.1138 0.0456 0.0744
Lung_Michigan – – – – – –
Lung_Ontario 0.0066 0.1128 0.0094 0.6733 0.1724 0.1718
Note: p ≥ 0.05 are presented in bold font.
Table 5: The Mix, Max, and Mean Number of selected features for the MAOA, AOA, and other
algorithms
Algorithm PSO GWO MFO WOA FFA AOA MAOA
Min 8956 6309 4908 6172 7009 7325 3253
Lung_Harvard1 Max 9248 6341 5730 6285 7109 7789 5276
Mean 9136.0 6327.7 5305.0 6241.3 7055.3 7581.7 3982.3
Min 8411 6097 5102 6156 6290 6188 1318
Lung_Harvard2 Max 8649 6201 6171 6167 6942 7248 1925
Mean 8494.7 6148.0 5487.0 6161.0 6510.7 6549.7 1691.0
Min 5075 3502 2864 3502 3891 4215 1323
Lung_Adenocarcinoma Max 5196 3613 4343 4834 3968 4370 2975
Mean 5141.3 3546.3 3592.3 3974.0 3938.0 4314.0 2226.5
Min 4729 3574 2588 3493 3501 4258 911
Lung_Michigan Max 5080 3628 3165 3500 3920 4300 1364
Mean 4914.0 3602.3 2820.3 3497.0 3656.0 4284.3 1130.3
Min 1995 1410 1140 1403 1564 1684 615
Lung_Ontario Max 2118 1441 1686 1423 1614 1792 1158
Mean 2061.0 1427.3 1338.3 1410.3 1593.0 1751.7 910.3
As shown in Table 5, MAOA outperformed AOA and other algorithms in all datasets because
it chose significantly fewer features. The superiority of MAOA is validated by a clear variation in
all data sets. These results obtained by MAOA emphasize that it perfectly initializes the solutions by
choosing the most informative features using the ranking method. Moreover, MAOA highlights its
stability to perform very well against high-dimensional datasets. Moreover, with each iteration, the
MAOA improves the solutions by selecting the relevant functionalities while removing the irrelevant
ones. This increased classification precision. It should be noted that the changes made to the AOA have
CMC, 2024, vol.79, no.3 5193
enhanced its mining and exploration abilities and given it more ability to search on high-dimension
datasets.
Table 6 provides a detailed breakdown of the proportion of features preserved under each
approach, addressing the aspect of feature complexity reduction.
In Table 6, each cell represents the percentage of features reserved for each algorithm. These
percentages show how different algorithms performed in the dimensionality reduction processes for
each dataset. MAOA significantly reduces the number of features compared to other algorithms for
most datasets.
Furthermore, the training accuracy and the number of selected features in each iteration are
utilized to produce the outcomes of the objective functions, as shown in Table 7.
Table 7: The Min, Mean and Std values for the Objective function in the MAOA, AOA, and other
algorithms
Algorithm PSO GWO MFO WOA FFA AOA MAOA
Min 0.055 0.072 0.067 0.061 0.061 0.059 0.054
Lung_Harvard1 Mean 0.056 0.077 0.071 0.063 0.061 0.062 0.056
Std 0.000 0.004 0.003 0.003 0.000 0.003 0.004
Min 0.005 0.005 0.005 0.005 0.005 0.005 0.001
Lung_Harvard2 Mean 0.005 0.009 0.005 0.005 0.005 0.005 0.001
Std 0.000 0.004 0.001 0.000 0.000 0.000 0.000
Min 0.239 0.267 0.268 0.223 0.224 0.239 0.210
Lung_Adenocarcinoma Mean 0.248 0.277 0.272 0.229 0.234 0.254 0.225
Std 0.001 0.017 0.008 0.008 0.008 0.014 0.014
Min 0.005 0.018 0.005 0.005 0.005 0.006 0.002
Lung_Michigan Mean 0.006 0.018 0.013 0.005 0.005 0.006 0.006
Std 0.001 0.000 0.007 0.000 0.000 0.000 0.007
Min 0.196 0.288 0.261 0.195 0.196 0.227 0.193
Lung_Ontario Mean 0.227 0.323 0.293 0.226 0.226 0.231 0.229
Std 0.207 0.303 0.283 0.216 0.206 0.229 0.217
5194 CMC, 2024, vol.79, no.3
Table 7 summarizes the minimum, mean, and standard deviation. All competing algorithms’
values for the objective function show that the MAOA outperforms them all in terms of generating the
minimum values for the objective function. As revealed in Table 7, MAOA outperforms other methods
(the best findings are shown in bold). These findings demonstrate the repeatability and stability of the
MAOA method on datasets with various degrees of dimension.
To further evaluate and scale the MAOA, Fig. 6 exhibits the mean convergence curves of the
competing algorithms across the five data sets, offering a visual representation of how the optimization
process evolves over iterations. MAOA continues to exhibit the best objective function values,
having the lowest values after 100 iterations. These findings demonstrate the appropriateness of the
incorporated amendment to the AIC. Therefore, from the experiments, MAOA always surpasses the
original AOA algorithm. This outperformance of the MAOA compared to the AOA derives from the
use of a local search algorithm based on two neighborhood operators to enhance the exploitation
capacity of the MAOA and prevent it from stagnating in the local optima.
Figure 6: (Continued)
CMC, 2024, vol.79, no.3 5195
Figure 6: Demonstrates the average convergence curves of the completed algorithms over the 5 datasets
Table 8 displays the average execution times of all competing methods to validate the impact of
the improvements on the MAOA’s execution time. It should be observed that the MAOA performed
better than all other algorithms in 4 out of 5 datasets in terms of average “execution time”. The MAOA
execution time proves to be lower than other algorithms, also, employing the local search algorithm
to reduce the selected features, leads to a reduction in the training time of the classifier, which makes
MAOA outperform other methods in terms of objective function values, feature reduction, accuracy,
and the statistical findings mentioned earlier. This combination of factors highlights the advantageous
balance achieved by MAOA.
Table 8: Comparison of Average execution time between the MAOA, AOA, and other algorithms
Algorithm PSO GWO MFO WOA FFA AOA MAOA
Lung_Harvard1 276.59 490.61 240.05 273.97 284.54 314.48 159.35
Lung_Harvard2 274.38 480.57 231.20 270.82 277.92 244.42 159.33
Lung_Adenocarcinoma 126.84 271.65 91.98 127.92 137.91 137.98 87.15
Lung_Michigan 128.14 259.90 92.08 130.58 138.47 136.67 93.94
Lung_Ontario 50.80 102.84 51.21 58.84 57.59 55.22 37.39
Lastly, to highlight the significance of the feature selection phase by the MAOA, Table 9 presents a
comparison between the MAOA accuracy with feature selection and the accuracy without utilizing the
feature selection produced by the best classifier taken from Table 2, which uses the full set of features.
The reported findings identify the significance of the feature selection phase, as it has led to improved
classification accuracy by reducing the dimensionality of such datasets and eliminating redundant and
irrelevant features. Considering all the above findings, MAOA has proven to be a competitor when used
to solve feature selection problems on high-dimension datasets.
5196 CMC, 2024, vol.79, no.3
Table 9: Mean accuracy using the full set of features of classification algorithms and with the MAOA
feature selection method
Dataset name Full set features Feature selection (MAOA)
(Best classifier)
Lung_Harvard1 KNN 92.59 97.56
Lung_Harvard2 KNN 100 100
Lung_Adenocarcinoma SVM 72.09 88.24
Lung_Michigan NB 98.95 100
Lung_Ontario RF 69.64 87.50
Acknowledgement: Many thanks to the Deanship of Scientific Research at the Imam Abdulrahman
Bin Faisal University for supporting this research.
Funding Statement: This work was supported by the Deanship of Scientific Research, at Imam
Abdulrahman Bin Faisal University. Grant Number: 2019-416-ASCS.
Author Contributions: Mutasem K. Alsmadi designed and implemented the proposed model, carried
out the experiments, and wrote the manuscript. The author read and approved the final manuscript.
Availability of Data and Materials: The data openly available in a public repository. The data that
support the findings of this study are available for lung adenocarcinoma at https://pubmed.ncbi.nlm.
CMC, 2024, vol.79, no.3 5197
nih.gov/12118244/, and the remaining datasets can be accessed from the DBCRepository at https://
leo.ugr.es/elvira/DBCRepository/LungCancer/.
Conflicts of Interest: The author declares that they have no conflicts of interest to report regarding the
present study.
References
[1] A. I. Kadhim, “Survey on supervised machine learning techniques for automatic text classification,” Artif.
Intell. Rev., vol. 52, no. 1, pp. 273–292, 2019. doi: 10.1007/s10462-018-09677-1.
[2] P. Mamoshina, A. Vieira, E. Putin, and A. Zhavoronkov, “Applications of deep learning in biomedicine,”
Mol. Pharm., vol. 13, no. 5, pp. 1445–1454, 2016. doi: 10.1021/acs.molpharmaceut.5b00982.
[3] C. Cao et al., “Deep learning and its applications in biomedicine,” Genom. Proteo. & Bioinform., vol. 16,
no. 1, pp. 17–32, 2018. doi: 10.1016/j.gpb.2017.07.003.
[4] S. M. D. A. C. Jayatilake and G. U. Ganegoda, “Involvement of machine learning tools in healthcare
decision making,” J. Healthc. Eng., vol. 2021, pp. 6679512, 2021. doi: 10.1155/2021/6679512.
[5] Y. Yang, H. Chen, A. A. Heidari, and A. H. Gandomi, “Hunger games search: Visions, conception,
implementation, deep analysis, perspectives, and towards performance shifts,” Expert. Syst. Appl., vol. 177,
no. 8, pp. 114864, 2021. doi: 10.1016/j.eswa.2021.114864.
[6] S. Li, H. Chen, M. Wang, A. A. Heidari, and S. Mirjalili, “Slime mould algorithm: A new
method for stochastic optimization,” Future Gener. Comput. Syst., vol. 111, pp. 300–323, 2020. doi:
10.1016/j.future.2020.03.055.
[7] G. G. Wang, “Moth search algorithm: A bio-inspired metaheuristic algorithm for global optimization
problems,” Memetic Comput., vol. 10, no. 2, pp. 151–164, 2018. doi: 10.1007/s12293-016-0212-3.
[8] S. Mirjalili, S. M. Mirjalili, and A. Lewis, “Grey wolf optimizer,” Adv. Eng. Softw., vol. 69, pp. 46–61, 2014.
doi: 10.1016/j.advengsoft.2013.12.007.
[9] I. Ahmadianfar, A. A. Heidari, A. H. Gandomi, X. Chu, and H. Chen, “RUN beyond the metaphor: An
efficient optimization algorithm based on Runge Kutta method,” Expert. Syst. Appl., vol. 181, no. 21, pp.
115079, 2021. doi: 10.1016/j.eswa.2021.115079.
[10] G. G. Wang, S. Deb, and Z. Cui, “Monarch butterfly optimization,” Neural Comput. Appl., vol. 31, no. 7,
pp. 1995–2014, 2019. doi: 10.1007/s00521-015-1923-y.
[11] I. Ahmadianfar, A. A. Heidari, S. Noshadian, H. Chen, and A. H. Gandomi, “INFO: An efficient
optimization algorithm based on weighted mean of vectors,” Expert. Syst. Appl., vol. 195, no. 12, pp.
116516, 2022. doi: 10.1016/j.eswa.2022.116516.
[12] A. A. Heidari, S. Mirjalili, H. Faris, I. Aljarah, M. Mafarja and H. Chen, “Harris hawks opti-
mization: Algorithm and applications,” Future Gener. Comput. Syst., vol. 97, pp. 849–872, 2019. doi:
10.1016/j.future.2019.02.028.
[13] X. Wang, L. Wang, and P. Zheng, “SC-dynamic R-CNN: A self-calibrated dynamic R-CNN model
for lung cancer lesion detection,” Comput. Math. Methods Med., vol. 2022, pp. 1–9, 2022. doi:
10.1155/2022/9452157.
[14] T. Sathiya, R. Reenadevi, and B. Sathiyabhama, “Lung nodule classification in CT images using grey wolf
optimization algorithm,” Ann. Romanian Soc. Cell Bio., vol. 25, no. 6, pp. 1495–1511, 2021.
[15] E. H. Houssein, H. N. Hassan, N. A. Samee, and M. M. Jamjoom, “A novel hybrid runge kutta optimizer
with support vector machine on gene expression data for cancer classification,” Diagnostics, vol. 13, no. 9,
pp. 1621, 2023. doi: 10.3390/diagnostics13091621.
[16] S. Murthy and P. M. K. Prasad, “Adversarial transformer network for classification of lung cancer
disease from CT scan images,” Biomed. Signal Process. Control, vol. 86, no. 1, pp. 105327, 2023. doi:
10.1016/j.bspc.2023.105327.
5198 CMC, 2024, vol.79, no.3
[17] E. H. Houssein, M. E. Hosney, W. M. Mohamed, A. A. Ali, and E. M. Younis, “Fuzzy-based hunger games
search algorithm for global optimization and feature selection using medical data,” Neural Comput. Appl.,
vol. 35, no. 7, pp. 5251–5275, 2023. doi: 10.1007/s00521-022-07916-9.
[18] Z. Guo, L. Xu, Y. Si, and N. Razmjooy, “Novel computer-aided lung cancer detection based on convo-
lutional neural network-based and feature-based classifiers using metaheuristics,” Int. J. Imaging Syst.
Technol., vol. 31, no. 4, pp. 1954–1969, 2021. doi: 10.1002/ima.22608.
[19] J. Ferlay et al., Global Cancer Observatory: Cancer Today. Lyon, France: International Agency for Research
on Cancer, 2022. Accessed: Jul. 27, 2022. [Online]. Available: https://gco.iarc.fr/today.
[20] C. de Martel , D. Georges, F. Bray, J. Ferlay, and G. M. Clifford, “Global burden of cancer attributable
to infections in 2018: A worldwide incidence analysis,” Lancet Glob. Health, vol. 8, no. 2, pp. e180–e190,
2020. doi: 10.1016/S2214-109X(19)30488-7.
[21] W. H. Organization, Assessing National Capacity for the Prevention and Control of Noncommunicable
Diseases: Report of the 2019 Global Survey, 2019. Accessed: Feb. 14, 2023. [Online]. Available: https://
www.who.int/publications/i/item/9789240002319
[22] P. Guan, D. Huang, M. He, and B. Zhou, “Lung cancer gene expression database analysis incorporating
prior knowledge with support vector machine-based classification method,” J. Exp. Clin. Cancer Res., vol.
28, no. 1, pp. 1–7, 2009. doi: 10.1186/1756-9966-28-103.
[23] D. G. Beer et al., “Gene-expression profiles predict survival of patients with lung adenocarcinoma,” Nature
Med., vol. 8, no. 8, pp. 816–824, 2002. doi: 10.1038/nm733.
[24] S. Ramaswamy, K. N. Ross, E. S. Lander, and T. R. Golub, “A molecular signature of metastasis in primary
solid tumors,” Nat. Genet., vol. 33, no. 1, pp. 49–54, 2003. doi: 10.1038/ng1060.
[25] G. Piatetsky-Shapiro and P. Tamayo, “Microarray data mining: Facing the challenges,” ACM SIGKDD
Explor. Newsl., vol. 5, no. 2, pp. 1–5, 2003. doi: 10.1145/980972.980974.
[26] I. Guyon, S. Gunn, M. Nikravesh, and L. A. Zadeh, Feature Extraction: Foundations and Applications,
Cham, Switzerland: Springer, vol. 207, 2008.
[27] B. Remeseiro and V. Bolon-Canedo, “A review of feature selection methods in medical applications,”
Comput. Biol. Med., vol. 112, no. 4, pp. 103375, 2019. doi: 10.1016/j.compbiomed.2019.103375.
[28] G. L. Kumari and N. N. M. Rao, “An efficient PSO based ensemble classification model on high
dimensional datasets,” Int. J. Soft Comput., vol. 8, no. 4, pp. 1–11, 2017. doi: 10.5121/ijsc.2017.8401.
[29] S. S. Shreem, S. Abdullah, M. Z. A. Nazri, and M. Alzaqebah, “Hybridizing ReliefF, MRMR filters and
GA wrapper approaches for gene selection,” J. Theor. Appl. Inform. Technol., vol. 46, no. 2, pp. 1034–1039,
2012.
[30] L. Abualigah, A. Diabat, S. Mirjalili, M. Abd Elaziz, and A. H. Gandomi, “The arithmetic opti-
mization algorithm,” Comput. Methods Appl. Mech. Eng., vol. 376, no. 2, pp. 113609, 2021. doi:
10.1016/j.cma.2020.113609.
[31] M. Alwohaibi, M. Alzaqebah, N. M. Alotaibi, A. M. Alzahrani, and M. Zouch, “A hybrid multi-
stage learning technique based on brain storming optimization algorithm for breast cancer recur-
rence prediction,” J. King Saud Univ.-Comput. Inf. Sci., vol. 34, no. 8, pp. 5192–5203, 2022. doi:
10.1016/j.jksuci.2021.05.004.
[32] M. Alzaqebah et al., “Hybrid feature selection method based on particle swarm optimization and
adaptive local search method,” Int. J. Elect. Comput. Eng., vol. 11, no. 3, pp. 2414, 2021. doi:
10.11591/ijece.v11i3.pp2414-2422.
[33] S. Sazzed, “Feature selection in gene expression profile employing relevancy and redundancy measures
and binary whale optimization algorithm (BWOA),” in Int. Conf. Adv. Data Min. Appl., Sydney, Australia,
Springer, 2022, pp. 45–60.
[34] C. Y. Lee, Y. J. Hsieh, and T. A. Le, “Induction motor fault classification based on combined genetic
algorithm with symmetrical uncertainty method for feature selection task,” Mathematics, vol. 10, no. 2,
pp. 230, 2022. doi: 10.3390/math10020230.
CMC, 2024, vol.79, no.3 5199
[35] M. Alzaqebah, N. Alrefai, E. A. Ahmed, S. Jawarneh, and M. K. Alsmadi, “Neighborhood search methods
with moth optimization algorithm as a wrapper method for feature selection problems,” Int. J. Elect. &
Comput. Eng., vol. 10, no. 4, pp. 3672, 2020. doi: 10.11591/ijece.v10i4.pp3672-3684.
[36] F. Bray, J. Ferlay, I. Soerjomataram, R. L. Siegel, L. A. Torre and A. Jemal, “Global cancer statistics 2018:
GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries,” CA: A
Cancer J. Clinic., vol. 68, no. 6, pp. 394–424, 2018. doi: 10.3322/caac.21492.
[37] D. Planchard et al., “Metastatic non-small cell lung cancer: ESMO clinical practice guidelines for diagnosis,
treatment and follow-up,” Ann. Oncol., vol. 29, pp. iv192–iv237, 2018. doi: 10.1093/annonc/mdy275.
[38] N. C. Institute, Non-Small Cell Lung Cancer Treatment (PDQ® )–Patient Version. USA: National Institutes
of Health, 2021. Accessed: Feb. 10, 2023. [Online]. Available: https://www.cancer.gov/types/lung/patient/
non-small-cell-lung-treatment-pdq
[39] P. Shi, S. Ray, Q. Zhu, and M. A. Kon, “Top scoring pairs for feature selection in machine learning
and applications to cancer outcome prediction,” BMC Bioinform., vol. 12, no. 1, pp. 1–15, 2011. doi:
10.1186/1471-2105-12-375.
[40] A. Bhattacharjee et al., “Classification of human lung carcinomas by mRNA expression profiling reveals
distinct adenocarcinoma subclasses,” Proc. Nat. Acad. Sci. USA, vol. 98, no. 24, pp. 13790–13795, 2001.
doi: 10.1073/pnas.191502998.
[41] X. Yu et al., “Synergizing the enhanced RIME with fuzzy K-nearest neighbor for diagnose
of pulmonary hypertension,” Comput. Biol. Med., vol. 165, no. 2023, pp. 107408, 2023. doi:
10.1016/j.compbiomed.2023.107408.
[42] Y. Zhang, D. Gong, Y. Hu, and W. Zhang, “Feature selection algorithm based on bare bones particle swarm
optimization,” Neurocomputing, vol. 148, no. 5, pp. 150–157, 2015. doi: 10.1016/j.neucom.2012.09.049.
[43] C. S. Yang, L. Y. Chuang, Y. J. Chen, and C. H. Yang, “Feature selection using memetic algorithms,” in
2008 Third Int. Conf. Converg. Hybrid Inform. Technol., Busan, Korea, IEEE, 2008, vol. 1, pp. 416–423.
[44] M. M. Kabir, M. Shahjahan, and K. Murase, “A new hybrid ant colony optimization algorithm for feature
selection,” Expert. Syst. Appl., vol. 39, no. 3, pp. 3747–3763, 2012. doi: 10.1016/j.eswa.2011.09.073.
[45] C. L. Huang and J. F. Dun, “A distributed PSO-SVM hybrid system with feature selection and parameter
optimization,” Appl. Soft Comput., vol. 8, no. 4, pp. 1381–1391, 2008. doi: 10.1016/j.asoc.2007.10.007.
[46] X. Wang, J. Yang, X. Teng, W. Xia, and R. Jensen, “Feature selection based on rough sets and particle swarm
optimization,” Pattern Recognit. Lett., vol. 28, no. 4, pp. 459–471, 2007. doi: 10.1016/j.patrec.2006.09.003.
[47] L. Y. Chuang, H. W. Chang, C. J. Tu, and C. H. Yang, “Improved binary PSO for feature
selection using gene expression data,” Comput. Biol. Chem., vol. 32, no. 1, pp. 29–38, 2008. doi:
10.1016/j.compbiolchem.2007.09.005.
[48] W. Ali and F. Saeed, “Hybrid filter and genetic algorithm-based feature selection for improving can-
cer classification in high-dimensional microarray data,” Processes, vol. 11, no. 2, pp. 562, 2023. doi:
10.3390/pr11020562.
[49] T. Elemam and M. Elshrkawey, “A highly discriminative hybrid feature selection algorithm for cancer
diagnosis,” The Scitif. World J., vol. 2022, no. 3, pp. 1–15, 2022. doi: 10.1155/2022/1056490.
[50] H. Almazrua and H. Alshamlan, “A comprehensive survey of recent hybrid feature selection methods in
cancer microarray gene expression data,” IEEE Access, vol. 10, pp. 71427–71449, 2022. doi: 10.1109/AC-
CESS.2022.3185226.
[51] M. Vahmiyan, M. Kheirabadi, and E. Akbari, “Feature selection methods in microarray gene expression
data: A systematic mapping study,” Neural Comput. Appl., vol. 34, no. 22, pp. 19675–19702, 2022. doi:
10.1007/s00521-022-07661-z.
[52] G. J. Gordon et al., “Translation of microarray data into clinically relevant cancer diagnostic tests using
gene expression ratios in lung cancer and mesothelioma,” Cancer Res., vol. 62, no. 17, pp. 4963–4967, 2002.
[53] D. A. Wigle et al., “Molecular profiling of non-small cell lung cancer and correlation with disease-free
survival,” Cancer Res., vol. 62, no. 11, pp. 3005–3008, 2002.
[54] H. Liu and H. Motoda, Feature Extraction, Construction and Selection: A Data Mining Perspective. Springer
New York: Springer Science & Business Media, 1998.
5200 CMC, 2024, vol.79, no.3
[55] A. W. Mohamed and H. Z. Sabry, “Constrained optimization based on modified differential evolution
algorithm,” Inf. Sci., vol. 194, no. 6, pp. 171–208, 2012. doi: 10.1016/j.ins.2012.01.008.
[56] E. Emary, H. M. Zawbaa, and A. E. Hassanien, “Binary ant lion approaches for feature selection,”
Neurocomputing, vol. 213, no. 6, pp. 54–65, 2016. doi: 10.1016/j.neucom.2016.03.101.
[57] P. Thanh Noi and M. Kappas, “Comparison of random forest, k-nearest neighbor, and support vector
machine classifiers for land cover classification using sentinel-2 imagery,” Sensors, vol. 18, no. 1, pp. 18,
2017. doi: 10.3390/s18010018.
[58] M. Tubishat et al., “Dynamic salp swarm algorithm for feature selection,” Expert. Syst. Appl., vol. 164, no.
1, pp. 113873, 2021. doi: 10.1016/j.eswa.2020.113873.
[59] A. E. Hegazy, M. Makhlouf, and G. S. El-Tawel, “Feature selection using chaotic salp swarm algorithm for
data classification,” Arab. J. Sci. Eng., vol. 44, no. 4, pp. 3801–3816, 2019. doi: 10.1007/s13369-018-3680-6.
[60] R. A. Khurma, I. Aljarah, A. Sharieh, and S. Mirjalili, “Evolopy-fs: An open-source nature-inspired
optimization framework in python for feature selection,” in S. Mirjalili, H. Faris, and I. Aljarah (Eds.),
Evolutionary Machine Learning Techniques: Algorithms and Applications. Singapore: Springer, 2022, pp.
131–173.
[61] J. Kennedy and R. Eberhart, “Particle swarm optimization,” in Proc. ICNN’95-Int. Conf. Neural Netw.,
Perth, Australia, IEEE, 1995, vol. 4, pp. 1942–1948.
[62] S. Mirjalili, “Moth-flame optimization algorithm: A novel nature-inspired heuristic paradigm,” Knowl.-
Based Syst., vol. 89, pp. 228–249, 2015. doi: 10.1016/j.knosys.2015.07.006.
[63] A. H. Gandomi, X. S. Yang, and A. H. Alavi, “Mixed variable structural optimization using firefly algo-
rithm,” Comput. & Struct., vol. 89, no. 23–24, pp. 2325–2336, 2011. doi: 10.1016/j.compstruc.2011.08.002.
[64] S. Mirjalili and A. Lewis, “The whale optimization algorithm,” Adv. Eng. Softw., vol. 95, no. 12, pp. 51–67,
2016. doi: 10.1016/j.advengsoft.2016.01.008.