The Spherical Search Algorithm For Bound-Constrained Global Optimization Problems
The Spherical Search Algorithm For Bound-Constrained Global Optimization Problems
The Spherical Search Algorithm For Bound-Constrained Global Optimization Problems
PII: S1568-4946(19)30515-0
DOI: https://doi.org/10.1016/j.asoc.2019.105734
Reference: ASOC 105734
Please cite this article as: A. Kumar, R.K. Misra, D. Singh et al., The spherical search algorithm for
bound-constrained global optimization problems, Applied Soft Computing Journal (2019), doi:
https://doi.org/10.1016/j.asoc.2019.105734.
This is a PDF file of an article that has undergone enhancements after acceptance, such as the
addition of a cover page and metadata, and formatting for readability, but it is not yet the definitive
version of record. This version will undergo additional copyediting, typesetting and review before it
is published in its final form, but we are providing this version to give early visibility of the article.
Please note that, during the production process, errors may be discovered which could affect the
content, and all legal disclaimers that apply to the journal pertain.
Abhishek Kumara , Rakesh Kumar Misraa , Devender Singha , Sujeet Mishrab , Swagatam Dasc,∗
a Department of Electrical Engineering, Indian Institute of Technology (BHU), Varanasi, Varanasi, 221005, India
b Chief Design Engineer (Electrical), Diesel Locomotive Works, Varanasi, India
c Electronics and Communication Sciences Unit, Indian Statistical Institute, Kolkata, India.
of
Abstract
pro
In this paper, a new optimization algorithm called Spherical Search (SS) is proposed to solve the bound-constrained
non-linear global optimization problems. The main operators of SS are the calculation of spherical boundary and
generation of new trial solution on the surface of the spherical boundary. These operators are mathematically mod-
eled with some more basic level operators: Initialization of solution, greedy selection and parameter adaptation,
and employed on the 30 black-box bound constrained global optimization problems. This study also analyzes the
applicability of the proposed algorithm on a set of real-life optimization problems. Meanwhile, to show the robust-
ness and proficiency of SS, the obtained results of the proposed algorithm are compared with the results of other
re-
well-known optimization algorithms and their advanced variants: Particle Swarm Optimization (PSO), Differential
Evolution (DE), and Covariance Matrix Adapted Evolution Strategy (CMA-ES). The comparative analysis reveals
that the performance of SS is quite competitive with respect to the other peer algorithms.
Keywords: Spherical Search Algorithm, Real-life Optimization Problems, Bound constrained optimization problem,
Optimization Algorithm, Global Optimization.
lP
1. Introduction
For over the last few decades, complexity of real-life optimization problems has been rapidly increasing with
the advent of latest technologies. Solving these optimization problems is an essential component of any engineering
design problem. So far numerous optimization techniques have been proposed and adapted to provide the optimal
a
solutions for different optimization problems. According to the nature of operators, these algorithms can be classified
into two classes: Deterministic techniques and Meta-heuristics. In deterministic techniques, the solution of the previ-
urn
ous iteration is used to determine the updated solution for the current iteration. Therefore, in the case of deterministic
techniques, the choice of the initial solution influences the final solution. Furthermore, the solutions can be a victim
of easily getting trapped into the local optima. Consequently, deterministic techniques are a less efficient and less
effective tool for solving multi-modal, highly complex, and high-dimensional optimization problems. As an alternate
technique, meta-heuristics have been preferred for solving global optimization problems. A lot of theoretical work on
these algorithms have been published in various popular journals thereby mainstreaming meta-heuristics. Principal
reasons for the popularity of the said algorithms over deterministic techniques are as follows:
Simplicity- Foremost characteristic of meta-heuristics is the simplicity of theories and techniques. Meta-heuristics
Jo
are basically inspired by the simple concepts of some biological or physical phenomena.
Flexibility- Meta-heuristics can easily be applied to the different optimization problems with no change or minor
changes in the basic structure of the technique. Most of the techniques of meta-heuristics assume the problem as a
black box requiring only input and output.
∗ Correspondingauthor
Email address: [email protected] (Swagatam Das )
Derivative free- The most important characteristic of these algorithms is a derivative-free mechanism. This means
that there is no need for derivatives to solve the real-life optimization problems having complex search space with
multiple local minima.
Local optima avoidance- Meta-heuristics have an in-built capability for local optima avoidance. Local optima
avoidance is required in the optimization of multi-modal problems. Meta-heuristics hence are preferred over conven-
tional techniques for finding global optima of multi-modal problems.
Researchers have introduced many meta-heuristics. Some of them are popular because of showing good efficiency
for most of the real-life optimization problems. The No-Free-Lunch theorem [1] logically proved that there is no uni-
of
versal method, which solves all type of problems efficiently. A particular method may show very efficient performance
on a particular type of problem, but, the same method may show very poor performance on a different problem.
Meta-heuristics are classified into two classes on the basis of population: one is single-agent based and other is
population-based. In single-agent based algorithms, the process starts with a single initial solution. The solution
is improved over the course of iterations. In the case of population-based algorithms, multiple initial solutions are
pro
generated. They are improved over the course of iterations by sharing the group information of search space. The
population-based algorithm has the following inherent advantages:-
1. Multiple solutions share their search information, by which they explore the promising areas of the search space
efficiently.
2. Multiple solutions may avoid the local optima of the search space by comparing other solutions.
3. Multiple solutions explore larger search space in one iteration than the single-agent algorithm but they require
a greater number of function evaluations.
re-
Some Algorithms of meta-heuristics are inspired by natural phenomenons. According to the source of inspiration,
these algorithms can be divided into two classes.
1. Evolutionary algorithms: The evolutionary algorithms (EAs) are motivated by process of natural evolution,
such as mutation, crossover, reproduction, and selection. Some popular EAs are Genetic Algorithm (GA) [2],
Evolution Strategies (ES) [3], Evolutionary Programming (EP) [4, 5], Genetic Programming (GP) [6], and
lP
Regardless of the differences among all the classes of meta-heuristics, two major characteristics are common for
all: exploration and exploitation. By using exploration characteristic of an algorithm, the individuals try to investi-
gate the promising areas of the search space. However, by using the exploitation characteristic of an algorithm, the
urn
individuals try to locally investigate the promising areas. To converge on a global minimum of the problem, a proper
balance between exploration and exploitation operator of an algorithm is necessary. For any new meta-heuristics
method, proper balancing between the exploration and exploitation operator is a challenging task. The wide spectrum
of research on and with such bio-inspired optimization algorithms have been extensively reviewed in [28].
Various parameters are introduced to effectively employ the exploration and exploitation characteristics in an
algorithm. These parameters are specifically tuned for a problem to effectively balance the exploration and exploitation
characteristics of an algorithm. Thus, the performance of meta-heuristics on a particular type of problem also depends
on the parameters they employ. The values of parameters for a particular problem are not known beforehand and one
Jo
has to employ some rule of thumb or heuristics or trial and error. One of the major contributions of the proposed
method is that it is a parameter-free search algorithm.
This work proposes a new swarm based metaheuristic to optimize bound-constraint optimization problems. The
main aim of this work is to introduce a new metaheuristic called Spherical Search (SS). The main characteristics of
the SS algorithm are as follows:
1. Less number of parameters.
2. Proper balance between exploration and exploitation.
2
Journal Pre-proof
of
7. Snap-drift Cuckoo Search [19]
8. Virus Colony Search [20]
9. Grasshopper Optimisation Algorithm [21]
10. Salp Swarm Algorithm [22]
pro
11. Squirrel Search Algorithm [23]
12. Interactive Search Algorithm [24]
13. Emperor Penguin Optimizer [25]
14. Coyote optimization algorithm [26]
15. Spotted Hyena Optimizer [27]
3. Rotational-Invariance.
re-
4. Maps the contour of search space.
5. Maintain high diversity during the optimization process.
The performance of the SS algorithm is assessed on two well-known benchmark suite viz. IEEE CEC 2014 and
IEEE CEC 2011 problem suite. The obtained outcomes of proposed algorithm are compared with the state-of-art
meta-heuristics. The comparative analysis proves that the SS algorithm operates better than the state-of-art meta-
heuristics. Other sections of this paper are organized as follows:
lP
In this section, the mathematical modeling of the SS algorithm is discussed and developed.
SS algorithm is a swarm based meta-heuristics proposed to solve the non-linear bound-constrained global opti-
mization problems. It shows some properties similar to other popular meta-heuristics, particularly PSO and DE. In
the SS algorithm, the search space is represented in the form of vector space where the location of each individual in
space is a position vector representing a candidate solution to the problem.
In a D-dimension search space, for each individual, (D − 1)-spherical boundary is prepared towards the target
Jo
direction in every iteration before generating the trial location of the individual. Here the target direction is the main
axis of the spherical boundary and the individual lies on the surface of the spherical boundary. An example of a 2-D
search space is depicted in figure (1). In this figure 1-spherical boundary for each of individual is shown as –, target
location has been shown by F F and each 1-spherical boundary is generated using axis obtained by individual location
and target locations. trial solutions appear on the 1-spherical boundary. Thus, in every iteration, the trial location for
each individual is generated on the surface of (D − 1)-spherical boundary. An objective function value determines
the fitness value of a location. On the basis of the fitness value of the trial locations, better locations pass on into the
next iteration as individual locations. In the SS algorithm, solution update procedure and spherical search movement
3
Journal Pre-proof
Individual Location
of
Trial Location
Target Location
pro
Circular Boundary
re-
Figure 1: Demonstrating the 1-spherical (circular) boundary of individuals of SS algorithm in 2-D search space
lP
balance the ability of exploration and exploitation. When the (D − 1)-spherical boundary is small, exploitation of
search-space is emphasized in the algorithm. On the other hand, in case of larger (D − 1)-spherical boundary, explo-
ration of the space gets emphasized. It is evident that when the target location of a respective individual is far-off the
individual has a tendency to explore, as the spherical boundary is large. This is advantageous as in such conditions
it is better to explore the larger search space. on the contrary, when the target locations of a respective individual are
nearby the individual has a tendency to exploit as the spherical boundary becomes small. this is advantageous as in
a
best solution reaches near to the predefined solution within a specified tolerance. For the reported experimental work
of this paper, both the stopping criteria have been used.
Nomenclature
Start
of
Trial Solution
Generation
pro
Calculation of Trial Solution
Selection of solution
re-
Stopping No
criteria ?
Yes
lP
Stop
(k)
Bi(k) = diag(b̄i ) a binary diagonal matrix at iteration k
(k)
b̄i consist of diagonal elements of matrix B(k)
i at iteration k for i−th solution of population
urn
(N×1)
c̄(k) ∈ R>0 a step-size control vector for iteration k
c(k)
i ∈ R>0 i−th element of step-size control vector c̄(k) , a step-size control parameter for i−th solution of
population
diag(b̄i ) a square diagonal matrix with the elements of vector b̄i on the diagonal
The main steps of SS algorithm depicted in the flowchart (figure (2)) are described as follows:
Jo
where
(k) (k) (k) T
x̄i (k) = [xi1 , xi2 , .....xiD ] (2)
5
Journal Pre-proof
Y
𝑟ഥ1
𝑟ഥ2
of
𝑥ҧ𝑡
𝑐𝑖 𝑧𝑖ҧ + 𝑥ҧ𝑖
pro
𝑧𝑖ҧ + 𝑥ҧ𝑖
O X
Figure 3: Demonstrating the solution update scheme of SS algorithm in 2-D search space
lP
here, xi j is the value of jth element (parameter) of ith solution and D is the total number of elements (parameters).
So, x̄i is actually representing a point on D−dimensional search space. Here, xi0j is initialized using random uniform
distribution between pre-specified lower and upper bounds of jth element as follows:
xi0j = (xh j − xl j ) ∗ rand(0, 1] + xl j , (3)
where xh j and xl j represent the upper and lower bounds of jth element respectively. Also, rand(0, 1] generates random
a
In the case of population-based optimization algorithms, in every iteration, there will be a need of calculation of
potential new solutions that compete with the old solution to become a part of the population for the next iteration. In
this algorithm, the name, trial solution, is used to represent these potential new solutions.
In SS algorithm, for each solution, a (D − 1)-spherical boundary is prepared where the search direction passes
through the main axis of boundary i.e search direction crosses the center of (D − 1)-spherical boundary. A simple
generation process of trial solutions, ȳi , is demonstrated on a 2-D search space in figure (3). It is seen from the figure
(3), that the (D − 1)-spherical boundary (locus of ȳi ) is a circle (1-sphere) with diameter of ci z¯i . Points represented
Jo
by vectors x̄i and ci z¯i + x̄i lie on the boundary of 1-sphere. Point represented by z̄i is the search direction determined
using vectors r̄1 , r̄2 and x̄t . The determination of search direction z̄i is described in Section-2.2.1
In SS algorithm, following equation is used to generate a trial solutions for i−th solution.
ȳi (k) = x̄i (k) + c(k) (k) (k)
i Pi z¯i , (4)
where, Pi is a projection matrix, which decides the value of ȳi k on the (D − 1)-spherical boundary. For a particular
¯ , different possible values of P(k) yield different values of ȳ(k) . Locus of ȳ(k) gives (D − 1) spherical
solution, xi(k) i i i
boundary.
6
Journal Pre-proof
To define all the iterative steps of the SS algorithm, the calculation procedure of z¯i (k) , ci(k) and P(k)
i are discussed in
following sections.
of
Search direction, z¯i (k) , should be generated in such a way that it guides the i−th solution towards the better solu-
tions. A simple illustration of calculation of search direction is shown in figure (3). From figure (3), it is clear that x̄t ,
r1 and r2 are needed to calculate the search direction as follows.
pro
where, x̄t is the target point. In equation (5), two random solutions r1 and r2 are selected from the current set of
solutions (population). So, the actual search direction deviates by some angle from target direction, as shown in figure
(3).
In this paper, two methods are introduced to calculate the search direction, namely towards-rand and towards-
best. Method towards-rand has a better exploration capability and towards-best improves the exploitation capability.
So to provide a good balance between the exploration and exploitation of search space, for the half population of
better solution, calculation of search direction can be done by towards-rand and for the rest half of the population,
re-
towards-best is used to calculate the search direction thereby forcing diversity in the set of better solutions and forcing
the inferior solutions to strive for improved fitness.
In towards-rand, the search direction, z¯i (k) , for ith solution at kth iteration is calculated using following equation.
where
pi , qi , and ri are randomly selected indices from among 1 to N such that pi , qi , ri , i.
While in towards-best, the search direction, z¯i (k) , for ith solution at kth iteration is calculated using following
equation.
z̄(k) (k) (k) (k) (k)
i = x̄ pbesti + x̄qi − x̄ri − x̄i , (7)
where x(k) pbesti represents the randomly selected individual from among the top p solutions searched so far.
a
Here, x̄ pi and x̄ pbesti represent the target points in towards-rand and towards-best respectively. Difference term ( x¯q −
x¯r ) is common in both towards-rand and towards-best which represents r̄1 −r̄2 (equation 5), which is an approximate of
distribution of difference of solutions in a population. In calculation of new search direction, difference term ( x¯q − x¯r )
urn
makes the population to evolve maintaining the diversity of solution thereby avoiding convergence to local minima.
combinations are infinite. Therefore, all the possible projections of ci z¯i + x̄i create a (D − 1)-spherical boundary on
search space. To illustrate the locus of all possible projection of a point on a 2-D search space, 10, 000 randomly
generated samples of projection matrix, P, transforming a point (1, 1) are plotted in figure (4). It is √ seen from the
figure (4) that locus of P × [1, 1]0 is a circular ring passing through (0, 0) and (1, 1) with a diameter of 2 and center
(0.5, 0.5).
Method to compute elements of P along with c has been illustrated as follows.
7
Journal Pre-proof
1.4
1.2
of
0.8
0.6
pro
0.4
0.2
-0.2
re-
-0.4
-0.4 -0.2 0 0.2 0.4 0.6 0.8 1 1.2 1.4
AA0 = I. (8)
Binary diagonal matrix, diag(b̄i ), are calculated randomly in such a way that,
urn
8
Journal Pre-proof
of
• Step 2: Calculate the objective function of each solution of P.
• Step 3: The best solution of population is selected as best solution.
• Step 4: Calculate the search direction for each solution of population P.
pro
• Step 5: Calculate the orthogonal matrix, A.
• Step 6: Calculate parameters: ci and rank of projection matrix.
• Step 7: Calculate trial solution for each solution of population P.
• Step 8: Update the population using greedy selection operator.
• Step 9: If the stopping criterion is satisfied then the algorithm will be stopped, otherwise it will return to Step
re-
3.
• Step 10: Return the best optimal solutions, after stopping criteria is satisfied.
Algorithm 1 SS
lP
12: end if
13: ȳi ← x̄i + ci A0 diag(b̄i )Az̄i
14: O fi ← ObjectiveFunction(ȳi )
15: FEs ← FEs + 1
16: x̄i ← Selection( x̄i , ȳi )
17: end for
18: P1 ← Sort(P1)
19: end while
20: end procedure
Jo
of
Hence, the overall time complexity of proposed algorithm is O(Maxiter × N × D × log(D)) = O(FEmax × D × log(D)).
pro
function[29] is simulated as a benchmark problem. For minimization of Griewank function, all algorithms are em-
ployed using 100 individuals for 200 iterations.
To describe the performance of SS algorithm with respect to PSO and DE, four qualitative metrics: Search
History, Search Trajectory of individual, Population diversity, and Convergence curve are used to describe the
performance of SS algorithm.
compared to PSO and DE. In the SS algorithm, the distribution of solutions is sparse in the worst region of search
space and is dense in the region closer to all local and global optimum.
uals shows large-scale variations in the early iterations as compared to PSO and DE. With the increase of iterations,
such variations reduce, and the positions of individuals mature uniformly and continuously tend to secure the location
urn
at global optimum without falling in local optimum in the later iterations. Consequently, solutions of PSO and DE
fails to secure global optimum position and slips to a local optimum. Apparently, the large variations in the earlier
iterations symbolize the exploration property, and the small variations in the following iterations show the exploitative
search for the better region of search-space. Exploitation and Exploration are cooperative during the optimization
process that is presented in the search histories and trajectory of a solution.
SS algorithm presents the better diversity than the PSO, and DE during the optimization process. Therefore, the
population diversity curve together with the search histories and trajectory of an individual exhibit that SS algorithm
attains a better balance between the exploration and exploitation properties as compared to PSO and DE.
10
Journal Pre-proof
of
very rough and drop slowly, which means that the exploration operators are dominant in PSO and DE as compared
to SS algorithm. Finally, in the case of PSO and DE, their convergence curves stagnate and cannot approximate the
global optimum in the final iterations. On the other hand, for the SS algorithm, the convergence performance and the
accuracy of the final approximation to the global optimal solution is better than the PSO, and DE.
Based on the above discussion, the SS algorithm establishes a better balance between the exploitation and explo-
pro
ration and search performance for the global optimum which is also satisfactory as compared to PSO and DE.
3. Parameter Adaptation
The performance of SS is highly dependent on the control parameters ci and rank and the size of population
N. Therefore, when SS is applied to the real-life optimization problems, there is the need for fine-tuning of these
parameters in order to obtain the better solution with fast convergence rate. Since this is a common practice, an online
re-
self-adaptive approach is considered as a better option to set the parameter values automatically during the search.
populations to match a exponential function where the population size at iteration 1 are N init and the population size
at the end of the run are N min . After each iteration k, the population size for the next iteration, N (k+1) , is computed
according to the formula:
!!k
N init − N min
N (k+1) = round N init 1 − (11)
n f esmax
N min is set to the smallest possible value such that the evolutionary operators can be applied. In the case of SS, N min
a
= 4 because the TowardsRand operator requires 4 individuals. n f esmax is the allowed maximum number of function
evaluations. Whenever N (k+1) < N (k) , the (N (k) − N (k+1) ) worst-ranking individuals are deleted from the populations.
urn
based parameter adaptation is proposed to adopt the crossover probability and scaling factor during the search, where
it creates a historical matrix, L, of size (2 × H) to save the H entries for both control parameters [31]. Similarly, in
SHPA, a historical matrix of size (2 × H) is maintained to save the learning values lr and lc for parameters rank and c,
respectively, of the last H iterations. Further this matrix L is used to generate the control parameters in calculation of
all trial solutions for the next iteration. Generation procedure of ranki and ci are discussed in following sections.
11
Journal Pre-proof
Table 2: Mean and SD of best error value obtained in 51 independent runs by SS, SStb, SStr, and SSr on 30-D CEC2014 problem suite (Mean:
Mean of best error, SD: Standard deviation of best error, W: result of Wilcoxon signed rank test).
of
+ + +
8 1.59E+02 1.00E+01 1.63E+02 1.13E+01 = 1.58E+02 9.05E+00 = 1.60E+02 8.54E+00 =
9 1.61E+02 1.04E+01 1.68E+02 9.19E+00 + 1.59E+02 1.04E+01 = 1.62E+02 9.55E+00 =
10 6.32E+03 2.55E+02 6.33E+03 2.80E+02 = 6.48E+03 2.71E+02 + 6.43E+03 3.21E+02 +
11 6.57E+03 3.04E+02 6.67E+03 3.11E+02 = 6.82E+03 3.26E+02 + 6.77E+03 2.46E+02 +
12 2.29E+00 3.10E-01 2.50E+00 2.52E-01 + 2.43E+00 2.77E-01 + 2.54E+00 2.74E-01 +
13 2.49E-01 3.21E-02 2.44E-01 4.98E-02 = 2.75E-01 3.69E-02 + 2.39E-01 3.21E-02 =
pro
14 2.51E-01 3.53E-02 2.47E-01 4.95E-02 = 2.68E-01 2.84E-02 + 2.65E-01 3.61E-02 +
15 1.39E+01 7.80E-01 1.43E+01 1.08E+00 + 1.39E+01 8.85E-01 = 1.39E+01 1.05E+00 =
16 1.19E+01 3.28E-01 1.23E+01 2.03E-01 + 1.25E+01 2.51E-01 + 1.24E+01 2.20E-01 +
17 4.80E+02 2.34E+02 6.52E+02 3.70E+02 + 1.12E+03 3.17E+02 + 9.58E+02 3.41E+02 +
18 7.33E+01 2.28E+01 7.72E+01 2.38E+01 = 6.76E+01 1.77E+01 = 5.42E+01 1.09E+01 -
19 5.37E+00 6.72E-01 5.17E+00 6.17E-01 = 5.16E+00 5.39E-01 = 5.02E+00 5.42E-01 -
20 5.77E+01 2.14E+01 4.13E+01 5.73E+00 - 4.39E+01 9.88E+00 - 3.51E+01 5.14E+00 -
21 4.85E+02 1.96E+02 6.36E+02 2.49E+02 + 7.86E+02 1.91E+02 + 7.02E+02 2.15E+02 +
22 2.29E+02 1.06E+02 3.31E+02 1.29E+02 + 4.15E+02 9.14E+01 + 3.64E+02 1.15E+02 +
23 3.15E+02 5.64E-13 3.15E+02 5.19E-13 + 3.15E+02 1.18E-04 + 3.15E+02 5.16E-13 +
24 2.21E+02 7.79E+00 2.24E+02 8.54E+00 + 2.27E+02 4.19E+00 + 2.24E+02 4.76E+00 +
25 2.03E+02 2.74E-01 2.03E+02 1.44E-01 = 2.03E+02 3.85E-01 + 2.03E+02 2.06E-01 =
re-
26 1.00E+02 3.24E-02 1.00E+02 3.98E-02 = 1.00E+02 3.55E-02 + 1.00E+02 3.41E-02 +
27 3.34E+02 4.34E+01 3.39E+02 4.19E+01 = 3.42E+02 2.63E+01 = 3.11E+02 2.13E+01 -
28 8.47E+02 9.35E+01 8.43E+02 4.58E+01 = 8.80E+02 4.33E+01 + 8.50E+02 4.37E+01 =
29 8.24E+02 6.91E+01 8.11E+02 6.51E+01 = 2.85E+03 4.00E+03 + 8.51E+02 2.74E+02 =
30 1.74E+03 8.95E+02 1.82E+03 7.31E+02 = 1.27E+03 6.33E+02 - 1.33E+03 5.64E+02 -
+/=/- 11/17/2 22/6/2 14/11/5
At each iteration k, in the calculation of ith trial-solution, the rank of projection matrix, ranki(k) , is independently
generated from binomial distribution with number of trials D and probability L1 j .
where Binornd represents the binomial distribution and j is randomly chosen from column of matrix L for every i
a
i
At each iteration k, a step-size control factor, ci , is generated from Cauchy distribution of mean L2 j and standard
deviation 0.1 as shown in equation 13.
12
Journal Pre-proof
P|S c(k) |
w(k)
h ch
(k)2
lc(k) = Ph=1(k) (15)
|S c |
h=1 wh(k) ch(k)
where |S r(k) | and |S c(k) | are the length of vectors S r(k) and S c(k) respectively. wh(k) is calculated using following equation.
fh(k) − fh(k−1)
wh(k) = P (k) (16)
( fg(k) − fg(k−1) )
of
|S r |
g=1
pro
In this section, the performance of SS algorithm is evaluated on the bound-constrained optimization problems.
The well-known 30 benchmark problems from IEEE CEC 2014s special session [32] are considered to analyze the
performance of SS algorithm as compared with the other state-of-art algorithms. Detailed information and charac-
teristics of these problems are available in [32]. To access the performance of SS algorithm, four experiments are
designed which are as follows.
1. Experiment 1: In first experiment, parameter sensitivity analysis is done.
2. Experiment 2: In second experiment, the effectiveness of different components of SS are studied in terms of
re-
exploitation, exploration and population diversity.
3. Experiment 3: In third experiment, the performance of basic SS (without self-adaptation of parameters) is
compared with the state-of-art algorithms.
4. Experiment 4: Finally, the performance of SASS (with self-adaptation of parameters) is compared with the best
performer of the IEEE CEC’s competition.
lP
Table 3: Experimental Results of SS with varying c on 30 test problems with 30D from IEEE CEC 2014 benchmark suite.
Table 4: Experimental Results of SS with varying rank on 30 test problems with 30D from IEEE CEC 2014 benchmark suite.
− 2 5 / 5 7
= 10 10 / 11 7
Table 5: Experimental Results of SS with varying p on 30 test problems with 30D from IEEE CEC 2014 benchmark suite.
All the experiments in this section are conducted on a personal computer with Intel(R) Core(TM) i5-3470 3.20
GHz processor and 10 GB of RAM using MATLAB 2015b. The source code of SS used in experiments can be
downloaded from this link: https://github.com/abhisheka456/spherical-search.
13
Table 6: Mean and SD of best error value obtained in 51 independent runs by SS, PSO, BB-PSO, CLPSO, APSO, and OLPSO on 30-D CEC2014 problem suite (Mean: Mean of best error, SD:
Nos
Mean SD
of
Standard deviation of best error, W: result of Wilcoxon signed rank test).
SS
Mean
PSO
SD W Mean
BB-PSO
SD W Mean
CLPSO
SD W Mean
APSO
SD W Mean
OLPSO
SD W
f01
f02
f03
f04
f05
8.75E+03
0.00E+00
1.03E-07
0.00E+00
2.09E+01 pro
6.70E+03
0.00E+00
2.81E-07
0.00E+00
5.88E-02
5.14E+07
2.42E+08
5.97E+04
1.66E+02
2.08E+01
2.24E+07
4.23E+08
1.07E+04
8.29E+01
9.01E+01
+
+
+
+
=
2.40E+06
1.84E+03
6.23E+03
5.99E+01
2.01E+01
1.97E+06
7.68E+03
1.40E+04
3.45E+01
5.20E-02
+
+
+
+
-
7.82E+06
8.93E+01
2.03E+02
6.92E+01
2.04E+01
2.44E+06
2.33E+02
1.85E+02
2.06E+01
4.13E-02
+
+
+
+
-
2.69E+09
1.02E+11
1.19E+06
2.49E+04
2.13E+01
3.28E+08
2.29E+09
1.25E+06
1.54E+03
5.61E-02
+
+
+
+
+
6.12E+06
1.28E+03
3.23E+02
8.64E+01
2.03E+01
3.58E+06
1.48E+03
5.69E+02
2.22E+01
1.28E-01
+
+
+
+
-
re-
f06 2.79E-01 5.86E-01 2.58E+01 6.40E+01 + 1.17E+01 2.12E+00 + 1.23E+01 1.47E+00 + 4.80E+01 1.79E+00 + 5.09E+00 1.48E+00 +
Journal Pre-proof
f07 0.00E+00 0.00E+00 9.28E+01 1.23E+01 + 2.73E-02 2.71E-02 + 3.43E-06 4.28E-06 + 1.06E+03 3.85E+01 + 1.02E-13 3.47E-14 =
f08 1.59E+02 1.00E+01 5.09E-02 1.11E+00 - 1.02E-03 8.96E-04 - 0.00E+00 0.00E+00 - 5.03E+02 3.02E+01 + 0.00E+00 0.00E+00 -
f09 1.61E+02 1.04E+01 3.18E+01 1.28E+02 - 3.16E+01 8.10E+00 - 5.22E+01 6.38E+00 - 4.78E+02 6.30E+00 + 4.06E+01 7.02E+00 -
f10 6.32E+03 2.55E+02 7.61E+02 7.61E+02 - 1.59E+00 1.42E+00 - 3.11E+00 1.60E+00 - 9.30E+03 5.68E+02 + 8.72E-02 2.04E-01 -
f11 6.57E+03 3.04E+02 6.83E+03 1.93E+03 + 2.11E+03 4.38E+02 - 2.33E+03 3.19E+02 - 9.24E+03 4.86E+02 + 2.28E+03 4.66E+02 -
f12 2.29E+00 3.10E-01 2.86E+00 2.77E+01 + 1.66E-01 5.61E-02 - 3.94E-01 5.53E-02 - 5.91E+00 1.32E+00 + 2.28E-01 6.38E-02 -
f13 2.49E-01 3.21E-02 6.40E-01 4.44E-01 + 2.38E-01 5.24E-02 =
lP
3.17E-01 3.75E-02 + 1.03E+01 7.53E-01 + 2.59E-01 3.20E-02 +
14
f14 2.51E-01 3.53E-02 4.21E-01 1.57E-01 + 5.75E-01 2.78E-01 + 2.51E-01 2.85E-02 = 3.95E+02 2.22E+01 + 2.41E-01 2.66E-02 =
f15 1.39E+01 7.80E-01 4.47E+00 1.92E+00 - 7.82E+00 3.03E+00 - 7.49E+00 1.02E+00 - 1.05E+06 0.00E+00 + 6.67E+00 1.62E+00 -
f16 1.19E+01 3.28E-01 1.29E+01 4.82E-01 + 9.13E+00 1.08E+00 - 1.04E+01 3.72E-01 - 1.42E+01 2.37E-01 + 1.17E+01 5.48E-01 =
f17 4.80E+02 2.34E+02 3.26E+05 2.57E+05 + 6.44E+05 6.17E+05 + 9.59E+05 4.41E+05 + 2.86E+08 1.28E+08 + 7.98E+05 4.13E+05 +
f18
f19
7.33E+01
5.37E+00
2.28E+01
6.72E-01
1.02E+06
2.82E+02
2.95E+06
1.98E+01
+
+
6.46E+03
7.43E+00
8.74E+03
1.29E+00
+
+
1.01E+02
7.39E+00
4.67E+01
6.27E-01
+
+
a8.75E+09
8.45E+02
3.11E+09
1.15E+02
+
+
3.58E+02
6.13E+00
5.12E+02
8.20E-01
+
+
urn
f20 5.77E+01 2.14E+01 1.08E+04 2.18E+03 + 3.02E+03 2.91E+03 + 3.14E+03 1.40E+03 + 1.59E+07 1.37E+07 + 5.58E+03 4.01E+03 +
f21 4.85E+02 1.96E+02 7.54E+05 4.21E+05 + 2.28E+05 3.29E+05 + 8.46E+04 5.35E+04 + 1.33E+08 7.50E+07 + 1.07E+05 8.33E+04 +
f22 2.29E+02 1.06E+02 8.24E+02 5.24E+02 + 2.43E+02 2.10E+02 + 2.08E+02 7.41E+01 + 1.31E+04 9.38E+03 + 2.20E+02 1.07E+02 =
f23 3.15E+02 6.61E-03 3.42E+02 6.24E+00 + 3.15E+02 1.87E-12 + 3.15E+02 2.02E-06 + 2.00E+02 0.00E+00 - 3.15E+02 1.23E-10 +
f24 2.21E+02 7.79E+00 2.05E+02 2.41E-01 - 2.28E+02 4.97E+00 + 2.25E+02 4.53E-01 + 2.00E+02 0.00E+00 - 2.24E+02 5.47E-01 +
f25 2.03E+02 2.74E-01 2.20E+02 4.51E+00 + 2.06E+02 2.27E+00 + 2.08E+02 1.06E+00 + 2.00E+02 0.00E+00 - 2.09E+02 1.75E+00 +
f26 1.00E+02 3.24E-02 1.00E+02 2.42E-01 1.00E+02 5.49E-02 1.00E+02 6.42E-02 1.86E+02 2.68E+01 1.00E+02 4.44E-02
f27
f28
f29
f30
3.34E+02
8.47E+02
8.24E+02
1.74E+03
+/=/-
4.34E+01
9.35E+01
6.91E+01
8.95E+02
2.51E+03
1.81E+03
8.77E+07
4.11E+05
4.82E+02
4.91E+02
3.24E+07
1.87E+04
23/2/5
=
+
+
+
+
6.53E+02
9.10E+02
1.70E+06
7.94E+03
5.06E+01
2.59E+01
3.47E+06
1.07E+03
21/1/8
+
+
+
+
+
4.14E+02
9.08E+02
9.71E+02
3.39E+03
4.92E+00
3.97E+01
9.14E+01
7.88E+02
20/2/8
+
=
+
+
+
2.00E+02
2.00E+02
2.00E+02
2.00E+02
0.00E+00
0.00E+00
0.00E+00
0.00E+00
23/0/7
+
-
-
-
-
Jo
3.26E+02
8.73E+02
1.36E+03
2.39E+03
3.80E+01
2.97E+01
2.82E+02
5.99E+02
18/5/7
+
=
+
+
+
Table 7: Mean and SD of best error value obtained in 51 independent runs by SS, CoBiDE, FCDE, RSDE, POBL ADE, and DE best on 30-D CEC2014 problem suite (Mean: Mean of best error,
SD: Standard deviation of best error, W: result of Wilcoxon signed rank test).
15
f14 2.51E-01 3.53E-02 2.65E-01 2.45E-02 + 4.44E-01 2.27E-01 + 2.36E-01 3.37E-02 = 2.26E-01 4.28E-02 - 3.90E-01 2.08E-01 +
f15 1.39E+01 7.80E-01 1.35E+01 1.24E+00 =
lP
1.54E+01 6.70E+00 + 5.92E+00 2.59E+00 - 7.73E+00 1.04E+00 - 1.66E+01 1.26E+00 +
f16 1.19E+01 3.28E-01 1.36E+01 1.85E-01 + 1.22E+01 5.74E-01 + 1.06E+01 7.70E-01 = 1.04E+01 4.58E-01 = 1.22E+01 2.73E-01 +
f17 4.80E+02 2.34E+02 1.68E+03 1.45E+02 + 6.54E+03 8.72E+03 + 1.24E+03 3.79E+02 + 1.10E+03 4.14E+02 + 7.64E+05 2.61E+05 +
f18 7.33E+01 2.28E+01 4.69E+01 6.34E+00 - 1.23E+02 6.70E+01 + 9.54E+01 4.34E+01 + 1.10E+02 3.81E+01 + 1.68E+03 1.30E+03 +
f19 5.37E+00 6.72E-01 1.25E+01 8.54E-01 + 1.32E+01 1.18E+01 + 5.65E+00 1.46E+00 + 8.88E+00 1.21E+01 + 6.29E+00 1.28E+03 +
f20 5.77E+01 2.14E+01 2.64E+01 3.45E+00 - 1.33E+02 7.99E+01 + 3.73E+01 2.55E+01 - 3.89E+01 2.21E+01 - 1.14E+02 1.65E+01 +
+ + = +
Journal Pre-proof
f21 4.85E+02 1.96E+02 7.16E+02 1.45E+02 3.01E+03 3.32E+03 4.71E+02 2.34E+02 3.86E+02 1.91E+02 - 3.45E+04 1.49E+04
re-
f22 2.29E+02 1.06E+02 2.49E+02 7.98E+01 + 4.57E+02 1.68E+02 + 1.91E+02 1.19E+02 - 2.31E+02 8.16E+01 + 1.55E+02 1.10E+02 -
f23 3.15E+02 6.61E-03 3.14E+02 2.85E-09 - 3.15E+02 1.24E-12 + 3.15E+02 1.40E-06 + 3.15E+02 5.74E-14 + 3.15E+02 4.02E-13 +
f24 2.21E+02 7.79E+00 2.43E+02 1.56E+01 + 2.50E+02 6.86E+00 + 2.24E+02 1.65E+00 + 2.22E+02 7.48E+00 + 2.23E+02 7.13E+00 +
f25 2.03E+02 2.74E-01 2.00E+02 6.58E-04 - 2.07E+02 4.01E+00 + 2.03E+02 1.17E-01 + 2.04E+02 3.22E+00 + 2.09E+02 1.77E+00 +
f26 1.00E+02 3.24E-02 1.00E+02 4.56E-02 + 1.01E+02 1.20E-01 + 1.00E+02 4.14E-02 + 1.39E+02 4.91E+01 + 1.00E+02 6.13E-02 +
f27 3.34E+02 4.34E+01 1.10E+03 6.71E+01 + 6.47E+02 2.50E+02 + 4.69E+02 9.46E+01 + 4.21E+02 4.64E+01 + 3.45E+02 4.44E+01 +
pro
f28 8.47E+02 9.35E+01 3.72E+02 6.37E-01 - 1.60E+03 5.92E+02 + 9.05E+02 1.21E+02 + 9.16E+02 1.63E+02 + 7.74E+02 1.03E+02 -
f29 8.24E+02 6.91E+01 2.16E+02 9.50E+01 - 7.55E+05 2.62E+06 + 6.52E+05 2.66E+06 + 3.39E+05 2.41E+06 + 3.57E+03 1.97E+03 +
f30 1.74E+03 8.95E+02 7.36E+02 1.02E+02 - 2.98E+03 1.26E+03 + 1.70E+03 8.67E+02 = 1.29E+03 5.14E+02 - 2.24E+03 6.54E+02 +
+/=/- 14/2/14 23/2/5 15/5/10 19/2/9 21/4/5
of
Table 8: Mean and SD of best error value obtained in 51 independent runs by SS, CMA-ES, I-POP-CMAES, LS-CMAES, CMSAES, and (1+1)cholesky CMAES on 30-D CEC2014 problem
Nos
Mean SD
of
suite (Mean: Mean of best error, SD: Standard deviation of best error, W: result of Wilcoxon signed rank test).
SS
Mean
CMA-ES
SD W
I-POP-CMAES
Mean SD W Mean
LS-CMAES
SD W Mean
CMSAES
SD W
(1+1)cholesky CMAES
Mean SD W
f01
f02
f03
f04
f05
8.75E+03
0.00E+00
1.03E-07
0.00E+00
2.09E+01 pro
6.70E+03
0.00E+00
2.81E-07
0.00E+00
5.88E-02
8.16E+04
4.59E+10
2.96E+03
4.71E+03
2.00E+01
2.66E+04
9.72E+09
1.29E+03
1.33E+03
3.70E-04
+
+
+
+
-
0.00E+00
0.00E+00
0.00E+00
0.00E+00
2.11E+01
0.00E+00
0.00E+00
0.00E+00
0.00E+00
7.69E-02
-
=
-
=
+
2.54E+08
6.98E+04
2.15E-02
5.50E+00
2.01E+01
1.37E+08
3.25E+04
1.90E-02
1.62E+01
2.18E-01
+
+
+
+
-
7.56E+08
6.85E+10
1.26E+05
1.01E+04
2.10E+01
1.40E+08
5.93E+09
1.75E+04
1.63E+03
4.38E-02
+
+
+
+
+
0.00E+00
0.00E+00
0.00E+00
1.27E+00
2.00E+01
0.00E+00
0.00E+00
0.00E+00
9.10E+00
5.85E-03
-
=
-
+
-
re-
f06 2.79E-01 5.86E-01 6.50E+01 1.57E+00 + 9.48E+00 7.65E+00 + 2.47E+01 3.55E+00 + 3.95E+01 1.04E+00 + 5.02E+01 4.68E+00 +
Journal Pre-proof
f07 0.00E+00 0.00E+00 1.12E+02 1.16E+01 + 0.00E+00 0.00E+00 = 4.45E-02 1.90E-02 + 6.01E+02 5.27E+01 + 1.11E-02 9.66E-03 +
f08 1.59E+02 1.00E+01 8.98E+02 1.83E+02 + 1.25E+02 1.20E+02 - 3.41E+02 1.02E+02 + 3.67E+02 1.66E+01 + 4.38E+02 7.76E+01 +
f09 1.61E+02 1.04E+01 9.14E+02 4.24E+00 + 3.70E+01 6.27E+01 - 3.85E+02 1.84E+02 + 4.31E+02 2.35E+01 + 6.14E+02 1.37E+02 +
f10 6.32E+03 2.55E+02 1.09E+03 5.78E+02 - 3.47E+03 2.02E+03 - 3.81E+03 5.94E+02 - 7.28E+03 2.80E+02 + 4.96E+03 7.40E+02 -
f11 6.57E+03 3.04E+02 2.16E+02 1.11E+02 - 2.74E+03 2.92E+03 - 3.38E+03 5.21E+02 - 7.19E+03 2.31E+02 + 5.19E+03 8.58E+02 -
f12 2.29E+00 3.10E-01 1.20E-01 1.16E+00 - 4.39E+00 1.18E+00 + 2.16E-01 5.88E-02 - 2.29E+00 2.91E-01 + 1.47E+00 7.07E-01 -
f13 2.49E-01 3.21E-02 1.31E+00 1.70E-01 + 4.67E-01 6.16E-01 +
lP
3.38E-01 4.88E-02 + 6.61E+00 4.79E-01 + 5.86E-01 1.44E-01 +
16
f14 2.51E-01 3.53E-02 1.57E+01 4.61E+00 + 8.89E-01 2.34E+00 + 1.79E-01 3.12E-02 - 2.01E+02 2.24E+01 + 4.21E-01 2.12E-01 +
f15 1.39E+01 7.80E-01 1.51E+03 1.48E+00 + 2.95E+05 2.04E+06 + 8.21E+00 1.97E+00 - 1.30E+06 4.74E+05 + 2.70E+01 1.34E+01 +
f16 1.19E+01 3.28E-01 1.62E+01 6.85E+00 + 1.09E+01 1.99E+00 - 1.28E+01 4.30E-01 + 1.31E+01 1.79E-01 + 1.43E+01 3.80E-01 +
f17 4.80E+02 2.34E+02 4.42E+03 9.19E+02 + 5.35E+06 2.59E+07 + 6.98E+04 3.19E+05 + 2.25E+07 8.63E+06 + 1.77E+03 3.91E+02 +
f18
f19
7.33E+01
5.37E+00
2.28E+01
6.72E-01
2.50E+03
2.04E+02
2.66E+02
2.19E+01
+
+
7.94E+02
3.68E+01
3.68E+03
1.01E+02
+
+
4.47E+04
9.82E+00
3.15E+05
1.42E+00
+
+
a9.95E+08
2.78E+02
3.53E+08
3.94E+01
+
+
1.43E+02
3.42E+01
4.82E+01
4.02E+01
+
+
urn
f20 5.77E+01 2.14E+01 2.76E+03 2.45E+02 + 1.92E+03 9.92E+03 + 5.40E+02 3.42E+02 + 9.63E+04 3.81E+04 + 3.84E+02 1.24E+02 +
f21 4.85E+02 1.96E+02 3.53E+03 1.22E+03 + 1.81E+06 1.24E+07 + 4.35E+04 2.65E+05 + 5.48E+06 2.46E+06 + 1.15E+03 3.80E+02 +
f22 2.29E+02 1.06E+02 5.17E+03 1.94E+02 + 2.19E+02 1.00E+02 - 4.98E+02 1.41E+02 + 1.43E+03 1.85E+02 + 8.63E+02 3.30E+02 +
f23 3.15E+02 6.61E-03 2.81E+02 1.43E+01 - 3.15E+02 2.30E-13 + 3.14E+02 8.99E+00 - 7.45E+02 7.97E+01 + 3.15E+02 4.09E-08 +
f24 2.21E+02 7.79E+00 2.76E+02 1.22E-01 + 2.28E+02 1.47E+01 + 2.16E+02 2.32E+01 - 4.06E+02 1.29E+01 + 4.78E+02 3.61E+02 +
f25 2.03E+02 2.74E-01 2.09E+02 9.14E+00 + 2.05E+02 3.61E+00 + 2.07E+02 9.31E+00 + 2.63E+02 7.69E+00 + 2.26E+02 1.61E+01 +
f26 1.00E+02 3.24E-02 2.75E+02 4.28E+01 1.24E+02 6.59E+01 1.00E+02 9.08E-02 1.06E+02 6.60E-01 1.93E+02 1.73E+02
f27
f28
f29
f30
3.34E+02
8.47E+02
8.24E+02
1.74E+03
+/=/-
4.34E+01
9.35E+01
6.91E+01
8.95E+02
4.50E+03
5.98E+03
1.27E+07
7.18E+05
8.76E+02
1.36E+02
6.28E+06
2.96E+05
25/0/5
+
+
+
+
+
3.02E+02
9.72E+02
2.42E+05
4.92E+04
2.40E+00
2.01E+02
1.71E+06
2.35E+05
18/3/9
+
-
+
+
+
4.19E+02
2.54E+03
3.59E+03
7.65E+03
1.40E+02
9.65E+02
2.89E+03
3.29E+03
22/0/8
+
+
+
+
+
8.82E+02
3.58E+03
8.34E+07
8.96E+05
8.61E+01
5.23E+02
3.53E+07
2.79E+05
30/0/0
+
+
+
+
+
Jo
9.03E+02
7.88E+03
4.19E+06
4.36E+03
4.26E+02
3.12E+03
1.17E+07
1.62E+03
23/1/6
+
+
+
+
+
Table 9: Mean and SD of best error value obtained in 51 independent runs by SS, GWO, GOA, MVO, and SCA on 30-D CEC2014 problem suite (Mean: Mean of best error, SD: Standard
deviation of best error, W: result of Wilcoxon signed rank test).
17
14 2.51E-01 3.53E-02 5.76E-01 5.66E-01 + 2.46E+02 2.05E+02 + 5.08E-01 2.95E-01 + 2.70E+01 2.24E+01 +
15 1.39E+01 7.80E-01 6.22E+01
lP
2.75E+02 + 1.16E+07 1.24E+07 + 1.09E+01 7.81E+00 - 1.86E+03 2.69E+03 +
16 1.19E+01 3.28E-01 1.07E+01 7.17E-01 - 1.29E+01 1.81E+00 + 1.12E+01 7.05E-01 - 1.19E+01 1.11E+00 =
17 4.80E+02 2.34E+02 1.06E+06 1.35E+06 + 1.91E+08 1.99E+08 + 4.33E+05 6.98E+05 + 4.22E+06 3.53E+06 +
18 7.33E+01 2.28E+01 5.69E+06 1.70E+07 + 5.34E+09 4.89E+09 + 1.42E+06 7.13E+06 + 8.57E+07 8.46E+07 +
19 5.37E+00 6.72E-01 2.59E+01 1.76E+01 + 6.81E+02 6.08E+02 + 1.74E+01 1.47E+01 + 6.19E+01 3.52E+01 +
20 5.77E+01 2.14E+01 1.28E+04 8.64E+03 + 9.40E+06 1.70E+07 + 5.95E+03 8.75E+03 + 1.34E+04 5.86E+03 +
21 4.85E+02 1.96E+02 5.42E+05 1.16E+06 + 9.40E+07 9.20E+07 + 2.85E+05 8.76E+05 + 1.14E+06 1.11E+06 +
Journal Pre-proof
re-
22 2.29E+02 1.06E+02 3.79E+02 1.87E+02 + 1.34E+04 2.32E+04 + 3.65E+02 1.47E+02 + 6.32E+02 2.64E+02 +
23 3.15E+02 6.61E-03 3.27E+02 4.28E+00 + 1.28E+03 8.31E+02 + 3.20E+02 5.67E+00 + 3.54E+02 2.64E+01 +
24 2.21E+02 7.79E+00 2.00E+02 1.01E-03 - 4.04E+02 1.67E+02 + 2.15E+02 1.49E+01 = 2.00E+02 2.69E-01 -
25 2.03E+02 2.74E-01 2.10E+02 3.07E+00 + 3.34E+02 1.07E+02 + 2.07E+02 3.24E+00 + 2.20E+02 1.01E+01 +
26 1.00E+02 3.24E-02 1.16E+02 3.66E+01 + 2.42E+02 1.35E+02 + 1.16E+02 3.66E+01 + 1.11E+02 2.96E+01 +
27 3.34E+02 4.34E+01 5.83E+02 1.16E+02 + 1.33E+03 6.36E+02 + 5.41E+02 1.24E+02 + 6.65E+02 2.52E+02 +
pro
28 8.47E+02 9.35E+01 9.61E+02 1.47E+02 + 5.30E+03 3.58E+03 + 9.97E+02 2.04E+02 + 1.58E+03 6.04E+02 +
29 8.24E+02 6.91E+01 8.97E+04 1.91E+05 + 4.89E+08 4.37E+08 + 9.42E+05 3.12E+06 + 8.13E+06 8.72E+06 +
30 1.74E+03 8.95E+02 2.95E+04 2.16E+04 + 6.43E+06 6.54E+06 + 1.54E+04 1.80E+04 + 1.57E+05 1.18E+05 +
+/-/= 22/1/7 28/2/0 21/1/8 23/4/2
of
Nos
of
Table 10: Mean and SD of best error value obtained in 51 independent runs by SS, SHO, SSA, SOA, and WOA on 30-D CEC2014 problem suite (Mean: Mean of best error, SD: Standard deviation
of best error, W: result of Wilcoxon signed rank test).
SS SHO SSA SOA WOA
pro
Mean STD Mean STD WT Mean STD WT Mean STD WT Mean STD WT
1 8.75E+03 6.70E+03 1.80E+07 2.65E+07 + 1.94E+07 2.56E+07 + 5.58E+07 1.15E+07 + 1.97E+07 1.81E+07 +
2 0.00E+00 0.00E+00 4.58E+08 1.17E+09 + 4.58E+08 1.17E+09 + 1.21E+08 1.36E+06 + 1.26E+08 1.20E+07 +
3 1.03E-07 2.81E-07 9.99E+03 1.24E+04 + 1.15E+04 1.13E+04 + 4.04E+04 1.35E+03 + 1.87E+04 2.03E+04 +
4 0.00E+00 0.00E+00 1.19E+02 6.55E+01 + 1.34E+02 5.71E+01 + 1.58E+02 4.68E-01 + 1.19E+02 5.68E+01 +
5 2.09E+01 5.88E-02 2.04E+01 4.58E-01 - 2.04E+01 4.28E-01 - 2.09E+01 3.12E-02 = 2.09E+01 7.09E-02 -
re-
Journal Pre-proof
6 2.79E-01 5.86E-01 2.20E+01 8.92E+00 + 1.63E+01 5.97E+00 + 1.24E+01 2.27E-02 + 1.74E+01 3.35E+00 +
7 0.00E+00 0.00E+00 3.00E+00 4.59E+00 + 2.99E+00 4.60E+00 + 7.65E+00 1.50E-02 + 1.88E+00 2.68E-02 +
8 1.59E+02 1.00E+01 8.80E+01 2.32E+01 - 9.90E+01 3.48E+01 - 8.16E+01 4.92E-01 - 1.48E+00 2.31E-01 -
9 1.61E+02 1.04E+01 1.25E+02 3.56E+01 - 1.13E+02 3.88E+01 - 8.68E+01 5.12E-01 - 9.63E+01 2.40E+01 -
10 6.32E+03 2.55E+02 1.98E+03 5.42E+02 - 2.99E+03 1.09E+03 - 2.08E+03 2.65E+00 - 1.34E+03 5.27E+02 -
11 6.57E+03 3.04E+02 3.37E+03 9.00E+02 - 3.34E+03 8.63E+02 - 2.11E+03 5.12E+00 - 2.46E+03 4.95E+02 -
12 2.29E+00 3.10E-01 1.05E+00 9.54E-01 - 1.02E+00 9.93E-01 - 1.54E-01 1.94E-03 - 1.56E-01 4.62E-02 -
13 2.49E-01 3.21E-02 4.38E-01 1.17E-01 + 4.51E-01
lP
1.23E-01 + 4.44E-01 9.16E-03 + 5.33E-01 1.27E-01 +
18
14 2.51E-01 3.53E-02 3.88E-01 2.16E-01 + 4.49E-01 2.52E-01 + 7.34E-01 4.21E-02 + 3.37E-01 1.03E-01 +
15 1.39E+01 7.80E-01 3.47E+01 1.92E+01 + 1.12E+01 7.76E+00 - 1.91E+01 1.02E-01 + 1.75E+01 1.02E+01 +
16 1.19E+01 3.28E-01 1.18E+01 1.02E+00 = 1.14E+01 8.39E-01 - 1.06E+01 1.57E-03 - 1.12E+01 6.77E-01 -
17 4.80E+02 2.34E+02 4.17E+05 7.01E+05 + 4.76E+05 6.84E+05 + 6.52E+05 8.23E+04 + 1.34E+06 2.13E+06 +
18
19
20
7.33E+01
5.37E+00
5.77E+01
2.28E+01
6.72E-01
2.14E+01
1.41E+06
2.18E+01
8.22E+03
7.13E+06
1.69E+01
7.59E+03
+
+
1.42E+06
1.92E+01
6.09E+03
7.13E+06
1.38E+01
8.66E+03
+
+
1.23E+04
1.23E+01
8.75E+03
a 3.84E+01
5.87E-03
5.84E+01
+
+
1.99E+03
2.67E+01
5.84E+03
3.70E+03
3.44E+01
1.93E+00
+
+
urn
+ + + +
21 4.85E+02 1.96E+02 2.81E+05 8.78E+05 + 3.00E+05 8.74E+05 + 6.95E+05 2.15E+04 + 8.60E+05 7.85E+05 +
22 2.29E+02 1.06E+02 5.73E+02 2.48E+02 + 3.91E+02 1.62E+02 + 3.65E+02 2.56E+01 + 5.69E+02 2.06E+02 +
23 3.15E+02 6.61E-03 3.19E+02 5.88E+00 + 3.20E+02 5.67E+00 + 3.54E+02 1.15E-01 + 3.16E+02 1.01E+00 +
24 2.21E+02 7.79E+00 2.17E+02 1.38E+01 = 2.17E+02 1.55E+01 = 2.00E+02 2.34E-04 - 2.15E+02 5.15E+00 +
25 2.03E+02 2.74E-01 2.11E+02 3.43E+00 + 2.11E+02 3.70E+00 + 2.09E+02 5.96E-03 + 2.11E+02 4.59E+00 +
26 1.00E+02 3.24E-02 1.49E+02 4.93E+01 + 1.10E+02 2.99E+01 + 1.00E+02 2.02E-02 = 1.51E+02 5.13E+01 +
27
28
29
30
3.34E+02
8.47E+02
8.24E+02
1.74E+03
+/-/=
4.34E+01
9.35E+01
6.91E+01
8.95E+02
7.31E+02
2.53E+03
5.78E+05
1.34E+04
3.32E+02
1.47E+03
2.14E+06
1.90E+04
22/2/6
+
+
+
+
6.40E+02
9.85E+02
1.05E+06
1.77E+04
1.53E+02
1.41E+02
3.49E+06
1.73E+04
21/1/8
+
+
+
+
5.62E+02
1.72E+03
9.68E+03
3.16E+04
1.04E+00
4.63E-01
1.43E+01
6.56E+02
21/2/7
+
+
+
+
7.62E+02
1.51E+03
8.68E+05
1.38E+04 Jo
1.98E+02
4.53E+02
2.67E+06
1.98E+02
23/0/7
+
+
+
+
Journal Pre-proof
Table 11: Ranking of Algorithm according to Friedman ranking based on mean error value. (FR: Friedman Ranking)
of
10 POBL ADE 7.4833 5 22 SSA 13.8333 15
11 DE best 9.7000 8 23 SOA 12.6500 13
12 CMA-ES 15.8333 20 24 WOA 13.3167 14
pro
The sensitivity of all parameters of SS is experimentally investigated. A set of 30 test problems with 30D from
IEEE CEC 2014 benchmark suite is utilized as the benchmark problems to study the sensitivity of parameters of SS.
Additionally, Wilcoxons signed rank test (WT) at the 0.05 significance level is employed to compare the performance
of algorithms.
The sensitivities of three important parameters such as c, rank, and p of SS algorithm are studied. In the case of
c, six different values of c in SS, i.e. c = 0.2, c = 0.4, c = 0.5, c = 0.6, c = 0.8, and c = 1.0 are considered in this
analysis. The outcomes of WT with respect to the SS with c = 0.5 is reported in Table 3. According to the WT’s
re-
outcomes, SS with c = 0.5 performs better than SS with c = 0.2, c = 0.4, c = 0.6, c = 0.7, and c = 0.9 on 21, 19,
six, seven, and 16 test problems, respectively. However, SS with c = 0.5 is outperformed by SS with c = 0.2, c = 0.4,
c = 0.6, c = 0.8, and c = 1.0 on three, two, three, five, and seven test problems, respectively. From this analysis, it
can be concluded that the performance of SS is highly sensitive to c and c = (0.5, 0.7) is recommended in this work.
For rank, five different values are considered, i.e. rank = 0.1D, rank = 0.2D, rank = 0.3D, rank = 0.5D,
rank = 0.7D, and rank = 0.9D to analyze the sensitivity of rank in the performance of the SS. The results of WT
lP
with respect to the SS with rank = 0.5D are reported in Table 4. It is seen from the Table 4 that the performance of
SS with rank = 0.5D is better than SS with rank = 0.1D, rank = 0.3D, rank = 0.7D, and rank = 0.9D on 18, 15, 14,
and 16 test problems out of 30 test problems, respectively. Although, the performance of SS with rank = 0.5D is not
better than SS with rank = 0.1D, rank = 0.3D, rank = 0.7D, and rank = 0.9D on 10, 10, 11, and seven test problems,
respectively. On the basis of the outcomes of WT, it can be concluded that the performance of SS is highly sensitive
to rank. The value of rank is set at 0.5D in SS.
Similarly, five different variants of SS with different values of p such as 0, 0.05, 0.1, 0.15, and 0.2 are utilized to
a
analyze the sensitivity of p on the performance of SS. The results of this analysis in terms of WT with respect to the
SS with p = 0.1 are reported in Table 5. As reported in Table 5, SS with p = 0.1 outperforms the SS with p = 0,
p = 0.05, p = 0.15, and p = 0.2 on 20, 18, 17, and 22 test problems, respectively. However, SS with p = 0.1 is
urn
outperformed by SS with p = 0, p = 0.05, p = 0.15, and p = 0.2 on seven, eight, ten and three, respectively. From
the outcomes of WT, we can conclude that the performance of SS is very sensitive to the value of p. This analysis
suggests the p = 0.1 in this work.
good balance between the exploitative and exploratory search, the half population of better solution utilizes towards-
rand to calculate search direction and rest half of the population uses towards-best to calculate the search direction
thereby driving diversity in the set of better solutions and pushing the inferior solutions to attempt for improved
objective values. In addition to this approach, three other combinations of towards-rand, and towards-best can be
incorporated in SS. To analyze the effectiveness of the proposed approach, the following combinations have been
compared with the SS on 30D problems of IEEE CEC14 problem suite.
1. SStr: SS with only towards-rand,
19
Journal Pre-proof
of
Furthermore, SS is outperformed by SStr, SStb, and SSr on two, two, and five out of 30 problems, respectively.
From this analysis, it can be concluded that the proposed approach provides a better balance between exploitation
and exploration of search space during the optimization process. Some important outcomes of this experiment are
summarized as follows.
pro
1. Configuration SStr is the least effective as compared with other configuration because it more emphasizes on
exploratory search which slows down the optimization process.
2. Configuration SStb provides a better balance between exploration and exploitation than SStr because it explores
solutions in the regions of search space containing better solutions in place of whole search space.
3. To improve the balancing characteristics of SS, SSr can be a good choice because it improves the performance
of the algorithm on composite problems but meanwhile it losses its performance on simpler hybrid problems.
Thus, the performance of SSr is almost similar to the SStb which shows that this configuration cannot utilize
re-
the features of towards-rand effectively during the run.
4. Proposed configuration of SS provides better performance than the other configuration which shows that this
configuration utilizes towards-rand and toeards-best effectively during the optimization process and provide
better balance between exploitation and exploration.
In this experiment, to analyze the performance, SS are benchmarked on 30 real-parameter single objective bound-
constrained optimization problems used in a special session of IEEE CEC-2014 [32]. Detailed information and
characteristics of these problems are available in [32]. To evaluate the performance of SS on the CEC 2014 problem
suite, the results are compared with other state-of-the-art algorithms. State-of-the-art algorithms are divided into four
groups:
1. Group-I:- Variants of PSO: Basic PSO[8], BB-PSO[33], CLPSO[34], APSO[35], OLPSO[36].
a
4. Group-IV:- Recently proposed optimization algorithms: GWO[11], GOA[21], MVO[46], SCA[47], SHO[48],
SSA[22], SOA[13], WOA[14].
PSO, DE, and CMA-ES are the popular classical Meta-heuristics. Popular variants of classical algorithms are also
taken from the literature to show the effectiveness of the SS.
In this experiment, the population size, N, is set to 80, the dimension of search space for all problems, D, is fixed to
30, and allowed maximum function evaluation, MaxFES, is fixed to 300,000 for 51 independent runs. The parameters
of other algorithms are set to their default values as reported in their referred paper.
Jo
Tables 6-10 summarize the mean and standard deviation (SD) of the error values obtained by the algorithms over
51 independent runs for each problem. We also performed the Wilcoxon Signed Ranks Test in this experiment. The
statistical results are also summarized in Tables 6-10, where ‘+’ denotes the performance of the SS is better than
other algorithms, ‘-’ denotes the performance of other method is better than the SS, and ‘=’ denotes that there is no
significant difference in performance. We also rank all algorithms along with the SS using Friedman ranking test
based on the mean of error values obtained by the algorithms over 51 independent runs. The statistical results of the
Friedman test are reported in Table 11.
20
Journal Pre-proof
1. SS vs Group-I’s algorithms: Table 6 summarizes the results obtained by algorithms SS, and Group-I’s algo-
rithms: PSO, BB-PSO, CLPSO, APSO, and OLPSO on the CEC-2014 problem suite. It is seen from Table
6 that the SS is showing better performance than PSO, BB-PSO, CLPSO, APSO and OLPSO on 23, 21, 20,
23, and 18 problems out of 30, respectively, performance of SS worse than PSO, BB-PSO, CLPSO, APSO and
OLPSO on five, eight, eight, seven, and seven problems respectively, and SS is significantly equal to the PSO,
BB-PSO, CLPSO, and OLPSO for 2, 1, 2, and 5 problems of CEC-2014 problem suite respectively.
2. SS vs Group-II’s algorithms: Table 7 summarizes the results of the algorithms SS, and Group-II’s algorithms:
CoBiDE, FCDE, RSDE, POBL ADE, and DE best. As shown in Table 7, SS have better performance than
of
CoBiDE, FCDE, RSDE, POBL ADE, and DE best on 14, 23, 15, 19, and 21 problems out of 30 respectively,
performance of SS worse than CoBiDE, FCDE, RSDE, POBL ADE, and DE best on 14, five, 10, nine, and five
problems respectively, and SS provides performance similar to the CoBiDE, FCDE, RSDE, POBL ADE, and
DE best for two, two, five, two and seven problems of CEC-2014 problem suite respectively.
pro
3. SS vs Group-III’s algorithms: Table 8 presents the results of the algorithms SS, and Group-III’s algorithms:
CMA-ES, I-POP-CMAES, LS-CMAES, CMSAES, and (1+1)cholesky-CMAES. when examined the last col-
umn of Table 8, SS shows better performs then CMA-ES, I-POP-CMAES, LS-CMAES, CMSAES, and (1+1)Cholesky-
CMAES on 25, 18, 22, 30, and 23 problems out of 30 respectively, performance of SSO worse than CMA-ES,
I-POP-CMAES, LS-CMAES, and (1+1)Cholesky-CMAES on five, nine, eight, and six problems respectively,
and SSO is significantly similar to the I-POP-CMAES, and (1+1)Cholesky-CMAES for three, and one problems
of CEC-2014 problem suite respectively.
4. SS vs Group-IV’s algorithms: In Tables 9 and 10, the outcomes of SS and Group-IVs algorithms are presented.
re-
The last rows of Tables 9 and 10 summarizes the results of WT. It is seen from Tables 9 and 10 that the
performance of SS is better than GWO, GOA, MVO, SCA, SHO, SSA, SOA, and WOA on 22, 28, 21, 23, 22,
21, 21, and 23 out of 30 problems, respectively. The performance of SS is outperformed by GWO, GOA, MVO,
SCA, SHO, SSA, SOA, and WOA on seven, zero, eight, three, six, eight, seven, and seven out of 30 problems ,
respectively.
In addition, the Friedman test (FT) is also used to detect the significant differences between SS and the other 23
lP
algorithms on all 30 problems of CEC-2014 problem suite. The detailed results of the FTfor all 24 algorithms are
shown in Table 11. From Table 11, it can be found that SS is ranked first by FT among all 24 algorithms. Variants of
PSO: BB-PSO, CLPSO, OLPSO, and Variants of DE: CoBiDE, RSDE, POBL-ADE, and DE-best are very competitive
with the SS but the performance of SS is slightly better than them. Similarly, variants of CMA-ES: I-POP-CMAES and
LS-CMAES are also well performed on the CEC-2014 problem suite, but they could not outperforms the performance
of SS. In the case of recently proposed algorithms, MVO, SOA, WOA, SSA, and GWO perform very well, but the
performance of SS is significantly better than them. Compared with the rest of other algorithms, SS is significantly
a
outperformed them.
urn
21
Journal Pre-proof
The same parameter setting is adopted for aformentioned algorithms as reported in their respective papers. Same
stopping criterion, 10,000D as maximum allowed number of function evaluations, is considered for all the algorithm
for 51 independent runs as suggested in [31]. In SASS, the parameters are set as follows: Nmin = 4, Nmax = 18D,
H = 6, L1 j = 0.5 for j = 1 to H and L2 j = 0.5 for j = 1 to H.
The mean and SD of the error value for all the algorithms, i.e. f (x∗) − f (xbest ), on D = 10, D = 30, D = 50, and
D = 100 problems are depicted in Tables 12-15, respectively. In addition, WT at 0.05 significance level is utilized
to check the significance of the performance of SASS as compared to others. In Tables 12-15, “+” sign denotes
the performance of SASS is better than the other contenders, “-” sign represents the SASS is outperformed by other
of
contenders, and “=” shows that the there is no significant difference in the performance of SASS as compared to other
contenders. The last rows of these tables summarize the outcomes of WT in terms of +(win)/=(tie)/-(lose).
From Table 12, it is found that the performance of SASS is better than the iCMAES-ILS, NBIPOP-aCMAES,
SHADE, LSHADE, EBOwithCMAR, and HSES on 15, 17, 17, 17, eight, 16 out of 30 problems of 10D, respectively.
SASS is outperformed by iCMAES-ILS, NBIPOP-aCMAES, SHADE, LSHADE, EBOwithCMAR, and HSES on
pro
six, five, zero, three, seven, and five problems of 10D, respectively. From the statistical outcomes of 10D problems,
it can be concluded that the performance of SASS is better than other contenders except for EBOwithCMAR. SASS
performs similar to the EBOwithCMAR on 10D problems.
An analysis of the results of WT on 30D problems (Table 13) shows that the SASS performs better than iCMAES-
ILS, NBIPOP-aCMAES, SHADE, LSHADE, EBOwithMAR, and HSES on 14, 14, 20, 11, 11, and 13 problems, re-
spectively. The performance of iCMAES-ILS, NBIPOP-aCMAES, SHADE, LSHADE, EBOwithCMAR, and HSES
is better than SASS on seven, seven, one, three, seven, and eight out of 30 problems of 30D, respectively. This
re-
comparative analysis reveals that the performance of SASS is comparatively better than the other algorithms on 30D
problems.
In the case of 50D problems, the SASS outperforms iCMAES-ILS, NBIPOP-aCMAES, SHADE, LSHADE,
EBOwithCMAR, and HSES on 13, 14, 19, 11, nine, and 12 problems as shown Table 14. The performance of SASS
is inferior to iCMAES-ILS, NBIPOP-aCMAES, SHADE, LSHADE, EBOwithCMAR, and HSES on eight, eight, six,
five, 10, and 13 out of 30 problems of 50D. These comparative outcomes revealed that SASS exhibits competitive
performance with comparison to EBOwithCMAR and HSES. Moreover, SASS provides better performance than the
lP
3, and 2.5 on 10D, 30D, 50D, and 100D problems, respectively. SASS is outperformed by EBOwithCMAR on 10D
problems, EBOwithCMAR, and HSES on 50D problems, and HSES on 100D problems on the basis of mean rank
of FT. The last column of Table 16 summarizes the overall ranking of algorithms obtained by FT. As shown in Table
16, SASS is the second best performing algorithm on IEEE CEC14 problem suite as compared to all contenders.
EBOwithCMAR provides the best performance. From the above analysis, it can be concluded that the performance
of SASS is superior or at least competitive with the best performers of IEEE CECs competition.
4.5. Application to Real-Life Complex Optimization Problems: IEEE CEC 2011 problem suite
Jo
SASS is also benchmarked in real-world optimization problems to analyze the effectiveness of SASS in solving
this type of optimization problems. Optimization problems of IEEE CEC’11 [54] problem suite are selected as test
problems. This problem suite contains the 22 challenging problems from different engineering fields viz. Optimal
power flow, economic dispatch problems, network expansion planning, radar system, spacecraft trajectory optimiza-
tion, and sound waves. The dimensions of these problems are very from six to 40.
The parametric setting is set at the same values as reported in the experiment 4.4. The outcomes of SASS are
compared statistically with six algorithms including the IEEE CEC competition’s winners. Each contender is em-
22
Journal Pre-proof
Table 12: Mean and SD of best error value obtained in 51 independent runs by SASS, iCMAES-ILS, NBIPOP-aCMA-ES, SHADE, LSHADE,
EBOwithCMAR, and HSES on 10-D CEC2014 problem suite (Mean: Mean of best error, SD: Standard deviation of best error, W: result of
Wilcoxon signed rank test).
of
SASS iCMAES-ILS NBIPOP-aCMA-ES SHADE
Prob Mean SD Mean SD W Mean SD W Mean SD W
F1 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F2 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F3 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F4 3.00E+01 1.21E+01 1.44E+01 1.60E+01 - 2.82E+00 8.24E+00 - 2.74E+01 1.43E+01 =
F5 1.45E+01 8.53E+00 1.47E+01 8.88E+00 = 1.81E+01 6.02E+00 + 1.80E+01 5.12E+00 +
pro
F6 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 3.30E-01 5.83E-01 + 0.00E+00 0.00E+00 =
F7 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 9.78E-03 1.47E-02 +
F8 0.00E+00 0.00E+00 2.54E-01 4.81E-01 + 3.70E+00 1.99E+00 + 0.00E+00 0.00E+00 =
F9 2.49E+00 8.03E-01 9.75E-02 2.99E-01 - 3.26E-01 6.41E-01 - 3.14E+00 8.99E-01 +
F10 2.45E-03 1.22E-02 1.22E+02 8.73E+01 + 9.16E+01 7.51E+01 + 3.67E-03 1.48E-02 =
F11 3.50E+01 4.37E+01 8.59E+00 8.36E+00 - 1.17E+02 1.25E+02 + 6.32E+01 5.37E+01 +
F12 6.59E-02 1.66E-02 6.50E-02 1.18E-01 = 1.01E-02 1.65E-02 - 1.42E-01 2.34E-02 +
F13 5.52E-02 1.11E-02 9.11E-03 4.40E-03 - 1.09E-02 6.21E-03 - 7.40E-02 1.43E-02 +
F14 7.72E-02 2.58E-02 1.55E-01 4.45E-02 + 2.82E-01 8.50E-02 + 1.06E-01 3.15E-02 +
F15 3.74E-01 6.64E-02 7.23E-01 1.93E-01 + 5.47E-01 1.28E-01 + 5.05E-01 8.84E-02 +
F16 1.31E+00 2.38E-01 1.91E+00 5.39E-01 + 2.53E+00 5.58E-01 + 1.56E+00 3.19E-01 +
F17 7.41E-01 7.81E-01 2.10E+01 1.71E+01 + 3.89E+01 3.97E+01 + 1.28E+00 1.92E+00 =
F18 1.41E-01 1.62E-01 5.26E-01 4.92E-01 + 3.58E+00 6.18E+00 + 2.26E-01 1.81E-01 =
F19 1.03E-01 5.37E-02 7.08E-01
re-
5.94E-01 + 8.28E-01 5.02E-01 + 2.26E-01 2.17E-01 +
F20 9.92E-02 5.85E-02 8.04E-01 5.01E-01 + 1.32E+00 2.28E+00 + 2.77E-01 1.49E-01 +
F21 3.36E-01 2.66E-01 3.21E+00 5.51E+00 + 1.67E+01 3.02E+01 + 3.76E-01 2.94E-01 =
F22 9.63E-02 3.48E-02 1.83E+01 7.04E+00 + 1.79E+01 6.76E+00 + 2.90E-01 1.13E-01 +
F23 3.29E+02 1.72E-13 2.50E+02 1.41E+02 - 3.16E+02 5.58E+01 - 3.29E+02 2.87E-13 =
F24 1.08E+02 1.48E+00 1.04E+02 4.48E+00 - 1.07E+02 2.98E+00 = 1.09E+02 1.37E+00 +
F25 1.28E+02 3.40E+01 1.26E+02 1.02E+01 = 1.39E+02 2.58E+01 + 1.50E+02 4.38E+01 +
F26 1.00E+02 1.81E-02 1.00E+02 4.42E-03 = 1.00E+02 5.56E-03 = 1.00E+02 1.53E-02 +
F27 2.31E+01 8.75E+01 1.05E+02 1.37E+02 + 2.02E+02 1.36E+02 + 1.50E+02 1.68E+02 +
F28 3.68E+02 4.88E+00 3.61E+02 5.18E+01 + 3.58E+02 5.34E+01 = 3.99E+02 4.75E+01 +
F29 2.22E+02 5.05E-01 2.22E+02 5.88E-01 + 2.20E+02 7.11E+00 = 2.22E+02 5.91E-01 =
F30 4.64E+02 7.45E+00 4.71E+02 1.77E+01 4.83E+02 2.30E+01 4.73E+02 2.23E+01
lP
+ + +
+/-/= 15/9/6 17/8/5 17/13/0
SASS LSHADE EBOwithCMAR HSES
Prob Mean SD Mean SD W Mean SD W Mean SD W
F1 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F2 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F3 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F4 3.00E+01 1.21E+01 3.08E+01 1.11E+01 + 8.10E+00 1.42E+01 - 3.48E+01 5.20E-03 +
F5 1.45E+01 8.53E+00 1.77E+01 6.24E+00 + 1.29E+01 9.31E+00 = 1.73E+01 6.95E+00 +
F6 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 3.09E-02 2.21E-01 =
F7 0.00E+00 0.00E+00 5.31E-03 1.13E-02 + 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F8 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 3.51E-01 5.91E-01 +
a
23
Journal Pre-proof
Table 13: Mean and SD of best error value obtained in 51 independent runs by SASS, iCMAES-ILS, NBIPOP-aCMA-ES, SHADE, LSHADE,
EBOwithCMAR, and HSES on 30-D CEC2014 problem suite (Mean: Mean of best error, SD: Standard deviation of best error, W: result of
Wilcoxon signed rank test).
of
SASS iCMAES-ILS NBIPOP-aCMA-ES SHADE
Prob Mean SD Mean SD W Mean SD W Mean SD W
F1 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 4.81E+02 7.86E+02 +
F2 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F3 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F4 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F5 2.01E+01 2.44E-02 2.00E+01 6.78E-06 - 2.05E+01 4.43E-01 + 2.01E+01 1.91E-02 -
pro
F6 3.26E-04 2.30E-03 4.00E-03 2.85E-02 + 7.14E-01 1.28E+00 + 5.29E-01 7.22E-01 +
F7 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 4.83E-04 1.97E-03 =
F8 0.00E+00 0.00E+00 2.42E+00 1.39E+00 + 9.98E+00 3.54E+00 + 0.00E+00 0.00E+00 =
F9 6.99E+00 1.66E+00 2.57E+00 1.24E+00 - 3.24E+00 1.47E+00 - 1.58E+01 3.17E+00 +
F10 1.22E-03 4.95E-03 1.45E+02 2.87E+02 + 6.36E+02 3.83E+02 + 1.27E-02 1.67E-02 +
F11 1.20E+03 2.40E+02 7.38E+01 9.85E+01 - 7.31E+02 4.48E+02 - 1.49E+03 1.93E+02 +
F12 1.54E-01 2.63E-02 2.83E-02 4.58E-02 - 1.32E-02 6.78E-03 - 1.65E-01 1.85E-02 =
F13 1.16E-01 1.57E-02 2.95E-02 7.22E-03 - 3.89E-02 1.26E-02 - 2.06E-01 3.64E-02 +
F14 2.38E-01 3.08E-02 1.70E-01 1.99E-02 - 3.28E-01 5.24E-02 + 2.30E-01 3.03E-02 =
F15 2.13E+00 2.95E-01 2.51E+00 4.06E-01 + 2.14E+00 3.87E-01 + 2.57E+00 3.14E-01 +
F16 8.62E+00 4.28E-01 1.09E+01 9.95E-01 + 1.06E+01 1.18E+00 + 9.19E+00 3.35E-01 +
F17 2.60E+02 1.52E+02 1.05E+03 2.91E+02 + 8.52E+02 2.96E+02 + 1.06E+03 3.18E+02 +
F18 8.19E+00 3.48E+00 9.61E+01 3.23E+01 + 1.15E+02 2.67E+01 + 6.15E+01 3.40E+01 +
F19 3.65E+00 5.79E-01 6.46E+00
re-
1.25E+00 + 5.70E+00 1.58E+00 + 4.43E+00 7.12E-01 +
F20 3.06E+00 1.39E+00 3.35E+01 4.49E+01 + 2.40E+01 3.57E+01 + 1.28E+01 7.69E+00 +
F21 9.67E+01 7.93E+01 6.44E+02 2.24E+02 + 4.91E+02 1.81E+02 + 2.63E+02 1.30E+02 +
F22 2.82E+01 1.73E+01 1.30E+02 8.16E+01 + 1.41E+02 5.16E+01 + 1.14E+02 5.57E+01 +
F23 3.15E+02 2.88E-13 3.15E+02 1.72E-13 = 3.15E+02 1.72E-13 = 3.15E+02 1.72E-13 =
F24 2.26E+02 3.43E+00 2.15E+02 1.38E+01 - 2.12E+02 1.31E+01 - 2.26E+02 3.34E+00 =
F25 2.03E+02 1.35E-01 2.03E+02 3.70E-02 = 2.03E+02 2.43E-02 = 2.03E+02 7.37E-01 +
F26 1.00E+02 2.08E-02 1.00E+02 7.85E-03 = 1.00E+02 6.69E-02 = 1.00E+02 3.54E-02 +
F27 3.00E+02 0.00E+00 3.00E+02 0.00E+00 = 3.00E+02 0.00E+00 = 3.25E+02 3.89E+01 +
F28 8.24E+02 2.18E+01 8.75E+02 2.17E+01 + 7.96E+02 4.83E+01 - 8.40E+02 3.16E+01 +
F29 7.17E+02 4.06E+00 7.27E+02 2.51E+02 + 6.66E+02 1.03E+02 = 7.28E+02 1.25E+01 +
F30 8.94E+02 3.60E+02 2.19E+03 6.27E+02 1.66E+03 3.83E+02 1.77E+03 7.68E+02
lP
+ + +
+/-/= 14/9/7 14/9/7 20/9/1
SASS LSHADE EBOwithCMAR HSES
Prob Mean SD Mean SD W Mean SD W Mean SD W
F1 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 3.72E-05 4.20E-05 + 0.00E+00 0.00E+00 =
F2 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F3 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F4 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F5 2.01E+01 2.44E-02 2.01E+01 3.68E-02 + 2.00E+01 4.57E-04 - 2.00E+01 3.44E-04 -
F6 3.26E-04 2.30E-03 1.38E-07 9.89E-07 = 1.64E-01 1.17E+00 + 1.08E+00 1.01E+00 +
F7 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F8 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 7.76E+00 2.23E+00 +
a
24
Journal Pre-proof
Table 14: Mean and SD of best error value obtained in 51 independent runs by SASS, iCMAES-ILS, NBIPOP-aCMA-ES, SHADE, LSHADE,
EBOwithCMAR, and HSES on 50-D CEC2014 problem suite (Mean: Mean of best error, SD: Standard deviation of best error, W: result of
Wilcoxon signed rank test).
of
SASS iCMAES-ILS NBIPOP-aCMA-ES SHADE
Prob Mean SD Mean SD W Mean SD W Mean SD W
F1 3.16E+02 6.46E+02 0.00E+00 0.00E+00 - 0.00E+00 0.00E+00 - 1.86E+04 1.36E+04 +
F2 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F3 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
F4 1.37E+01 2.96E+01 1.03E+01 2.44E+01 = 0.00E+00 0.00E+00 - 9.72E+00 2.94E+01 -
F5 2.03E+01 3.05E-02 2.00E+01 6.43E-05 - 2.08E+01 5.20E-01 + 2.01E+01 2.02E-02 -
pro
F6 6.61E-01 6.61E-01 2.24E-04 9.56E-04 - 1.31E+00 2.21E+00 + 5.17E+00 2.43E+00 +
F7 1.45E-04 1.04E-03 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 3.91E-03 7.41E-03 +
F8 0.00E+00 0.00E+00 5.76E+00 2.01E+00 + 2.52E+01 8.83E+00 + 0.00E+00 0.00E+00 =
F9 1.31E+01 1.85E+00 6.37E+00 1.87E+00 - 5.82E+00 2.96E+00 - 3.42E+01 6.16E+00 +
F10 3.80E-02 2.08E-02 1.60E+02 2.52E+02 + 2.04E+03 7.74E+02 + 1.03E-02 1.14E-02 -
F11 3.36E+03 3.01E+02 1.13E+02 1.27E+02 - 1.67E+03 8.77E+02 - 3.56E+03 3.27E+02 +
F12 2.20E-01 2.69E-02 1.03E-02 3.03E-02 - 1.31E-02 6.85E-03 - 1.64E-01 2.12E-02 -
F13 1.71E-01 1.73E-02 6.60E-02 1.86E-02 - 7.97E-02 2.95E-02 - 3.10E-01 4.55E-02 +
F14 3.07E-01 2.11E-02 2.43E-01 3.25E-02 = 3.18E-01 5.56E-02 + 2.95E-01 5.72E-02 -
F15 5.23E+00 4.30E-01 4.56E+00 4.07E-01 = 4.44E+00 5.55E-01 - 5.74E+00 6.15E-01 +
F16 1.70E+01 4.47E-01 1.99E+01 7.20E-01 + 1.89E+01 1.79E+00 + 1.74E+01 4.27E-01 +
F17 1.75E+03 3.48E+02 2.12E+03 4.02E+02 + 1.85E+03 2.83E+02 + 2.23E+03 7.81E+02 +
F18 1.01E+02 1.16E+01 1.72E+02 8.30E+01 + 2.47E+02 4.84E+01 + 1.54E+02 3.66E+01 +
F19 9.30E+00 1.74E+00 1.44E+01 2.29E+00 1.24E+01 2.12E+00 9.38E+00 3.25E+00
re- + + =
F20 1.43E+01 4.85E+00 2.93E+02 8.50E+01 + 2.43E+02 6.79E+01 + 1.98E+02 5.73E+01 +
F21 6.34E+02 2.40E+02 1.54E+03 3.31E+02 + 1.45E+03 3.06E+02 + 1.23E+03 4.06E+02 +
F22 1.32E+02 7.59E+01 2.01E+02 1.63E+02 + 2.71E+02 1.39E+02 + 3.68E+02 1.59E+02 +
F23 3.44E+02 3.59E-13 3.44E+02 5.74E-14 = 3.44E+02 5.74E-14 = 3.44E+02 5.74E-14 =
F24 2.76E+02 5.73E-01 2.62E+02 2.86E+00 - 2.71E+02 2.63E+00 - 2.75E+02 1.73E+00 -
F25 2.05E+02 4.38E-01 2.05E+02 3.05E-01 = 2.05E+02 2.70E-01 = 2.12E+02 6.92E+00 +
F26 1.00E+02 1.94E-02 1.00E+02 1.32E-01 = 1.00E+02 1.01E-01 = 1.04E+02 1.95E+01 +
F27 3.29E+02 3.07E+01 3.10E+02 1.73E+01 - 3.17E+02 2.23E+01 = 4.53E+02 5.70E+01 +
F28 1.12E+03 2.78E+01 1.25E+03 7.98E+01 + 1.15E+03 7.54E+01 + 1.14E+03 5.26E+01 +
F29 7.98E+02 3.42E+01 8.05E+02 4.74E+01 + 7.94E+02 2.96E+01 = 8.91E+02 5.54E+01 +
F30 8.80E+03 4.38E+02 9.48E+03 6.10E+02 1.04E+04 6.38E+02 9.38E+03 7.87E+02
lP
+ + +
+/-/= 13/9/8 14/8/8 19/5/6
SASS LSHADE EBOwithCMAR HSES
Prob Mean SD Mean SD W Mean SD W Mean SD W
F1 3.16E+02 6.46E+02 1.24E+03 1.52E+03 + 6.25E-03 1.98E-03 - 3.61E-09 1.74E-08 -
F2 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 3.82E-10 2.73E-09 +
F3 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 1.23E-09 5.35E-09 +
F4 1.37E+01 2.96E+01 5.89E+01 4.56E+01 + 2.89E+01 4.51E+01 + 0.00E+00 0.00E+00 -
F5 2.03E+01 3.05E-02 2.03E+01 3.09E-02 = 2.00E+01 1.46E-04 - 2.00E+01 1.12E-04 -
F6 6.61E-01 6.61E-01 3.58E-01 5.44E-01 - 1.06E-02 7.28E-02 = 3.23E-05 7.61E-05 -
F7 1.45E-04 1.04E-03 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 6.19E-10 3.29E-09 =
F8 0.00E+00 0.00E+00 2.58E-09 7.48E-09 + 0.00E+00 0.00E+00 = 3.57E+00 2.29E+00 +
a
25
Journal Pre-proof
Table 15: Mean and SD of best error value obtained in 51 independent runs by SASS, iCMAES-ILS, NBIPOP-aCMA-ES, SHADE, LSHADE,
EBOwithCMAR, and HSES on 100-D CEC2014 problem suite (Mean: Mean of best error, SD: Standard deviation of best error, W: result of
Wilcoxon signed rank test).
of
+
F7 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 = 1.98E-03 5.34E-03 +
F8 8.41E-04 4.65E-04 1.94E+01 3.05E+00 + 4.56E+01 1.39E+01 + 0.00E+00 0.00E+00 -
F9 5.81E+01 5.47E+00 2.01E+01 3.30E+00 - 6.93E+00 3.17E+00 - 1.14E+02 1.55E+01 +
F10 1.55E+01 3.73E+00 9.31E+02 6.51E+02 + 5.16E+03 2.53E+03 + 1.03E-02 7.58E-03 -
F11 1.10E+04 5.01E+02 1.51E+03 7.96E+02 - 3.72E+03 1.71E+03 - 9.80E+03 4.97E+02 -
F12 4.31E-01 4.18E-02 6.62E-04 5.94E-04 - 1.12E-02 5.16E-03 - 2.29E-01 2.44E-02 -
F13 2.55E-01 1.96E-02 2.39E-01 3.18E-02 = 2.70E-01 6.87E-02 + 4.00E-01 4.28E-02 +
pro
F14 2.31E-01 9.32E-03 1.35E-01 1.63E-02 - 1.18E-01 1.35E-02 - 1.20E-01 9.78E-03 -
F15 1.57E+01 1.26E+00 9.79E+00 8.37E-01 - 9.68E+00 1.00E+00 - 2.21E+01 2.87E+00 +
F16 3.93E+01 4.32E-01 4.25E+01 7.68E-01 + 4.29E+01 2.00E+00 + 3.96E+01 4.93E-01 +
F17 4.43E+03 6.79E+02 5.07E+03 7.47E+02 + 4.84E+03 6.24E+02 + 1.24E+04 4.80E+03 +
F18 2.15E+02 1.26E+01 5.43E+02 1.69E+02 + 6.60E+02 9.88E+01 + 5.07E+02 3.89E+02 +
F19 9.89E+01 3.08E+00 8.21E+01 3.35E+01 = 1.02E+02 1.61E+01 + 9.78E+01 1.55E+01 =
F20 2.25E+02 5.94E+01 6.53E+02 1.24E+02 + 6.16E+02 1.22E+02 + 4.92E+02 1.34E+02 +
F21 2.24E+03 5.11E+02 3.32E+03 7.39E+02 + 3.39E+03 6.66E+02 + 3.69E+03 1.62E+03 +
F22 1.09E+03 2.14E+02 5.38E+02 1.78E+02 - 7.62E+02 2.50E+02 - 1.40E+03 2.13E+02 +
F23 3.48E+02 2.49E-13 3.48E+02 5.74E-14 = 3.48E+02 5.74E-14 = 3.48E+02 5.74E-14 =
F24 3.95E+02 2.81E+00 3.53E+02 9.49E+00 - 3.88E+02 2.61E+00 = 3.94E+02 4.45E+00 =
F25 2.00E+02 4.92E-13 2.18E+02 1.86E+00 + 2.18E+02 2.19E+00 + 2.58E+02 5.87E+00 +
F26 2.00E+02 4.59E-13 1.91E+02 2.72E+01 + 1.49E+02 5.03E+01 - 2.00E+02 1.26E-02 +
F27 5.39E+02 6.26E+01 3.00E+02 1.93E-01 - 3.00E+02 2.77E+00 - 9.50E+02 8.84E+01 +
F28 2.21E+03 6.10E+01 2.49E+03 8.67E+01
re- + 2.46E+03 3.05E+02 + 2.32E+03 2.51E+02 +
F29 7.52E+02 4.03E+01 1.51E+03 7.18E+02 + 7.54E+02 3.03E+01 + 1.35E+03 1.49E+02 +
F30 7.09E+03 1.04E+03 8.71E+03 1.32E+03 + 9.04E+03 1.75E+03 + 8.50E+03 9.27E+02 +
+/-/= 13/6/11 14/10/6 18/5/7
SS LSHADE EBOwithCMAR HSES
Prob Mean SD Mean SD W Mean SD W Mean SD W
F1 1.72E+05 5.67E+04 1.72E+05 5.65E+04 = 5.17E-03 5.36E-04 - 1.28E-01 1.03E-01 -
F2 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 2.50E-09 7.01E-09 + 1.71E-09 6.21E-09 +
F3 0.00E+00 0.00E+00 0.00E+00 0.00E+00 = 3.89E-07 5.84E-07 + 1.18E-08 4.96E-08 +
F4 9.96E+01 4.72E+01 1.70E+02 3.06E+01 + 1.55E+02 2.82E+01 + 2.97E+01 5.29E+01 -
F5 2.06E+01 3.60E-02 2.06E+01 3.11E-02 = 2.00E+01 1.04E-04 - 2.00E+01 1.52E-04 -
lP
+ +
F23 3.48E+02 2.49E-13 3.48E+02 3.66E-13 = 3.48E+02 5.74E-14 + 3.48E+02 3.68E-03 +
F24 3.95E+02 2.81E+00 3.94E+02 2.87E+00 = 3.77E+02 4.77E+00 - 3.76E+02 2.26E+00 =
F25 2.00E+02 4.92E-13 2.00E+02 4.09E-13 + 2.18E+02 1.27E+00 + 2.04E+02 7.99E+00 +
F26 2.00E+02 4.59E-13 2.00E+02 6.27E-13 + 2.00E+02 1.46E-02 + 2.00E+02 3.81E-02 +
F27 5.39E+02 6.26E+01 3.77E+02 3.28E+01 - 3.01E+02 3.90E+00 - 3.00E+02 1.67E-05 -
F28 2.21E+03 6.10E+01 2.31E+03 4.65E+01 + 2.14E+03 2.99E+02 = 2.29E+03 2.41E+02 +
F29 7.52E+02 4.03E+01 7.98E+02 7.58E+01 + 9.51E+02 1.77E+02 + 7.53E+02 5.93E+01 +
F30 7.09E+03 1.04E+03 8.29E+03 9.63E+02 + 6.51E+03 9.53E+02 = 6.98E+03 6.35E+03 =
+/-/= 13/11/6 12/6/12 13/5/12
Table 16: Ranking of Algorithm according to Friedman ranking based on mean error value.(FR: Friedman Ranking)
Jo
26
Journal Pre-proof
ployed for 25 independent runs on each problem of IEEE CEC11 problem suite and 50, 000 function evaluations as a
stopping criterion as suggested in [54]. The following six algorithms are considered in this experiment.
1. GA-MPC: GA with a new multi-parent crossover (winner of the IEEE CEC’11 competition) [55].
2. SHADE: Success-history based parameter adaptation for differential evolution [31].
3. LSHADE: SHADE with Linear Population Size Reduction (winner of CEC’14 competition) [51] .
4. LSHADEEpSin: LSHADE with an ensemble sinusoidal parameter adaptation (winner of CEC’16 competetition)[56].
5. EBO: Effective Butterfly Optimizer [52].
of
6. EBOwithCMAR: EBO with covariance matrix adapted retreat phase (winner of CEC’17 competition)[52].
A comparison of SASS with the aforementioned contenders is performed and reported in Table 17. In Table 17,
mean and standard deviation (SD) of the returned best objective function values of each run for each algorithm is
shown. The Wilcoxon’s signed rank test (WT) is also utilized to check the significance of the outcomes of SASS with
pro
a comparison of other algorithms. In Table 17, the last row summarizes the outcome of WT in terms of a number of
superior, equal, and inferior performance (‘+’, ‘=’, and ‘-’ respectively) of SASS. The results of Table 17 show that
the SASS outperforms GA-MPC, SHADE, LSHADE, LSHADEEpSin, EBO, and EBOwithCMAR on 13, 14, 10, 10,
14, and nine out of 22 problems, respectively. However, SASS is also outperformed by GA-MPC, SHADE, LSHADE,
LSHADEEpSin, EBO, and EBOwithCMAR on five, three, four, five, one, and seven out of 22 problems, respectively,
nevertheless this clearly proves that the performance of SASS is better than the other aforementioned contenders over
IEEE CEC’11 problem suite. Moreover, a Friedman ranking is also done and depicted in Table 18 for each algorithm.
The SASS shows best mean rank equal to 3.5 followed by EBOwithCMAR. From the results of statistical tests and
re-
comparative analysis of the performance of all algorithms, it can be concluded that SASS is an effective and robust
optimization technique for real-world problems.
4.6. Discussions
In the context of this work, it can be concluded that the SS algorithm is free from the tuning of parameters that
depends upon the characteristics of optimization problems. The performance of this algorithm highly depends upon
lP
the projection matrix and search direction. Projection matrix improves the diversity of solutions in population, while
it requires a more extra computational burden. Two approaches are used to calculate search direction which provide
a good balance between exploration and exploitation. However, these approaches are somewhat influenced by step-
control parameter. In order to address this issue, a parameter adaptation is also proposed to tune the parameters online
during the optimization process. Thus, by analyzing the outcomes of detailed experiments and statistical tests on two
popular advanced benchmark suites, it can be concluded that the performance of the proposed algorithm is better or
a
highly competitive with the advanced variant of state-of-art algorithms. However, the performance of SASS (SS with
parameter adaptation) is outperformed by the advanced variant of state-of-art algorithms in the few cases.
urn
5. Conclusion
This paper proposed a new optimization technique, Spherical Search algorithm. The proposed method was tested
on modern unconstrained benchmark problems to analyses the performance. The experimental results show that it
provides the superior result on most of the benchmark problems compared to state-of-the-art meta-heuristics. The
results obtained from all experiments and comparison of characteristics with other algorithms conclude the following.
1. SS has few parameters to tune, and it is easy to implement for solving unconstrained optimization problems.
Jo
2. The quality and accuracy of obtained solutions, the rate of convergence, proficiency, and effectiveness of SS are
at an advanced degree as compared to its opponents.
3. SS possibly escapes from the situation of stagnation and benefits to get rid of stacking in local minima, due to
its projection property.
In the future, it is proposed to develop versions of SS for constrained, binary, and multi-objective optimization
problems.
27
Journal Pre-proof
Table 17: Comparison among the performance of SASS with the performance of other state-of-art algorithms on problem IEEE CEC 2011. Best
entries are shown in bold face. (Mean: Average of error achieved for 51 independent runs, Std: standard deviation of error achieved for 51
independent runs, WT: outcome of Wilcoxon signed rank test at 0.05 significance level)
of
T6 -2.91E+01 3.46E-01 -2.71E+01 2.44E+00 + -2.92E+01 4.24E-03 = -2.92E+01 4.78E-03 =
T7 1.16E+00 8.05E-02 2.06E+00 4.35E-02 + 6.11E-01 9.03E-02 - 1.17E+00 9.38E-02 +
T8 2.20E+02 0.00E+00 2.20E+02 0.00E+00 = 2.20E+02 0.00E+00 = 2.20E+02 0.00E+00 =
T9 2.88E+03 7.32E+02 2.27E+03 9.89E+02 - 1.98E+03 3.61E+02 - 2.65E+03 4.36E+02 -
T10 -2.16E+01 9.06E-02 -2.17E+01 1.27E-01 - -2.18E+01 9.91E-02 = -2.16E+01 8.70E-02 =
T11 5.19E+04 3.64E+02 5.38E+04 8.22E+03 + 5.25E+04 5.06E+02 + 5.20E+04 5.03E+02 +
T12 1.07E+06 1.41E+03 1.08E+06 1.19E+04 + 1.07E+06 4.43E+04 + 1.78E+07 5.68E+04 +
pro
T13 1.54E+04 1.30E-06 1.54E+04 8.75E-06 = 1.54E+04 1.09E+01 + 1.54E+04 6.16E-01 +
T14 1.81E+04 2.07E+01 1.83E+04 8.50E+01 + 1.81E+04 3.37E+01 + 1.81E+04 3.34E+01 +
T15 3.27E+04 9.25E-01 3.28E+04 3.86E+01 + 3.27E+04 1.16E+01 + 3.27E+04 3.31E-01 +
T16 1.24E+05 4.30E+02 1.36E+05 2.74E+03 + 1.30E+05 4.60E+02 + 1.24E+05 5.58E+02 +
T17 1.86E+06 1.26E+04 2.17E+06 4.80E+05 + 1.90E+06 1.22E+04 + 1.87E+06 9.83E+03 +
T18 9.33E+05 1.30E+03 1.03E+06 1.65E+05 + 9.40E+05 9.82E+02 + 9.32E+05 9.63E+02 =
T19 9.41E+05 8.85E+02 1.30E+06 1.99E+05 + 9.77E+05 1.15E+03 + 9.40E+05 9.00E+02 -
T20 9.33E+05 1.29E+03 1.04E+06 4.28E+04 + 9.40E+05 1.16E+03 + 9.32E+05 1.16E+03 -
T21 1.61E+01 6.37E-01 1.30E+01 3.41E+00 - 1.86E+01 7.75E-01 + 1.62E+01 7.61E-01 +
T22 1.37E+01 2.10E+00 1.05E+01 2.13E+00 - 1.49E+01 2.42E+00 + 1.34E+01 1.90E+00 -
+/=/- 13/4/5 14/5/3 10/8/4
SASS L-SHADE-EpSin
re- EBO EBOwithCMAR
Prob
Mean Std Mean Std WT Mean Std WT Mean Std WT
T1 1.30E-03 5.38E-03 9.73E-01 3.44E+00 + 1.66E-01 3.56E-01 + 7.08E-12 5.26E-12 -
T2 -2.59E+01 5.43E-01 -2.59E+01 5.81E-01 = -2.32E+01 1.31E+00 + -2.71E+01 6.94E-01 -
T3 1.15E-05 1.73E-21 1.15E-05 1.98E-19 = 1.15E-05 4.45E-20 = 1.15E-05 1.73E-21 =
T4 0.00E+00 0.00E+00 1.45E+01 1.08E+00 + 0.00E+00 0.00E+00 = 0.00E+00 0.00E+00 =
T5 -3.62E+01 7.67E-01 -3.65E+01 5.05E-01 = -3.59E+01 5.42E-01 + -3.59E+01 7.53E-01 +
T6 -2.91E+01 3.46E-01 -2.92E+01 7.77E-03 = -2.91E+01 3.46E-01 = -2.89E+01 1.27E+00 +
T7 1.16E+00 8.05E-02 1.15E+00 9.18E-02 = 1.31E+00 8.56E-02 + 7.36E-01 1.49E-01 -
T8 2.20E+02 0.00E+00 2.20E+02 0.00E+00 2.20E+02 0.00E+00 2.20E+02 0.00E+00
lP
= = =
T9 2.88E+03 7.32E+02 1.91E+03 3.41E+02 - 1.67E+03 3.54E+02 - 2.46E+03 5.22E+02 -
T10 -2.16E+01 9.06E-02 -2.16E+01 9.40E-02 + -2.15E+01 1.34E-01 + -2.16E+01 1.00E-01 =
T11 5.19E+04 3.64E+02 5.21E+04 6.05E+02 + 5.23E+04 2.68E+02 + 5.20E+04 4.63E+02 +
T12 1.07E+06 1.41E+03 1.78E+07 5.86E+04 + 1.07E+06 1.61E+03 = 1.07E+06 1.69E+03 =
T13 1.54E+04 1.30E-06 1.55E+04 1.34E+01 + 1.54E+04 5.13E-07 = 1.54E+04 7.43E-12 -
T14 1.81E+04 2.07E+01 1.81E+04 4.40E+01 + 1.81E+04 1.64E+02 = 1.81E+04 1.95E+01 =
T15 3.27E+04 9.25E-01 3.27E+04 8.29E+00 + 3.27E+04 1.34E+01 + 3.27E+04 3.25E+00 +
T16 1.24E+05 4.30E+02 1.24E+05 5.31E+02 + 1.26E+05 5.87E+02 + 1.26E+05 6.18E+02 +
T17 1.86E+06 1.26E+04 1.85E+06 1.00E+04 - 1.88E+06 1.11E+04 + 1.90E+06 2.63E+04 +
T18 9.33E+05 1.30E+03 9.30E+05 9.67E+02 - 9.37E+05 1.54E+03 + 9.38E+05 1.48E+03 +
a
References
[1] D. H. Wolpert, W. G. Macready, No free lunch theorems for optimization, Evolutionary Computation, IEEE Transactions on 1 (1997) 67–82.
[2] K. S. Tang, K. F. Man, S. Kwong, Q. He, Genetic algorithms and their applications, Signal Processing Magazine, IEEE 13 (1996) 22–37.
[3] I. Rechenberg, Evolution strategy: Natures way of optimization, in: Optimization: Methods and applications, possibilities and limitations,
Springer, 1989, pp. 106–126.
[4] X. Yao, Y. Liu, G. Lin, Evolutionary programming made faster, Evolutionary Computation, IEEE Transactions on 3 (1999) 82–102.
Jo
[5] L. J. Fogel, A. J. Owens, M. J. Walsh, Artificial intelligence through simulated evolution (1966).
[6] J. R. Koza, Genetic programming: on the programming of computers by means of natural selection, volume 1, MIT press, 1992.
[7] R. Storn, K. Price, Differential evolution–a simple and efficient heuristic for global optimization over continuous spaces, Journal of global
optimization 11 (1997) 341–359.
[8] J. Kennedy, Particle swarm optimization, in: Encyclopedia of Machine Learning, Springer, 2010, pp. 760–766.
[9] M. Dorigo, M. Birattari, Ant colony optimization, in: Encyclopedia of Machine Learning, Springer, 2010, pp. 36–39.
[10] D. Karaboga, B. Basturk, A powerful and efficient algorithm for numerical function optimization: artificial bee colony (abc) algorithm,
Journal of global optimization 39 (2007) 459–471.
[11] S. Mirjalili, S. M. Mirjalili, A. Lewis, Grey wolf optimizer, Advances in Engineering Software 69 (2014) 46–61.
28
Journal Pre-proof
Table 18: Ranking of Algorithm according to Friedman ranking based on mean error value.(FR: Friedman Ranking)
Ranking
Algorithm
FR Rank
SASS 3.5000 1
GA-MPC 5.1364 7
SHADE 4.1818 5
LSHADE 3.6591 4
LSHADEEpSin 3.5455 2
EBO 4.3636 6
EBOwithCMAR 3.6136 3
of
[12] X.-S. Yang, A new metaheuristic bat-inspired algorithm, in: Nature inspired cooperative strategies for optimization (NICSO 2010), Springer,
2010, pp. 65–74.
[13] G. Dhiman, V. Kumar, Seagull optimization algorithm: Theory and its applications for large-scale industrial engineering problems,
Knowledge-Based Systems 165 (2019) 169–196.
pro
[14] S. Mirjalili, A. Lewis, The whale optimization algorithm, Advances in engineering software 95 (2016) 51–67.
[15] S. Mirjalili, The ant lion optimizer, Advances in Engineering Software 83 (2015) 80–98.
[16] M. Yazdani, F. Jolai, Lion optimization algorithm (loa): a nature-inspired metaheuristic algorithm, Journal of computational design and
engineering 3 (2016) 24–36.
[17] S. Balochian, H. Baloochian, Social mimic optimization algorithm and engineering applications, Expert Systems with Applications (2019).
[18] A. Kaveh, A. Dadras, A novel meta-heuristic optimization algorithm: thermal exchange optimization, Advances in Engineering Software
110 (2017) 69–84.
[19] H. Rakhshani, A. Rahati, Snap-drift cuckoo search: A novel cuckoo search optimization algorithm, Applied Soft Computing 52 (2017)
771–794.
[20] M. D. Li, H. Zhao, X. W. Weng, T. Han, A novel nature-inspired algorithm for optimization: Virus colony search, Advances in Engineering
re-
Software 92 (2016) 65–88.
[21] S. Saremi, S. Mirjalili, A. Lewis, Grasshopper optimisation algorithm: theory and application, Advances in Engineering Software 105 (2017)
30–47.
[22] S. Mirjalili, A. H. Gandomi, S. Z. Mirjalili, S. Saremi, H. Faris, S. M. Mirjalili, Salp swarm algorithm: A bio-inspired optimizer for
engineering design problems, Advances in Engineering Software 114 (2017) 163–191.
[23] M. Jain, V. Singh, A. Rani, A novel nature-inspired algorithm for optimization: Squirrel search algorithm, Swarm and evolutionary compu-
tation 44 (2019) 148–175.
lP
[24] A. Mortazavi, V. Toğan, A. Nuhoğlu, Interactive search algorithm: a new hybrid metaheuristic optimization algorithm, Engineering Appli-
cations of Artificial Intelligence 71 (2018) 275–292.
[25] G. Dhiman, V. Kumar, Emperor penguin optimizer: A bio-inspired algorithm for engineering problems, Knowledge-Based Systems 159
(2018) 20–50.
[26] J. Pierezan, L. D. S. Coelho, Coyote optimization algorithm: a new metaheuristic for global optimization problems, in: 2018 IEEE Congress
on Evolutionary Computation (CEC), IEEE, pp. 1–8.
[27] G. Dhiman, V. Kumar, Spotted hyena optimizer: a novel bio-inspired based metaheuristic technique for engineering applications, Advances
in Engineering Software 114 (2017) 48–70.
[28] J. D. Ser, E. Osaba, D. Molina, X.-S. Yang, S. Salcedo-Sanz, D. Camacho, S. Das, P. N. Suganthan, C. A. C. Coello, F. Herrera, Bio-inspired
a
computation: Where we stand and what’s next, Swarm and Evolutionary Computation 48 (2019) 220 – 250.
[29] M. Locatelli, A note on the griewank test function, Journal of global optimization 25 (2003) 169–174.
[30] A. Kumar, R. K. Misra, D. Singh, S. Das, Testing a multi-operator based differential evolution algorithm on the 100-digit challenge for single
urn
objective numerical optimization, in: 2019 IEEE Congress on Evolutionary Computation (CEC), IEEE.
[31] R. Tanabe, A. Fukunaga, Success-history based parameter adaptation for differential evolution, in: Evolutionary Computation (CEC), 2013
IEEE Congress on, IEEE, pp. 71–78.
[32] J. Liang, B. Qu, P. Suganthan, Problem definitions and evaluation criteria for the cec 2014 special session and competition on single objective
real-parameter numerical optimization, Computational Intelligence Laboratory, Zhengzhou University, Zhengzhou China and Technical
Report, Nanyang Technological University, Singapore (2013).
[33] J. Kennedy, Bare bones particle swarms, in: Swarm Intelligence Symposium, 2003. SIS’03. Proceedings of the 2003 IEEE, IEEE, pp. 80–87.
[34] K. Mahadevan, P. Kannan, Comprehensive learning particle swarm optimization for reactive power dispatch, Applied soft computing 10
(2010) 641–652.
[35] Z.-H. Zhan, J. Zhang, Y. Li, H. S.-H. Chung, Adaptive particle swarm optimization, IEEE Transactions on Systems, Man, and Cybernetics,
Part B (Cybernetics) 39 (2009) 1362–1381.
Jo
[36] Z.-H. Zhan, J. Zhang, Y. Li, Y.-H. Shi, Orthogonal learning particle swarm optimization, IEEE transactions on evolutionary computation 15
(2011) 832–847.
[37] Y. Li, J. Feng, J. Hu, Covariance and crossover matrix guided differential evolution for global numerical optimization, SpringerPlus 5 (2016)
1176.
[38] Z. Li, Z. Shang, B. Y. Qu, J.-J. Liang, Differential evolution strategy based on the constraint of fitness values classification, in: Evolutionary
Computation (CEC), 2014 IEEE Congress on, IEEE, pp. 1454–1460.
[39] C. Xu, H. Huang, S. Ye, A differential evolution with replacement strategy for real-parameter numerical optimization, in: Evolutionary
Computation (CEC), 2014 IEEE Congress on, IEEE, pp. 1617–1624.
[40] Z. Hu, Y. Bao, T. Xiong, Partial opposition-based adaptive differential evolution algorithms: Evaluation on the cec 2014 benchmark set for
29
Journal Pre-proof
real-parameter optimization, in: Evolutionary Computation (CEC), 2014 IEEE Congress on, IEEE, pp. 2259–2265.
[41] N. Hansen, The cma evolution strategy: a comparing review, in: Towards a new evolutionary computation, Springer, 2006, pp. 75–102.
[42] A. Auger, N. Hansen, A restart cma evolution strategy with increasing population size, in: Evolutionary Computation, 2005. The 2005 IEEE
Congress on, volume 2, IEEE, pp. 1769–1776.
[43] A. Auger, M. Schoenauer, N. Vanhaecke, Ls-cma-es: A second-order algorithm for covariance matrix adaptation, in: International Conference
on Parallel Problem Solving from Nature, Springer, pp. 182–191.
[44] G. A. Jastrebski, D. V. Arnold, Improving evolution strategies through active covariance matrix adaptation, in: Evolutionary Computation,
2006. CEC 2006. IEEE Congress on, IEEE, pp. 2814–2821.
[45] C. Igel, T. Suttorp, N. Hansen, A computational efficient covariance matrix update and a (1+ 1)-cma for evolution strategies, in: Proceedings
of the 8th annual conference on Genetic and evolutionary computation, ACM, pp. 453–460.
of
[46] S. Mirjalili, S. M. Mirjalili, A. Hatamlou, Multi-verse optimizer: a nature-inspired algorithm for global optimization, Neural Computing and
Applications 27 (2016) 495–513.
[47] S. Mirjalili, Sca: a sine cosine algorithm for solving optimization problems, Knowledge-Based Systems 96 (2016) 120–133.
[48] E. Cuevas, F. Fausto, A. González, The selfish herd optimizer, in: New Advancements in Swarm Algorithms: Operators and Applications,
Springer, 2020, pp. 69–109.
[49] T. Liao, T. Stuetzle, Benchmark results for a simple hybrid algorithm on the cec 2013 benchmark set for real-parameter optimization, in:
pro
2013 IEEE Congress on Evolutionary Computation, IEEE, pp. 1938–1944.
[50] I. Loshchilov, Cma-es with restarts for solving cec 2013 benchmark problems, in: 2013 IEEE Congress on Evolutionary Computation, Ieee,
pp. 369–376.
[51] R. Tanabe, A. S. Fukunaga, Improving the search performance of shade using linear population size reduction, in: 2014 IEEE congress on
evolutionary computation (CEC), IEEE, pp. 1658–1665.
[52] A. Kumar, R. K. Misra, D. Singh, Improving the local search capability of effective butterfly optimizer using covariance matrix adapted
retreat phase, in: 2017 IEEE Congress on Evolutionary Computation (CEC), IEEE, pp. 1835–1842.
[53] G. Zhang, Y. Shi, Hybrid sampling evolution strategy for solving single objective bound constrained problems, in: 2018 IEEE Congress on
Evolutionary Computation (CEC), IEEE, pp. 1–7.
re-
[54] S. Das, P. N. Suganthan, Problem definitions and evaluation criteria for cec 2011 competition on testing evolutionary algorithms on real world
optimization problems, Jadavpur University, Nanyang Technological University, Kolkata (2010) 341–359.
[55] S. M. Elsayed, R. A. Sarker, D. L. Essam, Ga with a new multi-parent crossover for solving ieee-cec2011 competition problems, in: 2011
IEEE congress of evolutionary computation (CEC), IEEE, pp. 1034–1040.
[56] N. H. Awad, M. Z. Ali, P. N. Suganthan, R. G. Reynolds, An ensemble sinusoidal parameter adaptation incorporated with l-shade for solving
cec2014 benchmark problems, in: 2016 IEEE congress on evolutionary computation (CEC), IEEE, pp. 2958–2965.
a lP
urn
Jo
30
Journal Pre-proof
Y
0
Y
0
-2
-2 -2
-4
-4 -4
-6
-6 -6
of
-8
-8 -8
-10
-10 -10
-10 -8 -6 -4 -2 0 2 4 6 8 10
X -10 -8 -6 -4 -2 0 2 4 6 8 10 -10 -8 -6 -4 -2 0 2 4 6 8 10
Alg-DE, It-80
X X
pro
10
8 8
8
6 6
6
4 4
4
2 2
2
Y
0
Y
0
Y
0
-2 -2
-2
-4 -4
-4
-6 -6 -6
-8 -8 -8
0
Y
0
Y
0
-2
-2 -2
lP
-4
-4 -4
-6
-6 -6
-8
-8 -8
-10
-10 -10
-10 -8 -6 -4 -2 0 2 4 6 8 10
-10 -8 -6 -4 -2 0 2 4 6 8 10 -10 -8 -6 -4 -2 0 2 4 6 8 10
X
X X
6
6
4 4
4
2 2
2
Y
0
Y
0
Y
0
urn
-2 -2 -2
-4 -4 -4
-6 -6 -6
-8 -8 -8
-10 -10 -10
-10 -8 -6 -4 -2 0 2 4 6 8 10 -10 -8 -6 -4 -2 0 2 4 6 8 10 -10 -8 -6 -4 -2 0 2 4 6 8 10
X X X
8 8 8
6 6 6
4 4 4
Jo
2 2 2
Y
Y
0 0
Y
-2 -2 -2
-4 -4 -4
-6 -6 -6
-8 -8 -8
Figure 5: Search-History of individuals of SS, DE, and PSO with respect to iteration.
31
Journal Pre-proof
Y
0
Y
Y
0
-2
-2 -2
-4
-4 -4
-6
-6
of
-6
-8
-8
-8
-10
-10
-10
-10 -8 -6 -4 -2 0 2 4 6 8 10
-10 -8 -6 -4 -2 0 2 4 6 8 10
-10 -8 -6 -4 -2 0 2 4 6 8 10 X
X
X
pro
10
10 10
8
8 8
6
6 6
4 4 4
2 2 2
Y
0
Y
Y
0
-2 -2 -2
-4 -4 -4
-6 -6 -6
-8 -8 -8
-10 -8 -6 -4 -2 0 2 4 6 8 10 -10 -8 -6
re-
-4 -2 0 2 4 6 8 10 -10 -8 -6 -4 -2 0 2 4 6 8 10
X X X
0
lP
Y
0
Y
0
-2
-2 -2
-4
-4 -4
-6
-6 -6
-8
-8 -8
-10
-10 -10
-10 -8 -6 -4 -2 0 2 4 6 8 10
-10 -8 -6 -4 -2 0 2 4 6 8 10 -10 -8 -6 -4 -2 0 2 4 6 8 10
X
Alg-SS, Ind-74
X X
10
8 8
8
6 6
6
4 4
4
2 2
urn
2
Y
0
Y
0
Y
0
-2
-2
-2
-4
-4
-4
-6
-6
-6
-8
-8
-8
-10
-10 -10
-10 -8 -6 -4 -2 0 2 4 6 8 10
-10 -8 -6 -4 -2 0 2 4 6 8 10 -10 -8 -6 -4 -2 0 2 4 6 8 10
X
X X
8
6 6
6
4 4
4
2 2
2
Y
0
Y
0
Y
0
-2 -2
-2
-4 -4
-4
-6 -6 -6
-8 -8 -8
1 0
of
0 .1
P o p u la tio n D iv e r s ity
P S O
0 .0 0 1 D E
pro
S S
1 E -5
1 E -7
re-
1 E -9
5 0 1 0 0 1 5 0 2 0 0
N u m b e r o f Ite r a tio n s
lP
Figure 7: Population diversity of SS, DE, and PSO with respect to iteration.
a
Dimension-2 Dimension-20
1 1 0
B e s t O b je c tiv e F u n c tio n V a lu e
P S O
urn
B e s t O b je c tiv e F u n c tio n V a lu e
0 .0 1 D E 0 .1
S S
1 E -4 1 E -3
1 E -6 1 E -5
1 E -8 1 E -7 P S O
D E
S S
1 E -1 0 1 E -9
Jo
1 E -1 2 1 E -1 1
5 0 1 0 0 1 5 0 2 0 0 5 0 1 0 0 1 5 0 2 0 0 2 5 0 3 0 0
N u m b e r o f Ite r a tio n s N u m b e r o f Ite r a tio n s
Figure 8: Convergence of solutions of SS, DE, and PSO with respect to iteration.
33
Journal Pre-proof
Highlights
of
SS maintains high diversity during the optimization process compared to
PSO and DE.
Efficacy of SS is showcased through extensive experiments.
pro
re-
a lP
urn
Jo
Journal Pre-proof
Declaration of interests
☒ The authors declare that they have no known competing financial interests or personal relationships
that could have appeared to influence the work reported in this paper.
☐The authors declare the following financial interests/personal relationships which may be considered
of
as potential competing interests:
pro
re-
a lP
Swagatam Das,
urn