Hybridizing grey wolf optimization with neural network algorithm
Hybridizing grey wolf optimization with neural network algorithm
https://doi.org/10.1007/s00521-019-04580-4 (0123456789().,-volV)(0123456789().
,- volV)
ORIGINAL ARTICLE
Abstract
This paper proposes a novel hybrid algorithm, called grey wolf optimization with neural network algorithm (GNNA), for
solving global numerical optimization problems. The core idea of GNNA is to make full use of good global search ability
of neural network algorithm (NNA) and fast convergence of grey wolf optimizer (GWO). Moreover, both NNA and GWO
are improved to boost their own advantages. For NNA, an improved NNA is given to strengthen the exploration ability of
NNA by discarding transfer operator and introducing random modification factor. For GWO, an enhanced GWO is
presented, which adjusts the exploration rate based on reinforcement learning principles. Then the improved NNA and the
enhanced GWO are hybridized by dynamic population mechanism. A comprehensive set of 23 well-known unconstrained
benchmark functions are employed to examine the performance of GNNA compared with 13 metaheuristic algorithms.
Such comparisons suggest that the combination of the improved NNA and the enhanced GWO is very effective and GNNA
is clearly seen to be more successful in both solution quality and computational efficiency.
Keywords Artificial neural networks Reinforcement learning Grey wolf optimizer Numerical optimization
123
Neural Computing and Applications
algorithm (BA) [7], leadership hierarchy and hunting new hybrid method based on NNA and GWO to solve
mechanism of grey wolves in grey wolf optimizer global optimization problems.
(GWO) [8], social behaviour of fireflies based on their NNA is one of the newest metaheuristic algorithms,
flashing characteristics in firefly algorithm (FA) [9], which is inspired by biological nervous systems and arti-
swarming behaviour of salps in salp swarm algorithm ficial neural networks (ANNs). Benefiting from the unique
(SSA) [10] and social behaviour of humpback whales structure of ANNs, NNA shows the following outstanding
in whale optimization algorithm (WOA) [11]. characteristics: (1) strong portability. NNA can be easily
2. Evolutionary algorithms. These algorithms are inspired applied to different optimization problems without adjust-
from some behaviour of the biological evolution, such ing other parameters except common control parameters
as reproduction, mutation and selection. Genetic algo- like population size and stopping criterion. Here, it is worth
rithm (GA) [12] and differential evolution (DE) [13] mentioning that common control parameters are essential
are two classical evolutionary algorithms. What’s for all population-based optimization algorithms; (2)
more, biogeography-based optimization (BBO) is a powerful ability of global search. NNA can find the global
recent proposed evolutionary algorithm, which imitates optimum with the minimum possibility of getting trapped
the biological migration behaviour in nature [14]. in local minima compared with many existing meta-
3. Human activity-related algorithms. These algorithms heuristic algorithms [17]. Therefore, NNA is easy to
are inspired from human activities, such as traditional implement and has good global search ability. However,
teaching mode in teaching–learning-based optimiza- NNA has the slow convergence speed that restricts its
tion (TLBO) [15], vehicle passing mechanisms on two- applications in some specific types of optimization prob-
lane highways in passing vehicle search (PVS) [1], sine lems with limitations like computation resource constraints
and cosine waves in sine cosine algorithm (SCA) [16] and high accuracy.
and artificial neural networks in nervous network GWO is a recent proposed metaheuristic algorithm,
algorithm (NNA) [17]. which mimics the hunting mechanism of grey wolves.
4. Physics-based algorithms. These algorithms are GWO has fast convergence speed and good global search
inspired from some physical phenomenon, such as ability, which has been used to solve many optimization
the principle of solid annealing in simulated annealing problems [24–27]. However, GWO may tend to conver-
algorithm (SA) [18] and the water cycle process in gence to local minima in solving some complex opti-
water cycle algorithm (WCA) [4]. mization problems. Once GWO got trapped in local optima
at the end of optimization process, it is very difficult for
The no-free-lunch (NFL) theorem [19] was proposed in
GWO to escape from local optima due to the limited
1997, which is an important theoretical basis of develop-
exploration ability [8, 28–30].
ment of optimization algorithms. According to NFL, there
Motivated by the characteristics of NNA and GWO, this
is no metaheuristic best suited for solving all optimization
paper presents an effective hybrid algorithm called grey
problems. In other words, one metaheuristic algorithm may
wolf optimization with neural network algorithm (GNNA).
achieve very promising results on a set of optimization
The core idea of this hybrid technique is to make the best
problems while it may present poor performance on
of good global search ability of NNA and fast convergence
another set of optimization problems. Thus NFL encour-
of GWO. Note that GNNA is not a simple combination of
ages researchers to develop new optimization algorithms or
NNA and GWO. Firstly, an improved NNA is proposed to
improve some existed ones for solving a wider range of
strengthen exploration of NNA by discarding the transfer
optimization problems. A technique which combines two
operator and introducing the random modification factor,
or more metaheuristic algorithms to produce hybrid
and an enhanced GWO is presented to balance exploration
methods that contain advantages of all constitutive algo-
and exploitation of GWO by adjusting the exploration rate
rithms can be a possible solution [20]. In [21], a hybrid
according to reinforcement learning principles. Then the
GSA-GA was presented to solve constrained optimization
improved NNA and the enhanced GWO are hybridized
problems. Xiong et al. [22] proposed a hybrid differential
based on dynamic population mechanism. More specifi-
evolution with whale optimization algorithm for parameter
cally, in each iteration, population is divided into two
extraction of solar photovoltaic models. A novel hybrid
halves according to fitness values of individuals. Then the
between the electromagnetism-like algorithm and the fire-
best half of individuals are optimized by the enhanced
fly algorithm for discrete structural optimization was
GWO to improve the convergence speed of GNNA, and the
introduced in [23]. Note that all above mentioned hybrid
worst half of individuals are processed by the improved
algorithms have been proven to be more efficient compared
NNA to strengthen the ability of GNNA in escaping from
with corresponding original methods. Considering the
local optima. At last, all individuals construct a new pop-
effectiveness of the hybrid technique, this paper proposes a
ulation after one iteration and the best individual is selected
123
Neural Computing and Applications
from the new population as the current optimum solution. 2. Update weight matrix.
The proposed GNNA is examined using 23 well-known The renewal equation of the weight matrix can be
unconstrained benchmark test functions, and its efficiency expressed as:
and superiority are highlighted against 13 metaheuristic
t t
algorithms along with statistical tests. wtþ1
i ¼ wi þ 2 c w t
opt wi ; i ¼ 1; 2; . . .; N
The rest of this paper is organized as follows: Sect. 2 ð4Þ
describes NNA and GWO. Section 3 elaborates the
improved NNA, the enhanced GWO and the proposed where c is a random number between 0 and 1 and wtopt
GNNA in detail. The experimental results are illustrated in is the target weight vector. The target weight vector
Sect. 4. Section 5 presents conclusions and future work. wtopt and the target solution xtopt are updated simulta-
neously. If xtopt is equal to xts ðs 2 ½1; N Þ at time t,wtopt is
equal to wts .
2 Background 3. Bias operator.
In the bias operator, modification factor bt is used to
2.1 Neural network algorithm determine the bias proportion, and it is updated by [17]
NNA is inspired by the structure of ANNs and biological btþ1 ¼ 0:99 bt ð5Þ
nervous systems, which is a population-based optimization In NNA, bias operator consists of two parts: bias of
algorithm. In general, ANNs are used for prediction pur- population and bias of weight matrix. For bias of
poses in most cases, which try to close the gap between population, a random number Nbt is generated firstly
predicted solutions and given target solution by frequently according to the following equation:
adjusting the values of weight functions. However, an
optimization algorithm is aimed at finding an optimum Nbt ¼ dbt De ð6Þ
solution from the given search space of the optimization Then a set T is produced by randomly selecting Nbt
problem. In order to adapt ANNs to be suitable for using as integers between 0 and D. Thus the bias of population
an optimization technique, the current best solution is can be described as (Let l ¼ ðl1 ; l2 ; . . .; lN Þ and u ¼
viewed as the target solution in NNA. The core idea of ðu1 ; u2 ; . . .; uN Þ are the lower and the upper bounds of
NNA is to close the gap between the target solution and the design variables, respectively)
other solutions included in the population by adjusting the
xti;TðkÞ ¼ lTðkÞ þ uTðkÞ lTðkÞ u; k ¼ 1; 2; . . .; Nb
values of weight functions. NNA includes the following
four major stages: ð7Þ
1. Generate the new population. where u is a random number between 0 and 1, lTðkÞ and
In NNA, every individual i has its corresponding uTðkÞ are the upper bound and the lower bound of the
h i
weight vector wti ¼ wti;1 ; wti;2 ; . . .; wti;N , which can be TðkÞth variable. As for bias of weight matrix, a random
number Nwt is generated firstly by the following
defined as
equation:
X
N
wti;j ¼ 1; 0\wti;j \1; i ¼ 1; 2; . . .; N ð1Þ Nwt ¼ dbt N e ð8Þ
j¼1
Then Nwt integers between 0 and N are randomly
where N is population size and t is the current number selected, which form a set Q. Thus the bias of weight
of iterations. Thus the process of generating the new matrix can be defined as
population can be described as
wti;QðbÞ ¼ c; b ¼ 1; 2; . . .; Nw ð9Þ
X
N
xtþ1
new;j ¼ wti;j xti ; j ¼ 1; 2; . . .; N ð2Þ where c is a random number between 0 and 1.
i¼1 4. Transfer operator.
xtþ1 ¼ xti þ xtþ1 i ¼ 1; 2; . . .; N ð3Þ In NNA, the transfer operator is to transfer the current
i new;i ;
solution to a better solution towards the current best
where xti is solution of the ith individual at time t, wti;j is solution, which can be expressed as
the jth weight value of the ith individual at time t and
xtþ1
new;j is the weighted solution of the jth individual at
time t þ 1.
123
Neural Computing and Applications
xtþ1
i ¼ xtþ1
i þ 2 / xtopt xtþ1
i ; i ¼ 1; 2; . . .; N
ð10Þ
where / is a random number between 0 and 1.
These steps show NNA is very simple and easy to carry
out. In addition, Fig. 1 presents the pseudo code of NNA.
123
Neural Computing and Applications
123
Neural Computing and Applications
atþ1
i ¼ ati :min ati w þ s g fitþ1 fit ; w ; 3.3 The proposed GNNA
ð22Þ
i ¼ 1; 2; . . .; N
As mentioned previously, the proposed GNNA is aim at
where w is a constant, s is penalty factor, fit is the fitness combining the good global search of INNA and fast con-
value of individual i at time t and gðÞ is a control function. vergence of EGWO. In GNNA, the information exchange
gðÞ can be defined as (take the minimum problem of EGWO and INNA is based on dynamic population
example): mechanism. More specifically, this mechanism can be
8
< 1 if x\0 (Decrease action) summarized as follows: (1) the individuals processed by
gð x Þ ¼ 1 if x [ 0 (Increase action) ð23Þ EGWO and INNA are from the same population. More
: specifically, the optimized population is divided into two
0 if x ¼ 0 ðKeep action)
halves based on the fitness values of individuals. The best
Moreover, considering the value of at decreasing from 2 to half of population is processed by EGWO to accelerate
0 in the basic GWO, w is set to 2. As for penalty factor s, it convergence speed of GNNA, and INNA is used to opti-
should not be too big to avoid disturbing the performance mize the worst half of population to make it avoid getting
advantages of the basic algorithm and is set to 0.01 in our trapped into local optimum to some extent; (2) after the
experiment. Figure 5 presents the reinforcement learning- individuals are optimized by EGWO and INNA, they form
based EGWO schematic. In Fig. 5, the agent (EGWO) a new population; (3) INNA and EGWO share the current
received the state (the current fitness value of one indi- global optimal solution to accelerate the convergence speed
vidual at time t þ 1 is superior, inferior or equal to its of GNNA. Figure 7 presents the flowchart of GNNA and a
fitness value at time t) and a reward associated with the last step-wise procedure for execution of GNNA is given in the
sate transition (s g fitþ1 fit ) at time t. Then EGWO will following.
take an action (keeping, increasing or decreasing ati ) to
Step 1 Initialization information.
make a transition to new state, and the cycle will be
Initialize parameters. These parameters include the
repeated at time t þ 1. Figure 6 shows the pseudo code of
maximum number of function evaluations Tmax , popula-
the EGWO.
tion size N, the lower bounds of design variables l, the
upper bounds of design variables u, dimension of
problem D and fitness function f ðÞ. In addition, the
current of number iteration time t is 0, the current
number of function evaluations Tcurrent is set to 0, the
penalty factor s is set to 0.01, and the constant w is equal
to 2.
(1:1) Initialize population. A random population X t is
generated based on the initialization parameters,
which can be described as
123
Neural Computing and Applications
Start
Initialize parameters
No
Generate the new population by Eqs. (2) and (3) Calculate position vector of each grey wolf by Eq. (15-18)
Calculate the fitness value of population Update the exploration rate by Eqs. (25), (26) and (27)
The new population obtained by INNA The new population obtained by EGWO
Select the optimum solution and update the current number of function
evaluations by Eq. (29)
123
Neural Computing and Applications
where t is a random value between 0 and 1 and E 4 Numerical experiment and analysis
is equal to 0:5 N.
Step 2 Population evaluation. In this section, 23 well-known complex numerical opti-
The fitness values of individuals are calculated and the mization problems are used to examine the performance of
optimal solution Gt is selected. the proposed GNNA. This section is divided into the fol-
Step 3 Update the number of function evaluations. lowing three parts. Section 4.1 introduces the experiment
The current number of function evaluations Tcurrent is settings including benchmark test functions, parameters
updated by settings and evaluation metrics. Section 4.2 is to compare
the performance between GNNA and variants of NNA and
Tcurrent ¼ Tcurrent þ N ð29Þ GWO. The performance comparisons between GNNA and
Step 4 Termination criteria. several state-of-the-art metaheuristic algorithms are pre-
If the current number of function evaluations Tcurrent is sented in Sect. 4.3.
greater than the maximum number of function evalua-
tions Tmax , the algorithm stops; otherwise, go to Step 5. 4.1 Experiment settings
Step 5 Population optimization.
4.1.1 Benchmark test functions
(5:1) The population is divided into two halves based
on the fitness values of individuals. The best half The selected 23 well-known test functions have been listed
of population and the worst half of population are in Table 1, which have been employed to evaluate the
marked as X tbest and Xtworst , respectively. optimization performance of different algorithms
(5:2) Perform INNA for the population Xtworst . Firstly, [2, 28, 31, 33–35]. More specifically, these functions have
the worst solution of the population X tworst is different features including unimodal (F1–F4), multimodal
replaced with the optimal solution Gt . Secondly, (F5–F19) and rotated multimodal (F20–F23). In general,
target solution xtopt and the target weight wtopt are multimodal functions are more complicated than unimodal
selected. Thirdly, the population Xtworst and the functions, which have more than one local minimum.
weight matrix W t are updated according to Nineteen of 23 functions are multimodal functions as
Eqs. (2–3) and (4), respectively. Fourthly, for shown in Table 1, which provides a better examination of
123
Neural Computing and Applications
Table 1 Twenty-three benchmark test functions. ‘‘U’’, ‘‘M’’ and ‘‘R’’ stand for unimodal, multi-model and rotated, respectively
No. Name Definition Range Type
PD
F1 Sphere f ð xÞ ¼ 2
i¼1 xi
½100; 100 U
PD 2
F2 Step f ð xÞ ¼ i¼1 ðbxi þ 0:5cÞ ½100; 100 U
PD
F3 Noise f ð xÞ ¼ i¼1: i x4i
þ random½0; 1Þ ½1:28; 0:64 U
PD QD
F4 Schwefel 2.22 f ð xÞ ¼ i¼1: jxi jþ i¼1 jxi j
½10; 10 U
F5 Schwefel 2.21 f ð xÞ ¼ maxi fjxi j; 1 i Dg ½100; 100 M
PD 2 QD
F6 Griewank 1
f ð xÞ ¼ 4000 i¼1 xi
xiffi
i¼1 cos i þ 1
p ½600; 600 M
P
F7 Alpine f ð xÞ ¼ D i¼1 jxi sinðxi Þ þ 0:1xi j
½10; 10 M
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
P PD
F8 Ackley ½32; 32 M
f ð xÞ ¼ 20 exp 0:2 D1 D i¼1 xi
2 exp 1
D i¼1 cos 2pxi þ 20 þ e
PD
F9 Rastrigin f ð xÞ ¼ 2
i¼1 xi 10 cosð2pxi Þ þ 10 ½5:12; 5:12 M
qffiffiffiffiffiffiffiffiffiffiffiffiffi
P
F10 Schaffer sin2
D
x2 0:5
½100; 100 M
i¼1 i
f ð xÞ ¼ 0:5þ PD 2
1þ0:001 x2
i¼1 i
qffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffiffi
PD 2ffi P
F20 Rotated Ackley ½32; 32 R, M
f ð xÞ ¼ 20 exp 0:2 1
D i¼1 zi exp D1 D
i¼1 cos 2pzi þ 20 þ e; z ¼ x M
PD
F21 Rotated Rastrigin f ð xÞ ¼ 2
i¼1 zi 10 cosð2pzi Þ þ 10; z ¼ x M ½5:12; 5:12 R, M
PD 2 QD
F22 Rotated Griewank f ð xÞ ¼ 1
i¼1 xi
xiffi
i¼1 cos i þ 1; z ¼ x
p M ½600; 600 R, M
4000
(
F23 Rotated noncontinuous PD yi ; jyi j\0:5 ½5:12; 5:12 R, M
2
Rastrigin f ð xÞ ¼ i¼1 zi 10 cosð2pzi Þ þ 10; zi ¼ ;y ¼ x M
0:5ðroundð2yi ÞÞ; jyi j 0:5
optimization performance of the proposed method in grey wolf optimizer (IGWO) [25]. The parameters of the
solving complex optimization problems. In addition,M is compared algorithms were taken from the original litera-
an orthogonal matrix in Table 1. All test functions should ture and have been shown in Table 2. Moreover, some
be minimized, which have a global minimum at f ðxÞ ¼ 0. parameters of GWO, SSA, SCA, WOA, IGWO and
Besides, the test dimension of all functions is set to 30. AGWO are related to the maximum number of function
evaluations, which are not listed in Table 2 and can be
4.1.2 Parameters settings found in the corresponding references.
In addition, considering the results of a single run might
In order to better show the efficiency and robustness of the be unreliable, all algorithms applied were executed in 30
proposed algorithm in solving the complex optimization independent runs for the same test function. Then the
problems, 13 optimizers are selected to compare with average values of the obtained results were recorded. The
GNNA. These algorithms include DE, CS, GWO, SCA, maximum number of function evaluations was considered
NNA, TLBO, WOA, SSA, PSO, INNA, EGWO, aug- as stopping condition, which was 5000 multiples by
mented grey wolf optimizer (AGWO) [24] and improved dimension size for all functions [17].
123
Neural Computing and Applications
4.1.3 Evaluation metrics rank test with a significance level a = 0.05 are produced by
using the mean optimal solutions of 30 runs of GNNA and
To compare the optimization performance among different the compared algorithms to solve benchmark test functions.
algorithms, three quality indicators are used in our exper- Tables 4 and 6 present the test results. In these tables, ‘‘H’’
iments, which can be described as follows. is marked as ‘‘1’’, which means there is significant differ-
The first one is the value-based method. Mean value and ence between GNNA and the compared algorithm; ‘‘H’’ is
standard deviation are good indicators to evaluate accuracy marked as ‘‘0’’, which indicates there is no significant
and stability of one algorithm. The smaller the mean value difference between GNNA and the compared algorithm.
is, the stronger the global optimization ability of the ‘‘S’’ is marked as ‘‘?’’, which imply the proposed method
algorithm is; the smaller the standard deviation is, the more is superior to the compared algorithm; ‘‘S’’ is marked as
stability the algorithm is. Tables 3 and 5 show the statis- ‘‘=’’, which indicates the proposed method has the same
tical results of mean value and standard deviation. In these performance with the compared algorithm. Moreover, the
tables, ‘‘ME’’, ‘‘Std’’ and ‘‘TR’’ stand for mean value, stand last column of the each of these tables under the heading
deviation and tied rank, respectively. Moreover, the best w/t/l represents the win, tie and lose counts of the proposed
results are highlighted in bold. GNNA over the compared algorithms.
The second one is the rank-based method. Tied rank
(TR) [31, 34] is employed to compare the performance 4.2 GNNA comparison with variants of GWO
among the applied algorithms. More specifically, TR and NNA
assigns rank 1 to the algorithm with the best mean value;
rank 2 to the second best and rank M (the number of the GNNA is inspired from basic GWO and NNA. Thus it is
applied algorithms) to the Mth best. If several algorithms necessary to compare the optimization performance among
have the same mean values, they will share the average of GNNA, NNA, GWO, and some recent variants of NNA
ranks. Tables 3 and 5 show the results of tied rank. and GWO. In this section, the applied algorithms include
Moreover, the last column of the each of these tables under GNNA, NNA, GWO, INNA, EGWO, IGWO and AGWO.
the heading average rank represents the mean tied rank of The experimental results obtained by the applied algo-
every algorithm for 23 benchmark test functions. rithms on 23 benchmark test functions with 30-dimensional
The third one is the statistical test-based method. Wil- are presented in Table 3, and the best results are high-
coxon signed-rank test [37] is used to compare the per- lighted in boldface. Figure 8 gives the statistical results
formance between the proposed GNNA and the other based on Table 3. In Fig. 8, ‘‘Win’’ means the number of
algorithms, which has been widely used to compare the functions that GNNA can offer better mean values than the
performance between two optimization algorithms compared algorithm; ‘‘Tie’’ indicates the number of func-
[38–41]. In this research, the results of Wilcoxon signed- tions that GNNA can achieve the same mean values with
123
Neural Computing and Applications
Table 3 The experimental results obtained by seven different algorithms for 23 benchmark functions
No. Metric GWO EGWO AGWO IGWO NNA INNA GNNA
123
Neural Computing and Applications
Table 3 (continued)
No. Metric GWO EGWO AGWO IGWO NNA INNA GNNA
the compared algorithm. According to Table 3 and Fig. 8, algorithms in terms of global search ability and stability.
GNNA is superior or equal to the compared algorithms in This indicates the combination of EGWO and INNA is
terms of mean value. More specifically, INNA is inferior to very successful in solving optimization problems. How-
GNNA on all test functions. GNNA is superior to NNA on ever, we also can find an interesting phenomenon that the
all test functions except F2 and F20. GNNA outperforms proposed EGWO outperforms the basic GWO while the
IGWO on all test functions except F2, F13, F15, F17 and basic NNA surpasses the proposed INNA according to the
F20. In addition, GWO, EGWO and AGWO show out- sorted results. We can explain this phenomenon from two
standing global search ability, which can offer the same aspects as follows. Firstly, compared with the basic GWO,
results with GNNA on F1, F2, F6, F7, F9, F11, F12, F13, F15, the proposed EGWO adjusts the exploration rate according
F17, F18 and F20 while they still cannot surpass GNNA on to the reinforcement learning principles, which can better
the another 11 test functions. Besides, based on results of balance exploration and exploitation of the basic GWO.
the average rank, the applied algorithms can be sorted into The experimental results demonstrate the improved strat-
the following order from good to bad: GNNA, AGWO, egy in EGWO is helpful to enhance the global search
EGWO, GWO, IGWO, NNA and INNA. capability of the basic GWO. Secondly, the basic NNA
Table 4 gives the statistical results produced by Wil- balance its exploration and exploitation by bias operator
coxon signed-rank test. As can be seen from Table 4, and transfer operator, respectively. However, the proposed
GNNA outperforms INNA on all test functions. GNNA is INNA discards the transfer operator to make the most of
superior or equal to GWO, IGWO and NNA on all test exploration of NNA. In other words, exploration and
functions. It also may be noted that EGWO and AGWO exploitation are serious imbalance in the INNA. Thus NNA
show strong global search ability. EGWO and AGWO are can show better optimization performance than INNA.
equal to GNNA on 12 and 13 functions, respectively. In order to compare the convergence performance
However, EGWO and AGWO are still inferior to GNNA among different algorithms, Fig. 9 shows several typical
on 11 and 10 functions, respectively. averaged curves over 30 independent runs. The selected
Based on the above analysis, the proposed GNNA shows functions include 2 unimodal functions (i.e. F1 and F4), 10
better optimization performance than the compared multimodal functions (i.e. F5, F6, F7, F9, F10, F11, F12, F13,
123
Neural Computing and Applications
Table 4 Wilcoxon signed-rank test results between GNNA and six compared algorithms
No. GNNA versus
GWO EGWO AGWO IGWO NNA INNA
P value H S P value H S P value H S P value H S P value H S P value H S
123
Neural Computing and Applications
Fig. 9 Several typical convergence curves obtained by GNNA and six compared algorithms
123
Neural Computing and Applications
Table 5 The experimental results obtained by GNNA and seven state-of-the-art algorithms for 23 benchmark functions
No. Metric DE PSO CS SCA TLBO WOA SSA GNNA
123
Neural Computing and Applications
Table 5 (continued)
No. Metric DE PSO CS SCA TLBO WOA SSA GNNA
123
Table 6 Wilcoxon signed-rank test results between GNNA and seven state-of-the-art algorithms
No. GNNA versus
DE PSO CS SCA TLBO WOA SSA
P value H S P value H S P value H S P value H S P value H S P value H S P value H S
Neural Computing and Applications
123
Neural Computing and Applications
Fig. 11 Several typical convergence curves obtained by GNNA and seven state-of-the-art algorithms
123
Neural Computing and Applications
towards the global optimal solution but also offers higher References
accuracy. On the contrary, the compared algorithms con-
verge easily to the local optima in solving complex opti- 1. Savsani P, Savsani V (2016) Passing vehicle search (PVS): A
mization problems such as F5, F6, F10 and F23. In addition, novel metaheuristic algorithm. Appl Math Model 40:3951–3978.
https://doi.org/10.1016/j.apm.2015.10.040
GNNA and WOA can converge to the optimum solutions 2. Zhang J, Xiao M, Gao L, Pan Q (2018) Queuing search algo-
on F1, F7, F9, F11, F12, F13, F17 and F18, while GNNA can rithm: a novel metaheuristic algorithm for solving engineering
converge faster than WOA on these functions. optimization problems. Appl Math Model 63:464–490. https://
doi.org/10.1016/j.apm.2018.06.036
3. Rao RV, Savsani VJ, Vakharia DP (2012) Teaching–learning-
based optimization: an optimization method for continuous non-
5 Conclusions linear large scale problems. Inf Sci 183:1–15. https://doi.org/10.
1016/j.ins.2011.08.006
In order to solve the global numerical optimization problems, 4. Eskandar H, Sadollah A, Bahreininejad A, Hamdi M (2012)
Water cycle algorithm—a novel metaheuristic optimization
this paper presents an efficient hybrid method called grey wolf method for solving constrained engineering optimization prob-
optimization with neural network algorithm (GNNA). The lems. Comput Struct 110–111:151–166. https://doi.org/10.1016/j.
core idea of GNNA is to make full use of the good global compstruc.2012.07.010
search ability of NNA and fast convergence of GWO. Firstly, 5. Kennedy J, Eberhart R (1995) Particle swarm optimization. In:
Proceedings of ICNN’95—international conference on neural
NNA and GWO are improved to enhance the optimization networks, vol 4, pp 1942–1948
performance and INNA and EGWO are obtained. Then INNA 6. Yang X, Deb S (2009) Cuckoo search via Lévy flights. In: 2009
and EGWO are hybridized based on dynamic population world congress on nature biologically inspired computing
mechanism. In the optimization process, the population is (NaBIC), pp 210–214
7. Yang X-S (2010) A new metaheuristic bat-inspired algorithm. In:
divided into two halves according to the fitness values. The González JR, Pelta DA, Cruz C et al (eds) Nature inspired
best half of population is optimized by EGWO and the worst cooperative strategies for optimization (NICSO 2010). Springer,
half of population is processed by INNA. Berlin, pp 65–74
A comprehensive set of 23 well-known unconstrained 8. Mirjalili S, Mirjalili SM, Lewis A (2014) Grey wolf optimizer.
Adv Eng Softw 69:46–61. https://doi.org/10.1016/j.advengsoft.
benchmark functions are employed to examine the perfor- 2013.12.007
mance of GNNA compared with 13 metaheuristic algorithms 9. Gandomi AH, Yang X-S, Alavi AH (2011) Mixed variable
including two basic algorithms (i.e. NNA and GWO), four structural optimization using firefly algorithm. Comput Struct
variants of the basic algorithms and seven state-of-the-art 89:2325–2336. https://doi.org/10.1016/j.compstruc.2011.08.002
10. Mirjalili S, Gandomi AH, Mirjalili SZ et al (2017) Salp swarm
algorithms. According to the experimental results, the pro- algorithm: a bio-inspired optimizer for engineering design prob-
posed GNNA is superior or equal to the compared algorithms lems. Adv Eng Softw 114:163–191. https://doi.org/10.1016/j.
for all test functions in terms of mean values. Note that the advengsoft.2017.07.002
proposed GNNA can find global optimal solutions on 17 11. Mirjalili S, Lewis A (2016) The whale optimization algorithm.
Adv Eng Softw 95:51–67. https://doi.org/10.1016/j.advengsoft.
functions including three unimodal functions (i.e. F1, F2, F4), 2016.01.008
11 multimodal functions (i.e. F5, F6, F7, F9, F11, F12, F13, F15, 12. Goldberg DE, Holland JH (1988) Genetic algorithms and
F17, F18, F19) and three rotated functions (i.e. F21, F22 and machine learning. Mach Learn 3:95–99. https://doi.org/10.1023/
F23). Such comparisons suggest that GNNA shows good A:1022602019183
13. Rahnamayan S, Tizhoosh HR, Salama MMA (2008) Opposition-
global search ability and fast convergence speed and is clearly based differential evolution. IEEE Trans Evol Comput 12:64–79.
seen to be more successful in both solution quality and com- https://doi.org/10.1109/TEVC.2007.894200
putational efficiency compared with other algorithms. 14. Simon D (2008) Biogeography-based optimization. IEEE Trans
The future research will focus on the following two Evol Comput 12:702–713. https://doi.org/10.1109/TEVC.2008.
919004
aspects. On the one hand, we intend to apply GNNA to 15. Rao RV, Savsani VJ, Vakharia DP (2011) Teaching–learning-
some practical optimization problems such as multilevel based optimization: a novel method for constrained mechanical
image segmentation and economic dispatch problems. On design optimization problems. Comput Aided Des 43:303–315.
the other hand, we try to improve the performance of other https://doi.org/10.1016/j.cad.2010.12.015
16. Mirjalili S (2016) SCA: a sine cosine algorithm for solving
metaheuristic algorithms by the proposed reinforcement optimization problems. Knowl Based Syst 96:120–133. https://
learning mechanism. doi.org/10.1016/j.knosys.2015.12.022
17. Sadollah A, Sayyaadi H, Yadav A (2018) A dynamic meta-
heuristic optimization model inspired by biological nervous
systems: neural network algorithm. Appl Soft Comput
Compliance with ethical standards 71:747–782. https://doi.org/10.1016/j.asoc.2018.07.039
18. Kirkpatrick S, Gelatt CD, Vecchi MP (1983) Optimization by
simulated annealing. Science 220:671–680. https://doi.org/10.
Conflict of interest The authors declare that they have no conflict of
1126/science.220.4598.671
interest.
123
Neural Computing and Applications
19. Wolpert DH, Macready WG (1997) No free lunch theorems for 32. Kaelbling LP, Littman ML, Moore AP (1996) Reinforcement
optimization. IEEE Trans Evol Comput 1:67–82. https://doi.org/ learning: a survey. J Artif Intell Res 4:237–285
10.1109/4235.585893 33. Salgotra R, Singh U, Saha S (2018) New cuckoo search algo-
20. Zang H, Zhang S, Hapeshi K (2010) A review of nature-inspired rithms with enhanced exploration and exploitation properties.
algorithms. J Bionic Eng 7:S232–S237. https://doi.org/10.1016/ Expert Syst Appl 95:384–420. https://doi.org/10.1016/j.eswa.
S1672-6529(09)60240-7 2017.11.044
21. Garg H (2019) A hybrid GSA-GA algorithm for constrained 34. Sun Y, Wang X, Chen Y, Liu Z (2018) A modified whale opti-
optimization problems. Inf Sci 478:499–523. https://doi.org/10. mization algorithm for large-scale global optimization problems.
1016/j.ins.2018.11.041 Expert Syst Appl 114:563–577. https://doi.org/10.1016/j.eswa.
22. Xiong G, Zhang J, Yuan X et al (2018) Parameter extraction of 2018.08.027
solar photovoltaic models by means of a hybrid differential 35. Long W, Jiao J, Liang X, Tang M (2018) An exploration-en-
evolution with whale optimization algorithm. Sol Energy hanced grey wolf optimizer to solve high-dimensional numerical
176:742–761. https://doi.org/10.1016/j.solener.2018.10.050 optimization. Eng Appl Artif Intell 68:63–80. https://doi.org/10.
23. Le DT, Bui D-K, Ngo TD et al (2019) A novel hybrid method 1016/j.engappai.2017.10.024
combining electromagnetism-like mechanism and firefly algo- 36. Wang H, Wu Z, Rahnamayan S et al (2011) Enhancing particle
rithms for constrained design optimization of discrete truss swarm optimization using generalized opposition-based learning.
structures. Comput Struct 212:20–42. https://doi.org/10.1016/j. Spec Issue Interpret Fuzzy Syst 181:4699–4714. https://doi.org/
compstruc.2018.10.017 10.1016/j.ins.2011.03.016
24. Qais MH, Hasanien HM, Alghuwainem S (2018) Augmented 37. Derrac J, Garcı́a S, Molina D, Herrera F (2011) A practical
grey wolf optimizer for grid-connected PMSG-based wind energy tutorial on the use of nonparametric statistical tests as a
conversion systems. Appl Soft Comput 69:504–515. https://doi. methodology for comparing evolutionary and swarm intelligence
org/10.1016/j.asoc.2018.05.006 algorithms. Swarm Evol Comput 1:3–18. https://doi.org/10.1016/
25. Long W, Liang X, Cai S et al (2017) A modified augmented j.swevo.2011.02.002
Lagrangian with improved grey wolf optimization to constrained 38. Mafarja M, Aljarah I, Heidari AA et al (2018) Binary dragonfly
optimization problems. Neural Comput Appl 28:421–438. https:// optimization for feature selection using time-varying transfer
doi.org/10.1007/s00521-016-2357-x functions. Knowl Based Syst 161:185–204. https://doi.org/10.
26. Khairuzzaman AKM, Chaudhury S (2017) Multilevel threshold- 1016/j.knosys.2018.08.003
ing using grey wolf optimizer for image segmentation. Expert 39. Sun G, Ma P, Ren J et al (2018) A stability constrained adaptive
Syst Appl 86:64–76. https://doi.org/10.1016/j.eswa.2017.04.029 alpha for gravitational search algorithm. Knowl Based Syst
27. Sahoo A, Chandra S (2017) Multi-objective grey wolf optimizer 139:200–213. https://doi.org/10.1016/j.knosys.2017.10.018
for improved cervix lesion classification. Appl Soft Comput 40. Martı́nez-Peñaloza M-G, Mezura-Montes E (2018) Immune
52:64–80. https://doi.org/10.1016/j.asoc.2016.12.022 generalized differential evolution for dynamic multi-objective
28. Zhang X, Kang Q, Cheng J, Wang X (2018) A novel hybrid environments: an empirical study. Knowl Based Syst
algorithm based on biogeography-based optimization and grey 142:192–219. https://doi.org/10.1016/j.knosys.2017.11.037
wolf optimizer. Appl Soft Comput 67:197–214. https://doi.org/ 41. Yi J, Gao L, Li X et al (2019) An on-line variable-fidelity sur-
10.1016/j.asoc.2018.02.049 rogate-assisted harmony search algorithm with multi-level
29. Lu C, Gao L, Li X, Xiao S (2017) A hybrid multi-objective grey screening strategy for expensive engineering design optimization.
wolf optimizer for dynamic scheduling in a real-world welding Knowl Based Syst 170:1–19. https://doi.org/10.1016/j.knosys.
industry. Eng Appl Artif Intell 57:61–79 2019.01.004
30. Emary E, Zawbaa HM, Grosan C (2018) Experienced gray wolf
optimization through reinforcement learning and neural net- Publisher’s Note Springer Nature remains neutral with regard to
works. IEEE Trans Neural Netw Learn Syst 29:681–694. https:// jurisdictional claims in published maps and institutional affiliations.
doi.org/10.1109/TNNLS.2016.2634548
31. Rakhshani H, Rahati A (2017) Snap-drift cuckoo search: a novel
cuckoo search optimization algorithm. Appl Soft Comput
52:771–794. https://doi.org/10.1016/j.asoc.2016.09.048
123