A_Fast_Dynamic_Evolutionary_Multiobjective_Algorithm_via_Manifold_Transfer_Learning
A_Fast_Dynamic_Evolutionary_Multiobjective_Algorithm_via_Manifold_Transfer_Learning
still suffers from two significant defects that prevent its Definition 2 [Dynamic Pareto-Optimal Set (DPOS)]: Both
application in real-world tasks. First, the TL technique [7] x and x∗ are decision vectors, and if a decision vector x∗ is
solves the Non-IID problem by searching for a latent space, said to be nondominated at time t if and only if there is no
and the determination of the optimal latent space involves other decision vector x such that x t x∗ at time t. The DPOS
choosing the optimal values of numerous hyperparameters. is the set of all Pareto-optimal solutions at time t, that is
This process of parameter tuning usually requires trial and DPOS = x∗ | ∃x, x t x∗ .
error and often takes a significant amount of time. Second,
even if the optimal latent space is found, Tr-DMOEA still Definition 3 [Dynamic Pareto-Optimal Front (DPOF)]: At
needs to invoke the optimization algorithm to produce the time t, the DPOF is the corresponding objective vectors of the
initial population, which also consumes excessive computing DPOS
resources. DPOF = F x∗ , t | x∗ ∈ DPOS .
The motivation of this article is to accelerate DMOPs-
The purpose of a DMOA is to produce a set of solutions as
solving without deteriorating the solution quality. We propose
close as possible to the changing POF while maximizing the
a method for solving DMOPs by combining a memory mech-
diversity of the solution.
anism with manifold TL. To the best of our knowledge, this
is the first attempt to explore an integrative approach of
B. Manifold-Based Transfer Learning
both the techniques. These two approaches complement each
other and this integration provides two benefits. First, the 1) Manifold and Geodesic: A manifold is a topological
proposed method compresses high-dimensional data in a low- space that is locally homeomorphic to a Euclidean space. In
dimensional space with a special TL technique and it requires other words, at any point of a d-dimensional manifold, it is
a few hyperparameters. Second, computing resources are spent locally homeomorphic to the Euclidean space Rd . For exam-
on knowledge transfer of elite individuals via the memory ple, the surface of the Earth as a 2-D manifold, and any part
mechanism, which greatly improves the speed of finding the of it can be considered as a 2-D plane. A geodesic is a smooth
solutions in evolutionary optimization. curve on the manifold, that is locally the shortest curve con-
The remainder of this article is organized as follows. necting two points. A geodesic can be considered as a “bent”
Section II introduces the fundamental concepts of dynamic straight line on a manifold. Calculating the distance between
optimization problems and discusses the existing works. two points on the manifold is to calculate the geodesic dis-
Section III presents background on manifold TL and explains tance. For more details about manifold learning and geodesic,
the details of our method memory-driven manifold TL- please refer to [8].
based EA for dynamic multiobjective optimization (MMTL- 2) Sample Geodesic Flow: The manifold-based TL algo-
DMOEA). In Section IV, we present the experimental results rithm used in this article is sample geodesic flow (SGF) [9].
of the proposed algorithm on different test functions and The SGF algorithm is inspired by incremental learning: a
compare with state-of-the-art algorithms. Section V draws a learning system can continuously learn from the samples and
summary of this article and outlines future research directions. keep most knowledge of them. In order to apply the knowl-
edge learned from the source domain to the target domain,
the SGF algorithm constructs a geodesic path between the
II. P RELIMINARIES AND R ELATED W ORKS two domains and then transports the knowledge through this
A. Dynamic Multiobjective Optimization Problems path to the target domain. In other words, the source and target
The essential characteristic of DMOPs is that the objectives domains are mapped to the starting and ending points, and the
change over time or under different environments. Without algorithm constructs a geodesic flow between the two points
loss of generality, the general form of DMOPs is defined as in the manifold. After that, it will select k intermediate points
on the geodesic flow, and finally, transform the data of the
min F(x, t) = <f1 (x, t), f2 (x, t), . . . , fm (x, t)> source and target domains to these intermediate points.
s.t. x ∈ X (1) Fig. 1 presents an illustrative diagram of the SGF method.
First, the algorithm maps the source and target domains to
where x = x1 , x2 , . . . , xn is the n-dimensional decision the two points S0 and S1 on a Grassmann manifold G(d, n),
variable, and X ∈ Rn is the decision space. t is the time and the algorithm will construct a geodesic flow from S0 to
(environment) variable, f1 , f2 , . . . , fm are objective functions, S1 (shown as a dotted line in the figure). Next, it will select
and m is the number of objectives. k intermediate points, the red dot S0.2 , S0.4 , S0.6 , S0.8 on the
A set of solutions is defined as a dynamic Pareto-optimal set geodesic flow, and transform the training and testing data to
(POS) if this solution set exists in a DMOP at a given time t, these intermediate subspaces. The data onto these subspaces
and the decision vector in the set cannot be further optimized will be analyzed to perform classification.
in any dimension.
Definition 1 (Dynamic Decision Vector Domination): At C. Related Works
time t, a decision vector x1 Pareto dominates another vector x2 ,
denoted by x1 t x2 , if and only if Existing DMOAs can be divided into the following
categories: maintaining diversity methods, multipopulation-
∀i = 1, . . . , m, fi (x1 , t) ≤ fi (x2 , t) based methods, memory-based methods, and prediction-based
(2)
∃i = 1, . . . , m, fi (x1 , t) < fi (x2 , t). methods.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
JIANG et al.: FAST DYNAMIC EVOLUTIONARY MULTIOBJECTIVE ALGORITHM VIA MANIFOLD TL 3419
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
3420 IEEE TRANSACTIONS ON CYBERNETICS, VOL. 51, NO. 7, JULY 2021
the wrong direction, which means actual results will be worse Process 1 MMTL-DMOEA
than a method that does not use the prediction technique. Input: The dynamic optimization function: F(x, t); the size of
TL is a powerful weapon that solves the dilemma caused by the population: N; an external storage P and its capacity
the Non-IID hypothesis, and recent research has also proved C; the number of clusters: L.
that transferring acquired knowledge in the optimization pro- Output: The Solutionst at time t.
cess is promising [28], especially for solving dynamic [6] 1: Initialization;
or multiple optimization tasks simultaneously [29]–[31]. 2: Solutions0 = SMOA(POP0 , F(x, 0), N);
Jiang et al. [6] proposed a framework called Tr-DMOEA for 3: Store Solutions0 into P;
solving DMOPs. This framework integrates between TL and 4: while Change detected do
classical evolutionary multiobjective optimization algorithms. 5: t = t + 1;
TL effectively generates a high-quality initial population pool 6: LastBestSol = FindBestSol(P, F(x, t), N);
via reusing past experience to speed up the evolutionary 7: TransSol = Transfer(LastBestSol, F(x, t), N, L);
process. 8: POPt = LastBestSol ∪ TransSol;
Bali et al. [31] proposed an online transfer parameter 9: (Solutionst ) = SMOA(POPt , F(x, t), N);
estimation approach for the multitasking EA. The proposed 10: if size(P ∪ Solutionst ) ≤ C then
algorithm evaluates the similarities between different task 11: Store Solutionst into P;
instances, therefore, the extent of knowledge transfer can be 12: else
adapted based on the optimal mixture of probabilistic mod- 13: Replace the earliest stored individuals with
els. Da et al. [32] demonstrated a framework for black-box Solutionst ;
transfer optimization. The framework measures the similari- 14: end if
ties between the source instance and the target instance and 15: end while
reveals latent synergies during optimizing. When faced with 16: return Solutionst ;
multiple sources, the positive and negative transfer is automat-
ically identified so as to curb the negative transfer, thereby the
excellent performance is kept. of this method is to accelerate the process of solution search-
Inspired by the recent autoencoding evolutionary ing while ensuring the solution quality. The main contribution
search [29], Zhou et al. [33] proposed an improved of MMTL-DMOEA is the integration of manifold TL with
adaptive indicator-based EA. The proposed algorithm is the elite memory mechanism, to produce an initial population
capable of transferring the past search information by for the next generation. This initial population can help any
adapting some obtained solutions with high hypervolume population-based multiobjective optimization algorithms find
value. Min et al. [34] designed a transfer stacking of the the POS of the optimization problem faster.
Gaussian process surrogate models by reusing the acquired Fig. 2 shows the algorithm flow of MMTL-DMOEA. This
knowledge. The adaptive knowledge gained from expensive algorithm first stores the optimal solutions obtained at differ-
problem-solving information is introduced to improve the ent times in external memory. Once the environment changes
performance of the optimization process of the target task. are detected, the algorithm selects a certain number of indi-
Liu et al. [35] proposed a neural-network-based information viduals from external storage according to predefined criteria.
transfer method to reuse past solutions. When the environment In the third step, the algorithm predicts individuals through
changes, the proposed algorithm collects the solutions from the principal component analysis (PCA) and the manifold TL
both the previous environment and the new environment. algorithm. After generating the predicted individuals, the elite
Then, the neural-network transfers acquired solutions for individuals selected from the memory are merged with them
guiding the search in the new environment. to constitute an initial population.
Introducing TL into the field of dynamic optimization The detailed steps of MMTL-DMOEA are shown in
shows its potential. However, at the same time, some new Process 1. It is worth pointing out that any population-based
problems arise. For example, Tr-DMOEA often leads to static multiobjective algorithm (SMOA) can be used as a rou-
diminishing diversity in solutions. Researchers proposed a tine for processing static multiobjective optimization problems
dynamic multiobjective estimation of distribution algorithm, in our proposed algorithm.
called DANE-EDA [36]. This design combines the Monte Remark 1: The parameter C refers to the maximum number
Carlo method with the TL technique, and this combination of individuals that can be stored in the external memory.
maintains a tradeoff of exploration–exploitation from temporal Remark 2: When the external memory overflows, the algo-
and spatial perspectives. Although researchers made progress rithm replaces the earliest stored individuals with the newly
in the DMOP domain, existing methods are generally time generated Solutionst .
consuming and not suitable for some applications in the real Our proposed algorithm follows the elite reservation mech-
world. anism. In other words, we will preserve the optimal solutions
obtained at previous moments. Although solutions at different
times are varying, there should be some inherent relationship
III. MMTL-DMOEA A LGORITHM between these solutions. Therefore, it is possible to obtain a
This article proposes the method of memory-driven mani- better initial population by reserving the best individuals and
fold TL (MMTL-DMOEA) to address DMOPs. The purpose making a reasonable transformation of these individuals.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
JIANG et al.: FAST DYNAMIC EVOLUTIONARY MULTIOBJECTIVE ALGORITHM VIA MANIFOLD TL 3421
Fig. 2. Main idea of MMTL-DMOEA. Step 1: MMTL-DMOEA will preserve the nondominated solutions obtained from time 0 to time t. Step 2: Some
individuals are selected from the stored elite set according to the predefined mechanism. Step 3: The selected individuals are mapped to the predicted individuals
by using the manifold-based TL method. Step 4: MMTL-DMOEA generates an initial population by merging predicted and selected individuals.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
3422 IEEE TRANSACTIONS ON CYBERNETICS, VOL. 51, NO. 7, JULY 2021
Process 3 Transfer constructing the geodesic flow spends O(d2 ) and O(1), respec-
Input: The elite individuals: LastBestSol; tively. The computational cost of finding solutions by using
The dynamic optimization function: F(x, t); the interior-point algorithm consumes O(m3 n) [6]. Because
Number of individuals: N; the number of clusters is a constant set by the user, therefore,
Clustering number: L. the computational complexity of Transfer is O(m3 n). The total
Output: Predicted individuals: TransSol. computational complexity of the proposed prediction is thus
1: Initialization: dimension d = m − 1; TransSol = ∅; O(m3 n).
2: Clustering LastBestSol: LastBestSol1 , ..., LastBestSolL by
LPCA; IV. E XPERIMENTS
3: for j = 1 to L do
In order to verify the effectiveness of MMTL-DMOEA, we
4: Use PCA for LastBestSolj to get PS ;
carried out three sets of experiments. The first experiment
5: Generate set T containing N individuals of F(x, t);
compares the proposed MMTL-DMOEA with the state-of-the-
6: Use PCA for T to get PT ∈ Rd×N ;
art DMOAs. The second experiment verifies the effectiveness
7: Construct the geodesic flow φ(k) = PS U1 (k) −
of integrating the components of memory and manifold TL.
RS U2 (k), k ∈ (0, 1);
The third experiment compares with Tr-DMOEA [6] and
8: for x ∈ LastBestSolj do
analyzes performance improvement.
9: Project x to φ(·) and get x;
10: x = arg minx xT φ(·) − x;
11: TransSol = x ∪ TransSol; A. Comparison With Other DMOPs
12: end for The test functions used are FDA series [38], dMOP
13: end for series [14], and DF series [39]. The time parameter in the
14: return TransSol; test functions is t = (1/nt )τ/τt , where nt , τt , and τ are the
severity of the time change, the frequency of change, and the
maximum generation of the problems, respectively.
The algorithms used for comparison in this experiment are
The detailed steps for Transfer are shown in Process 3.
as follows: MOEA/D-SVR [40], Tr-DMOEA [6], MOEA/D-
We believe that the estimated nondominated solutions
KF [23], PPS [22], and the baseline algorithm MOEA/D [24],
LastBestSol can be viewed as m − 1-dimensional seg-
which are modified as RI-MOEA/D to adapt to dynamic
mented manifolds [37], where m is the number of objec-
change, that is, 10% of the population is randomly reinitial-
tives. Hence, first, the LastBestSol is clustered by LocalPCA
ized when the environment changes. For a fair comparison,
(LPCA) [37] into m − 1-dimensional L segmented manifolds:
the baseline algorithm in all the compared algorithms are
LastBestSol1 , . . . , LastBestSolL in line 2. For each cluster,
replaced by MOEA/D [24], and these compared algorithms
the geodesic flow φ(·) are constructed in lines 4–7. Then,
are denoted as KT-MOEA/D, SVR-MOEA/D, Tr-MOEA/D,
x ∈ LastBestSolj is mapped into the geodesic flow in line 9,
KF-MOEA/D, and PPS-MOEA/D, respectively, and most of
and we can obtain the mapped data x. In the dynamic environ-
the parameters of these algorithms are set according to their
ment, the manifolds LastBestSolj in the source domain may
original references.
be similar to those in the target domain, so the mapped tar-
There are three metrics used in this group of comparison
get domain data xT φ(·) is similar to x. Therefore, in line 10,
experiments, and three metrics are listed as follows.
we find a solution x , such that the mapped data xT φ(·) on
Inverted generational distance (IGD) measures the conver-
the geodesic flow is closet to x. This is a single-objective
gence and diversity of the solutions. A smaller value of IGD
optimization problem, and any single-objective optimization
indicates a better convergence of the solution obtained by
algorithm can be applied to solve the problem. In this article,
the algorithm and the higher the diversity. This metric is
we use the interior-point method to solve the problem.
computed as
1
A. Computational Complexity IGD = min p − p∗2 (5)
nPt t
p∗ ∈P∗
p∈P
In the FindBestSol, constructing the SVR estimator with ns
samples needs O(n2s d), where d is the dimension of decision where represents the true POF of the problem, P∗ represents
Pt
variables. Using SVR to estimate the objective values con- the obtained POF by testing algorithms, and nPt represents the
sumes O(N 2 d), where N is the population size. According to number of solutions in the true POF.
the estimated objective values, the cost of finding nondomi- The MIGD metric is a variant of IGD and is defined as the
nated solutions by the fast nondominated sorting is O(N 2 m), mean IGD values in time steps
where m is the number of objectives. Because N ns 1
and d m, therefore, the computational complexity of MIGD POF∗t , POFt = IGD POF∗t , POFt (6)
|T|
FindBestSol is O(N 2 d). t∈T
The computational cost of Transfer involves clustering, con- where T is a set of discrete-time points in a run and |T| is the
structing the geodesic flow, and finding solutions by using cardinality of T.
the interior-point algorithm. Clustering LastBestSol by LPCA Schott’s spacing metric (SP) measures the solution unifor-
calls for O(d2 ). For each cluster, mapping data via PCA and mity. A smaller SP value reflects a more uniform distribution
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
JIANG et al.: FAST DYNAMIC EVOLUTIONARY MULTIOBJECTIVE ALGORITHM VIA MANIFOLD TL 3423
TABLE I
M EAN AND S TANDARD D EVIATION VALUES OF MIGD M ETRIC O BTAINED BY C OMPARED A LGORITHMS FOR D IFFERENT DYNAMIC T EST S ETTINGS
of the solution obtained by the algorithm. This metric is these tables are the best metrics obtained by the compared
calculated as algorithm.
The experimental results of Table I show that the MMTL-
nPOF∗
1 2 MOEA/D algorithm has achieved the best results in 20 out of
SP = Di − D (7)
nPOF − 1
∗ 24 test cases based on the MIGD values, indicating that the
i=1
optimal solution obtained by MMTL-MOEA/D is better than
where Di represents the Euclidean distance between the ith the existing algorithm in terms of convergence. Fig. 3 plots the
individual in POF∗ and its closest solution, and D represents IGD values obtained by different algorithms after each change.
the mean of Di . This metric is similarly modified to the IGD It shows that the curves obtained by the proposed method are
to act as a performance metric for evaluating DMOAs. at the bottom in most cases, and the curves of the proposed
Maximum spread (MS) measures the coverage extent of method are smoother, which means that the method not only
obtained solutions in the true POF. A larger MS value indi- performs better but also is more stable. Table II presents the
cates a higher coverage by the obtained solution in the real average SP metric values on the FDA and dMOP series of
POF. This metric is calculated as test functions. The results show that MMTL-MOEA/D is far
⎤2 better than the other five algorithms in 15 out of 24 test cases.
⎡
∗ ∗
1 ⎣ min Pk , Pk − max Pk , Pk ⎦
M t t This good performance shows that the distribution of solutions
MS = obtained by MMTL-MOEA/D is relatively uniform. Table III
M
k=1 Ptk − Ptk shows the average MS of the solutions obtained for the six
algorithms. The results show that the MMTL-MOEA/D has a
where Ptk and Ptk represent the maximum and minimum of the higher degree of coverage of the true POF than the other five
kth objective in true POF, respectively, and P∗k and P∗k rep- algorithms on most of the test cases.
resent the maximum and minimum of the kth objective, in To see whether the proposed prediction model could be
the obtained POF, respectively. This metric is similarly modi- integrated into other MOEAs, we also replace the baseline
fied to the IGD to act as a performance metric for evaluating SMOA by RM-MEDA [37] and these compared algorithms
DMOAs. are denoted as MMTL-RM-MEDA, SVR-RM-MEDA, Tr-RM-
The experiment parameters are set as follows. The size of MEDA, KF-RM-MEDA, and PPS-RM-MEDA, respectively.
population N is 100; and the external storage size C is set to The experimental results are presented in Tables I and II in the
10 × N. Most of the parameters of the SVR in FindBsetSol supplementary material. From these tables, we can obverse that
are set by default [41]; the number of sampling individuals ns our proposed model is still competitive in terms of MIGD, SP,
in FindBsetSol is 30; the number of manifold segments L in and MS metric values. We also plot the IGD values obtained
Transfer is 4; and the number of intermediate subspaces p in by different algorithms in Fig. 1 in the supplementary material.
Transfer is 5. We conducted experiments on the DF series test functions
The experimental results are shown in Tables I–III. The data under the configuration of nt = 10 and τt = 10. All DF
in the table are the MIGD, SP, and MS values of compared functions have a changing POS, which belong to type I (POS
algorithms under different configurations. The bold data in changes, but POF does not change) or type II (POS and POF
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
3424 IEEE TRANSACTIONS ON CYBERNETICS, VOL. 51, NO. 7, JULY 2021
TABLE II
M EAN AND S TANDARD D EVIATION VALUES OF SP M ETRIC O BTAINED BY C OMPARED A LGORITHMS FOR D IFFERENT DYNAMIC T EST S ETTINGS
change as well) DMOPs and meet the Non-IID problems. B. Ablation Study
The statistical results of MIGD, MS, and SP are shown in The main contribution of MMTL-DMOEA is to use the
Table IV in the supplementary material. It can be observed optimal solution information obtained from the previous
that in 20 out of 42 cases, MMTL-MOEA/D performs the best moment to generate the initial population at the next moment.
among all the algorithms, eight for RI-MOEA/D, six for KF- Specifically, our method produces the initial population by
MOEA/D, four for SVR-MOEA/D, three for Tr-MOEA/D, and two different mechanisms: 1) memory-based one and 2) TL-
only one for PPS-MOEA/D. The comparative studies on DF based one. In order to verify the effectiveness of these two
test functions can also verify the effectiveness of our proposed mechanisms, we designed an ablation experiment.
MMTL-DMOEA in solving Non-IID problems. In this experiment, we propose two modified algorithms
From the experimental results, it is obvious that MMTL- by using only one out of the two mechanisms mentioned
MOEA/D performs not very well on the FDA5 problem. above. For the cases in which the memory or the TL mecha-
FDA5 is triobjective problem and the POS and the POF nism alone generates all individuals of the initial population,
change as well, so FDA5 has more complicated POF and we respectively refer the case as MMTL-MOEA/DM or
POS. Hence, clustering solutions FDA5 into low-dimensional MMTL-MOEA/DT . During the operation of the algo-
segmented manifolds meets challenges and leads to a poor rithm, if the number of the predicted initial individuals
performance of Transfer. less than the population size, the MMTL-MOEA/DM and
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
JIANG et al.: FAST DYNAMIC EVOLUTIONARY MULTIOBJECTIVE ALGORITHM VIA MANIFOLD TL 3425
TABLE III
M EAN AND S TANDARD D EVIATION VALUES OF MS M ETRIC O BTAINED BY C OMPARED A LGORITHMS FOR D IFFERENT DYNAMIC T EST S ETTINGS
TABLE IV
M EAN AND S TANDARD D EVIATION VALUES OF MIGD, SP, AND MS M ETRICS O BTAINED BY
MMTL-MOEA/D, MMTL-MOEA/DM , MMTL-MOEA/DT , AND RI-MOEA/D
MMTL-MOEA/DT both add random individuals to replenish we can find that when the initial population of the next
the population so as to ensure sufficient diversity of the initial moment is generated using the TL or the memory-based
population. mechanism, the metric of MIGD is significantly improved,
We compare these two algorithms with MMTL-MOEA/D compared with the RI-MOEA/D algorithm. It shows that when
and the baseline algorithm RI-MOEA/D. The settings of the problem environment changes, the two mechanisms can
experimental parameters are consistent with the above exper- adapt the evolution algorithm to the new environment faster
iments, and the FDA series and dMOP series test functions and eventually improve the convergence and diversity of the
are used. solution.
Table IV lists the average MIGD, MS, and SP metric val- At the same time, the comparison among the three
ues for the experiment. Comparing the three metrics of the columns of MMTL-MOEA/D, MMTL-MOEA/DM , and
MOEA/D, the MMTL-MOEA/DM , and the MMTL-MOEA/DT , MMTL-MOEA/DT show that the combination of the two
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
3426 IEEE TRANSACTIONS ON CYBERNETICS, VOL. 51, NO. 7, JULY 2021
TABLE V
RUNNING T IME OF T WO A LGORITHMS ON D IFFERENT B ENCHMARK F UNCTIONS (U NIT: S ECONDS )
TABLE VI
RUNNING T IME OF DMOA S ON D IFFERENT F UNCTIONS (U NIT: S ECONDS )
mechanisms is superior to the use of only one mechanism. Tables I–IV in the supplementary material record the MIGD,
One thing we need to point out is that in this experi- SP, and MS values of the Tr-MOEA/D and MMTL-MOEA/D
ment, the MMTL-MOEA/DM algorithm is superior to the on FDA, dMOP, and DF problems. The above experimen-
MMTL-MOEA/DT algorithm. tal results show that the manifold transfer-based population
We think one of the reasons for this difference is that the prediction is effective in solving DMOPs. The reasons for the
MMTL-MOEA/DT is more sensitive to parameters, and it is superior performance of the proposed MMTL-MOEA/D are
possible to improve the performance of the MMTL-MOEA/DT two-fold. First, it compresses high-dimensional variables in a
by setting reasonable parameters, but how to find the optimal low-dimensional space, so MMTL-MOEA/D is more effective;
parameters is a tough problem. second, the proposed design can transfer information gradually
over geodesic flow so that it can improve its adaptability to
solution space under the new environment. Besides, from the
C. Compared With Tr-DMOEA previous ablation study, the memory-based mechanism and the
The MMTL-DMOEA stems from the framework of the Tr- TL-based mechanism can both help improve performance, so
DMOEA, therefore we want to prove that MMTL-DMOEA the combination of these two mechanisms can achieve better
outperforms the original algorithm in terms of the evolution performance in solving DMOPs.
speed and convergence. Both MMTL-MOEA/D and Tr-MOEA/D are prediction
In order to compare evolution speed, we recorded the run- methods based on TL. Compared with Tr-MOEA/D, one of the
ning time of the proposed Tr-MOEA/D and MMTL-MOEA/D differences is that Tr-MOEA/D requires a significant amount
under the same hardware configuration.1 Table V shows the of computation to obtain latent space to find optimal solutions,
mean running time of each environment of Tr-MOEA/D and while our MMTL-MOEA/D directly predicts the optimal solu-
MMTL-MOEA/D in solving the FDA series and dMOP series tions in the decision space. Therefore, MMTL-MOEA/D takes
test problems. The experimental results show that the proposed less time but achieves better performance.
method has greatly improved the speed of operation, and the
acceleration rate is hundreds of times.
Two reasons are accountable for this performance improve- D. Running Time Cost
ment. First, the TL method in the proposed method spent The good performance of MMTL-MOEA/D comes with
much less time than the Tr-MOEA/D algorithm. Second, a small price. We compare the mean running time cost of
MMTL-MOEA/D directly predicts the optimal solutions in each environment of different algorithms and present the
the decision space, while the Tr-MOEA/D method uses the results in Table VI. As shown in Table VI, KF-MOEA/D
POF to make the prediction. After locating the individual in and PPS-MOEA/D consume less time than the other com-
the objectives space, Tr-MOEA/D needs extra time to find the pared algorithms. MMTL-MOEA/D is faster than the other
corresponding individuals in the decision space. To ensure the compared algorithms on FDA5 and dMOP1 test functions. In
fairness of the comparison, we use the same settings mentioned other functions, MMTL-MOEA/D is only 1–2 s slower than
above to conduct the experiment. PPS-MOEA/D or KF-MOEA/D. But considering the good
performance of MMTL-MOEA/D in MIGD, SP and MS met-
1 The implementation environment is as follows: 1.6-GHz Intel Core i5, rics, MMTL-MOEA/D is still very competitive as compared
8-GB 1600-MHz DDR3. to other algorithms.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
JIANG et al.: FAST DYNAMIC EVOLUTIONARY MULTIOBJECTIVE ALGORITHM VIA MANIFOLD TL 3427
V. C ONCLUSION [16] R. Chen, K. Li, and X. Yao, “Dynamic multiobjectives optimization with
a changing number of objectives,” IEEE Trans. Evol. Comput., vol. 22,
TL-based methods have shown their promising applica- no. 1, pp. 157–171, Feb. 2018.
tions in solving DMOPs. However, existing methods are faced [17] D. Gong, B. Xu, Y. Zhang, Y. Guo, and S. Yang, “A similarity-
with various challenges particularly on the speed of conver- based cooperative co-evolutionary algorithm for dynamic interval multi-
objective optimization problems,” IEEE Trans. Evol. Comput., vol. 24,
gence. To overcome the issue, a memory-driven manifold no. 1, pp. 142–156, Feb. 2020.
TL algorithm, called MMTL-DMOEA, has been proposed. [18] Y. Wang and B. Li, “Multi-strategy ensemble evolutionary algorithm for
The MMTL-DMOEA combines the manifold-based TL algo- dynamic multi-objective optimization,” Memetic Comput., vol. 2, no. 1,
pp. 3–24, 2010.
rithm with a memory-based prediction method to generate [19] M. Rong, D. Gong, Y. Zhang, Y. Jin, and W. Pedrycz, “Multidirectional
the initial population during the evolution. We conducted sys- prediction approach for dynamic multiobjective optimization problems,”
tematic experiments and showed that the solutions obtained IEEE Trans. Cybern., vol. 49, no. 9, pp. 3362–3374, Sep. 2019.
[20] W. T. Koo, C. K. Goh, and K. C. Tan, “A predictive gradient strategy for
by the proposed MMTL-DMOEA are better than state-of-the- multiobjective evolutionary algorithms in a fast changing environment,”
art algorithms in terms of convergence, diversity, uniformity, Memetic Comput., vol. 2, no. 2, pp. 87–110, 2010.
and convergence speed in particular. The integration of both [21] M. Rong, D. Gong, W. Pedrycz, and L. Wang, “A multi-model prediction
method for dynamic multi-objective evolutionary optimization,” IEEE
memory and manifold TL can retain the effective strategy of Trans. Evol. Comput., vol. 24, no. 2, pp. 290–304, Apr. 2020.
elitism in traditional EA, while at the same time making full [22] A. Zhou, Y. Jin, and Q. Zhang, “A population prediction strategy for evo-
use of the advantages in TL. lutionary dynamic multiobjective optimization,” IEEE Trans. Cybern.,
vol. 44, no. 1, pp. 40–53, Jan. 2014.
[23] A. Muruganantham, K. C. Tan, and P. Vadakkepat, “Evolutionary
dynamic multiobjective optimization via Kalman filter prediction,” IEEE
R EFERENCES Trans. Cybern., vol. 46, no. 12, pp. 2862–2873, Dec. 2016.
[24] Q. Zhang and H. Li, “MOEA/D: A multiobjective evolutionary algorithm
[1] R. Azzouz, S. Bechikh, and L. B. Said, “Dynamic multi- based on decomposition,” IEEE Trans. Evol. Comput., vol. 11, no. 6,
objective optimization using evolutionary algorithms: A survey,” in pp. 712–731, Dec. 2007.
Recent Advances in Evolutionary Multi-Objective Optimization. Cham, [25] J. Zou, Q. Li, S. Yang, H. Bai, and J. Zheng, “A prediction strat-
Switzerland: Springer, 2017, pp. 31–70. egy based on center points and knee points for evolutionary dynamic
[2] Z. Yang, Y. Jin, and K. Hao, “A bio-inspired self-learning coevolutionary multi-objective optimization,” Appl. Soft Comput., vol. 61, pp. 806–818,
dynamic multiobjective optimization algorithm for Internet of Things Dec. 2018.
services,” IEEE Trans. Evol. Comput., vol. 23, no. 4, pp. 675–688, [26] B. Xu, Y. Zhang, D. Gong, Y. Guo, and M. Rong, “Environment
Aug. 2019. sensitivity-based cooperative co-evolutionary algorithms for dynamic
[3] J. Xu, P. B. Luh, F. B. White, E. Ni, and K. Kasiviswanathan, “Power multi-objective optimization,” IEEE/ACM Trans. Comput. Biol.
portfolio optimization in deregulated electricity markets with risk Bioinformat., vol. 15, no. 6, pp. 1877–1890, Nov. 2018.
management,” IEEE Trans. Power Syst., vol. 21, no. 4, pp. 1653–1662, [27] Y. Woldesenbet and G. Yen, “Dynamic evolutionary algorithm with vari-
Nov. 2006. able relocation,” IEEE Trans. Evol. Comput., vol. 13, no. 3, pp. 500–513,
[4] A. Simões and E. Costa, “Prediction in evolutionary algorithms for Jun. 2009.
dynamic environments,” Soft Comput, vol. 18, no. 8, pp. 1471–1497, [28] A. Gupta, Y. Ong, and L. Feng, “Insights on transfer optimization:
Oct. 2013. Because experience is the best teacher,” IEEE Trans. Emerg. Topics
[5] C. Rossi, M. Abderrahim, and J. C. Díaz, “Tracking moving optima Comput. Intell., vol. 2, no. 1, pp. 51–64, Feb. 2018.
using Kalman-based predictions,” Evol. Comput., vol. 16, no. 1, [29] L. Feng et al., “Evolutionary multitasking via explicit autoencoding,”
pp. 1–30, Mar. 2008. IEEE Trans. Cybern., vol. 49, no. 9, pp. 3457–3470, Sep. 2019.
[6] M. Jiang, Z. Huang, L. Qiu, W. Huang, and G. G. Yen, “Transfer [30] L. Zhou et al., “Toward adaptive knowledge transfer in multifactorial
learning-based dynamic multiobjective optimization algorithms,” IEEE evolutionary computation,” IEEE Trans. Cybern., early access, Mar. 6,
Trans. Evol. Comput., vol. 22, no. 4, pp. 501–514, Aug. 2018. 2020, doi: 10.1109/TCYB.2020.2974100.
[7] S. J. Pan, I. W. Tsang, J. T. Kwok, and Q. Yang, “Domain adaptation [31] K. K. Bali, Y. Ong, A. Gupta, and P. S. Tan, “Multifactorial evolutionary
via transfer component analysis,” IEEE Trans. Neural Netw., vol. 22, algorithm with online transfer parameter estimation: MFEA-II,” IEEE
no. 2, pp. 199–210, Feb. 2011. Trans. Evol. Comput., vol. 24, no. 1, pp. 69–83, Feb. 2020.
[32] B. Da, A. Gupta, and Y. Ong, “Curbing negative influences online
[8] T. Lin and H. Zha, “Riemannian manifold learning,” IEEE Trans. Pattern
for seamless transfer evolutionary optimization,” IEEE Trans. Cybern.,
Anal. Mach. Intell., vol. 30, no. 5, pp. 796–809, May 2008.
vol. 49, no. 12, pp. 4365–4378, Dec. 2019.
[9] R. Gopalan, R. Li, and R. Chellappa, “Domain adaptation for object [33] W. Zhou et al., “A preliminary study of adaptive indicator based
recognition: An unsupervised approach,” in Proc. IEEE Int. Conf. evolutionary algorithm for dynamic multiobjective optimization via
Comput. Vis. (ICCV), 2011, pp. 999–1006. autoencoding,” in Proc. IEEE Congr. Evol. Comput. (CEC), 2018,
[10] K. Deb, U. N. B. Rao, and S. Karthik, “Dynamic multi-objective pp. 1–7.
optimization and decision-making using modified NSGA-II: A case [34] A. T. W. Min, Y. Ong, A. Gupta, and C. Goh, “Multiproblem surrogates:
study on hydro-thermal power scheduling,” in Proc. Int. Conf. Evol. Transfer evolutionary multiobjective optimization of computationally
Multi Crit. Optim., 2007, pp. 803–817. expensive problems,” IEEE Trans. Evol. Comput., vol. 23, no. 1,
[11] Q. Chen, J. Ding, S. Yang, and T. Chai, “A novel evolution- pp. 15–28, Feb. 2019.
ary algorithm for dynamic constrained multiobjective optimization [35] X.-F. Liu et al., “Neural network-based information transfer for dynamic
problems,” IEEE Trans. Evol. Comput., early access, Dec. 6, 2019, optimization,” IEEE Trans. Neural Netw. Learn. Syst., early access,
doi: 10.1109/TEVC.2019.2958075. Jul. 19, 2019, doi: 10.1109/TNNLS.2019.2920887.
[12] G. Ruan, G. Yu, J. Zheng, J. Zou, and S. Yang, “The effect of diver- [36] M. Jiang, L. Qiu, Z. Huang, and G. G. Yen, “Dynamic multi-objective
sity maintenance on prediction in dynamic multi-objective optimization,” estimation of distribution algorithm based on domain adaptation and
Appl. Soft Comput., vol. 58, pp. 631–647, Sep. 2017. nonparametric estimation,” Inf. Sci., vol. 435, pp. 203–223, Apr. 2018.
[13] S. Jiang and S. Yang, “A steady-state and generational evolutionary [37] Q. Zhang, A. Zhou, and Y. Jin, “RM-MEDA: A regularity model-based
algorithm for dynamic multiobjective optimization,” IEEE Trans. Evol. multiobjective estimation of distribution algorithm,” IEEE Trans. Evol.
Comput., vol. 21, no. 1, pp. 65–82, Feb. 2017. Comput., vol. 12, no. 1, pp. 41–63, Feb. 2008.
[14] C.-K. Goh and K. C. Tan, “A competitive-cooperative coevolutionary [38] M. Farina, K. Deb, and P. Amato, “Dynamic multiobjective optimization
paradigm for dynamic multiobjective optimization,” IEEE Trans. Evol. problems: Test cases, approximations, and applications,” IEEE Trans.
Comput., vol. 13, no. 1, pp. 103–127, Feb. 2009. Evol. Comput., vol. 8, no. 5, pp. 425–442, Oct. 2004.
[15] D. Yazdani, M. N. Omidvar, J. Branke, T. T. Nguyen, and X. Yao, [39] S. Jiang, S. Yang, X. Yao, K. Tan, M. Kaiser, and N. Krasnogor,
“Scaling up dynamic optimization problems: A divide-and-conquer “Benchmark problems for cec2018 competition on dynamic
approach,” IEEE Trans. Evol. Comput., vol. 24, no. 1, pp. 1–15, multiobjective optimisation,” in Proc. CEC2018 Competition, 2018,
Feb. 2020. pp. 1–8.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.
3428 IEEE TRANSACTIONS ON CYBERNETICS, VOL. 51, NO. 7, JULY 2021
[40] L. Cao, L. Xu, E. D. Goodman, C. Bao, and S. Zhu, “Evolutionary Shihui Guo (Member, IEEE) received the B.S.
dynamic multiobjective optimization assisted by a support vector regres- degree in electrical engineering from Peking
sion predictor,” IEEE Trans. Evol. Comput., vol. 24, no. 2, pp. 305–319, University, Beijing, China, in 2010, and the Ph.D.
Apr. 2020. degree in computer animation from the National
[41] C.-C. Chang and C.-J. Lin, “LIBSVM: A library for support vector Centre for Computer Animation, Bournemouth
machines,” ACM Trans. Intell. Syst. Technol., vol. 2, no. 3, pp. 1–27, University, Poole, U.K., in 2015.
May 2011. He is an Associate Professor with the School of
Informatics, Xiamen University, Xiamen, China. His
research interests include computational intelligence
and human–computer interaction.
Authorized licensed use limited to: BEIHANG UNIVERSITY. Downloaded on January 24,2024 at 06:18:07 UTC from IEEE Xplore. Restrictions apply.