Topology Optimization Via Machine Learning and Dee
Topology Optimization Via Machine Learning and Dee
1736–1766
DOI: 10.1093/jcde/qwad072
Advance access publication date: 13 July 2023
Review article
Received: October 16, 2022. Revised: June 30, 2023. Accepted: July 1, 2023
© The Author(s) 2023. Published by Oxford University Press on behalf of the Society for Computational Design and Engineering. This is an Open Access article
distributed under the terms of the Creative Commons Attribution-NonCommercial License (https://creativecommons.org/licenses/by-nc/4.0/), which permits
non-commercial re-use, distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact
[email protected]
Journal of Computational Design and Engineering, 2023, 10(4), 1736–1766 | 1737
performance. Single-stage MLTO consists of one network, whereas volume fraction and loading and boundary conditions for the in-
multistage MLTO consists of two or more networks. put of deep CNN but also the strain field information through FEA
to generalize this method to various loading and boundary con-
2.2.1. Single-stage process ditions and various initial designs. Similarly, Nie et al. (2021) pro-
First, as a single-stage method to predict optimization results, posed cGAN-type TopologyGAN, in which the strain energy den-
Abueidda et al. (2020) presented a CNN (ResUnet) that di- sity and von Mises stress values obtained from FEA are used as
rectly derives an optimized shape with 2D loading and bound- the condition, along with the volume fraction, displacement, and
ary conditions, and volume fraction as the input for three loading and boundary conditions.
material properties, such as linear elasticity with small de- Cang et al. (2019) introduced a “theory-driven” mechanism
formation (i.e., without non-linear constraints), non-linear hy- based on adaptive sampling rather than a batch-mode training
perelasticity with geometric non-linearity, and linear elasticity method to reduce computational costs in the optimization pro-
with stress constraint (i.e., non-linear constraint). Sharpe and cess. The deviation of the topology predicted by the NN from the
Seepersad (2019) derived the optimized topology in real-time optimality conditions for TO is calculated and used to select a new
by inputting conditions, such as volume fraction, loading lo- data point (optimization problem set) for the next training. This
cation, and material type, by using a conditional generative method allows for the better prediction of designs with optimal
adversarial network (cGAN). Kollmann et al. (2020) predicted structural compliance for unseen loadings than using static data
the optimal topology of 2D meta-material via encoder and collection under the same computational budget.
decoder from ResUNet by using volume fraction, filter ra-
dius, and design objective (e.g., maximum bulk modulus, max- 2.2.2. Multistage process
imum shear modulus, or minimum Poisson’s ratio) as the in- Scholars have also proposed a multistage non-iterative process
put. to predict the optimization result with better engineering perfor-
Similarly, by using 2D optimization results obtained from the mance by integrating two or more networks.
SIMP method under specific loading and boundary conditions as First, Yu et al. (2019) performed TO in a non-iterative man-
the reference data, Rawat and Shen (2018) used Wasserstein gen- ner and then upscaled the predicted low-resolution image into
erative adversarial network (WGAN) to predict the optimal shape high-resolution. This study was the first to propose a near-
in real-time and then additionally predicted the optimization con- optimal optimization process without repeated iterations. The
dition of the generation result. This approach allows the gener- low-resolution optimal images are first predicted by inputting x-,
ated shape to be mapped to the corresponding optimization con- and y-directional loading and boundary conditions as the input,
ditions, such as the volume fraction, penalty, and radius of the and then the final optimal topology in high-resolution can be pre-
smoothing filter through CNN. Extending this scheme to the 3D dicted with the low-resolution image. Li et al. (2019a) also applied
domain, Rawat and Shen (2019a) predicted the 3D optimal topol- a non-iterative method similar to that of Yu et al. (2019) to a con-
ogy through WGAN, and the optimization conditions (e.g., volume ductive heat transfer problem. Low-dimensional topology images
fraction, penalty, and radius of smoothing filter) are predicted by are generated through the generative adversarial network (GAN)
CNN. Rawat and Shen (2019b) attempted to predict quasi-optimal based on the heat sink position, the heat source position, and the
topology by using volume fraction as a condition to conditional mass fraction information, and upscale the result through super-
WGAN. resolution GAN (SRGAN). Keshavarzzadeh et al. (2021) constructed
Furthermore, researchers have attempted to predict the opti- a prediction model that approximates a high-resolution optimal
mized shape instantly aided by physics information from FEA for design, given training data of multiple resolutions. To this end,
higher accuracy. For example, Zhang et al. (2019) used not only they proposed a model integrating the disjunctive normal shape
Journal of Computational Design and Engineering, 2023, 10(4), 1736–1766 | 1739
model (DNSM) at the end of a fully connected DNN to construct sity sequence (DS) prediction network that predicts the initial den-
a multiresolution network with both low-resolution and high- sity by using the initial compliance and the volume fraction ini-
resolution designs. After mapping low-dimensional parameters tialization to predict the final optimal density. The second method
such as loading, boundary conditions, and volume fraction to the is coupled density and compliance sequence (CDCS) prediction
DNSM coefficient by using the DNN, the DNSM model is used to network, which uses the intermediate density and the intermedi-
predict the optimal topologies of various resolutions. ate compliance information together to predict the optimal den-
Rade et al. (2020) proposed two methods to explore high- sity though five repetitive processes. By using the first 30 iteration
resolution TO (HRTO) for 2D and 3D features. The first is the den- results from ESO, Qiu et al. (2021a) trained the U-net CNN model to
1740 | Topology optimization via artificial intelligence: a review
Acceleration of iteration Banga et al., 2018; Bi et al., 2020; Kallioras & Lagaros, 2021a; Kallioras et al., 2020; Lin et al., 2018;
Sosnovik & Oseledets, 2019
Non-iterative optimization Abueidda et al., 2020; Almasri et al., 2020; Cang et al., 2019; Keshavarzzadeh et al., 2021; Kollmann et al.,
2020; Lew & Buehler, 2021; Li et al., 2019a; Nie et al., 2021; Qiu et al., 2021a; Rade et al., 2020; Rawat &
Shen, 2018, 2019a, b; Sharpe & Seepersad, 2019; Yu et al., 2019; Zhang et al., 2019
Meta-modeling Asanuma et al., 2020; Aulig & Olhofer, 2014; Chi et al., 2021; Deng et al., 2022; Doi et al., 2019;
Keshavarzzadeh et al., 2020; Kim et al., 2021; Lee et al., 2020; Li et al., 2019b; Patel & Choi, 2012; Qian &
Ye, 2021; Sasaki & Igarashi, 2019a, b; Takahashi et al., 2019; White et al., 2019; Xia et al., 2017; Zheng et
al., 2021; Zhou & Saitou, 2017
Dimensionality reduction Guo et al., 2018; Li et al., 2019c; Ulu et al., 2016
of design space
Improvement of optimizer Bujny et al., 2018; Chandrasekhar & Suresh, 2021a, b, c; Deng & To, 2020; Halle et al., 2021; Hoyer et al.,
predict the intermediate topologies and then used the evolution of ods to proceed immediately without undertaking FEA. Deng et al.
the predicted topologies to learn the subsequent structure trans- (2022) also used DNN to approximate the objective function.
formations using long short-term memory (LSTM). This method Kim et al. (2021) used the representative volume element (RVE)
required less training data but could offer much broader appli- method for TO of functionally graded composite structure design,
cability. Lew and Buehler (2021) utilized representations in latent constructing a DNN-based continuous model to predict the effec-
space to explore the latent design space via the variational au- tive elasticity tensor (a material property) for given design param-
toencoder (VAE)-LSTM model. The LSTM was trained to learn the eters to replace the FEA process in the iterative TO process. This
trajectories in a latent space that could compress the optimiza- methodology is applicable to both 2D and 3D problems.
tion routes, allowing the system to learn the valid optimization In the work of White et al. (2019), which deals with the prob-
progression. lem of multiscale TO, the computational process for effective elas-
In a study that additionally considered the manufacturing con- tic stiffness of meta-material is replaced by the NN surrogate
ditions, Almasri et al. (2020) used a conditional convolutional dual- model. Through this NN, microscale geometric parameters can
discriminator GAN to accelerate TO, considering both a single gen- be mapped to macroscale stiffness. Li et al. (2019b) used feed-
eral manufacturing condition and a mechanical condition. In the forward NNs (FFNN) to map macroscale strain components and
prediction of the 2D optimized shape, both geometrical (e.g., the macroscale stress components. Using the series of stress values
total number of bars) and mechanical (e.g., boundary conditions, predicted by FFNN, they predicted whether the microstructure is
loading configuration, and volume fraction) constraints are used damaged or not through CNN to provide a criterion for sensitivity
as the input to the generator, and two discriminators ensure that analysis.
both mechanical and geometrical constraints are both satisfied.
Furthermore, to evaluate the performance of the generated topol- 2.3.2. Sensitivity
ogy, a CNN-based network is used to predict the compliance of the
As a study to predict sensitivity, Aulig and Olhofer (2014) replaced
generated structure.
sensitivity analysis using a regression model . The material den-
sity of an element, which is a local state feature, and the displace-
2.3. Meta-modeling ment values of the element nodes obtained through FEA are used
In the TO process, the calculation of objective function and sen- to predict the sensitivity. Xia et al. (2017) proposed a TO method
sitivity analysis process through FEA occupies a high computa- applied with a constrained maximum-weight connected graph
tional cost. Therefore, some studies that accelerate the TO process (CMWG) to remove the checkerboard pattern from the optimiza-
through ML are conducted to accelerate the optimization process tion results and used kriging models based on the support vector
by replacing the solvers, such as FEA. machine (SVM) to predict the sensitivity. Through the CMWG al-
gorithm, the result of the previous ten iterations and the data of
the surrounding elements are combined and used as training data
2.3.1. Objective function to calculate the sensitivity value of the element.
As a representative study to predict the objective function Takahashi et al. (2019) attempted to find the global optimum
value, Lee et al. (2020) attempted to accelerate the optimiza- by calculating the sensitivity through CNN in the TO process. Ad-
tion process by replacing the objective function calculation pro- ditionally, filtering was also performed at the end to enable the
cess through FEA with CNN . The compliance and volume optimization results to be more meaningful from an engineering
fraction are predicted by inserting the intermediate material perspective. Chi et al. (2021) also trained the ML model to directly
layout as an input, allowing for the optimality criteria (OC) meth- predict sensitivity without solving the state equations. However,
Journal of Computational Design and Engineering, 2023, 10(4), 1736–1766 | 1741
in contrast to the work of Takahashi et al. (2019), they proposed and Igarashi (2019a) predicted the average torque and torque rip-
an online training concept in the optimization process and used ple value of the interior permanent magnet (IPM) motor through
the history data of the optimization process as a training sample. CNN to calculate the objective function. According to the pre-
In providing sufficient information for training while avoiding the dicted objective function value, probability P of FEA is calculated
most time-consuming step of solving the state equation, TO for- to determine whether to execute FEA or not. The probability P
mulas with two discretization levels, fine-scale mesh and coarse- helps to skip unnecessary FE processes because FE calculations
scale mesh, are used. Design variables are used from the fine- are only performed when P is large. In the existing genetic al-
scale mesh, and the strain vectors are used from the coarse-scale gorithm (GA), many candidates are randomly generated. How-
mesh. ever, as these initial shapes do not all satisfy the torque perfor-
mance, CNN predicts the torque performance to reduce the num-
2.3.3. Objective function and sensitivity ber of samples to be subjected to FEM to save time and comput-
As a study for predicting both objective function calculation and ing cost. Sasaki and Igarashi (2019b) also reduced the FE com-
sensitivity analysis, Qian and Ye (2021) used dual model NN for putation through a classification model that predicts the per-
more accurate sensitivity analysis. The initial design obtained formance of IPM motors. In contrast to the work of Sasaki and
using this method, the optimal topology can be efficiently found uated the displacement of the topology via physics simulation
within the global optimization. and updated the network parameters based on the physics-based
loss function by using the standard gradient-based algorithm. In
2.5. Improvement of the optimizer this manner, the optimal topology can be obtained with the de-
sired compliance. This process highlights the structural optimiza-
2.5.1. Optimization parameter tuning
tion method, which can be generalized to various optimization
Scholars have also supplemented conventional TO methods by problems. Similarly, Zhang et al. (2021a) proposed TO via a neu-
tuning the optimization parameters based on ML to quickly and ral reparameterization framework that would not need to con-
efficiently search for parameters used in the optimization process. struct a dataset in advance and would not suffer from structural
The parameters applied to the MMC are tuned by experience disconnection for various problems, including non-linear TO
and can obtain an infeasible solution. To solve this problem, Lei problems. Physics information was introduced into the loss func-
et al. (2019) combined ML to optimize parameters used in the MMC tion and the design constraint processing method, which was
method. After obtaining the eigenvalue and eigenvector of the pa- based on the volume-preserving Heaviside function with sigmoid
rameter vector through PCA, non-linear regression was performed transformation.
using support vector regression (SVR) and K-nearest neighbor
avoided. Although the proposed method requires more computa- mal topology), to learn the corresponding failure distribution of
tional time than the conventional FEM methods, this mechanism the optimal topology. To generate only high-performance shapes,
offers a fully mesh-free approach to TO for 2D and 3D problems. training was conducted by updating the dataset with samples
Additionally, researchers have suggested integrating the meth- with a good performance through FEM among those generated in
ods above with generative models to generate various optimal the previous iteration. Through a data-driven approach, Yamasaki
topologies with high engineering performance. This topic is elab- et al. (2021) produced new material distribution with good perfor-
orated on in Section 2.6. mance by constructing VAE, a deep generative model, based on
current high-rank data, which is the best design evaluated by FEM
2.6. Generative design among initial material distribution data from the SIMP method.
There are studies for generative design that quickly and diversely By integrating this into the current high-rank data, the pro-
generate new optimal designs. cess of screening new high-rank data is iteratively repeated, and
when the elite data converges, the final optimal topology can be
2.6.1. Naïve approach derived.
Engineering performance aided by NN
The most naïve method is to train a generative model using the ex-
mization frameworks. First, a model consisting of conditional AAE this method, not only the training sample and condition but also
(c-AAE) and regression network is presented, secondly, c-AAE + DE a random condition was additionally inputted to the discrimina-
framework, which uses differential evolution (DE) global opti- tor, and the minibatch discrimination technique was applied to al-
mizer for c-AAE, and finally, c-AAE + rDE framework that also uses low more stable convergence and prevent the mode collapse phe-
important physics-driven regularization. Among them, the third nomenon in which the generator continues to generate similar
model allows the physical regularization of the compressed de- shapes.
sign space to adjust the design space configuration to perform GO
exploration better. This method utilizes the physics information 2.6.5. Manufacturing constraints
inherent in the optical performance of meta-devices with com- In general, many cases claim that manufacturing is impossible
plex material composition. by simply referring to the topological optimization results. To ap-
Engineering performance aided by clustering ply manufacturing constraints to TO, Greminger (2020) used syn-
In the previous section, the studies used engineering perfor- thetic 3D voxel training set to represent the distribution of a set
mance information during the training process. Next, a study con- of geometries that can be manufactured by a specific manufac-
sidered engineering performance by applying clustering to the turing technique to generate specific manufacturable 3D designs
Therefore, by studying ML to obtain high-resolution optimized im- sists of three steps: clustering, optimization, and connectiv-
ages, Napier et al. (2020) used ANN to upscale a low-dimensional ity. This study proposed a density–strain-based method with
input to a high-dimensional one. In this study, the FFNN is trained a fixed clustering pattern according to the static density and
to upscale the low-resolution data to high-resolution data by us- stress state . However, fixed clustering patterns can limit the
ing the preprocessed data of high-resolution beam optimization performance. Consequently, Qiu et al. (2021b) excluded den-
data to low resolution through averaging. Similarly, Wang et al. sity from CA to optimize cellular structures with uniform
(2021) used CNN to map the relationship between low-resolution porosity and used the direction and ratio of principal stress
and high-resolution structures in TO problems. This approach is for clustering. By introducing a stress ratio into clustering,
helpful for large-scale structural design. the stiffness performance and connectivity of the microstruc-
In addition, Xue et al. (2021) proposed a super-resolution CNN ture can be improved. The clustering pattern should be up-
(SRCNN) framework with superior applicability to improve the dated by the current stress tensor in grouping the microstruc-
resolution of TO designs and the HRTO method established by im- ture by using the dynamic clustering method. When clus-
plementing SRCNN and pooling strategy. SRCNN is a network that tering, the stress direction-based cluster analysis is first per-
maps low-resolution images to high resolution, and is available formed, and then the elements of each direction cluster are
Various studies have replaced the inefficient process in itera- variables through NN. The design parameters predicted by
tive TO with ML. For example, the studies presented in Section Keshavarzzadeh et al. (2021) correspond to the DNSM co-
2.1 were aimed at accelerating TO by combining ML at the end efficient and are predicted to map DNSM’s coefficient for
of the iterative TO, thereby ensuring accuracy and improving ef- design parameters (load and boundary conditions and vol-
ficiency. The studies in Section 2.3 replaced the time-consuming ume fraction) via NN through a DNSM metric model. In
simulation stage in the iterative TO with an ML-based meta-model addition, in the work of Yildiz et al. (2003), the design pa-
to reduce the computational cost of the conventional TO. Some rameters correspond to the class for the shape of the hole.
studies in Section 2.4 utilized the latent design space to efficiently Yildiz et al. (2003) predicted which class the hole of the
seek the optimal solution. The studies attempted to accelerate shape corresponds to for the optimized feature and per-
the TO while maintaining the iterative process to ensure accu- formed TO based on the results. Second, in studies pre-
racy. Meanwhile, the studies in Section 2.2 were aimed at gen- dicting the performance of a shape (Guo et al., 2021), NN is
erating the optimal topology in one shot by using ML without used to replace some calculations in the TO process. Deng
the need for implementing a time-consuming iterative process. et al. (2022) used DNN to replace the calculation of the ob-
This method can instantly provide optimal topologies once the jective function during the optimization process. Chi et al.
Supervised learning Neural network Prediction Aulig & Olhofer, 2014; Bujny et al., 2018; Chi et al., 2021;
Deng & To, 2020; Deng et al., 2022; Keshavarzzadeh et al.,
2021; Patel & Choi, 2012; White et al., 2019; Yildiz et al.,
2003; Zheng et al., 2021
Optimization Bi et al., 2020; Chandrasekhar & Suresh, 2021a, b, c; Halle
et al., 2021; Kim et al., 2021; Napier et al., 2020; Zehnder et
al., 2021
Convolutional neural Prediction Asanuma et al., 2020; Doi et al., 2019; Gaymann &
network Montomoli, 2019; Lee et al., 2020; Li et al., 2019b; Sasaki &
Igarashi, 2019a, b; Takahashi et al., 2019
Optimization Abueidda et al., 2020; Banga et al., 2018; Hoyer et al., 2019;
Kollmann et al., 2020; Lin et al., 2018; Rade et al., 2020;
sign density by combining design conditions and FEA infor- reflect compliance in the loss function. Finally, Xue et al.
mation. Rade et al. (2020) used the CDCS method to train (2021) and Wang et al. (2021) used CNN to upscale optimal
three encoder–decoder-shaped CNN models. The compli- designs.
ance prediction model could predict the next compliance
LSTM
by using the inputs of initial compliance and current den-
LSTM is a useful method for predicting a series of processes
sity. The density prediction model could also predict the
(e.g., the TO process). It is often used in conjunction with other
next density by using the inputs of current compliance and
methodologies for TO optimization:
current density. During the five loops, the two models uti-
lized each other’s outputs as inputs. Finally, with the out- 1. Optimization: Qiu et al. (2021a) utilized the LSTM model
put density determined after five loops, final density pre- to predict the next sequence of TO and then performed
diction model predicts optimal density. In addition, Hoyer DLTO integrated with ESO. An initial 1st–10th sequence im-
et al. (2019) updated the objective function for the pre- age was predicted via CNN(U-net). From the 11th sequence,
dicted design via the deep image prior to the implemen- optimization was performed by predicting the shape of the
tation of CNNS, reflected this information using gradient- next sequence, in which the shapes of the previous 1st to
methods, and allowed the design to be generated to sat- 10th sequences were inputted. Similarly, Rade et al. (2020)
isfy the objective function. Similarly, Zhang et al. (2021a) used DS method with CNN-LSTM and SIMP method to pre-
proposed the TONR method and introduced physics infor- dict the density of subsequent sequences with the input
mation into the loss function. With the noise input, the shapes of the initial two iterations. Lew and Buehler (2021)
model could predict the quasi-density and conduct FEA to trained a 2D convolutional VAE with the shapes of TO se-
1748 | Topology optimization via artificial intelligence: a review
quences to encode an image into a 2D latent vector. With 3.1.2. Unsupervised learning
two sets of initial latent vectors, the LSTM model could pre- The use of unsupervised learning for TO usually heavily uti-
dict the time series latent vectors for the subsequent 147 lizes the generative model and generally has a design exploration
sequences. Kallioras and Lagaros (2021b) used LSTM for the purpose similar to the original purpose of the generative model
time-series classification for generative design. With mul- methodology.GAN
tiple mesh domain, after 20 iterations of SIMP, the density
values of each element were input into the LSTM model 1. Exploration: Greminger (2020) generated feasible data by
to derive the final density class of every element of the FE using a 3D GAN. Li et al. (2021) evaluated the performance
discretization. Then, multiple image filters were applied to of designs generated by GAN through FEM, and Sim et al.
the output shape to generate various optimized shapes. Fi- (2021) clustered the designs generated by GAN model into
nally, 20 SIMP iterations were performed in the final desired k-means clustering. Besides simply generating shapes, Oh
mesh size. et al. (2018, 2019) refined the design through the process of
re-optimizing the GAN-generated structures as an initial
SVM design. Rawat and Shen (2018, 2019a) used GAN to gen-
SVMs are also often used for postprocessing and prediction in erate structures and then replace FEA with meta-models
model for the performance of the sample data to proceed For design optimization, value-based RL is used. Hayashi and
with optimization. Ohsaki (2020) derived designs corresponding to design parame-
ters on a graph basis by using Q-learning (Watkins & Dayan, 1992).
Clustering
Brown et al. (2022) reformulated TO problems to a sequential RL
1. Optimization: As a methodology of using clustering al- task, a Markov decision process (MDP; Bellman, 1957), and opti-
gorithm, scholars have added clustering for shapes, de- mized elementally discretized 2D topologies by using the double-
sign characteristics, or FEM results to an iterative TO step deep Q-learning algorithm (Van Hasselt et al., 2016).
to reduce the calculation costs (Jiang et al., 2020; Kumar For design exploration, Sun and Ma (2020) used the e-greedy
& Suresh, 2020; Liu et al., 2015; Qiu et al., 2021b). Among policy to disturb the search direction, UCB to add a sensitivity fea-
them, Jiang et al. (2020) trained an ensembled ETs surro- ture, TS and IDS to direct the search. The optimal shape changes
gate model that determines whether it is feasible through were compared on the basis of the parameter changes. Jang et al.
the results of clustering. (2022) conducted a generative design method on designs to ex-
plore and eventually maximize design diversity by using the prox-
3.1.3. Semi-supervised learning imal policy optimization (PPO) algorithm (Schulman et al., 2017).
as design parameters. Lynch et al. (2019) recommended the tuning discretization provided by the initial steps of the SIMP method.
of parameters for the corresponding design problem during opti- Although Hoyer et al. (2019) and Zhang et al. (2021a) used noise
mization and uses Bayesian optimization to tune the parameters as an input, not the intermediate optimization results, it deter-
and proceed with optimization. Lei et al. (2019) used PCA dimen- mines the performance of the shape generated from the noise and
sion reduction for design parameters to obtain eigenvector and optimizes the design by allowing the performance to affect net-
eigenvalue. Using this information, the design is derived through work parameter updates with the backwards pass. Liu et al. (2015)
non-linear regression. Ulu et al. (2016) predicted the PCA dimen- classified the shape of intermediate iteration into k number of ma-
sionally reduced value of the topology by using the design param- terials through K-means clustering. On this basis, optimization is
eter (i.e., loading configuration) as an input. When a new design re-progressed and obtained optimal material distribution.
parameter is introduced, the shape corresponding to the predicted
PCA value is derived. 3.2.3. Final optimization result
Exploration Studies with the final shape as an input can be largely divided into
As for the studies on design exploration, Jiang et al. (2019), Shen three purposes: the purpose of prediction, the purpose of design
and Chen (2019), Wen et al. (2019), Jiang and Fan (2019b), Rawat exploration, and the purpose of postprocessing.
(2021), Xue et al. (2021), and Napier et al. (2020) predicted high- of deriving optimal designs. Qiu et al. (2021a) also used the design
resolution shapes when optimal low-resolution shapes are input. parameter (load condition) and the shapes of previous iterations
In addition, Yildiz et al. (2003) predicted the shape of the hole (tri- as inputs. Rade et al. (2020) implemented two methodologies, DS
angle, trapezoid, rectangle, etc.) of the design by using topology and CDCS prediction, as mentioned in Section 2.2, in which the
optimized shapes as input, thereby enabling appropriate shape DS methodology was used to predict optimal designs by using ini-
optimization to proceed. tial design parameters and FEA information. The CDCS methodol-
ogy can predict optimal designs through intermediate optimiza-
3.2.4. Structural information (FEA information) tion results and intermediate FEA information. Qian and Ye (2021)
The two main purposes in which FEA information is used as input replaced compliance and sensitivity calculations by inputting de-
are prediction and design optimization. sign parameters and intermediate optimization results for meta-
Prediction modeling purposes.
As a study aimed at FEA meta-modeling, Aulig and Olhofer
(2014) replaced sensitivity calculations with predictive models by 3.3. Loss function
entering density FEA information (displacement) of shapes, and The form of loss function of DL model in TO is defined as follows
and Suresh (2021a, b, c) defined the loss function to reflect the 3.3.5. Adding two types of loss functions for multiple mod-
objective function and constraint values calculated through FEA els
of shape in each iteration to optimize the structure. Hoyer et al. In studies involving two or more networks, the use of more than
(2019), Jiang and Fan (2019a, b), Qian and Ye (2021), and Zhang two loss functions for each network is reasonable. Rawat and Shen
et al. (2021a) also incorporated the performance values obtained (2018, 2019a) generated shapes by WGAN by using Wasserstein
through FEA into loss function and repeat the process of updating distance (Section 3.3.1) and predicted the parameters of the gen-
the variables by calculating gradients through backpropagation to erated shapes by using the MSE loss (Section 3.3.2). Similarly, Ha
optimize the shape. et al. (2021) used MSE loss for a generative network to generate
shapes and used BCE to determine the feasibility of generated
shapes through 3D CNN. Li et al. (2019a) used standard GAN loss
to generate low-resolution shapes over GAN and added the MSE
loss between real, predicted shapes in SRGAN for upscaling (Sec-
tion 3.3.1). Yu et al. (2019) selected mean absolute error (MAE)
3.3.4. Integrating additional constraints
loss of real and predicted low-resolution shapes on CNN network
The loss function to add additional constraints is defined in the (Section 3.3.1) and added MAE loss term on the standard GAN
form of adding a specific constraint loss term to the existing loss. loss to resemble the real shape on the cGAN network (Section
Zhang et al. (2019) used Kullback–Leible divergence (Kingma & 3.3.1). Li et al. (2019b) used MSE loss to map macroscale strain and
Welling, 2013) to obtain the difference between the prediction and macroscale stress (FFNN) and predict the feasibility of microstruc-
the real shapes corresponding to Section 3.3.1, and further defines ture with a stress value (Section 3.3.2). Rade et al. (2020) utilized a
additional loss in the form of L2 regulation term to prevent overfit- different type of loss for two methodologies. The study proposed a
ting. Guo et al. (2018) added a reconstruction error corresponding DS method with MSE loss to calculate the differences between real
to standard VAE loss, a style loss (Gatys et al., 2015), which eval- and predicted shapes (Section 3.3.1) and a CDCS prediction net-
uates how well the generated shape match the style of the real work with loss that combines MAE loss that calculates the differ-
shape, and a mode-collapse loss (Zhao et al., 2016) to prevent only ence between real and predicted shapes in the early stages (Sec-
similar samples from being generated. Nie et al. (2021) defined BCE tion 3.3.1) while the BCE loss used to predict compliance in terms
of standard GAN loss, L2 loss term between real and fake shapes of optimal design convergence (Section 3.3.2).
corresponding to Section 3.3.1 and further added an absolute er-
ror term between the target volume and the volume of predicted
shape. Sosnovik and Osledets (2019), Lin et al. (2018), and Banga 3.4. Physics
et al. (2018) defined the BCE loss to determine the difference be- FEA is used in DLTO to ensure a much more efficient convergence
tween predicted and real shapes (Section 3.3.1) and add MSE loss of optimal designs and produce meaningful results. This concept
for the volume constraint term to the loss. is similar to a physics-guided NN and a physics-informed NN
Journal of Computational Design and Engineering, 2023, 10(4), 1736–1766 | 1753
2D Abueidda et al., 2020; Almasri et al., 2020; Aulig & Olhofer, 2014; Bi
et al., 2020; Brown et al., 2022; Bujny et al., 2018; Cang et al., 2019;
Chu et al., 2016; Deng et al., 2022; Gillhofer et al., 2019; Halle et al.,
2021; Hayashi & Ohsaki, 2020; Hoyer et al., 2019; Jiang et al., 2020;
Kallioras & Lagaros, 2020, 2021a, b; Kallioras et al., 2020;
2D design domain Keshavarzzadeh et al., 2020; Lee et al., 2020; Lei et al., 2019; Lew &
(beam, bridge, truss, Buehler, 2021; Li et al., 2019a; Liu et al., 2015; Lynch et al., 2019;
bracket, etc.) Malviya, 2020; Napier et al., 2020; Nie et al., 2021; Patel & Choi, 2012;
Qian & Ye, 2021; Qiu et al., 2021a; Rawat & Shen, 2018, 2019b;
Sharpe & Seepersad, 2019; Shen & Chen, 2019; Sim et al., 2021;
Sosnovik & Oseledets, 2019; Takahashi et al., 2019; Ulu et al., 2016;
Wang et al., 2021; Yamasaki et al., 2021; Yu et al., 2019; Zehnder et
Nanophotonics, Blanchard-Dionne & Martin, 2021; Jiang & Fan, 2019a, b; Jiang et al.,
spinodoid, 2019; Kim et al., 2021; Kollmann et al., 2020; Kudyshev et al., 2021;
meta-surface, periodic Kumar & Suresh, 2020; Li et al., 2019b, c, 2021; Qian & Ye, 2021; Qiu
structure et al., 2021b; Wang et al., 2020; Wen et al., 2019, 2020
Motor Asanuma et al., 2020; Doi et al., 2019; Sasaki & Igarashi, 2019a, b
Meta-material Karlsson et al., 2021; White et al., 2019; Zheng et al., 2021
Application to 3D design problems bility of DLTO. Therefore, we explore the studies that enable the
DLTO still lacks an expansion mechanism for the 3D design do- change in design domains. Hayashi and Ohsaki (2020) expressed
main. However, as expansion to 3D is essential for future usability, the 2D truss structure with nodes and edges, allowing the size of
studies with 3D applications were reviewed. the design domain to be adjusted. The RL agent learned general-
Rawat and Shen (2019a) optimized for 3D beams but did not izations for the design domain and could discover optimal truss
consider various load and boundary conditions. Meanwhile, Banga topologies for various load cases. Brown et al. (2022) trained the
et al. (2018), Rade et al. (2020), and Keshavarzzadeh et al. (2021) gen- RL agent to design optimal topologies (i.e., 6 × 6 in size) with di-
eralized for various load and boundary conditions and proceeded verse load and boundary conditions. Then, they tested a multistep
with extensions to 3D objects. Chi et al. (2021) updated the strain progressive refinement approach to prove that the RL agent could
information about coarse mesh online and optimizes fine mesh. adopt a generalized design strategy.
The study calculated strain for various load and boundary con- Bi et al. (2020) applied the approach to general problems by up-
ditions and applied it to 3D. Kim et al. (2021) performed the op- dating the corresponding gradient online for a given 2D shape
timization of structures considering materials in the problem of rather than using the trained model. Kumar and Suresh (2020) and
2D and 3D structure optimization through FEA meta-modeling of Qiu et al. (2021b) optimized the 2D design domain by clustering mi-
periodic structure. Extensions to 3D are also applied in various ar- crostructure with strain and stress information, respectively. The
eas. Greminger (2020) conducted a study of generating synthetic clustering process may be applied to various design domains.
3D voxel dataset. White et al. (2019) replaced the process of calcu- While some studies have expanded their focus to various
lating effective elasticity of 3D metal with NN, and Ha et al. (2021) 2D design domains, other studies have expanded to 3D. Chan-
optimized 3D robot grippers for 3D optimized shapes. drasekhar and Suresh (2021a, b, c), Zehnder et al. (2021), and Deng
Application to any design domain and To (2020) used the network itself as an optimization calcula-
The most important point of expanding the application area is tion tool, thereby allowing the network to be used for any design
whether the design domain can be set flexibly beyond (1) and (2) domain and any load and boundary condition, and it can even
in Section 3.5.2, which is the most important issue in the applica- be expanded to 3D. The DLTOP methodology (Kallioras et al., 2020)
Journal of Computational Design and Engineering, 2023, 10(4), 1736–1766 | 1755
and other applied versions (Kallioras & Lagaros, 2020, 2021a, b) are of high computational cost and time consumption (Blanchard-
also applicable to any design domain and various load and bound- Dionne & Martin, 2021; Doi et al., 2019; Gaymann & Montomoli,
ary conditions when optimization progression is combined with 2019; Sasaki & Igarashi, 2019b; Zhang et al., 2021a). For example,
SIMP and DL. Zheng et al. (2021) matched the stiffness tensors of Chandrasekhar and Suresh (2021a) demonstrated that their FEA
3D macroscale structures with the predicted stiffness tensors of consumed approximately 50% of the computational cost in the
microscale spinodoid meta-materials to optimize the spinodoid case of a 60 × 30 mesh size, and the percentage would likely in-
topologies of macrostructures with various load and boundary crease with increasing mesh size. The design generated in this
conditions. Xue et al. (2021) generalized the scheme to all differ- method guaranteed engineering performance and increase the
ent load and boundary conditions, and even 3D design domains, accuracy of the network, but it imposed a burden on the computa-
for the upscaling at high resolution. tional cost; this scenario implies a more significant disadvantage
when used in large-scale problems, such as 3D design problems.
Consequently, some studies have attempted to introduce meth-
4. Limitations and Research Opportunities ods such as parallel FEM computation (Deng et al., 2022), although
The previous sections have verified that studies are being con- more research is needed to utilize FEA while attempting to effec-
Improvement
Research Element size TO time Training Data generation MLTO (times faster) Computational resources
2020; Chandrasekhar & Suresh, 2021b, c; Halle et al., 2021; Jiang 4.2.2. Boundary conditions
& Fan, 2019a; Kumar & Suresh, 2020; Liu et al., 2015; Napier et al., Many studies are applicable to only one design problem (Asanuma
2020; Nie et al., 2021; Oh et al., 2019; Qian & Ye, 2021; Qiu et al., et al., 2020; Ha et al., 2021; Hayashi & Ohsaki, 2020; Jang et al., 2022;
2021b; Sharpe & Seepersad, 2019; Ulu et al., 2016; Wang et al., 2020, Jiang & Fan, 2019a, b; Jiang et al., 2019; Wang et al., 2020, 2021; Wen
2021; Yildiz et al., 2003; Yu et al., 2019; Zhou & Saitou, 2017). In other et al., 2019; Zhang et al., 2019; Zheng et al., 2021; Zhou & Saitou,
words, the works mentioned above are not yet applicable in the ac- 2017), and in most cases, the initial design domain has to be fixed
tual product development stage (Bujny et al., 2018; Oh et al., 2019; (Nie et al., 2021; Sharpe & Seepersad, 2019) and cannot be applied
Wang et al., 2021). In reality, the 3D design domain must be applied, to new boundary conditions or loading conditions (Bujny et al.,
especially in large-scale problem. Furthermore, the TO process is 2018; Halle et al., 2021; Hayashi & Ohsaki, 2020; Lew & Buehler,
more time-consuming, and the computational cost is expensive, 2021; Malviya, 2020; Rawat & Shen, 2018; Zheng et al., 2021). Fur-
so the application of MLTO is an urgent research topic. However, as thermore, additional studies on complex problems (Kallioras &
the 3D MLTO problem entails numerous design parameters, more Lagaros, 2021a; Li et al., 2019b, 2021; Lynch et al., 2019), such as
training data are required, which causes the problem of computa- multi-objective problems or various objective functions (Bujny et
tional cost attributable to data collection and simulation analysis. al., 2018; Chandrasekhar & Suresh, 2021a, c; Jiang & Fan, 2019a;
Therefore, MLTO can be used while supplementing this problem. Kumar & Suresh, 2020; Qiu et al., 2021b; Sun & Ma, 2020), robust
Journal of Computational Design and Engineering, 2023, 10(4), 1736–1766 | 1757
TO (Chi et al., 2021), non-linear TO (Halle et al., 2021; Qian & Ye, fore, more studies are needed to predict manufacturable opti-
2021), and MMTO (Kim et al., 2021; Li et al., 2019c; Napier et al., mization results considering complex manufacturing constraints
2020), are needed. (Karlsson et al., 2021; Wang et al., 2020; Zhang et al., 2019), such
as AM conditions (Almasri et al., 2020) and injection molding
(Greminger, 2020).
4.2.3. Industrial applicability
Scalability is a crucial factor directly related to the usability of
4.3.3. Robust learning performance
MLTO in industrial applications. From an industrial perspective,
an adaptive tool with excellent generalizability to solve various Whether the network used for MLTO has robust quality should
problems is urgently needed. Most MLTO methods require a large be verified and improved. This issue is also linked to the cherry-
amount of pre-optimized data, such as those in the studies pre- picking problem, showing only one-time good results frequently
sented in Section 2.2. However, collecting sufficient data, espe- occurring in ML/DL research. As some of the networks are unsta-
cially those used in the industry, takes time. Furthermore, the ble, additional research on the convergence of the network (Gill-
computational complexity of supervised MLTO limits its use in hofer et al., 2019; Li et al., 2021; Takahashi et al., 2019) is needed. Fur-
the industry, as explained in Section 4.1. Considering that these thermore, scholars have attempted to additionally apply various
Design (CSCWD)(pp. 10–15). IEEE. https://doi.org/10.1109/CSCWD. Ha, H., Agrawal, S., & Song, S. (2021). Fit2Form: 3D generative model
2016.7565954. for robot gripper form design. In Proceedings of the Conference on
Cover, T., & Hart, P. (1967). Nearest neighbor pattern classification. Robot Learning(pp. 176–187). PMLR.
IEEE Transactions on Information Theory, 13(1), 21–27. https://doi.or Halle, A., Campanile, L. F., & Hasse, A. (2021). An artificial
g/10.1109/TIT.1967.1053964. intelligence–Assisted design method for topology optimization
Deng, H., & To, A. C. (2020). Topology optimization based on deep rep- without pre-optimized training data. Applied Sciences, 11(19), 9041.
resentation learning (DRL) for compliance and stress-constrained https://doi.org/10.3390/app11199041.
design. Computational Mechanics, 66(2), 449–469. https://doi.org/10 Hayashi, K., & Ohsaki, M. (2020). Reinforcement learning and
.1007/s00466- 020- 01859- 5. graph embedding for binary truss topology optimization
Deng, C., Wang, Y., Qin, C., Fu, Y., & Lu, W. (2022). Self-directed online under stress and displacement constraints. Frontiers in
machine learning for topology optimization. Nature Communica- Built Environment, 6, 59. https://doi.org/10.3389/fbuil.2020.00059.
tions, 13(1), 388. https://doi.org/10.1038/s41467- 021- 27713- 7. Hoyer, S., Sohl-Dickstein, J., & Greydanus, S. (2019). Neural repa-
Doi, S., Sasaki, H., & Igarashi, H. (2019). Multi-objective topology op- rameterization improves structural optimization. arXiv preprint
timization of rotating machines using deep learning. IEEE Trans- arXiv:1909.04240.
actions on Magnetics, 55(6), 1–5. https://doi.org/10.1109/TMAG.201 Huang, X., & Xie, Y. M. (2007). Convergent and mesh-independent
Keshavarzzadeh, V., Alirezaei, M., Tasdizen, T., & Kirby, R. M. (2021). Li, H., Kafka, O. L., Gao, J., Yu, C., Nie, Y., Zhang, L., Tajdari, M., Tang,
Image-based multiresolution topology optimization using deep S., Guo, X., Li, G., Tang, S., Cheng, G., & Liu, W. K. (2019b). Clus-
disjunctive normal shape model. Computer-Aided Design, 130, tering discretization methods for generation of material perfor-
102947. https://doi.org/10.1016/j.cad.2020.102947. mance databases in machine learning and design optimization.
Keshavarzzadeh, V., Kirby, R. M., & Narayan, A. (2020). Stress-based Computational Mechanics, 64, 281–305. https://doi.org/10.1007/s004
topology optimization under uncertainty via simulation-based 66- 019- 01716- 0.
Gaussian process. Computer Methods in Applied Mechanics and En- Liang, Q. Q., Xie, Y. M., & Steven, G. P. (2000). Optimal topology se-
gineering, 365, 112992. https://doi.org/10.1016/j.cma.2020.112992. lection of continuum structures with displacement constraints.
Kim, J. E., Cho, N. K., & Park, K. (2022). Computational homoge- Computers & Structures, 77(6), 635–644. https://doi.org/10.1016/S0
nization of additively manufactured lightweight structures with 045- 7949(00)00018- 3.
multiscale topology optimization. Journal of Computational Design Limkilde, A., Evgrafov, A., & Gravesen, J. (2018). On reducing compu-
and Engineering, 9(5), 1602–1615. https://doi.org/10.1093/jcde/qwa tational effort in topology optimization: We can go at least this
c078. far!. Structural and Multidisciplinary Optimization, 58, 2481–2492.
Kim, C., Lee, J., & Yoo, J. (2021). Machine learning-combined topology https://doi.org/10.1007/s00158- 018- 2121- 1.
optimization for functionary graded composite structure design. Lin, Q., Hong, J., Liu, Z., Li, B., & Wang, J. (2018). Investigation into
Mlejnek, H. P. (1992). Some aspects of the genesis of structures. Struc- ential equations. Journal of Computational Physics, 378, 686–707.
tural Optimization, 5, 64–69. https://doi.org/10.1007/BF01744697. https://doi.org/10.1016/j.jcp.2018.10.045.
Mnih, V., Badia, A. P., Mirza, M., Graves, A., Lillicrap, T., Harley, T., Silver, Rawat, S., & Shen, M. H. (2018). A novel topology design approach
D., & Kavukcuoglu, K. (2016). Asynchronous methods for deep re- using an integrated deep learning network architecture. arXiv
inforcement learning. In Proceedings of the International Conference preprint arXiv:1808.02334.
on Machine Learning(pp. 1928–1937). PMLR. Rawat, S., & Shen, M. H. (2019a). Application of adversarial networks
Mnih, V., Kavukcuoglu, K., Silver, D., Graves, A., Antonoglou, I., Wier- for 3D structural topology optimization. SAE Technical Paper (No.
stra, D., & Riedmiller, M. (2013). Playing Atari with deep reinforce- 2019-01-0829). SAE International. https://doi.org/10.4271/2019-0
ment learning. arXiv preprint arXiv:1312.5602. 1-0829.
Mnih, V., Kavukcuoglu, K., Silver, D., Rusu, A. A., Veness, J., Bellemare, Rawat, S., & Shen, M. H. H. (2019b). A novel topology optimiza-
M. G., Graves, A., Riedmiller, M., Fidjeland, A. K., Ostrovski, G., Pe- tion approach using conditional deep learning. arXiv preprint
tersen, S., Beattie, C., Sadik, A., Antonoglou, I., King, H., Kumaran, arXiv:1901.04859.
D., Wierstra, D., Legg, S., & Hassabis, D. (2015). Human-level con- Ringnér, M. (2008). What is principal component analysis?. Nature
trol through deep reinforcement learning. Nature, 518(7540), 529– Biotechnology, 26(3), 303–304. https://doi.org/10.1038/nbt0308-30
533. https://doi.org/10.1038/nature14236. 3.
Sigmund, O. (2001b). Design of multiphysics actuators using topology Watkins, C. J., & Dayan, P. (1992). Q-learning. Machine Learning, 8, 279–
optimization–Part II: Two-material structures. Computer Methods 292. https://doi.org/10.1007/BF00992698.
in Applied Mechanics and Engineering, 190(49-50), 6605–6627. https: Wen, F., Jiang, J., & Fan, J. A. (2019). Progressive-growing of generative
//doi.org/10.1016/S0045- 7825(01)00252- 3. adversarial networks for metasurface optimization. arXiv preprint
Sigmund, O. (2007). Morphology-based black and white filters for arXiv:1911.13029.
topology optimization. Structural and Multidisciplinary Optimization, Wen, F., Jiang, J., & Fan, J. A. (2020). Robust freeform metasurface
33(4-5), 401–424. https://doi.org/10.1007/s00158- 006- 0087- x. design based on progressively growing generative networks. ACS
Sigmund, O., & Torquato, S. (1997). Design of materials with extreme Photonics, 7(8), 2098–2104. https://doi.org/10.1021/acsphotonics.0
thermal expansion using a three-phase topology optimization c00539.
method. Journal of the Mechanics and Physics of Solids, 45(6), 1037– White, D. A., Arrighi, W. J., Kudo, J., & Watts, S. E. (2019). Multiscale
1067. https://doi.org/10.1016/S0022- 5096(96)00114- 7. topology optimization using neural network surrogate models.
Sim, E. A., Lee, S., Oh, J., & Lee, J. (2021). GANs and DCGANs for gener- Computer Methods in Applied Mechanics and Engineering, 346, 1118–
ation of topology optimization validation curve through cluster- 1135. https://doi.org/10.1016/j.cma.2018.09.007.
ing analysis. Advances in Engineering Software, 152, 102957. https: Wu, Z., Wang, S., Xiao, R., & Yu, L. (2020). A local solution approach for
//doi.org/10.1016/j.advengsoft.2020.102957. level-set based structural topology optimization in isogeometric
Zhang, C., Liu, J., Yuan, Z., Xu, S., Zou, B., Li, L., & Ma, Y. (2021b). A
novel lattice structure topology optimization method with ex-
treme anisotropic lattice properties. Journal of Computational De-
sign and Engineering, 8(5), 1367–1390. https://doi.org/10.1093/jcde
/qwab051.
Zhang, Y., Peng, B., Zhou, X., Xiang, C., & Wang, D. (2019). A deep con-
volutional neural network for topology optimization with strong
generalization ability. arXiv preprint arXiv:1901.07761.
Zhao, J., Mathieu, M., & LeCun, Y. (2016). Energy-based generative ad-
versarial network. arXiv preprint arXiv:1609.03126.
Zheng, L., Kumar, S., & Kochmann, D. M. (2021). Data-driven topology
optimization of spinodoid metamaterials with seamlessly tun-
able anisotropy. Computer Methods in Applied Mechanics and Engi-
neering, 383, 113894. https://doi.org/10.1016/j.cma.2021.113894.
Zhou, Y., & Saitou, K. (2017). Topology optimization of composite
tive function and constraints for the design variable is required in from the structure to find the optimal structure. BESO, an ad-
each iteration. This optimization algorithm for maximizing stiff- vanced method of the ESO, was introduced in Yang et al. (1999),
ness performs sensitivity analysis to determine the effect of ele- which removes unnecessary materials while adding necessary
materials. The optimization problem that minimizes compliance
ment density changes on the objective function and constraints.
with the most basic ESO is defined as follows:
Therefore, sensitivity analysis can be expressed as a differential of
min c (x ) = U T KU
the objective function and the constraint for the element density V (x ) (A.11)
s.t. V0
=f KU = F xe = [ 0, 1] , e = 1, . . . , Ne
under the setting that all elements have unit volume.
At this time, the criteria for removing elements are based on the
∂C
= − p(xe ) p−1 [ue ]T [Ke ] [ue ] (A.4) sensitivity number, and the sensitivity number can be obtained
d∂
N through the following equation (Liang et al., 2000).
∂V ∂ e
= xe ve = 1 (A.5) 1 T
∂ xe ∂ xe αie =
u Ki ui (A.12)
e=1 2 i
The OC method is a method commonly used to update design Through this, the sensitivity number is calculated to remove
variables in TO problems, and is most commonly applied to the elements with elements lower than the criteria.
density-based method. The method of updating design variables
where k is an arbitrary value containing a zero function value and variables with di = (x0i, y0i , Li , ti , θi )T , i = 1, . . . nc. Ud is the admissi-
x is an arbitrary coordinate on the curved surface φ. As the struc- ble sets that d belongs to. Moreover, gi , i = 1, ..., l are the consid-
tural optimization changes over time, Equation (A.17) can be ex- ered constraint functionals.
pressed as level set function dynamically changing as follows:
method of grouping data with similar features into the same clus- training the decoder to be used as a generative model, training is
ter. As an example, K-means is the simplest and most popular al- performed by attaching an encoder.
gorithm for iterative and hill climbing clustering algorithms. The
K-means is an algorithm that groups given data into k clusters, A2.3. Reinforcement learning
which are clustered based on the center of each cluster and the RL is a type of ML that makes decisions through the interaction
average distance of the data. K-means were originally designed to of agents and environments (Sutton & Barto, 2018). RL is usually
find the optimal partition for dividing data, but k-means generally based on the MDP, but it does not require explicit models.
do not converge to the optimal partition (MacQueen, 1967). The agent takes an action at in the current time t, which returns
Dimensionality reduction is the process of extracting high- state st and reward Rt from the environment ε. The algorithm is
dimensional data into low-dimensions, which still compresses trained through trial-and-error so that the agent determines the
meaningful features of the data. For example, PCA is a mathemat- action in the direction of maximizing the total reward and returns
ical algorithm that reduces the dimension of data while main-
Rt = γ k · Rt+k received by the agent in the state of the given en-
taining the variations of most of the data, and the dimension is k
vironment (where γ ∈ (0, 1 ), γ is a discount factor that reflects
reduced by finding a basis that maximizes the variance (Jolliffe,
future rewards in the current state). The two main approaches
Received: October 16, 2022. Revised: June 30, 2023. Accepted: July 1, 2023
© The Author(s) 2023. Published by Oxford University Press on behalf of the Society for Computational Design and Engineering. This is an Open Access article distributed
under the terms of the Creative Commons Attribution-NonCommercial License (https://creativecommons.org/licenses/by-nc/4.0/), which permits non-commercial re-use,
distribution, and reproduction in any medium, provided the original work is properly cited. For commercial re-use, please contact [email protected]