Learning Two-phase Microstructure Evolution Using Neural Operators and Autoencoder Architectures
Learning Two-phase Microstructure Evolution Using Neural Operators and Autoencoder Architectures
com/npjcompumats
ARTICLE OPEN
Phase-field modeling is an effective but computationally expensive method for capturing the mesoscale morphological and
microstructure evolution in materials. Hence, fast and generalizable surrogate models are needed to alleviate the cost of
computationally taxing processes such as in optimization and design of materials. The intrinsic discontinuous nature of the physical
phenomena incurred by the presence of sharp phase boundaries makes the training of the surrogate model cumbersome. We
develop a framework that integrates a convolutional autoencoder architecture with a deep neural operator (DeepONet) to learn the
dynamic evolution of a two-phase mixture and accelerate time-to-solution in predicting the microstructure evolution. We utilize the
convolutional autoencoder to provide a compact representation of the microstructure data in a low-dimensional latent space. After
DeepONet is trained in the latent space, it can be used to replace the high-fidelity phase-field numerical solver in interpolation tasks
or to accelerate the numerical solver in extrapolation tasks.
npj Computational Materials (2022)8:190 ; https://doi.org/10.1038/s41524-022-00876-7
1234567890():,;
1
School of Engineering, Brown University, Providence, RI, USA. 2Division of Applied Mathematics, Brown University, Providence, RI, USA. 3Center for Integrated Nanotechnologies,
Sandia National Laboratories, Albuquerque, NM 87185, USA. ✉email: [email protected]; [email protected]
Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences
V. Oommen et al.
2
1234567890():,;
Fig. 1 Schematic representation of DeepONet with convolutional autoencoder. Step 1 involves training of the convolutional autoencoder
to minimize Lae . The encoder learns a suitable transformation from the high-dimensional microstructure to a low-dimensional latent space
through a series of convolution (blue layers) and MaxPooling (green layers) operations. The decoder remaps the latent representation of the
microstructure back to the original, real space by performing transpose convolution (orange layers) operations. A detailed description of the
architecture is provided in Table 1. In step 2, we train the DeepONet in the latent space to minimize Ld . The entire history of 80 steps is
encoded by the pre-trained convolutional encoder as Φ. ~ DeepONet learns to predict ϕðtÞ
~ at any desired time t, fed to the trunk network. The
latent representation of the microstructure predicted by DeepONet is then re-mapped back to the primitive space by the transpose
convolutional decoder.
function G : u ! ϕ such that, (ordinary differential equations (ODEs) and PDEs) of such problems
by learning the underlying operator, a mapping from functions to
Gðuðx; y; tÞÞ ¼ ϕðx; y; tÞ; (1)
functions, from the available datasets for a broad range of
where u is the history of the microstructure evolution and ϕ(x, y, t) problems28,42. We show that such an architecture is more robust
is the state of the microstructure at time t. We develop a than the RNN-based architecture in terms of training, computational
framework that integrates a convolutional autoencoder architec- efficiency, and sensitivity to noise. The decoder part of the
ture with a Deep Operator Network41 (DeepONet) to learn this convolutional autoencoder can efficiently reconstruct the time-
mapping. Figure 1 illustrates the complete end-to-end workflow of evolved microstructure from the DeepONet predictions bypassing
the proposed algorithm. We utilize a convolutional autoencoder to the challenges associated with reconstruction-induced errors when
provide a compact representation of the microstructure data in a using statistical functions to represent the microstructure for
low-dimensional, latent space. This convolutional autoencoder instance. Overall, the trained autoencoder–DeepONet framework
approach is then combined with the DeepONet architecture to can then be used to replace the high-fidelity phase-field numerical
learn the dynamics of two-phase microstructures in the auto- solver in interpolation tasks for parameters inside the distribution of
encoder latent space. The DeepONet architecture has demon- inputs used during training or to accelerate the numerical solver in
strated its ability to model the governing differential equations extrapolation tasks for parameters outside this distribution.
npj Computational Materials (2022) 190 Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences
V. Oommen et al.
3
RESULTS information about the history through the branch network. In
Training and optimization of neural operators and addition, due to the lack of temporal dependencies, the fully
autoencoder architectures connected layers in the trunk network and convolutional layers in
We first investigated the impact of the size of the latent dimension the branch network of the DeepONet architecture can be easily
of the autoencoder, ld, on the model performance. To this end, we parallelized, unlike LSTMs. This makes training and inference of
trained five autoencoder models with ld = 9, 25, 64, 100, and 196 DeepONet significantly faster than the RNN architectures.
respectively. Details of the hyper-parameters used in these five We carried out additional simulations to analyze the sensitivity of
convolutional autoencoders are provided in Table 1. For any given the proposed approach to the number of samples used for training.
time step during the evolution of the microstructure, the encoder We considered training datasets with 25%, 50%, 75% and all the
reduced a 128 × 128 microstructure ϕ(x, y, t) to a latent vector of 1600 training samples. We adopted the same methodology
size ld. The decoder mapped the microstructural latent space proposed in Methods and trained separate autoencoder–DeepONet
representation back to a 128 × 128 microstructure ϕðx; ^ y; tÞ (see models on each of these datasets. Details can be found in
Methods for more details). Each autoencoder training took Supplementary Note 2. The model performance was evaluated on
approximately 33 h on one NVIDIA GeForce RTX 3090 GPU. Next, the basis of forecasting errors on test data, shown in Supplementary
we trained the DeepONet model for 120,000 epochs on the latent Figure 2. As expected, we observe better accuracy in the model
space learned by the convolutional encoder for each of the five predictions when increasing the number of training samples. The
trained autoencoder models. The last layer of the branch and model trained with 1200 data samples shows similar accuracy to the
trunk networks for all the models uses a linear activation function. best model trained with 1,600 data samples, indicating convergence
The output of the DeepONet model was then sent to the trained of the training procedure.
convolutional decoder, which performed a mapping from the Finally, we also evaluated the effect of using different loss
latent space back to the original microstructure space, ϕðx;^ y; tÞ. functions on training the autoencoder models. Specifically, we
We evaluated the effect of the size of the latent dimension of trained various autoencoders by minimizing L1 loss, relative L1
each of the models on the basis of the relative L2 norm computed loss, L2 loss, relative L2 loss, and mixed loss (L2 loss for the initial
across the training and testing dataset for all the time steps, 5000 epochs and L1 loss for the remaining epochs). The choice of a
including the forecasting time frames, t = {t90, … , t99} not seen by loss function determines the landscape in a hyperspace for the
the surrogate model (Note that one time frame is equal to 500,000 optimizer to traverse in pursuit of global minima/best local
time steps, t = 500,000Δt, see Methods for additional details). All minima and avoiding the saddle points. For this task, we used a
the details of this survey analysis, including the DeepONet DeepONet model with architecture 1 (Table 2) on each of the
architecture, the L2 norm of relative error on train and test learned latent microstructure data and re-transformed the
datasets, and the computational time taken for training the DeepONet predictions using a pre-trained decoder to retrieve
DeepONet model are reported in Table 2. From this survey, we the microstructure. We analyzed the model performance by
observe that the model predictions improve when we increase the computing the forecasting error, Dtest ðtÞ, on unseen test data, as
size of the latent dimension. In general, DeepONet models with shown in Supplementary Note 3. We observed that models for
tanh and sin activation functions performed better compared to which the autoencoder is trained on L2 loss performed better than
models with a ReLU activation for this particular class of problems. the one which used L1 loss. When there are no outliers as
As such, our best model consists of a convolutional autoencoder solutions, L2 loss is expected to perform better than L1. In the
with ld = 196 and a DeepONet model with architecture 1 and sin presence of outliers, L2 squares them as compared to linear
activation function (shown in Table 2). Although the training contribution in the L1 norm. Similarly, mean values of relative L1
dataset consists of 1,600 different microstructure-evolution and L2 are a better choice for autoencoder loss, Lae , than the
trajectories, each represented by over 80 snapshots from t = {t10, mean of L1 and L2, respectively. The relative loss values are always
t11, … , t89}, the DeepONet training is faster compared to popular of Oð1Þ and help in achieving convergence faster as the learning
RNN architectures such as the Long Short-Term Memory (LSTM) or rate is of Oð103 Þ. We have observed such an improvement in
Gated Recurrent Unit (GRU) networks38,42,43. Since DeepONet does convergence in other problems, e.g. electro-convection, where we
not have recurrent connections, there are no temporal depen- had sharp interfaces and multiscale dynamics44. Overall, all the
dencies during the training or at the inference stage. Instead, the models performed consistently well. As such, our autoencoder
network relies on the convolution operations that encode architecture of choice is an autoencoder with ld = 196 using a
Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences npj Computational Materials (2022) 190
V. Oommen et al.
4
Table 2. Detailed survey of different latent dimension size, ld, network architecture, and non-linear activation functions.
relative L2 loss function. Taken together, these results demonstrate high for the initial time steps, where features span multiple length
not only the ability of our framework to accurately provide a scales and evolve rapidly with time. However, the predictions
compact representation of the microstructure data in a low- improve over time when the evolution process slows down and
dimensional latent space, but they also illustrate the robustness of the microstructure features coarsen. The time steps shown in
the training of this framework. Fig. 3a were used during the training of the model.
Next, we evaluated the capability of the model to forecast time
Performance accuracy and forecasting ability frames t = {t90, t91, … , t99}. From Fig. 3b, the error is seen to
increase gradually when the model extrapolates at unseen time
A comparison of the predictions from our accelerated framework
instances. A closer look at the forecasting predictions offers
with that from high-fidelity phase-field simulations for a represen-
further insights into the DeepONet predictive performance. We
tative case of microstructure evolution at three different time steps
computed the mean of DðtÞ across the training and testing
is shown in Fig. 2. During the initial time steps, the microstructure is
datasets for all the models given in Table 2. We also plotted these
rich with multiple features and evolves rapidly with respect to time.
values in Supplementary Fig. 4 with additional details in
Our autoencoder–DeepONet as a surrogate model is able to
Supplementary Note 4. We observe that the mean relative L2
successfully predict the larger features and the overall morphology
error reduces when increasing the latent dimension of the
of the microstructure. The point-wise error snapshots suggest that
autoencoder model. In other words, the model with a larger
the model fails to identify the relatively smaller features in the
latent space is able to predict the evolution of the microstructure
microstructure and contains significant errors along the sharp
in forecasting mode. This is intuitive because a larger dimension of
boundaries. In other words, the spatial gradient of the phase
the latent space implies that there are more basis functions to
concentration is not as sharp as that of the true microstructure
express the encoded information about the microstructure and its
obtained from high-fidelity, phase-field simulations.
evolution, and therefore the network has an improved represen-
From Fig. 2, we qualitatively get the intuition that the predicted
tation capability. However, this trend seems to saturate beyond
microstructures contain errors at the earlier time steps because of
ld = 100. For the model with ld = 100 or ld = 196, the forecasting
the missing, small-size features and at the later time steps due to
error is always less than 6%. The logarithm of the relative L2 error
smoother boundaries predicted by the model. To confirm and
linearly increases for ld = 64,100 and 196 for the forecasting time
quantify this notion, we computed the L2 norm of the relative
error at each time step, DðtÞ, defined as: step, whereas for ld = 9 and ld = 25, the error is high and remains
constant.
PP 2
^
x y ϕðx; y; tÞ ϕðx; y; t; θÞ
DðtÞ ¼ PP 2 ; t 2 ½t10 ; t 11 ; ¼ ; t99 : Robustness of the surrogate DeepONet framework: Sensitivity
x y ϕð x; y; t Þ
to noise
(2) We evaluated the accuracy and robustness of the predictions from
To analyze the accuracy of the prediction at each time step, we our surrogate model by systematically increasing the noise levels
calculated DðtÞ across the samples in the training and testing in the model input. For this analysis, we considered the best
datasets, and created a boxplot as shown in Fig. 3. The error is model with ld = 196 and DeepONet architecture 1 (see Table 2)
npj Computational Materials (2022) 190 Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences
V. Oommen et al.
5
Fig. 2 Predictions of microstructure evolutions. The true (top row), predicted (middle row), and point-wise error (bottom row) for a
microstructure realization evolving in time. The snapshots at time frames t = t10, t30, t99 are shown here. The network used for this simulation
has ld = 196, and uses the following network architecture: Branch network -- 2 × [conv(128, (3, 3))] + [3920]; Trunk network -- 2 × [100] + [3920].
with sin activation functions. We added Gaussian white noise to are sufficient to capture the salient features of the microstructure
our microstructure data with zero mean and standard deviations, in latent space, such representation does not uniquely map back
σ = 0.5%, 1%, 2%, 3%, 4%, 5%, 10%. To evaluate the model to the true microstructure as it is an ill-posed, inverse problem.
performance, we used the relative L2 norm, D, as defined in Eq. Here, our results essentially bypass such a challenge by taking
(8). The forecasting error, D, is calculated across the samples advantage of the fact that autoencoders are robust to corruption
present in the test dataset. Details can be found in Supplementary in the representations they learn. The denoising nature of a
Note 5. trained autoencoder enables the encoder to learn a stable and
From Supplementary Table 1 and Supplementary Figure 5, the consistent mapping to the latent space. This makes training of the
relative L2 norm does not increase noticeably when noise is added DeepONet much more stable and results in accurate predictions
to the model input. In fact, the surrogate is almost invariant to of the microstructure at any desired time step.
noise up to 10% Gaussian white noise, as presented in
Supplementary Figure 5. Previous studies45–47 illustrated the
Effect of time resolution
capability of autoencoders to denoise noisy images. The
transformation to a low-dimensional latent space forces the The high-fidelity phase-field forward numerical solver (MEMPHIS)
autoencoder to retain the dominant features alone while discretizes the time with a time step Δt = 1 × 10−4 (see Methods
discarding unnecessary noise. The convolutional autoencoder for additional details). The stability of this numerical integration
used in our approach does exactly that by denoising the noisy scheme can be achieved by strictly following the Courant-
microstructure input. The encoder filters out noise and only Friedrichs-Lewy (CFL) condition to solve the Cahn-Hilliard
retains the dominant energy modes of the microstructure data. equation for 50M time steps. The solver saves snapshots of the
The output of the convolutional encoder is almost in its pure form, solution at every 500,000th time step resulting in 100 micro-
free from noise and therefore it enables the DeepONet to make structure time frames for each realization. We initially utilized 80
stable predictions. The decoder accurately reconstructs the equally spaced time frames between the 10th and 90th time
microstructure from the predictions made by DeepONet in the frames for training the surrogate model. Therefore, for the
latent space. This performance illustrates the robustness and surrogate DeepONet model, each time step was 500000 × Δt = 50.
efficacy of the present framework as compared to other machine- To investigate the effect of different spacing of physical time on
learned frameworks that use statistical functions to encode the the surrogate DeepONet model, we performed a sensitivity study
microstructure representation40. Indeed, it has been illustrated by using data that are spaced differently in time to train the model.
Herman and coworkers40 and others48 that, while statistical Specifically, we trained DeepONet models on datasets with a time
functions such as the microstructure auto-correlation functions spacing of 500kΔt, 2 × 500kΔt, 5 × 500kΔt, and 10 × 500kΔt. The
Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences npj Computational Materials (2022) 190
V. Oommen et al.
6
npj Computational Materials (2022) 190 Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences
V. Oommen et al.
7
Fig. 5 Capturing small microstructural features at earlier time steps. a True microstructure at t = t10. b Predicted microstructure without
emphasizing earlier time steps during training. c Microstructure predicted by DeepONet trained on a dataset where the earlier time steps
where repeated to increase the importance given to earlier time steps.
‘leap in time’. The algorithm is presented in Algorithm 1. Here we For instance, one could easily consider to use very short time steps
choose to split the time evolution predicted between the high- within the high-fidelity phase-field solver window to course-
fidelity simulation and those of the autoencoder–DeepONet to be correct the physical predictions and much longer time steps when
equal to one another. Each solver within this integrated scheme using DeepONet to accelerate the time evolution predictions. This
sequentially predicts 10-time frames, which corresponds to type of time splitting integration scheme would dramatically
10 × 500k = 5M time steps for the high-fidelity phase-field solver increase the speedup even further. Additionally, although we
alone. A schematic of the approach and results are shown in Fig. 6 showed the microstructure evolution only until t = t105, such a
for the training data. Results for the test data are provided in hybrid time integration strategy can be adopted to forecast time
Supplementary Fig. 6. The discontinuity along the centerlines of evolution of the microstructure for time windows that can be
the predictions made by the autoencoder–DeepONet model arises much longer, for instance as long as the input time history used to
from splitting each realization into four different realizations, as train the DeepONet model while still keeping a good accuracy,
discussed in the ‘Microstructure-evolution dataset’ subsection. leading to substantial savings in CPU hours.
Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences npj Computational Materials (2022) 190
V. Oommen et al.
8
Fig. 6 Schematic of our hybrid approach for integrating DeepONet with the numerical solver MEMPHIS to accelerate phase-field
predictions. The computational time corresponding to the autoencoder--DeepONet model and the MEMPHIS solver for one realization in the
training dataset is reported in this figure. The error is shown on the third column.
npj Computational Materials (2022) 190 Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences
V. Oommen et al.
9
high-gradient regions in the microstructure data (see Fig. 7a) can Second, even though the workflow presented in this work is
make the training of the model challenging. However, the encoder focused on two-dimensional (2D) microstructure data, it can easily
transforms microstructure data to a latent space (Fig. 7b), where be extended to three-dimensional (3D) microstructure data. For
the gradients are not as high and more gradual . In other words, the 3D microstructure evolution case, each realization can be
the encoder learns a non-linear mapping of the microstructure represented by a sequence of 3D tensor data structures. Hence,
data coming from an untrainable distribution in the primitive we could use an autoencoder with 3D convolution layers in the
space, as shown in Fig. 7c, to a trainable distribution in the low- encoder and 3D transpose convolution layers in the decoder and
dimensional latent space, as shown in Fig. 7d. Such a transforma- learn a suitable non-linear mapping to a low-dimensional latent
tion from data with a high gradient to data with gradual, manifold. Following the same approach, a DeepONet model can
smoother gradient facilitates the training of the DeepONet model. be trained to learn the dynamics in the latent space, and be then
Although we have trained our own autoencoder model in this remapped to primitive space using the already trained decoder.
work, we believe that fine-tuning any autoencoder pre-trained on As shown in a recent theoretical paper, DeepOnet can tackle the
existing image datasets with similarities to our microstructure curse of dimensionality in the input space, so training it in
images will be suitable for this task. Reusing such readily available highdimensions is not a prohibitive issue49.
Third, the autoencoder and DeepONet are trained solely from
pre-trained autoencoders can further save on the computational
data, making the proposed approach purely data-driven, inde-
cost of our workflow.
pendent of the boundary conditions. The boundary conditions are
never explicitly assumed as input data to the framework at any
stage. They are implicitly fed through the latent representation of
Table 3. A comparison between computational time for high-fidelity
phase-field simulations (MEMPHIS) and proposed hybrid model
the microstructure history data inputted to the branch network of
(Hybrid) for a single microstructure evolution realization from time
DeepONet. Therefore, any information regarding a change in the
frame t1 to time frame t105. boundary condition will be available in the latent microstructure
history fed to the branch network, enabling the DeepONet model
Computational time to predict the dynamics accordingly in the latent manifold. In this
manner, the purely data-driven nature makes the proposed
Time frames MEMPHIS Hybrid autoencoder–DeepONet framework agnostic to any changes in
t1 to t30 135 mins 135 mins boundary conditions within the considered history. We also need
to clarify that periodic boundary conditions were imposed at all
t30 to t40 45 mins 2 sec
four boundaries of the computational domain while generating
t40 to t60 90 mins 90 mins the data from the numerical solver MEMPHIS. We note that
t60 to t70 45 mins 2 sec DeepONet can be trained to map boundary conditions to an
t70 to t90 90 mins 90 mins output field if so desired for a very general set of variable
t90 to t100 45 mins 2 sec boundary conditions.
t100 to t105 23 mins 23 mins There are several extensions to the present framework that can
be implemented in order to improve the accuracy, predictability,
Total time 473 mins 338.1 mins
and acceleration performance. These improvements are related to
Fig. 7 Representation of microstructure data and statistical insights. a represents a 3D visualization of the function to be approximated by
the surrogate model. The presence of several high-gradient regions at every time-step, makes it challenging for neural network models to
learn the evolution dynamics of microstructures. Panel (b) represents a smoother latent-microstructure learned by the encoder during the
autoencoder training. c The microstructure data, ϕ(x, y, t), is predominantly represented by 1s or 0s. d The encoder transforms ϕ to a latent
~ where deep neural networks can learn easily. The curves in (c) and (d) represent the smoothed density estimates of the histogram.
space, ϕ,
Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences npj Computational Materials (2022) 190
V. Oommen et al.
10
the training of the model and extension to a multi-fidelity present model, both the mobility and the interfacial energy are taken to
implementation. The first topic is related to improving the be isotropic and ωc and κc are stet to unity for simplicity. The evolution of
accuracy of the model with physics constraints in order to better one phase is expressed as a symmetric double-well potential, with minima
capture non-linearities in the model evolution. The second topic is at ϕ ± 1.
related to fusing different sources of data within our dataset, e.g.,
‘low’ fidelity from simulation and ‘high’ fidelity from physical Microstructure-evolution dataset
experiments of the same nature. The phase-field model described above is implemented using Sandia’s in-
Regarding a physics-informed implementation, Wang et al.50 for house multi-physics phase-field modeling code MEMPHIS10,54. In order to
instance put forward a physics-informed DeepONet, where the generate a diverse and large set of simulation results exhibiting a rich
PDE of the underlying system is added as a soft constraint to the variety of microstructure features, we independently sampled the phase
loss functions. In the present study, the training framework is fraction ϕA, such that each phase has at least a minimum concentration of
purely data-driven and we are learning the dynamical system in 0.15 (note that ϕB = 1 − ϕA), and the phase mobilities MA and MB of
species ‘A’ and ‘B’. Phase mobilities are sampled independently to vary in
the latent space defined by non-primitive coordinates except for
the range [0.01, 100]. In total we generated 500 triplets (ϕA, MA, MB) using
time, which is fed as an input to the trunk net. However, similar to Latin Hypercube Sampling. In the simple case of the spinodal decomposi-
Wang et al., some physical constraints could be injected into the tion, only the tuple (ϕA, MA/MB) is necessary. As demonstrated in other
current framework by using, for instance, the fact that mass ϕ is studies10,40 that share similar microstructure evolution as the spinodal
conserved at all times. decomposition, it is, however, necessary to handle (ϕA, MA, MB) separately
Regarding the multi-fidelity implementation, the present since the ratio MA/MB by itself will not be sufficient anymore to characterize
approach can be extended to incorporate experimental data the dynamics of the microstructure evolution. Herein, we frame the
coming from similar processes. For instance, recently, several present work in a broader context for generality. All the simulations were
researchers51–53 explored diverse ways of exploiting the inherent performed using a two-dimensional (2D) square domain Ω = [0, 1] × [0, 1],
correlations between datasets coming from different sources of discretized with 512 × 512 grid points, with a dimensionless spatial
data with different levels of fidelity and obtained optimal discretization of unity in either direction, and a temporal discretization
of Δt = 1 × 10−4. The simulation domain’s composition field is initialized
predictions. In this context, the data obtained from phase-field using truncated random Gaussian distribution in the range [ − 1, 1] with
models using a numerical solver can be considered as a low- μ = ϕA, and σ = 0.35. The microstructure was allowed to evolve and grow
fidelity dataset and the limited amounts of experimental for 50,000,000 time steps, saving the state of the microstructural domain
microstructure image data from similar processes can be treated every 500,000 time steps, hence a total of 100-time frames were saved
as a high-fidelity dataset. The autoencoder–DeepONet framework from each simulated case.
proposed here can be extended to generate accurate predictions In order to use the data in the proposed algorithm, we down-sampled
from a limited number of high-fidelity experimental microstruc- each snapshot of our 512 × 512 domain into four images of 256 × 256, and
ture microscopy image data, by utilizing the high correlation with later used cubic interpolation55 to further reduce the resolution to
the surplus low-fidelity phase field data. The assimilation of 128 × 128. Hence, from the 500 microstructure evolution samples, we were
experimental data in the present DeepONet architecture can be able to generate 2000 microstructure evolution samples of 128 × 128
resolution. From this dataset, we have used 1600 cases for training the
concatenated with numerical data as another realization and the DeepONet and 400 cases for testing the network accuracy. Since the
proposed workflow will remain unchanged. Merging and taking compositional field is randomly distributed spatially, the microstructure
advantage of both experimental and modeling efforts is a future has no recognizable features at the first frame t0. The quick development
direction of our research. of subdomains is then observed between frames t0 and t10, followed by a
To summarize, we developed and applied a machine-learned smooth and steady coalescence and growth of the microstructure from
framework based on neural operators and autoencoder architec- time frames t10 to t100. We have trained our proposed model based on this
tures to efficiently and rapidly predict complex microstructural observation, starting at time frame, t10, when the microstructure had
evolution problems. Such an architecture is not only computa- reached a slow and steady development regime.
tionally efficient and accurate, but it is also robust to noisy data.
The demonstrated performance makes it an attractive alternative Training the autoencoder: learning the latent microstructure
to other existing machined-learned strategies to accelerate the representation
predictions of microstructure evolution. It opens up a computa- In this work, each microstructure evolution is represented by
tionally viable and efficient path forward for discovering, under- (NT, Nx, Ny) = (80 × 128 × 128), with NT representing the number of snap-
standing, and predicting materials processes, where evolutionary shots and Nx × Ny denoting the spatial resolution along x− and y−
mesoscale phenomena are critical, such as in the optimization and direction, respectively. To handle the entire feature space (R128 ´ 128 )
design of materials problems. 16,384 distinct features are required to represent the microstructure at
each time step. Subsequently, to compute the prediction for all 1,600
microstructure evolutions, we will have 1600 × 80 × 16, 384 (≈2.5 Billion)
32-bit floating data points. Learning microstructure dynamics from such a
METHODS
high-dimensional dataset is challenging.
Phase-field model of the spinodal decomposition of a two- To circumvent issues pertaining to the data dimensionality and
phase mixture preparing the phase-field microstructure data for DeepONet training, we
We illustrate our accelerated phase-field workflow on the simplest case of explored a couple of options. First, we tried using Principal Component
the spinodal decomposition of a two-phase mixture. This model is highly Analysis (PCA) with a linear kernel, for reducing the dimensionality of the
relevant to many phase-field models. In the spinodal decomposition of a data21,56,57. The low-dimensional representation of the data obtained from
two-phase mixture uses a single order parameter, ϕ(x, t) to describe the PCA is a linear transformation of the high-dimensional data and discards
atomic fraction of solute diffusing within a matrix. The free energy of the the insignificant modes (eigen/singular) corresponding to the lower eigen/
system is expressed by the Cahn-Hilliard equation based on the Onsager singular values (λi). However, the system considered here is non-diffusive,
force-flux relationship such that which is confirmed by cumulative explained variance and energy
distribution of the system over principal modes. Therefore, using PCA for
∂ϕ
¼ ∇ Mc ðϕÞ∇½ωc ðϕ3 ϕÞ þ κ c ∇2 ϕ ; (3) reducing the dimensionality of the microstructure description could result
∂t in the loss of valuable information, if only a convenient low number of
where ωc is the height of the energy barrier between the two phases, κc is principal components are considered. A detailed explanation on PCA of
the gradient energy coefficient, and Mc denotes the concentration the microstructure dataset is presented in Supplementary Note 1. Learning
dependent mobility, with Mc = s(ϕ)MA + (1 − s(ϕ))MB. The function s a non-linear mapping from a high-dimensional to a low-dimensional latent
defines a smooth interpolation to switch from phase ‘A’ to phase ‘B’. This space is one way to compress data without losing as much information as
interpolation function is defined as sðϕÞ ¼ 14 ð2 ϕÞð1 þ ϕÞ2 . In the in the PCA. An autoencoder precisely does this by learning a non-linear
npj Computational Materials (2022) 190 Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences
V. Oommen et al.
11
transformation to a low-dimensional latent space using an encoder. The DeepONet, θd, are obtained by minimizing a loss function, ↕d, defined as:
decoder learns the mapping to retrieve initial high-dimensional data from 2
Ld ¼ min ϕðtÞ ~
~ GðΦÞðt; θd Þ2 ; (7)
its latent representation. θd
In this study, we have used a convolutional autoencoder58 with
~
where ϕðtÞ is the ground truth for the low-dimensional phase field
convolutional layers in the encoder and transpose convolutional layers in
the decoder as shown in Fig. 1. The encoder learns a nonlinear mapping of representation at time, t obtained from the convolutional encoder. The
trained DeepONet is used to predict ϕðtÞ ~ 2 Rld . The output of the
the high-dimensional microstructure data, ϕ(x, y, t), to a low-dimensional
~
latent space represented by ϕðtÞ and is expressed as DeepONet is fed into the transposed convolutional decoder to predict,
^ 2 R128 ´ 128 . The DeepONet is trained using the Adam optimizer59. The
ϕðtÞ
~
αθenc : ϕðx; y; tÞ ! ϕðtÞ; (4) implementation has been carried out using the TensorFlow framework60.
We use Xavier Initialization61 to initialize the weights of all the models.
~ ! ϕðx;
βθdec : ϕðtÞ ^ y; tÞ; (5)
Error metrics
where α and β represent the mappings performed by the encoder and the The L2 norm of relative error, D, is used as the evaluation metric to analyze
decoder, respectively. In Eq. 4, the encoder takes ϕðx; y; tÞ 2 R128 ´ 128 as the performance of each model considered in this study. D is defined as:
~ 2 Rld , where ld is the dimension of the latent
input, and maps it to ϕðtÞ
space. θenc represents the trainable parameters of the convolutional P P P P ðnÞ 2
^ ðnÞ
n x y t ϕ ðx; y; tÞ ϕ ðx; y; t; θÞ
encoder. Equation 5 represents the decoder network, which takes the D¼ P P P P ðnÞ ; (8)
~ 2 Rld as the input and predicts the
latent dimensional representation, ϕðtÞ 2
n x y t ϕ ðx; y; tÞ
^ y; tÞ 2 R128 ´ 128 , using transpose convolu-
primitive microstructure, ϕðx;
tional operations. The details of the autoencoder architecture are provided where n corresponds to the nth sample of the given dataset.
in Table 1.
θae represents the trainable parameters of the autoencoder. These
parameters are learned by minimizing the loss function, Lae , which reads DATA AVAILABILITY
The data that support the findings of this study are available from the corresponding
Lae ¼ min ^ y; t; θae Þk :
kϕðx; y; tÞ ϕðx;
2
(6) authors upon reasonable request.
2
θae ¼fθenc ;θdec g
Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences npj Computational Materials (2022) 190
V. Oommen et al.
12
15. Sun, Z. Z. A second-order accurate linearized difference scheme for the two- 40. Herman, E., Stewart, J. A. & Dingreville, R. A data-driven surrogate model to
dimensional Cahn-Hilliard equation. Math. Comput. 64, 1463–1471 (1995). rapidly predict microstructure morphology during physical vapor deposition.
16. Liu, C. & Shen, J. A phase field model for the mixture of two incompressible fluids Appl. Math. Model. 88, 589–603 (2020).
and its approximation by a Fourier-spectral method. Phys. D: Nonlinear Phenom. 41. Lu, L., Jin, P., Pang, G., Zhang, Z. & Karniadakis, G. E. Learning nonlinear operators
179, 211–228 (2003). via DeepONet based on the universal approximation theorem of operators. Nat.
17. Barrett, J. W., Blowey, J. F. & Garcke, H. Finite element approximation of the Cahn- Mach. Intell. 3, 218–229 (2021).
Hilliard equation with degenerate mobility. SIAM J. Numer. Anal. 37, 286–318 42. Lin, C. et al. Operator learning for predicting multiscale bubble growth dynamics.
(1999). J. Chem. Phys. 154, 104118 (2021).
18. Gómez, H., Calo, V. M., Bazilevs, Y. & Hughes, T. J. Isogeometric analysis of the 43. Osorio, J. D. et al. Forecasting solar-thermal systems performance under transient
Cahn-Hilliard phase-field model. Computer Methods Appl. Mech. Eng. 197, operation using a data-driven machine learning approach based on the deep
4333–4352 (2008). operator network architecture. Energy Convers. Manag. 252, 115063 (2022).
19. Chan, C. L., Anitescu, C. & Rabczuk, T. Strong multipatch C1-coupling for iso- 44. Cai, S., Wang, Z., Lu, L., Zaki, T. A. & Karniadakis, G. E. Deepm&mnet: Inferring the
geometric analysis on 2D and 3D domains. Computer Methods Appl. Mech. Eng. electroconvection multiphysics fields based on operator approximation by neural
357, 112599 (2019). networks. J. Computational Phys. 436, 110296 (2021).
20. Alikakos, N. D., Bates, P. W. & Chen, X. Convergence of the Cahn-Hilliard equation 45. Vincent, P., Larochelle, H., Bengio, Y. & Manzagol, P.-A.Extracting and composing
to the Hele-Shaw model. Arch. Ration. Mech. Anal. 128, 165–205 (1994). robust features with denoising autoencoders. In Proceedings of the 25th Inter-
21. Brough, D. B., Kannan, A., Haaland, B., Bucknall, D. G. & Kalidindi, S. R. Extraction of national Conference on Machine Learning, 1096–1103 (2008).
process-structure evolution linkages from X-ray scattering measurements using 46. Vincent, P. et al. Stacked denoising autoencoders: Learning useful representa-
dimensionality reduction and time series analysis. Integrating Mater. Manuf. tions in a deep network with a local denoising criterion. Journal of Machine
Innov. 6, 147–159 (2017). Learning Research 11 (2010).
22. Pfeifer, S., Wodo, O. & Ganapathysubramanian, B. An optimization approach to 47. Gondara, L.Medical image denoising using convolutional denoising auto-
identify processing pathways for achieving tailored thin film morphologies. encoders. In 2016 IEEE 16th International Conference on Data Mining Workshops
Computational Mater. Sci. 143, 486–496 (2018). (ICDMW), 241–246 (IEEE, 2016).
23. Teichert, G. H. & Garikipati, K. Machine learning materials physics: Surrogate 48. Bostanabad, R. et al. Computational microstructure characterization and recon-
optimization and multi-fidelity algorithms predict precipitate morphology in an struction: Review of the state-of-the-art techniques. Prog. Mater. Sci. 95, 1–41 (2018).
alternative to phase field dynamics. Computer Methods Appl. Mech. Eng. 344, 49. Lanthaler, S., Mishra, S. & Karniadakis, G. E. Error estimates for DeepONets: A deep
666–693 (2019). learning framework in infinite dimensions. Trans. Math. Its Appl. 6, tnac001 (2022).
24. Lin, C., Maxey, M., Li, Z. & Karniadakis, G. E. A seamless multiscale operator neural 50. Wang, S., Wang, H. & Perdikaris, P. Learning the solution operator of parametric
network for inferring bubble dynamics. J. of Fluid Mechanics. 929 (2021). partial differential equations with physics-informed deeponets. Sci. Adv. 7,
25. Zhang, X. & Garikipati, K. Machine learning materials physics: Multi-resolution eabi8605 (2021).
neural networks learn the free energy and nonlinear elastic response of evolving 51. De, S., Hassanaly, M., Reynolds, M., King, R. N. & Doostan, A.Bi-fidelity modeling of
microstructures. Computer Methods Appl. Mech. Eng. 372, 113362 (2020). uncertain and partially unknown systems using deeponets. Preprint at https://
26. Goswami, S., Yin, M., Yu, Y. & Karniadakis, G. E. A physics-informed variational arxiv.org/abs/2204.00997 (2022).
deeponet for predicting crack path in quasi-brittle materials. Computer Methods 52. Howard, A. A., Perego, M., Karniadakis, G. E. & Stinis, P.Multifidelity deep operator
Appl. Mech. Eng. 391, 114587 (2022). networks. Preprint at https://arxiv.org/abs/2204.09157 (2022).
27. Li, Z. et al. Fourier neural operator for parametric partial differential equations. Preprint 53. Lu, L., Pestourie, R., Johnson, S. G. & Romano, G. Multifidelity deep neural
at: https://arxiv.org/abs/2010.08895 (2020). operators for efficient learning of partial differential equations with application to
28. Goswami, S., Anitescu, C., Chakraborty, S. & Rabczuk, T. Transfer learning fast inverse design of nanoscale heat transport. Phys. Rev. Res. 4, 023210 (2022).
enhanced physics informed neural network for phase-field modeling of fracture. 54. Dingreville, R. P. M., Stewart, J. A., Chen, E. Y. & Monti, J. M.Benchmark problems
Theor. Appl. Fract. Mech. 106, 102447 (2020). for the Mesoscale Multiphysics Phase Field Simulator (MEMPHIS). Tech. Rep.,
29. Kunselman, C., Attari, V., McClenny, L., Braga-Neto, U. & Arroyave, R. Semi- Sandia National Laboratories, Albuquerque, NM (United States) (2020).
supervised learning approaches to class assignment in ambiguous micro- 55. Süli, E. & Mayers, D. F. An Introduction to Numerical Analysis (Cambridge University
structures. Acta Materialia 188, 49–62 (2020). Press, 2003).
30. Haghighat, E., Raissi, M., Moure, A., Gomez, H. & Juanes, R. A physics-informed 56. Niezgoda, S. R., Kanjarla, A. K. & Kalidindi, S. R. Novel microstructure quantification
deep learning framework for inversion and surrogate modeling in solid framework for databasing, visualization, and analysis of microstructure data.
mechanics. Computer Methods Appl. Mech. Eng. 379, 113741 (2021). Integrating Mater. Manuf. Innov. 2, 54–80 (2013).
31. Zhang, K. et al. High-throughput phase-field simulations and machine learning of 57. Gupta, A., Cecen, A., Goyal, S., Singh, A. K. & Kalidindi, S. R. Structure–property
resistive switching in resistive random-access memory. npj Computational Mater. linkages using a data science approach: application to a non-metallic inclusion/
6, 1–10 (2020). steel composite system. Acta Materialia 91, 239–254 (2015).
32. Goswami, S., Anitescu, C. & Rabczuk, T. Adaptive fourth-order phase field 58. Lee, K. & Carlberg, K. T. Model reduction of dynamical systems on nonlinear
analysis using deep energy minimization. Theor. Appl. Fract. Mech. 107, 102527 manifolds using deep convolutional autoencoders. J. Computational Phys. 404,
(2020). 108973 (2020).
33. Attari, V. & Arroyave, R. Machine learning-assisted high-throughput exploration of 59. Kingma, D. & Ba, J. Adam: a method for stochastic optimization. International
interface energy space in multi-phase-field model with calphad potential. Mater. Conference on Learning Representations. (2014).
Theory 6, 1–20 (2022). 60. Abadi, M. et al. TensorFlow: Large-scale machine learning on heterogeneous
34. Samaniego, E. et al. An energy approach to the solution of partial differential distributed systems. Software available from tensorflow.org. 1 (2015).
equations in computational mechanics via machine learning: Concepts, imple- 61. Glorot, X. & Bengio, Y. Understanding the difficulty of training deep feedforward
mentation and applications. Computer Methods Appl. Mech. Eng. 362, 112790 neural networks. Proceedings of the Thirteenth International Conference on Artificial
(2020). Intelligence and Statistics, 249–256 (2010).
35. Shukla, K., Jagtap, A. D., Blackshire, J. L., Sparkman, D. & Em Karniadakis, G. A
physics-informed neural network for quantifying the microstructural properties
of polycrystalline nickel using ultrasound data: A promising approach for solving ACKNOWLEDGEMENTS
inverse problems. IEEE Signal Process. Mag. 39, 68–77 (2022). R.D. and G.E.K. acknowledge funding under the BeyondFingerprinting Sandia Grand
36. Perera, R., Guzzetti, D. & Agrawal, V. Graph neural networks for simulating crack Challenge Laboratory Directed Research and Development (GC LDRD) program. V.O.
coalescence and propagation in brittle materials. Computer Methods Appl. Mech. and K.S. acknowledge Dr. Elton Chen from Sandia National Laboratories for their
Eng. 395, 115021 (2022). valuable discussions that enabled us to better understand the MEMPHIS code. V.O. and
37. Montes de Oca Zapiain, D., Stewart, J. A. & Dingreville, R. Accelerating phase-field- K.S. acknowledge Dr. Aniruddha Bora for his insightful suggestions and valuable
based microstructure evolution predictions via surrogate models trained by discussions at various stages of this project. The phase-field framework is supported by
machine learning methods. npj Computational Mater. 7, 1–11 (2021). the Center for Integrated Nanotechnologies (CINT), an Office of Science user facility
38. Hu, C., Martin, S. & Dingreville, R. Accelerating phase-field predictions via recur-
operated for the U.S. Department of Energy. This research was conducted using
rent neural networks learning the microstructure evolution in latent space.
computational resources and services at the Center for Computation and Visualization,
Computer Methods Appl. Mech. Eng. 397, 115128 (2022).
Brown University. Sandia National Laboratories is a multi-mission laboratory managed
39. Fullwood, D. T., Niezgoda, S. R. & Kalidindi, S. R. Microstructure reconstructions
and operated by National Technology and Engineering Solutions of Sandia, LLC., a
from 2-point statistics using phase-recovery algorithms. Acta Materialia 56,
wholly owned subsidiary of Honeywell International, Inc., for the U.S. Department of
942–948 (2008).
Energy National Nuclear Security Administration under contract DE-NA0003525. This
npj Computational Materials (2022) 190 Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences
V. Oommen et al.
13
paper describes objective technical results and analysis. Any subjective views or Reprints and permission information is available at http://www.nature.com/
opinions that might be expressed in the paper do not necessarily represent the views of reprints
the U.S. Department of Energy or the United States Government.
Publisher’s note Springer Nature remains neutral with regard to jurisdictional claims
in published maps and institutional affiliations.
AUTHOR CONTRIBUTIONS
R.D. and G.E.K. designed the study and supervised the project. V.O. and K.S.
developed the autoencoder–DeepONet architecture and the hybrid model. V.O., K.S.
and S.G. analyzed the results. V.O., K.S., S.G., R.D. and G.E.K. wrote the manuscript. Open Access This article is licensed under a Creative Commons
Attribution 4.0 International License, which permits use, sharing,
adaptation, distribution and reproduction in any medium or format, as long as you give
appropriate credit to the original author(s) and the source, provide a link to the Creative
COMPETING INTERESTS
Commons license, and indicate if changes were made. The images or other third party
The authors declare no competing interests. material in this article are included in the article’s Creative Commons license, unless
indicated otherwise in a credit line to the material. If material is not included in the
article’s Creative Commons license and your intended use is not permitted by statutory
ADDITIONAL INFORMATION regulation or exceeds the permitted use, you will need to obtain permission directly
Supplementary information The online version contains supplementary material from the copyright holder. To view a copy of this license, visit http://
available at https://doi.org/10.1038/s41524-022-00876-7. creativecommons.org/licenses/by/4.0/.
Published in partnership with the Shanghai Institute of Ceramics of the Chinese Academy of Sciences npj Computational Materials (2022) 190