Applications of Machine Learning To One Dimensional Problems of Mechanics
Applications of Machine Learning To One Dimensional Problems of Mechanics
2022)
International Journal of Mechanical Engineering
Abstract:
A new trend in machine learning research is neural networks incorporating physical governing equations as constraints. In this
vein, we provide a neural network constraint based on the governing equation for a deep learning model for one-dimensional
consolidation. Prior research is reviewed and discussed first. Automatic differentiation is used by the deep learning model to
restrict the application of the governing equation. Analytical and model-predicted solutions, as well as constraints, are used to
calculate total loss (a requirement to satisfy the governing equation). Forward and inverse difficulties are both taken into account.
For one-dimensional consolidation issues, the forward tasks show how well a neural network model with physical constraints
performs in prediction. The coefficient of consolidation may be predicted using inverse problems. As an example, we employ
Terzaghi's problem with shifting boundary conditions, and the deep learning model displays a fantastic performance in both the
forward- and inverse-problem scenarios. A deep learning model with physical law integration may be useful for a wide range of
applications, including faster real-time numerical prediction for digital twins, reproducibility of numerical models, and
optimization of constitutive model parameters, although this particular application is a simple one-dimensional consolidation
problem.
1. Introduction
Even outside of the realm of pure computer science, the application possibilities for machine learning have expanded rapidly in
recent years. Artificial neural networks, a specific form of deep learning, are being used successfully in a variety of fields of
science and industry. Recent years have seen a rise in the use of deep learning in the context of partial differential equations
(PDEs). Several researchers are involved in this endeavour, and the use of deep learning in conjunction with physical systems
regulated by PDEs is referred to by a variety of names. Physics-informed neural networks, theory-guided data science, and deep
hidden physics models are just a few of the prevalent designations. Numerical techniques for solving partial differential equations
(PDEs) may be more efficient, accurate, and generalizable with the help of these applications.
Deep learning and machine learning may be used to solve one-dimensional consolidation difficulties. Fluid movement and excess
pore water pressure dissipation in porous media are described in the issue. The problem's governing equation is briefly explained.
The governing partial differential equation's deep learning model is then explained. For both forward and inverse tasks, outcomes
are reported.
Consolidation theory explains how compressive stress delays porous media deformation by dissipating fluid from the porous
medium. One-dimensional consolidation is governed by the equation
(1)
The vertical effective stress (zz) is equal to the vertical effective pressure (p) plus Biot's coefficient (), and the pore space storage
(S) is equal to the restricted compressibility (mv) of the porous medium. In the conventional one-dimensional consolidation issue,
a compressive force is applied at time t=0 and the load is sustained for time t>0. A continuous tension zz is seen for time intervals
longer than zero. It's therefore possible to simplify the general formula (as in (1))
t=0 is the starting point for the 1D consolidation issue, when the complete vertical load is borne by the pore fluid and there is no
dissipation from the porous medium. The dissipation rate of the pore fluid is controlled by the second equation, which takes into
account time as well as spatial dimension. Analytical or numerical approaches may be used to solve this equation for varying
drainage boundary conditions at the top and bottom of the porous material. Analytical solutions are shown here for two alternative
drainage boundary conditions, which will be discussed in more depth in a subsequent section.
2. Literature Survey
There is a shift away from conventional statistical approaches and mechanistic models toward ML and DL, such as deep neural
networks (DNNs), across many commercial applications and sectors, such as education, natural science, medical research,
engineering, as well as social science. mechnical engineering has typically relied on mechanistic models [1]. A common criticism
of ML approaches is that they seem to be "black boxes," meaning they take in inputs and produce outputs but do not disclose
information that can be understood by the user[2,3]. Black-box models have been widely criticised for their opacity, and some
scientists have constructed physics-based ML in response. Even academics who construct the algorithms that produce the
mechanical engineering ML models do not comprehend how variables are integrated to make predictions in these models. No
matter how many input variables are included in a black-box predictive ML model, no researcher can grasp how the variables are
linked together to arrive at the final prediction. High-data-demanding ML models, for example, have trouble estimating structural
damage since they are related to processes that aren't fully understood. Thus, their large data requirements, difficulty in delivering
physically consistent results, and inability to generalise to out-of-sample events [4,5]. ML and DL model[6]s are tested on large,
curated datasets with well-defined, accurately labelled categories. These issues can be handled successfully by DL since it
presume that the world is steady. While these classifications are continually developing in the actual world, particularly in
mechanical engineering, We can only detect the issue after doing extensive testing on ML responses to diverse visual cues.
mechanical engineering applications such as earthquake risk reduction, irrigation management, structural design and analysis, and
structural health monitoring need physics-based numerical models. High-performance computers have enabled mechanical
engineers and scientists to run ultra-realistic simulations with millions of degrees of freedom in their models for use in the real
world. Simulating in the mechanical engineering industry is too time-consuming for an iterative design approach to integrate.
Although they are often used during testing and certification, they are commonly used only in the last phases of development.
Because numerical tools may be used throughout the design process if they can be accelerated, this is an essential challenge to
solve [7,8]. Model complexity has hindered the development of innovative applications, such as enhancing construction
productivity, that might benefit from the development of numerical techniques for quick simulations. An further key example of
analysis that may be possible if simulation expenses were much reduced is uncertainty quantification. In fact, the variables of
interest tracked by numerical simulations are affected by the actual system environment, which is often unknown. It may be
necessary to estimate probability distributions for the quantities of interest to ensure the dependability of the product if these
uncertainties have a substantial influence on simulation outcomes. For complex scientific and technological applications, neither
an ML-only nor a scientific knowledge-only approach may suffice. To better understand the continuum between mechanistic and
machine learning models, researchers are trying to combine scientific knowledge with data.
A physical restriction based on the governing one-dimensional consolidation equation is addressed in this part, along with the
neural network design. Hyper-parameters that are controlled during training are also described, as is the model training technique.
Using the excess pore pressure training data, the neural network with the required number of hidden layers and hidden units
predicts the excess pore pressure[11]. Also included is a physical constraint, based on the governing one-dimensional
consolidation equation, which is assessed via automated differentiation, which is briefly addressed in a subsequent sub-section
below. An artificial neural network is built to minimise the training loss while still accommodating a physical limitation.
3.2 Automatic Differentiation
Automatic differentiation is a critical component of the deep learning model used to solve this issue. Automated differentiation
should not be confused with other techniques for calculating derivatives in computer systems. Derivatives may be computed in
four ways: manually, using finite difference approximations, numerically, using computers to approximate derivatives, and
automatically, using computers to calculate symbolic derivatives and then evaluate them using algebraic expressions. Automatic
differentiation, like the other approaches, gives numerical values of derivatives where these are derived using the principles of
symbolic differentiation, but instead of producing the final expressions, they maintain track of the derivative values. Using this
way of monitoring derivative values, automated differentiation is superior to the two most prevalent methods of calculating
derivatives, numerical and symbolic differentiation[12,13]. There are just a few simple arithmetic operations and elementary
function evaluations involved in each derivative calculation, no matter how sophisticated. The chain rule is continually used until
the required derivative is calculated. Using this methodology, automated differentiation may be performed at machine accuracy
and is far less computationally intensive than previous approaches. TensorFlow's automated differentiation functionality is used to
evaluate the derivatives in the controlling one-dimensional equation[14]. By recording all operations and calculating the gradients
of the recorded calculations using reverse mode differentiation, TensorFlow offers an API for automatically differentiating
models.
A one-dimensional model with a drained top border, a model height of 1 m and a coefficient of consolidation of cv=0.6 m2/yr is
studied here in an inverted situation with the same geometry and material/model characteristics. Using Nz=100 and Nt=100, the
analytical solution is produced again. This means that the precise solution has a total of 10000 points in z, t, and p. The
architecture of the neural network is designed to include 10 hidden layers, each having 20 hidden units. The neural network is
trained using a random selection of 2000 points from the analytical solution data (which has a total of 10000 points). A batch size
of 200 is used to shuffle and split the training data[15]. Initialization of wcv=0, the trainable weight corresponding to the
consolidation coefficient, implies an initial consolidation coefficient of 1.0 m2/yr. Minimize training and constraint losses by
training with a learning rate of 0.0001 and optimising using Adam.
Using the inverse analysis, the outcomes of the issue with the drained upper border are shown in the graph below. As you can see
in the top colour plot, white dots represent the randomly picked training data points. The deep learning model predicts the extra
pore pressure well from a little training sample data, much as in the forward problem. Another example of the outstanding
performance of the physical constraint produced by automated differentiation is provided here as well.
(a) (b)
Fig. 4. There are two plots here: (a) a plot showing the projected consolidating coefficient with time and (b) one of the mean
squared error vs the number of training epochs.
Fig.6. (a) Evolution of the anticipated coefficient of consolidation for drained top and bottom bounds as a function of training
epochs; (b) Mean squared error as a function of training epochs.
According to the deep learning model, cv=0.0994 m2/yr, which is an absolute inaccuracy of 6.010-4, is the projected
consolidation coefficient. These findings may be seen on the left plot of the image above, which depicts the expected coefficient
of consolidation over time as a function of training epochs. To get an idea of how many epochs it takes to train a neural network,
we plotted the mean squared errors (training and constraint) as a function of epochs.
6. Conclusions
The governing partial differential equation is utilised as a constraint in a deep learning model for one-dimensional consolidation.
Researchers in the machine learning field have lately begun to investigate physics-constrained neural networks, and the work
References
1. Momeny, M., et al., A noise robust convolutional neural network for image classification. Results in Engineering, 2021. 10: p.
100225.
2. Malami, S.I., et al., Implementation of hybrid neuro-fuzzy and self-turning predictive model for the prediction of concrete
carbonation depth: A soft computing technique. Results in Engineering, 2021: p. 100228.
3. Baloyi, V.D. and L. Meyer, The development of a mining method selection model through a detailed assessment of multi-
criteria decision methods. Results in Engineering, 2020: p. 100172.
4. Sharma, D., et al. Deep learning Applications to classify Cross-Topic Natural Language Texts Based on Their Argumentative
Form. in 2021 2nd International Conference on Smart Electronics and Communication (ICOSEC). 2021. IEEE.
5. Vadyala, S.R. and S.N. Betgeri, Predicting the spread of COVID-19 in Delhi, India using Deep Residual Recurrent Neural
Networks. arXiv preprint arXiv:2110.05477, 2021.
6. Vadyala, S.R. and E.A. Sherer, Natural Language Processing Accurately Categorizes Indications, Findings and Pathology
Reports from Multicenter Colonoscopy. arXiv preprint arXiv:2108.11034, 2021.
7. Santhosh, A.J., et al., Optimization of CNC turning parameters using face centred CCD approach in RSM and ANN-genetic
algorithm for AISI 4340 alloy steel. Results in Engineering, 2021. 11: p. 100251.
8. Inazumi, S., et al., Artificial intelligence system for supporting soil classification. Results in Engineering, 2020. 8: p. 100188.
9. Chakraborty, D., I. Awolusi, and L. Gutierrez, An explainable machine learning model to predict and elucidate the
compressive behavior of high-performance concrete. Results in Engineering, 2021. 11: p. 100245.
10. Di Ciaccio, F. and S. Troisi, Monitoring marine environments with autonomous underwater vehicles: a bibliometric analysis.
Results in Engineering, 2021: p. 100205.
11. Vadyala, S.R., et al., Prediction of the number of covid-19 confirmed cases based on k-means-lstm. arXiv preprint
arXiv:2006.14752, 2020.
12. Vadyala, S.R. and S.N. Betgeri, Physics-Informed Neural Network Method for Solving OneDimensional Advection
Equation Using PyTorch. arXiv preprint arXiv:2103.09662, 2021.
13. Sai Nethra Betgeri, J.C.M., David B. Smith. Comparison of Sewer Conditions Ratings with Repair Recommendation
Reports. in North American Society for Trenchless Technology (NASTT) 2021. 2021.
https://member.nastt.org/products/product/2021-TM1-T6-01.
Copyrights @Kalahari Journals Vol. 7 (Special Issue, Jan.-Feb. 2022)
International Journal of Mechanical Engineering
463
14. V Yugandhar, B.P., BS Nethra. Statistical Software Packages for Research In Social Sciences. in Recent Research
Advancements in Information Technology. 2014.
15. McGovern, A., et al., Making the black box more transparent: Understanding the physical implications of machine learning.
Bulletin of the American Meteorological Society, 2019. 100(11): p. 2175-2199.