Physics-Informed Machine Learning For Structural Health Monitoring
Physics-Informed Machine Learning For Structural Health Monitoring
Structural Health
Monitoring Based
on Data Science
Techniques
Structural Integrity
Volume 21
Series Editors
José A. F. O. Correia, Faculty of Engineering, University of Porto, Porto, Portugal
Abílio M. P. De Jesus, Faculty of Engineering, University of Porto, Porto, Portugal
Advisory Editors
Majid Reza Ayatollahi, School of Mechanical Engineering, Iran University of
Science and Technology, Tehran, Iran
Filippo Berto, Department of Mechanical and Industrial Engineering, Faculty of
Engineering, Norwegian University of Science and Technology, Trondheim,
Norway
Alfonso Fernández-Canteli, Faculty of Engineering, University of Oviedo, Gijón,
Spain
Matthew Hebdon, Virginia State University, Virginia Tech, Blacksburg, VA, USA
Andrei Kotousov, School of Mechanical Engineering, University of Adelaide,
Adelaide, SA, Australia
Grzegorz Lesiuk, Faculty of Mechanical Engineering, Wrocław University of
Science and Technology, Wrocław, Poland
Yukitaka Murakami, Faculty of Engineering, Kyushu University, Higashiku,
Fukuoka, Japan
Hermes Carvalho, Department of Structural Engineering, Federal University of
Minas Gerais, Belo Horizonte, Minas Gerais, Brazil
Shun-Peng Zhu, School of Mechatronics Engineering, University of Electronic
Science and Technology of China, Chengdu, Sichuan, China
Stéphane Bordas, University of Luxembourg, ESCH-SUR-ALZETTE,
Luxembourg
Nicholas Fantuzzi , DICAM Department, University of Bologna,
BOLOGNA, Bologna, Italy
Luca Susmel, Civil Engineering, University of Sheffield, Sheffield, UK
Subhrajit Dutta, Department of Civil Engineering, National Institute of Technology
Silchar, Silchar, Assam, India
Pavlo Maruschak, Ternopil IP National Technical University, Ruska, Ukraine
Elena Fedorova, Siberian Federal University, Krasnoyarsk, Russia
The Structural Integrity book series is a high level academic and professional series
publishing research on all areas of Structural Integrity. It promotes and expedites
the dissemination of new research results and tutorial views in the structural
integrity field.
The Series publishes research monographs, professional books, handbooks,
edited volumes and textbooks with worldwide distribution to engineers,
researchers, educators, professionals and libraries.
Topics of interested include but are not limited to:
– Structural integrity
– Structural durability
– Degradation and conservation of materials and structures
– Dynamic and seismic structural analysis
– Fatigue and fracture of materials and structures
– Risk analysis and safety of materials and structural mechanics
– Fracture Mechanics
– Damage mechanics
– Analytical and numerical simulation of materials and structures
– Computational mechanics
– Structural design methodology
– Experimental methods applied to structural integrity
– Multiaxial fatigue and complex loading effects of materials and structures
– Fatigue corrosion analysis
– Scale effects in the fatigue analysis of materials and structures
– Fatigue structural integrity
– Structural integrity in railway and highway systems
– Sustainable structural design
– Structural loads characterization
– Structural health monitoring
– Adhesives connections integrity
– Rock and soil structural integrity.
** Indexing: The books of this series are submitted to Web of Science, Scopus,
Google Scholar and Springerlink **
This series is managed by team members of the ESIS/TC12 technical
committee.
Springer and the Series Editors welcome book ideas from authors. Potential
authors who wish to submit a book proposal should contact Dr. Mayra Castro,
Senior Editor, Springer (Heidelberg), e-mail: [email protected]
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature
Switzerland AG 2022
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Switzerland AG
The registered company address is: Gewerbestrasse 11, 6330 Cham, Switzerland
Foreword
One can speculate that there has been an interest in detecting damage in engineered
systems since man has used tools. Over time, early ad hoc qualitative damage detec-
tion procedures, many of which were vibration-based, evolved into more refined
approaches that became what we know today as non-destructive evaluation (NDE)
methods. One drawback of most NDE methods is that the system being inspected
must be taken out of service and often disassembled before such methods can be
used. Structural health monitoring (SHM) attempts to address this shortcoming by
developing more continuous, automated in situ damage detection capabilities that
strives to minimize the human-in-loop aspect of the assessment process.
The term structural health monitoring begins to appear regularly in the tech-
nical literature around the late 1980s and early 1990s. These early studies focused
primarily on deterministic, inverse physics-based modeling approaches that identi-
fied the presence, location, and extent of damage. When researchers and practitioners
attempted to apply such methods to in situ structures, various limitations were iden-
tified including difficulties handling the mismatch between measured and analytical
degrees of freedom, the almost exclusive use of linear models when simulating both
the undamaged and damaged system response, and the inability to handle the opera-
tional and environmental variability that all real-world systems experience. The latter
limitation associated with the fact that operational and environmental variability will
cause changes in the SHM system sensor readings and these changes must be distin-
guished from changes in sensor reading caused by damage has proven to be one of the
most significant challenges associated with transitioning SHM research to practice.
In the late 1990s and early 2000s, various research groups started to recog-
nize that SHM is not a deterministic problem. Instead, they proposed to address
SHM through more data-driven approaches based on general statistical-pattern-
recognition-based methodologies. Although many variations of this statistical pattern
recognition approach have been proposed in different SHM studies; almost all encom-
pass three common components: 1. A deployed sensing system typically monitoring
kinematic response quantities; 2. the extraction of damage-sensitive features from
the raw sensor data; and 3. the statistical classification of those features into damage
and undamaged categories. A common misconception with these approaches is that
v
vi Foreword
they preclude the use of physics-based models when, in fact, the pattern recognition
will always be improved when it is based on knowledge of the physics governing the
system response in both its undamaged and damaged states.
This paradigm shift from inverse deterministic modeling to statistical-pattern-
recognition-based SHM began the process of adopting many data-driven algorithms
from disparate fields such as radar and sonar detection, machine learning, speech-
pattern recognition, statistical decision theory and econometrics to the SHM problem.
In aggregate, these fields represent components of the more general field referred to
as data science. This focus on applying elements of data science to SHM has mostly
replaced the earlier deterministic inverse modeling approaches. Furthermore, data
science offers approaches that can better address the random and systematic changes
in sensor measurements caused by operational and environmental variability and can
produce a quantified probability of detection measure. Both attributes are essential
for the adaptation of SHM by asset owners and regulatory agencies.
Currently, all scientific and engineering fields are benefitting from the rapid
advances in data science and the associated availability of general software tools
for implementing these algorithms. The field of structural health monitoring is one
such beneficiary. However, as in other technical fields, innovation and application-
specific knowledge are required to effectively adapt these general tools to domain-
specific problems. This book provides numerous examples from aerospace, civil, and
mechanical engineering applications that demonstrate how SHM researchers have
taken the tools of data science and creatively adapted them to address many prob-
lems that have been limiting the more widespread adaptation of SHM by industry.
The chapters in this book show the breadth of data science methodologies that can
be applied to SHM. Furthermore, these chapters demonstrate that advances in data
science can impact every aspect of a SHM process. As such, this book will provide
experienced researchers new to the data science field an overview of how such tools
can be used in a damage detection context. Additionally, this book will provide those
just beginning their technical careers with ideas for new research and application
directions to pursue and the associated technologies they will need to learn that will
be the foundation for making future advances in SHM.
Structural health monitoring (SHM) may be defined as the general process of making
an assessment, based on appropriate analyses of in situ measured data, about the
current ability of a structural component or system to perform its intended design
function(s) successfully. A successful SHM strategy may enable significant owner-
ship cost reduction in a life cycle perspective through maintenance optimization,
performance maximization during operation, unscheduled downtime minimization,
and/or enable significant life safety advantage through catastrophic failure mitigation.
Broadly speaking, SHM strategies for most applications necessarily integrate real-
time data acquisition, feature extraction from the acquired data, statistical modeling
of the features, and classification of the features to make informed decisions; the ulti-
mate global goal of SHM systems is to direct economically efficient and/or safety-
maximized structural health decision making for the general purpose of long-term
effective life cycle management.
An explosion of approaches that address some or part of this overall SHM strategy
has occurred in recent years, across many different structural applications ranging
from civil to aerospace to industrial/mechanical systems. A significant fraction of
this growth has been fueled by ubiquitous “Internet of Things (IoT)” data streams
from diverse sources, advances in computing such as cloud computing, and the
adoption and development of advanced analytics techniques drawn from machine
learning and data science. This domain of advancement in SHM can address some
of the paramount challenges in long-term monitoring of civil structures such as,
but not limited to, (i) structural complexity, (ii) operational and environmental vari-
ability (e.g., loading conditions, operating environment), (iii) complex, intercon-
nected degradation and failure modes, (iv) challenges in monitoring very large-scale
structures with potentially localized failure modes (e.g., pitting corrosion), and (v)
data reliability and security, including long-term functionalities of sensor networks.
Damage identification as well as continuous condition monitoring are among the
most important aspects related to proper operation of structural systems to ensure their
integrity, safety, and desirable operational properties. In recent years, an exponential
development of damage identification methods as well as condition monitoring has
been observed. The degradation process of structural systems is usually due to a
vii
viii Preface
The editors would like to express their sincere gratitude to several professionals
and their organizations that directly or indirectly contribute for the realization of this
book; to the Springer Structural Integrity Series Editors, Dr. José A. F. O. Correia and
Prof. Abílio M. P. De Jesus for all the motivation and encouragement to submit a book
proposal in a new and challenging thematic area for structural health monitoring; to
the Senior Editor from Springer, Dr. Mayra Castro, as well as the Coordinators of
the Springer Production team, Vijay Kumar and Augustus Vinoth, for the prompt
support in addressing all questions during the book preparation; and to all authors
of the book chapters for their extraordinary commitment with the objectives of the
publication and their enthusiasm to be part of this project. The authors come from
several companies/institutions from both academia and industry around the world.
Editor Alexandre Cury would like to thank his family, colleagues, and under-
graduate and graduate students which were and are part of the Graduate Programs
in Civil Engineering at the University of Juiz de Fora (PEC/UFJF) and Univer-
sity of Ouro Preto (PROPEC/UFOP). Their constant support and encouragement
were essential to all the academic accomplishments that drove him to be a part of
this project. Special thanks to colleague and friend Prof. Flávio Barbosa (UFJF)
for presenting the initial concepts of SHM and structural damage detection; to Dr.
Christian Cremona (Bouygues Construction) for being a deeply personal and scien-
tific inspiration; and last but not least, to the financial support agencies FAPEMIG,
CNPq and CAPES for their pivotal role in funding research and innovation in Brazil
during these challenging economic times.
Editor Diogo Ribeiro would like to express his gratitude to some colleagues
and institutions that provided the necessary support and motivation to conclude this
editorial project; to the School of Engineering of Polytechnic of Porto (ISEP) and
CONSTRUCT R&D unit for providing all the conditions to conduct research in the
fields of SHM and damage identification; to the friends and professors from ISEP,
Maria da Fátima Portela, Maria do Rosário Oliveira, Ricardo Santos and Teresa Neto,
for the permanent companionship and mutual help, particularly in assuming several
background activities that provide me the necessary time for the book project; to
my Ph.D. student and friend Andreia Meixedo for all the fruitful discussions and
xi
xii Acknowledgements
mutual learning during the last years in the thematic of damage identification; and to
Prof. Rui Calçada from Faculty of Engineering of University of Porto (FEUP) for the
constant scientific inspiration as well as for the encouraging words at the beginning
of this editorial project.
Editor Filippo Ubertini would like to express his gratitude to all colleagues, post-
doctoral scholars, Ph.D. students, and technicians currently working within his “SHM
Lab” group at University of Perugia, as well as to the formerly graduated Ph.D.
students, Master students, and international visiting students, some of which are now
building their own careers in academia at an international level. Special thanks also go
to worldwide distinguished collaborators and particularly to Prof. Simon Laflamme
from Iowa State University, with whom a quite fruitful scientific collaboration in
SHM and a sincere friendship have started more than 10 years ago.
Michael D. Todd would like to thank first and foremost all of the students and
postdoctoral scholars whom he has had the good fortune to mentor throughout his
career. Working with them and helping them launch and develop their careers are the
most gratifying part of being an academic. He also thanks his many collaborators
worldwide for a very satisfying and productive career, but he would like to express
special gratitude to Chuck Farrar at the Los Alamos National Laboratory, who intro-
duced him to structural health monitoring more than 20 years ago. He has truly been
a mentor and a friend.
Finally, the editors would like to thank all the readers of this book for their
interest and passion in shaping the future of data science applied to structural health
monitoring.
Contents
xiii
xiv Contents
n
K= αi Ki , (1)
i=1
where Ki is the ith element stiffness matrix that can be obtained from the finite
element model of a structure, αi is the ith element stiffness parameter to be updated
according to the observed data, and n is the number of structural elements. The
change in mass is assumed to be negligible when damage occurs.
Suppose that the element stiffness parameter reduces to α i , the stiffness reduction
factor (SRF) is defined as follows [27]:
α i − αi
θi = . (2)
αi
SRF, as the damage index, indicates damage location and damage severity. Given
that structural damage typically occurs at several sections or members only, θ is a
sparse vector with only several nonzero components at the damaged locations.
The rth structural eigenvalue and the corresponding mode shape are governed by
the following eigenvalue equation:
(K − λr M)φr = 0. (3)
4 R. Hou et al.
Suppose that N m modes of vibration have been identified from modal testing, the
identified eigenvalues and mode shapes can be expressed as follows:
λ̂ = λ̂1 , λ̂2 , . . . , λ̂ Nm , (4)
ψ̂ = φ̂ 1 , φ̂ 2 , . . . , φ̂ Nm , (5)
where φ̂ r ∈ R Np denotes the identified mode shape of the rth mode at the Np
measurement points. A set of modal data is expressed as follows:
D = λ̂, ψ̂ . (6)
Bayes’ theorem is used to develop a PDF for the damage indexes θ , conditional on
the measured modal data and chosen class of models [28]:
where p(θ |D, M) is the posterior PDF of the damage indexes given the modal data D
and model class M; c = p(D|M) is a normalizing constant referred to as evidence;
p(D|θ, M) is the likelihood function representing the PDF of modal data given the
damage indexes; and p(θ |M) is the prior PDF of the damage indexes. To simplify
the notation, the dependence of the PDF on M is dropped thereafter.
According to the axioms of probability, PDF of the modal data p(D|θ) in Eq. (7)
can be expressed as follows [28]:
Nm
p(D|θ) = p λ̂r |θ p φ̂ r |θ . (8)
r =1
Given the damage index, the modal parameters are deemed independent mode by
mode.
The measurement errors εr and er are assumed to follow the Gaussian distribution
with a zero mean and diagonal variance matrix [29], which are, respectively, given
Vibration-Based Structural Damage Detection Using Sparse … 5
as follows:
λ̂r − λr
εr = ∼ N 0, β −1 , (9)
λ̂r
er = φ̂ r − φ r ∼ N 0, γ −1 I , (10)
where the hyper-parameters β and γ equal the reciprocal of the variance, representing
the precision of the measured eigenvalue and mode shapes, respectively.
The resulting likelihood function of θ based on the measured eigenvalues λ̂ and
φ̂ are given as follows:
⎧ 2 ⎫
β
Nm
2 ⎨ β Nm
λ̂r − λr (θ) ⎬
p λ̂|θ, β = exp − , (11)
2π ⎩ 2 λ̂r ⎭
r =1
Nm
γ Np ·Nm
γ 2
2
p φ̂|θ, γ = exp − φ̂ − φ r (θ) . (12)
2π 2 r =1 r 2
In the Bayesian framework, the damage indexes θ are assigned with a prior distribu-
tion according to engineering judgments and knowledge. Moreover, the suggestion is
to adopt the conjugate prior to simplify the calculation. SBL proposes to use the auto-
matic relevance determination (ARD) prior to incorporate a preference for sparser
parameters [17, 18, 30]. ARD prior following the Gaussian distribution is conjugated
to the likelihood function in this study and adopted as the prior, which is expressed
as follows:
n n
n
1 2 1 1
p(θ |α) = p(θi |αi ) = αi exp − αi θi ,
2 2
(13)
i=1
2π i=1
2
From Eq. (7), the posterior PDF of the damage indexes θ is expressed as follows:
6 R. Hou et al.
p θ |λ̂, ψ̂, α, β, γ = c−1 p λ̂, ψ̂|θ , β, γ p(θ |α)
= c−1 p λ̂|θ , β p ψ̂|θ , γ p(θ|α)
n n
2 γ 1 2 21
Nm Np ·Nm
−1 β 2
=c αi
2π 2π 2π i=1
⎧ 2
⎨ β Nm
λ̂r − λr (θ) γ
Nm 2 1 n
exp − − φ̂ r − φ r (θ) − αi θi2 ,
⎩ 2 λ̂r 2 2 2
r =1 r =1 i=1
(14)
with the distributions on the right-hand sideas defined by Eqs. (11), (12), and (13).
The evidence is calculated by integrating p λ̂, ψ̂, θ |α, β, γ with respect to θ :
c = p λ̂, ψ̂|α, β, γ = p λ̂, ψ̂, θ |α, β, γ dθ. (15)
The maximum posterior (MAP) estimate of the damage indexes θ̃ in Eq. (14) can
be calculated by maximizing ln p θ |λ̂, ψ̂, α, β, γ , or equivalently minimizing the
following objective function:
2 Nm
Nm
λ̂r − λr (θ) 2 n
J (θ ) = β +γ φ̂ r − φ r (θ ) + αi θi2 . (16)
r =1 λ̂r r =1
2
i=1
The items unrelated with θ are omitted in the preceding equation. Note that
Eq. (16) is similar to the regularization technique, in which the first two terms are
equivalent to the data-fitting terms with different weights and the third term to the
regularization term with α as the regularization parameter. We will demonstrate that
Eq. (16) closely resembles the l0 regularization (Refer to Eq. 20).
The minimization of Eq. (16) cannot be directly obtained because parameters
{α, β, γ } are unknown.
MPVs of the hyper-parameters {α, β, γ } can be calculated
by maximizing p α, β, γ |λ̂, ψ̂ . In particular,
p λ̂, ψ̂|α, β, γ p(α, β, γ )
p α, β, γ |λ̂, ψ̂ = ∝ p λ̂, ψ̂|α, β, γ , (17)
p λ̂, ψ̂
mode shapes are in nonlinear relations with θ . Therefore, analytical and numerical
techniques are required. In the following sections, EM, Laplace’s approximation
and variational Bayesian inference (VBI), and delayed rejection adaptive metropolis
(DRAM) algorithm will be introduced to solve this problem.
1 1
n n
n 1
+ ln + ln αi − αi E θi2 . (19)
2 2π 2 i=1 2 i=1
In practice, the complete dataset is not available, and the latent variable θ is given
by the posterior distribution p θ |λ̂, ψ̂, α, β, γ . In the E step, given the current
values
of the hyper-parameters {α, β, γ }old , the posterior distribution of θ given by
p θ |λ̂, ψ̂, α old , β old , γ old is used to determine the expectation of the complete-data
E ln p θ , λ̂, ψ̂|α, β, γ . In the subsequent M step, the new estimate {α, β, γ }new is
obtained by maximizing the expectation with respect to {α, β, γ }. By differentiating
Eq. (19) with respect to α, β, and γ , and setting these derivatives to zero thereafter,
the hyper-parameters are solved as follows:
8 R. Hou et al.
1
αi = , (20)
E θ θi2
Nm
β= 2 , (21)
Nm λ̂r −λr (θ )
Eθ r =1 λ̂r
Np · Nm
γ = 2 , (22)
Nm
Eθ r =1 φ̂ r − φ r (θ )
2
We calculate the variance for each damage index θi independently [28]. There-
after, we generate samples from the posterior PDF p θ|λ̂, ψ̂, α, β, γ , and the
probabilistic information encapsulated in p θ |λ̂, ψ̂, α, β, γ is characterized by
the posterior samples θ (k) , k = 1, . . . , K . The expectations in Eqs. (20), (21), and
(22) are eventually approximated by Eqs. (24), (25), and (26), respectively:
1
K
2
E θ θi2 = θi2 p θi |λ̂, ψ̂, α, β, γ dθi ≈ (θi )(k) , (24)
K k=1
⎧ 2 ⎫ N 2
⎨ Nm
λ̂r − λr (θ ) ⎬ m
λ̂r − λr (θ)
Eθ = p θi |λ̂, ψ̂, α, β, γ dθ
⎩ λ̂r ⎭ λ̂r
r =1 r =1
Vibration-Based Structural Damage Detection Using Sparse … 9
2
1 λ̂r − λr θ (k)
K Nm
≈ , (25)
K k=1 r =1 λ̂r
N N
m 2 m 2
Eθ φ̂ r − φ r (θ) = φ̂ r − φ r (θ ) p θi |λ̂, ψ̂, α, β, γ dθ
2
r =1 r =1
Nm 2
1
≈ φ̂ r − φ r θ (k) . (26)
K r =1
⎧ ⎧ ⎡ ⎤2 ⎫
2 ⎫ ⎪ ⎪
⎨ Nm
λ̂r − λr (θ) ⎬ ⎨ λ̂r − λr θ̃
N m ⎬
E θ̃ =E ⎣ ⎦ , (28)
⎩ λ̂r ⎭ ⎪
⎩ r =1 λ̂r ⎪
⎭
r =1
N N
m 2 m 2
E θ̃ φ̂ r − φ r (θ ) = E φ̂ r − φ r θ̃ . (29)
2 2
r =1 r =1
3.3 Summary
Each iteration uses the estimates of {α, β, γ } to determine the posterior distribution of
the latent variable θ . The current distribution of θ is utilized to improve the estimates
of {α, β, γ }. The proposed EM algorithm is implemented as follows:
1. Initialize the hyper-parameters {α, β, γ }(0) and latent variable θ (0) .
10 R. Hou et al.
The Hessian matrix is decomposed into three items (i.e., W is a diagonal matrix
with entries Wii = αi ), and H and P are given as follows:
⎡ ⎤2
∂ 1
Nm λ̂r − λr θ̃
Hi j = ⎣ ⎦
∂θi ∂θ j 2 r =1 λ̂r
⎡ ⎤
Nm
1 ∂λ r θ̃ ∂λ r θ̃ Nm λ̂r − λ r θ̃ ∂λ θ̃
⎦· 1 ·
r
= · · − ⎣ , (32)
r =1 λ̂r
2 ∂θi ∂θ j r =1 λ̂r λ̂r ∂θi ∂θ j
Vibration-Based Structural Damage Detection Using Sparse … 11
Np 2
∂ 1
Nm
Pi j = φ̂k,r − φk,r θ̃
∂θi ∂θ j 2 r =1 k=1
Np ∂φ Np ∂φk,r θ̃
k,r θ̃ ∂φk,r θ̃
Nm Nm
= · − φ̂k,r − φk,r θ̃ · . (33)
r =1 k=1
∂θi ∂θ j r =1 k=1
∂θi ∂θ j
The derivatives of the eigenvalues and eigenvectors with respect to the stiffness
parameter can be calculated using the Nelson’s method [33] or the substructural
approach [34].
Given that the integral of a normalized Gaussian distribution equals 1, the evidence
p λ̂, ψ̂|α, β, γ can be solved as follows:
p λ̂, ψ̂|α, β, γ = p λ̂, ψ̂, θ |α, β, γ dθ
∼ 1 T
= p λ̂, ψ̂, θ̃ |α, β, γ · exp − θ − θ̃ A θ − θ̃ dθ
2
(2π ) n/2
= p λ̂, ψ̂, θ̃ |α, β, γ , (34)
|A|1/2
In this way, the asymptotic analytical expression of the evidence is obtained. For
convenience of calculation, the logarithm form is used and shown as follows:
n 1
ln p λ̂, ψ̂|α, β, γ = ln p λ̂, ψ̂, θ̃ |α, β, γ + ln(2π ) − ln|A|
2 2
= ln p λ̂|θ̃ , β ln p λ̂|θ̃ , β + ln p ψ̂|θ̃ , γ
n 1
+ ln p θ̃ |α + ln(2π ) − ln|A|
2 2
Nm β Np Nm γ n 1
= ln + ln + ln
2 2π 2 2π 2 2π
⎡ ⎤2
n n Nm λ̂ − λ θ̃
1 1 β ⎣
r r
⎦
+ ln αi − αi θ̃i2 −
2 i=1 2 i=1 2 r =1 λ̂r
Nm 2 n
γ 1
− φ̂ r − φ r θ̃ + ln(2π ) − ln|A|. (35)
2 r =1 2 2 2
MPVs of {α, β, γ } can be estimated by maximizing p λ̂, ψ̂|α, β, γ . From
Eq. (35), setting the derivative of p λ̂, ψ̂|α, β, γ with respect to α, β, γ equal
to zero results in the following equations:
12 R. Hou et al.
1
αi = ' ( , (36)
θ̃i2 + (W+βH + γ P)−1 ii
Nm
β= 2 , (37)
Nm λ̂r −λr θ̃ n ' −1
(
r =1 λ̂r
+ i=1 (W+βH + γ P) H ii
Nm Np
γ = 2 ( . (38)
n '
+ γ P)−1 P
Nm Np
r =1 k=1 φ̂k,r − φk,r θ̃ + i=1 (W+βH ii
Note that the solutions of θ̃ (Eq. 16) and {α, β, γ } are coupled. Therefore, they
can be determined from an iterative process of Eqs. (16) and (36)–(38).
The preceding formulations can be summarized as follows:
1. Initialize the hyper-parameters {α, β, γ }(0) and θ (0) .
( j)
2. For j = 1, 2, . . ., update θ̃ using Eq. (16) with α ( j−1) , β ( j−1) , γ ( j−1) .
( j) ( j)
3. Update αi using Eq. (36) with θ̃ , β ( j−1) , γ ( j−1) .
( j) ( j)
4. Update β ( j) using Eq. (37) with θ̃ , αi , γ ( j−1) .
( j) ( j)
5. Update γ ( j) using Eq. (38) with θ̃ , αi , β ( j) .
6. Let j = j + 1, repeat Steps 2–5 until the convergence criterion is satisfied (i.e.,
( j) ( j−1) ( j)
θ̃ − θ̃ 2 /θ̃ 2 ≤ Tol).
5.1 VBI
The mechanism of VBI is to propose a tractable PDF to approximate the target PDF
(i.e., posterior PDF p θ |λ̂, ψ̂, α, β, γ ) [31].
The evidence can be calculated as follows:
p λ̂, ψ̂, θ |α, β, γ
c = p λ̂, ψ̂|α, β, γ = . (39)
p θ |λ̂, ψ̂, α, β, γ
Taking the logarithm of the two sides in Eq. (39), the formulation changes as
follows:
p λ̂, ψ̂, θ |α, β, γ p θ |λ̂, ψ̂, α, β, γ
ln p λ̂, ψ̂|α, β, γ = ln − ln , (40)
Q Q
Vibration-Based Structural Damage Detection Using Sparse … 13
where Q(θ , α, β, γ ) is the proposed PDF to approximate p θ |λ̂, ψ̂, α, β, γ , which
is simplified as Q.
Given that ln p λ̂, ψ̂|α, β, γ is irrelevant to θ, taking the expectation of the two
sides in Eq. (40) with respect to Q yields the following equations:
ln p λ̂, ψ̂|α, β, γ = L( Q) + DKL Q|| p θ |λ̂, ψ̂, α, β, γ , (41)
⎡ ⎤
p λ̂, ψ̂, θ |α, β, γ
L( Q) = E Q ⎣ln ⎦, (42)
Q
Q
DKL Q|| p θ |λ̂, ψ̂, α, β, γ = Q ln dθ , (43)
p θ|λ̂, ψ̂, α, β, γ
where L( Q) represents the lower bound of Q and DKL is the KL divergence [32]
between Q and posterior PDF p θ |λ̂, ψ̂, α, β, γ .
Moreover, DKL ≥ 0 has been proven [31]. DKL = 0 when Q =
p θ |λ̂, ψ̂, α, β, γ . Therefore, increasing the proximity of Q to the posterior PDF
is equivalent to minimizing DKL . Given that the posterior PDF p θ |λ̂, ψ̂, α, β, γ
is unknown, DKL cannot be calculated directly. According to Eq. (41), minimizing
DKL is equivalent to maximize L( Q) [31]. To obtain the independent posterior PDF
of the damage index and parameters, Q is factorized into two components based on
mean field theory [35] as follows:
n
q(α) = q(αi ). (47)
i=1
Therefore,
n
ln q(α, β, γ ) = ln q(αi ) + ln q(β) + ln q(γ ). (48)
i=1
According to Eqs. (19) and (45), the logarithm of the posterior PDF of the
individual parameter is obtained thereafter, and all parameters follow the gamma
distribution:
1
α
i
q(αi ) ∝ (αi ) 2 · exp − E θ θi2 , (49)
2
⎧ ⎛ 2 ⎞⎫
⎨ β
Nm
λ̂ − λ (θ) ⎬
q(β) ∝ (−β) 2 · exp − E θ ⎝ ⎠ ,
Nm r r
(50)
⎩ 2 λ̂r ⎭
r =1
N
m 2
Np Nm γ
q(γ ) ∝ (γ ) 2 · exp − E θ φ̂ r − φ r (θ ) . (51)
2 r =1
2
Therefore, the mean and variance of each parameter are calculated as follows:
3 6
E(αi ) = ; Var(αi ) = ' (2 , (52)
E θi2
E θi2
Nm + 2 2Nm + 4
E(β) = 2 ; Var(β) = 2 2 , (53)
Nm λ̂r −λr (θ ) Nm λ̂r −λr (θ)
Eθ r =1 λ̂r Eθ r =1 λ̂r
Np Nm + 2 2Np Nm + 4
E(γ ) = 2 ; Var(γ ) = 2 2 .
Nm Nm
Eθ φ̂
r =1 r − φ r (θ ) Eθ r =1 φ̂ r − φ r (θ )
2 2
(54)
Factor
q(θ ) can be similarly derived by calculating the expectation of
ln p λ̂, ψ̂, θ |α, β, γ with respect to {α, β, γ }:
ln q(θ) = E α,β,γ ln p λ̂, ψ̂, θ |α, β, γ + const, (55)
where the items independent with θ are merged into the constant item. Therefore,
Vibration-Based Structural Damage Detection Using Sparse … 15
⎧ 2
⎨ E(β)
Nm
λ̂r − λr (θ)
q(θ ) ∝ exp −
⎩ 2 r =1 λ̂r
Nm 2 n
E(γ ) E(αi ) 2
− φ̂ − φ r (θ ) − · θi . (56)
2 r =1 r 2
i=1
2
Equations (52)–(54) and (56) are coupled and should be iteratively calculated.
The iterations are corresponding to the variational Bayesian EM step [31]. Given
that VBI does not guarantee to converge to the global optimum, the parameters are
initialized differently at the beginning of the iteration. The iterative process requires
2
Nm λ̂r −λr (θ)
the calculation of three expectations, namely E θi2 , E θ r =1 λ̂r
, and
2
Nm
Eθ r =1 φ̂ r − φ r (θ ) , which is a full Bayesian analysis that considers poste-
2
rior uncertainties of θ . However, the specific distribution of θ cannot be directly
recognized from Eq. (56) because of the nonlinear relationship between θ and the
modal parameters. In the next subsection, the numerical DRAM algorithm is used
to obtain the statistical distribution of θ .
5.3 Summary
The damage index θ is sampled using the DRAM sampling technique according
to Eqs. (16) and (56), and MPVs θ are calculated from the obtained samples. The
parameters {α, β, γ } are updated according to Eqs. (52)–(54). The proposed VBI-
DRAM algorithm is implemented as follows:
1. Initialize the hyper-parameters {α, β, γ }(0) and latent variable θ (0) .
2. At the jth iteration,
Solve θ̃ by minimizing J (θ) in Eq. (16);
Let μ0 = θ (1) = θ̃ , generate n s samples of θ θ (1) ∼ θ (Ns ) based on the DRAM
( j)
algorithm according to Eq. (56), and identify θ MAP ;
Based on the generated samples, calculate E(αi )( j) and Var(αi )( j) from Eq. (52);
Calculate E(β)( j) and Var(β)( j) from Eq. (53);
Calculate E(γ )( j) and Var(γ )( j) from Eq. (54).
3. Repeat Step 2 for the (j + 1)th iteration until the convergence criterion is met
( j) ( j−1) ( j)
(i.e., θ MAP − θ MAP 2 /θ MAP 2 ≤ Tol).
6 Numerical Example
6.0
Clamped
50.75
900
stiffness while mass remains unchanged. Element 1 at the clamped end and Element
23 at the mid-span (Fig. 1) are damaged by 50% (i.e., SRF(1) = SRF(23) = −50%).
Natural frequencies are used for damage detection. Items associated with mode
shapes are excluded, and the objective function in Eq. (16) is simplified as follows:
2
Nm
λ̂r − λr (θ )
n
J1 (θ) = β + αi θi2 . (57)
r =1 λ̂r i=1
The first six natural frequencies of the beam before and after damage are listed in
Table 1.
Damage index and hyper-parameters should be initialized first. Modal experiment
experiences have shown that natural frequencies may comprise 1% noise [38]. There-
fore, a noise level of 1% is assigned to the frequencies (i.e., β (0) = 1/(1%)2 = 1 ×
104 ). The uncertainty level of damage index is assumed as 10% of the exact damage
index. In particular, the initial value αi(0) = 1/(10%)2 = 100 (i = 1, 2, . . . , 45).
Initial damage indexes are set as θ (0) = {0, . . . , 0}T , indicating a lack of damage.
The convergence criterion for the iteration process is set as T ol = 0.01. Identification
error δ is defined as follows:
SRF
SRF
SRF
-
.
. θ̃ − θ 2
/
δ= 2
, (58)
n
The reason is that the numerical sampling of the damage index in the VBI-DRAM
algorithm is directly conducted on PDF (Eq. 56).
7 Experimental Study
Hou et al. [39] tested a three-story steel frame (Fig. 3). Each story is 0.5 m high,
and the span is 0.5 m. Beams and columns have the same cross-section dimension of
75.0 mm × 5.0 mm. The Young’s modulus of the steel is 2.0 × 1011 N/m2 , and mass
density is 7.92 × 103 kg/m3 . The frame was excited with an instrumented hammer.
The structure was impacted eight times and each impact lasted for 30 s. The eight
signals were averaged to improve the modal identification accuracy. Measurement
points were chosen every 100 mm, resulting in a total of 39 measurement points.
Bruel & Kjaer accelerometers with a magnetic base were firmly mounted on the
frame to measure the vertical acceleration of the beam and horizontal acceleration
of the column. Sampling frequency was set as 2000 Hz.
The frame is divided into 225 Euler–Bernoulli beam elements, with each
measuring 20-mm long. Two cuts were introduced to the frame model. Cuts 1 and 2
were located at the column end and beam/column joint, corresponding to elements
1 and 176, respectively. Saw cuts have the same length of 20 mm and depth of
22.5 mm, leading to a reduction in the moment of inertia of the cut sections by 60%
Cut 2 (60%)
5
1500
Accelerometer
5100
100
Cut 1 (60%)
20 R. Hou et al.
(i.e., SRF(1) = SRF(176) = −60%). The measured first seven frequencies and MAC
of the frame structure before and after damage are listed in Table 2.
Natural frequencies and mode shapes are used for damage identification. Given
that natural frequencies can be measured more accurately than mode shapes, their
uncertainty levels are set as 1% and 5%, respectively. Thus, the hyper-parameters
are initialized as β (0) = 1 × 104 , γ (0) = 400 and αi(0) = 100 (i = 1, 2, . . . , 225).
The initial damage indexes and convergence criterion are the same as those in the
numerical example.
EM Algorithm. Given that only one set of measurement modal data is available,
50 sets of modal data D j = [λ̂( j) , ψ̂ ( j) ] ( j = 1, 2, . . . , 50) are generated through
( j) ( j)
numerical simulation λ̂r ∼ N λ̂r , (0.01λ̂r )2 and φ̂r ∼ N φ̂r , (0.05)2 I . MPVs
of the damage indexes are calculated by minimizing the objective function in Eq. (16)
iteratively using the likelihood sampling. The process converges after five iterations,
and only the results in the last iteration are shown in Fig. 4a for brevity. The two
damaged elements are located successfully, while the severity of the damage at
column end is larger than the true value (i.e., with 60% reduction).
Laplace Approximation. Following the iterative procedures summarized in Sect. 4,
the proposed Laplace approximation algorithm is applied to detect damage of the
frame. Convergence is achieved within five iterations, and the results are shown in
Fig. 4b. Upon convergence, two damaged elements are accurately detected with no
false identifications.
Vibration-Based Structural Damage Detection Using Sparse … 21
(a)
-0.2
SRF
-0.4
-0.6
-0.8
20 40 60 80 100 120 140 160 180 200 220
Element No.
(b)
-0.2
SRF
-0.4
-0.6
-0.8
20 40 60 80 100 120 140 160 180 200 220
Element No.
(c)
-0.2
SRF
-0.4
-0.6
-0.8
20 40 60 80 100 120 140 160 180 200 220
Element No.
100 200
80
60
α 100
α1 40
150
20
100
0 1 2 3 4 5 0 1 2 3 4 5
Iteration No. Iteration No.
Fig. 5 Variation of hyper-parameters during the iterative process for the Laplace approximation
those in the numerical example. The possible reason is that the number of unknowns
in the experimental example is substantially larger than that in the numerical one.
Similarly, the EM technique requires the most computational time. However, the
VBI-DRAM algorithm is more efficient than the Laplace approximation for this
case because the calculation of the large-scale Hessian matrix is extremely time-
consuming.
In summary, the main advantages of the EM algorithm are its simplicity and ease
of implementation. Laplace approximation is the most efficient for low-dimensional
problem. The VBI-DRAM algorithm can provide a full characterization of the poste-
rior PDF instead of merely focusing on MPVs. The EM technique and VBI-DRAM
algorithm are suitable for standard and nonstandard probability distributions. Thus,
they are widely applicable.
9 Conclusions
This study proposes an SBL model for probabilistic structural damage detection using
modal parameters. Sparsity of structural damage is exploited as an important prior
information from the Bayesian perspective. Analytical approximation and numer-
ical sampling techniques are introduced to deal with the computationally prohibited
integration in the evidence.
A numerical cantilever beam and an experimental three-story frame are utilized
to verify their effectiveness. The results show that damage location and severity can
be accurately detected using the three techniques, even when the measurement data
are considerably fewer than the damage indexes. Comparative studies indicate that
each algorithm has its own advantages.
References
1. Hemez FM (2005) Uncertainty quantification and the verification and validation of compu-
tational models. In: Inman DJ (ed) Damage prognosis: for aerospace, civil and mechanical
systems. Wiley, West Sussex, pp 201–220
2. Hou RR, Xia Y (2021) Review on the new development of vibration-based damage
identification for civil engineering structures: 2010–2019. J Sound Vib 491:115741
3. Xia Y, Hao H (2003) Statistical damage identification of structures with frequency changes. J
Sound Vib 263(4):853–870
4. Sohn H, Dzwonczyk M, Straser E et al (1999) An experimental study of temperature effect on
modal parameters of the Alamosa Canyon Bridge. Earthq Eng Struct D 28(8):878–897
5. Xia Y, Hao H, Zanardo G et al (2006) Long term vibration monitoring of an RC slab: temperature
and humidity effect. Eng Struct 28(3):441–452
6. Hua XG, Ni YQ, Chen ZQ et al (2008) An improved perturbation method for stochastic finite
element model updating. Int J Numer Meth Eng 73(13):1845–1864
7. Yeo I, Shin S, Lee HS et al (2000) Statistical damage assessment of framed structures from
static responses. J Eng Mech 126(4):414–421
8. Gul M, Catbas FN (2009) Statistical pattern recognition for structural health monitoring
using time series modeling: theory and experimental verifications. Mech Syst Signal Process
23(7):2192–2204
9. Beck JL, Katafygiotis LS (1998) Updating models and their uncertainties. I: Bayesian statistical
framework. J Eng Mech 124(4):455–461
24 R. Hou et al.
10. Ching J, Beck JL (2004) Bayesian analysis of the phase II IASC-ASCE structural health
monitoring experimental benchmark data. J Eng Mech 130(10):1233–1244
11. Zhu YC, Au SK (2018) Bayesian operational modal analysis with asynchronous data, part I:
most probable value. Mech Syst Signal Process 98:652–666
12. Yan WJ, Katafygiotis LS (2019) An analytical investigation into the propagation properties of
uncertainty in a two-stage fast Bayesian spectral density approach for ambient modal analysis.
Mech Syst Signal Process 118:503–533
13. Huang Y, Shao CS, Wu B et al (2019) State-of-the-art review on Bayesian inference in structural
system identification and damage assessment. Adv Struct Eng 22(6):1329–1351
14. Tipping ME (2001) Sparse Bayesian learning and the relevance vector machine. J Mach Learn
Res 1:211–244
15. Malioutov D, Cetin M, Willsky AS (2005) A sparse signal reconstruction perspective for source
localization with sensor arrays. IEEE Trans Signal Process 53(8):3010–3022
16. Fang J, Shen YN, Li FW (2015) Support knowledge-aided sparse Bayesian learning for
compressed sensing. In: Proceedings of IEEE international conference on acoustics. IEEE,
Brisbane
17. Huang Y, Beck JL (2015) Hierarchical sparse Bayesian learning for structural health monitoring
with incomplete modal data. Int J Uncertain Quantification 5(2):139–169
18. Huang Y, Beck JL et al (2017) Hierarchical sparse Bayesian learning for structural damage
detection: theory, computation and application. Struct Saf 64:37–53
19. Huang Y, Beck JL, Li H (2017) Bayesian system identification based on hierarchical sparse
Bayesian learning and Gibbs sampling with application to structural damage assessment.
Comput Meth Appl Mech Eng 318:382–411
20. Hou RR, Xia Y, Zhou XQ et al (2019) Sparse Bayesian learning for structural damage detection
using expectation–maximization technique. Struct Control Health Monit 26(5):e2343
21. Wang XY, Hou RR, Xia Y et al (2020) Laplace approximation in sparse Bayesian learning for
structural damage detection. Mech Syst Signal Process 140:106701
22. Wang XY, Hou RR, Xia Y et al (2020) Structural damage detection based on variational
Bayesian inference and delayed rejection adaptive Metropolis algorithm. Struct Health Monit.
https://doi.org/10.1177/1475921720921256
23. Beal MJ (2003) Variational algorithms for approximate Bayesian inference. Dissertation,
University of London
24. Rouder JN, Lu J (2005) An introduction to Bayesian hierarchical models with an application
in the theory of signal detection. Psychon Bull Rev 12:573–604
25. Lam HF, Yang JH, Au SK (2015) Bayesian model updating of a coupled-slab system using field
test data utilizing an enhanced Markov chain Monte Carlo simulation algorithm. Eng Struct
102:144–155
26. Lam HF, Yang JH, Au SK (2018) Markov chain Monte Carlo-based Bayesian method for
structural model updating and damage detection. Struct Control Health Monit 25(4):e2140
27. Zhou XQ, Xia Y, Weng S (2015) L1 regularization approach to structural damage detection
using frequency data. Struct Health Monit 14(6):571–582
28. Vanik MW, Beck JL, Au SK (2000) Bayesian probabilistic approach to structural health
monitoring. J Eng Mech 126(7):738–745
29. Jaynes ET (2003) Probability theory: the logic of science. Cambridge University Press,
Cambridge
30. Mackay DJC (1992) Bayesian methods for adaptive models. Dissertation, California Institute
of Technology
31. Bishop CM (2006) Pattern recognition and machine learning. Springer, New York
32. Dempster AP, Laird N, Rubin D (1997) Maximum likelihood for incomplete data via the EM
algorithm. J R Stat Soc B 39:1–38
33. Nelson RB (1976) Simplified calculation of eigenvector derivatives. AIAA J 14(9):1201–1205
34. Weng S, Zhu HP, Xia Y et al (2013) Substructuring approach to the calculation of higher-order
eigensensitivity. Comput Struct 117:23–33
35. Parisi G (1988) Statistical field theory. Addison-Wesley, Boston
Vibration-Based Structural Damage Detection Using Sparse … 25
36. Haario H, Laine M, Mira A et al (2006) DRAM: efficient adaptive MCMC. Stat Comput
16:339–354
37. Wan HP, Ren WX (2016) Stochastic model updating utilizing Bayesian approach and Gaussian
process model. Mech Syst Signal Process 70–71:245–268
38. Mottershead JE, Friswell MI (1993) Model updating in structural dynamics: a survey. J Sound
Vib 167(2):347–375
39. Hou RR, Xia Y, Zhou XQ (2018) Structural damage detection based on l 1 regularization using
natural frequencies and mode shapes. Struct Control Health Monit 25(3):e2017
Bayesian Deep Learning
for Vibration-Based Bridge Damage
Detection
1 Background
Damaged bridges are not a thing of the past. On August 1, 2007, the I-35W bridge
in Minneapolis collapsed, killing 13 and injuring 145. On August 14, 2018, a major
bridge in Genoa, Italy, partially collapsed, killing 43 people. One can wonder if a
modern, data-based structural health monitoring system alongside strict maintenance
monitoring could have prevented these events.
This chapter presents an end-to-end method of detecting damage in a real-life
bridge using only sensor measurements, based on the previous work of [1–3]. Even
though the results are a promising step into a more modern structural health moni-
toring of bridges, we believe that we are far from being able to replace standard moni-
toring methods of bridges. Using the presented method as a replacement for standard
bridge inspection could be dangerous and possibly life-threatening. What this chapter
presents is a method that could complement standard procedure and possibly improve
the speed and accuracy of damage detection. If the method confidently predicts a
damaged state, the prediction will have to be confirmed manually.
There are two main approaches for structural health monitoring (SHM) of bridges: a
model-based approach and a data-based approach. In both cases, sensors of different
kinds are attached to the structure to measure its response to external stimulation. In
a model-based approach, a finite element analysis is done specifically for each bridge
which is calibrated with the sensor data. This analysis produces a baseline for values
of displacement, strain, and vibration at each point. The real sensor measurements
are then compared to the calibrated finite element model, and damage is assessed.
In a data-based approach, only the sensor data is used for damage assessment. One
aspect that data-based approaches to SHM have in common is a comparison of
a current state to a baseline state. Sensor data is collected over a certain period,
considered to be the structurally healthy baseline, and deviations from the baseline
are assessed as structural damage. One of the biggest challenges of bridge SHM is the
natural environmental response. Significant changes in a bridge’s natural frequency
and stiffness can occur due to temperature changes, making a modal approach to
SHM impractical [4, 5]. This variability calls for more complex modeling. Vibration-
based methods address some weaknesses of traditional modal methods and can be
more robust to environmental variability and more sensitive to damage [6–8]. The
use of artificial intelligence for SHM has recently shown great promise in many
different applications [9], and machine learning methods are becoming established
as an efficient alternative approach to classical modeling techniques in structural
engineering [10].
Gonzalez and Karoumi [2] use bridge vibrational data and a specialized bridge
weigh in motion system that records the bridges’ load position, magnitude, and speed.
Bayesian Deep Learning for Vibration-Based Bridge Damage … 29
A neural network was trained to predict the bridges’ vibrations at each time step. A
Gaussian process was then used to characterize the prediction errors. Neves et al. [3]
extend this idea from [2] without the specialized weigh in motion system.
There is a prominent need for addressing uncertainty in a data-based approach to
SHM [11]. The value of addressing uncertainty has been shown as being more robust
than plain neural networks for damage detection in a composite airfoil structure [10].
Uncertainty is therefore a central theme of this chapter.
Here, some fundamental concepts are reviewed from the field of machine learning
that will be useful to understand the method section.
Linear regression is the most basic building block of a neural network, a linear
combination of D + 1-dimensional input variables X = x0 , . . . , x D and weight
parameters w = w0 , . . . , w D . A bias term x0 is added to the D-dimensional input,
which makes the input into the model D + 1-dimensional. The output prediction ŷ
is on the form of a matrix dot product, Eq. (1).
ŷ = w T X (1)
Given N observations of (X, y) pairs, there exists a closed form solution to the w
that minimizes the error function. For linear regression, the error function is usually
the mean squared error function which corresponds to a maximum likelihood solution
if the error is assumed to be normally distributed with zero mean.
Logistic regression is very similar in form, except that the output prediction is a
binary classification. A nonlinear activation function σ (x) outputs a value between
0 and 1 that can be interpreted as the probability of a class being 1, Eq. (2).
ŷ = σ w T X (2)
Fig. 1 A single forward pass of a neural network with ReLU activation function and dropout 0.5,
where a node connection is randomly dropped with 50% probability. Dropped nodes are depicted
as dotted, and observable nodes are shaded gray
Fig. 2 An example of an autoencoder network that encodes the six-dimensional input to a two-
dimensional code. This compact representation is then decoded to reconstruct the original six-
dimensional input. This example has three hidden layers, and ReLU activation function and the
encoder and decoder have a mirrored structure
2 Method
The need for uncertainty in SHM is explained in Fig. 3. The collected training data
for bridge structural health monitoring cannot be truly complete in a reasonable time
frame, due to potentially large variations in regular environmental variables. Such a
complete recording could possibly take many years. Therefore, a predictive model
is nearly always going to be trained on a limited healthy-state dataset. It is important
to note that a healthy state can still contain old damages even though this base state
is considered as undamaged. It is of course not possible to detect damages induced
in the bridge before the sensors were installed.
32 D. S. Ásgrímsson et al.
A crucial observation is what happens when the model is faced with unfamiliar
input data, depicted as points B and C in Fig. 3. The predictive error for test point
A close to the training dataset should be lower than for a point B far away. Using
only high predictive error for damage detection would likely falsely flag point B as
damaged. By including model uncertainty, the reconstruction of point B will have
large error, but the model should be able to explain the large error by having large
uncertainty and therefore not label the large error as anomalous. If the true values
fall outside the predictive uncertainty bounds of the model as in point C, the model
cannot explain the large difference between real and predicted values and therefore
correctly reject test point C.
The method to detect change in structural condition presented in this chapter builds
upon the work of [1–3]. A Bayesian deep autoencoder is trained on the collected
healthy bridge vibrational and environmental sensor data to reconstruct the given
healthy-state input data. The autoencoder neural network reconstructs the input data
with an uncertainty interval by using MC dropout. The reconstruction error of the
unseen healthy-state data is then quantified, and a 95th percentile rejection threshold
is determined. If the total sequence reconstruction error of a sequence in an unknown
state falls outside the determined rejection threshold, then the sequence is rejected.
If enough sequences are rejected, the bridge is determined to be in a damaged state.
2.1 Assumptions
A model is a set of assumptions about the data, and the following are the most
important assumptions of the modeling. Some of them are limiting, and some make
the problem more feasible. These assumptions are general enough so the method can
be applied to a similar bridge in structure, and no assumption was created specifically
to match the particular Z24 bridge structure.
Vibration sensor data is correlated.
Since the sensors are all connected to the same continuous bridge structure, it is
reasonable to assume that the sensor recordings are correlated. When one sensor
is detecting high vibration, it is assumed that the other sensors are detecting high
Bayesian Deep Learning for Vibration-Based Bridge Damage … 33
vibrations too. The assumption that the vibration sensor data is correlated makes
the prediction problem easier than if they were uncorrelated, the model will learn
the interplay of the sensors and how the sensors react to stimulation in unison in a
healthy state. If the bridge is damaged near a particular sensor, the interplay between
the sensors will change, with higher error rates, and therefore makes it easier to detect
damage. Due to this assumption, it is important that the input to the model contains
data of all sensors simultaneously on the bridge. This type of input is referred to
as a sliding window input in time. The sliding window input needs to contain the
simultaneous sensor measurements of all the sensors. If each sensor measurement
would be reconstructed separately, this critical information of how the sensors react
in unison to stimulation would be lost.
Enough information is contained within a short time frame.
The chosen neural network model architecture can only take a fixed, limited size
input. A reasonable input to the model would be a sliding window of sensor data
in the order of seconds. This is a limitation to the model since the true behavior
could possibly not be captured in such a short time frame. To address this issue,
multiple consecutive sliding window inputs are reconstructed. This is referred to as a
sequence, and the total summed reconstruction error is used to either accept or reject
a sequence. The sequence size is in the order of minutes. Note that more complex
recurrent neural network architectures exist that can take the full sequence as input
by using internal state to model the temporal dynamic behavior.
Vibrations are dependent on environmental variables.
Environmental variables such as air temperature, soil temperature, humidity, wind
speed, as well as vehicle passing have a big impact on the response of a bridge. It
is therefore very important to include these variables as an input to the predictive
model.
The vibration data is hierarchical.
The nature of a deep neural network is in the cascading composition of simpler
elements, as described above. The vibrations of a bridge are the result of many
interplaying variables such as weather, winds, and loads. It is therefore assumed that
the data is inherently compositional, and a deep neural network would be a good fit
for this problem.
A vibration sensor measures local condition.
One of the pillars of bridge structural health monitoring is the localization of damage.
To address this, it is assumed that if the real sensor values for a given sensor are far
away from the model’s predictive uncertainty of that sensor, the bridge is damaged
in the region of the physical location of the sensor. This assumption holds in many
damage scenarios, but for a damage in a main longitudinal load carrying beam can
also be seen in sensors located far from the damage location compared to a damage
of a secondary load carrying member such as a crossbeam which will only influence
local measurements. This is out of the reach of the suggested approach.
34 D. S. Ásgrímsson et al.
For the damage detection, three algorithms are presented. Note that all three algo-
rithms use the exact same trained model. This makes it possible to deploy all three
in a real-time system without significant increase in computational load. All three
methods reconstruct a given 10-min-long sensor sequence, in a succession of non-
overlapping 2 s sliding window inputs, with the autoencoder neural network trained
on the healthy-state data. The reconstruction error for the whole sequence is summed
and compared to the distribution of total sequence error for the healthy-state test data.
If the total reconstruction error of the given sequence is over the 95% cutoff rejection
threshold, the sequence is rejected. If more than 5% of sequences are rejected, the
bridge is determined to be in a damaged state.
The standard method reconstructs the input in one single forward pass of
the neural network. The method does not take into consideration any notion of
uncertainty. This is a standard way of prediction in a neural network.
The mean method uses MC Dropout [26] to stochastically reconstruct the input
100 times and to use the mean of the 100 reconstructions as the final prediction. This
method gives a Maximum A Posteriori estimation.
The uncertainty method uses MC Dropout [26] to stochastically reconstruct the
input 100 times. True values that fall within 2 standard deviations from the mean of
the predictions are then considered to have 0 error. Only values that fall outside the
uncertainty range are taken into the reconstruction error sum. The reasoning behind
this approach is to be able to disregard reconstruction error that falls within the range
of expected values. This method therefore only reports errors that are considerably
different from the true values.
Bayesian Deep Learning for Vibration-Based Bridge Damage … 35
Fig. 4 Dataset visualized. Only the healthy bridge state data is used to train the models. The best
model is picked by performance on the validation data. The errors are quantified on the unseen test
data. The true damage detection performance is then assessed
2.3 Validation
The central idea in this method is to reject a small part of the healthy-state sequences
to obtain a rejection criterion. The sequences with the highest reconstruction error
are used as a benchmark for comparison sequences of a bridge in an unknown state.
This threshold is chosen to be the nth percentile value, where n is chosen to be one
of 95th, 99th, or 99.9th. The chosen percentile is not optimized in this study to find
the n with the best damage detection performance, to reflect how this method would
be used in a real-life SHM scenario where you do not have any damaged state data.
If significantly more sequences than the chosen percentile value in an unknown
state are rejected, it is assumed that the state of the bridge has changed and therefore
it is determined to be damaged. The higher the chosen percentile value, the more
sequences are needed to make a statistically significant determination of the bridge
state. Due to limitations of the chosen dataset of the Z24 bridge, namely the small
number of sequences in each damaged state, we chose to use the 95th percentile as
our rejection threshold to be able to confidently assess each bridge state.
Sequences with reconstruction error above the 95th percentile are rejected in
the healthy-state test set. This threshold gives a baseline of comparison of summed
sequence error of a bridge in an unknown state. By doing so, we know that 5% of
36 D. S. Ásgrímsson et al.
never seen, healthy-state sequences are rejected. If a significantly higher portion than
5% of sequences of a bridge in an unknown state are rejected, the bridge is considered
to be damaged. An example of the determined threshold can be seen in the Results
chapter (Fig. 6).
3 Experiments
One of the challenges of a machine learning-based bridge SHM is that there are very
few datasets that contain data of a bridge in a healthy state and in a damaged state.
The Z24 bridge benchmark [33, 34] has sensor recordings of a real full-sized bridge
that was monitored for almost a year before being deliberately damaged in a realistic
and controlled way. The damage was introduced in 15 steps, each step damaging the
bridge further in a realistic manner. This is useful to assess the model’s sensitivity, to
see at what stage the model can determine a damage state. The Z24 bridge was built in
1963—a typical posttensioned box girder highway bridge (Fig. 5) with a main span
of 30 m over an underlying road and two connecting side spans of 14 m [33]. Bridge
sensor measurements were recorded such as bridge vibrations, air temperature, soil
temperature, humidity, wind speed, and vehicle passing. Most of the environmental
sensors measured temperature. Ten-minute samples with a sampling rate of 100 Hz
were recorded by seven accelerometers.
The Z24 benchmark contains data when the temperature drops below freezing,
which has a big effect on the behavior of the bridge [4]. It is a challenge to assess
environmental effects versus damage events, as bridges can present drastic seasonal
changes [35]. In this method, events such as freezing temperatures receive no special
treatment. These events are included in the training and test datasets, and there-
fore included in the 95% rejection threshold. The dataset presents an opportunity to
determine a model’s ability able to assess such events and better assess environmental
changes versus damage cases. The Z24 bridge data is split into two parts, long-term
continuous sensor monitoring for a period of one year and a short-term progressive
Fig. 5 Side view of a three-span highway beam bridge, where the main span goes over a road. For
a more detailed view of the Z24 bridge, refer to [33]
Bayesian Deep Learning for Vibration-Based Bridge Damage … 37
Fig. 6 Distribution of healthy-state test set summed sequence errors, showing the determined 95%
rejection threshold for the uncertainty method model
damage with much higher spatial resolution instrumentation. The presented method
only uses data from the long-term sensor measurements.
3.2 Training
The healthy bridge state data is cut up into 10-min-long sequences. The sequences
are randomly sampled without substitution so that 70% belong to the training set,
15% to the validation set, and 15% to the test set. See Fig. 4 for a visualization of
the dataset split.
A feed-forward autoencoder neural network with ReLU activation function
(Fig. 2) is trained using dropout [27] on the healthy-state test dataset to reconstruct
the healthy-state bridge sensor data. The autoencoder reconstructs vibration sensor
data from all sensors at once. Environmental sensor data is also included in the input.
The model is trained on non-overlapping sliding time window inputs corresponding
to a 2 s time interval. The Adam optimizer [16] is used to train the model. The model
was trained on a cloud computer with 16 CPUs, 50 GB memory, and a 16 GB GPU
NVIDIA Tesla P100. The full training data was around 50 GB. The final parameters
of the trained model after hyperparameter search can be seen in Table 1.
38 D. S. Ásgrímsson et al.
4 Results
The main results can be seen in Table 2. Each row in the table represents a damage
case. Gray cells mark when over 5% of the 10-min sequences are rejected and there-
fore the bridge is determined to be in a damage state. Since there are varying number
of sequences available for each damage case, the results must be interpreted care-
fully. For example, damage case 2 had over 100 recorded sequences, and damage
case 13 had only 10.
The presented method can confidently detect damage cases 5–7, 9–11, and 13–
14, where all three methods determine the state to be damaged. Note that the three
different methods of detecting damage are all using the same trained model. The
sequence in Fig. 7 shows a comparison of the three presented reconstruction methods
for a healthy and damaged state.
Grayed cells mark where the bridge state is determined to be damaged, since over 5% of sequences
are classified as damaged
Bayesian Deep Learning for Vibration-Based Bridge Damage … 39
Fig. 7 All three model methods compared for a given input. The plotted data only shows one
sensor and a 50-ms window for clarity. Damage state 7 below, healthy state above. Take note of the
increasing uncertainty in the damage scenario. The true values in the damage scenario fall outside
the 2 standard deviation ranges of the predictions at around 20 and 40 ms
5 Discussion
The method of using the same model in several ways is a novel approach to structural
health monitoring, utilizing the trained model to a fuller extent. The standard method
is more performant in detecting damage, while the uncertainty method is much more
cautious in the damage assessment. The standard method is much better at detecting
damage, but it does not take uncertainty into account in any way. Therefore, in an
extreme weather event that is not present in the training data, this model would more
likely wrongly determine the bridge to be in a damage state. This is a known issue in
SHM, false damage detection. When the uncertainty method determines the bridge to
be in a damaged state, it is a more certain determination. This is important in real-life
usage of a structural health monitoring system to reduce false damage classification.
An interesting result is that being better at reconstruction does not equal being
better at detecting damage. The mean method achieves lower reconstruction error
40 D. S. Ásgrímsson et al.
than the standard method in all cases, healthy or damaged. This does not mean that
the mean method is better at detecting damage; in fact, it is the opposite that is true.
The standard method reconstructs healthy sequences adequately, but reconstructs
damaged sequences poorly and is therefore better at detecting the difference. This
does not mean that the standard method overfitted on the training data, since it does
not achieve remarkably lower reconstruction error on the training dataset compared to
the testing and validation datasets. The mean model and uncertainty model generalize
better to the damaged sequences and therefore achieve worse damage detection.
Why is the method failing for certain damage cases? As seen in Table 2, the
damage detection did not confidently reject cases 1–4, 8, and 12. Note that the
standard method rejected all damage cases, but the uncertainty was too great for the
uncertainty method to reject these cases.
The lowering of the pier is a rare occurrence in a real-life situation, but not unheard
of. The severity of such an event depends on the specific design of the structure. In a
three-span bridge with reasonable stiffness and length, as the Z24 bridge, a ~40 mm
lowering will not induce a significant damaging force. The lowering of a pier could
still be a dangerous scenario, but this will be heavily dependent on the structural
design of the bridge. The damage detection does not confidently reject this scenario.
It is interesting to note that the damage detection confidently rejects cases 6–7,
spalling of concrete at soffit. Such events are very common and are not immediately
dangerous as they do not directly affect the bearing capacity of the concrete deck. It
is possible that the sensors were located closely to the affected area, and therefore
the damage detection was more sensitive to the changes.
The method does not confidently reject damage case 8, “landslide of 1 m at abut-
ment.” Such an event will have a specific effect on the bridge response in particular
cases since the landfill supports the bridge only in a certain load scenario. During
the damage cases, the bridge was closed to passing traffic. This could explain that
the damage was not detected, since the agitation of the bridge was only coming from
traffic passing under the bridge which is not as substantial.
Damage case 12, “rupture of 2/16 tendons,” was not rejected confidently. The
rupture of tendons is rare and possibly dangerous but is very hard to detect from
standard visual inspection. Therefore, it might be more common than we think. The
main reason that this damage scenario was not detected is that there were very few
sequences, only 10—corresponding to only 120 min. It is therefore reasonable to say
that the damage detection could have been more accurate given a greater number of
sequences.
The presented method faces a well-known challenge in bridge structural health moni-
toring that damage is best assessed when the bridge is actively vibrating. This is
clearly shown in Fig. 8, where the reconstruction error drops during the night when
the bridge is in low use.
Bayesian Deep Learning for Vibration-Based Bridge Damage … 41
Fig. 8 Reconstruction error of a healthy-state data. Here, it is clearly shown that the reconstruction
error is dependent on activity, as the error drops notably down overnight
The healthy-state training data was captured, while the bridge was in full use.
Vehicles could drive over the bridge and under. When the bridge was damaged, it was
closed for traffic. For every damage case, the bridge was only excited by underneath
traffic which results in much lower vibration levels. Therefore, the damage detection
is even more challenging and requires a more sensitive damage detection.
The timescale of damage detection is more nuanced since the damage detection is
based on evaluating multiple sequences. The damage sensitivity is also dependent on
bridge activity. In the case of considerable damage, the model would start rejecting the
10-min sequences in succession. For the model to assess damage with more certainty,
it would need multiple 10-min sequences. Therefore, the timescale of reliable damage
detection is in the order of hours, possibly days.
6 Conclusion
The results are promising for the field of bridge structural health monitoring. The
autoencoder neural network model trained on the complete sensor data was able to
detect several realistic damage cases on the Z24 bridge benchmark [34], which has
been used in research for over 20 years. Due to the size of the full data and compu-
tational limitations, few researchers have presented methods of structural health
monitoring on the full data. To our knowledge, this is the first time a model has been
trained on the full healthy-state data and successfully detected damage cases while
taking no special care of freezing temperature variations of the healthy-state data.
The method is performant and is a prominent step toward modern infrastructure.
The methods’ ability to detect damage in realistic damage scenarios of a real bridge
has many economic, social, and ethical aspects. Such a system could possibly detect
damage before a traditional bridge inspector could and prevent a costly repair. The
model can reconstruct a full 10-min sequence and classify it in the order of seconds,
on a standard modern personal computer. This means that it is computationally very
fast and can be used in real time. The training of the model on the other hand requires
considerable computing power.
Acknowledgements We thank Dr. Edwin Reynders at KU Leuven for providing the Z24 benchmark
dataset; without the contribution, this work would not have been possible.
42 D. S. Ásgrímsson et al.
References
Abstract Due to the aging of civil infrastructure and the associated economic
impact, there is an increasing need to continuously monitor structural and non-
structural components for system life-cycle management, including maintenance
prioritization. For complex infrastructure, this monitoring process involves different
types of data sources collected at different timescales and resolutions, including but
not limited to abstracted rating data from human inspections, historical failure record
data, uncertain cost data, high-fidelity physics-based simulation data, and online
high-resolution structural health monitoring (SHM) data. The heterogeneity of the
data sources poses challenges to implementing a diagnostic/prognostic framework
for decision making for life-cycle actions such as maintenance. Using quoin blocks
components of a miter gate as an example, this chapter presents a holistic Bayesian
data analytics and machine learning (ML) framework to demonstrate how to inte-
grate various data sources using Bayesian and ML methods for effective SHM, and
Prognostics and Health Management (PHM). In particular, this chapter discusses
how Bayesian data analytics and ML methods can be applied to (1) diagnosis of
bearing loss-of-contact degradation in quoin blocks; (2) optimized sensor placement
for SHM on the gate; (3) fusion of various data sources for effective PHM; and
(4) deciding maintenance strategies by considering the behavioral aspect of human
decision making under uncertainty.
1 Introduction
Fig. 1 Diagnosis, prognosis, and maintenance decision-making framework for civil infrastructure
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 47
Fig. 2 FEM modeling of a leaf of a miter gate including gap length deterioration
For civil systems, the approach is usually carried out by using a physics-based model
(e.g., finite element (FE) model) of the structure [2–4]. It is fundamentally an inverse
problem because the system parameters are estimated from measured response quan-
tities. Generally, a physics-based model is a “forward” problem where the system
responses (e.g., FE output response) are predicted as a function of the (known) system
parameters (e.g., FE inputs). Synthetic system parameters may be used to obtain the
FE system response, whose responses can be compared to the “true” system as shown
in Fig. 2.
where P(·) is a probability operator, XtR = AtR , BtR , CtR , DtR , FtR , CFtR are the
damage ratings at time step t, and “|” is a conditional operator.
The above transition matrix obtained from inspection data can be used to estimate
the possible condition (and failure condition) after n time steps as follows:
R n
P Xn+k = P XkR · PReport , (2)
R R R R R R
where P XkR = P X 1,k , P X 2,k , P X 3,k , P X 4,k , P X 5,k , P X 6,k are the
condition probabilities at the current time or initial time (i.e., k = 0) where X 1,k
R
and
R R R
X 6,k represents Ak and CFk , respectively. The condition CF represents the complete
failure
R condition.
PReport is the transition matrix built from the discrete ratings, and
P Xn+k are the condition probabilities at n time steps in the future. The current
time P XkR may be either obtained from current inspections or using SHM data.
Human error can greatly affect the reliability of the inspection assessment. For
example, human psychology influences inspectors to make conservative or non-
conservative assessment that can greatly influence maintenance decisions. Bench-
mark data may be available to account for the accuracy of the assessment given the
inspector qualification, training, and certification [24]. A human observation error
matrix can be obtained/estimated as follows to probabilistically measure the human
error
⎡ h ⎤
h
P11 h
P12 ··· P16
⎢ Ph h
P22 ··· h ⎥
P26
⎢ 21 ⎥
Phuman = ⎢. .. .. .. ⎥
⎣ .. . . . ⎦
h
P61 h
P62 · · · P66
h
50 M. A. Vega et al.
⎡ R tr ⎤
P At At P AtR Bttr · · · P AtR CFtrt
⎢ P B R Atr P BtR Bttr · · · P BtR CFtrt ⎥
⎢ t t ⎥
= ⎢. .. .. .. ⎥, (3)
⎣ .. . . . ⎦
tr
P CF R Atr
t t P CF R B tr · · ·
t t P CF CF
t
R
t
The above Phuman models the behavior of an inspector that regularly tends to assess
condition than reality. Accounting for Phuman allows to
a component to be in a better
estimate thePtrue and P Xktr from PReport as will be discussed in Sect. 4.2. The terms
Ptrue and P Xktr are the true transition matrix and the true condition probabilities,
respectively, at the current time.
For damage diagnosis with limited SHM data, inverse modeling via physics-based
models such as finite element (FE) models have been used with data from SHM
systems to estimate parameters that infer some form of damage state. A fast, efficient
simulation of complex FE models is essential for appropriately fast damage diagnosis.
Depending on the dimension space of the inputs and outputs of interest, different
machine learning techniques can be chosen to build “cheap” yet accurate surrogate
models of the physics-based FE model. Two of the commonly used techniques,
artificial neural networks and Gaussian process regression, are briefly summarized
as below.
Artificial Neural Network (ANN). ANNs are an attractive option for surrogate
emulation of FE models. This type of supervised learning model works well with
classification (for discrete classes) and regression (for continuous processes) prob-
lems. However, ANNs are effective when a large amount of data is available, and
they are built to create point estimates rather than probabilistic estimates. Some
researchers [25–28] have used Bayesian inference to estimate the ANN’s weight and
model parameters, which has been referred to as Bayesian neural networks (BNN).
BNNs are good for high-dimensional spaces and better handle the issue of limited
data availability.
Gaussian Process Regression (GP). GP (or Kriging) models are an attractive option
for surrogate architectures because they are built to quantify the uncertainty in the esti-
mations rather than simply point-based estimates, as most other supervised learning
models (e.g., ANNs, support vector machines) do. Several researchers use GP regres-
sion to build Bayesian prediction models for civil engineering structures [23, 29]. A
GP surrogate model, γ̂ j = ĝ j (x) is defined as
where μ j (x) and σ j2 (x) are the mean and variance of the prediction of γ j , respectively,
for the input x.
In applications such as the miter gate presented in this chapter, the output space
of the simulation or the sensors available (e.g., hundreds or thousands of nodes in
the FE model) can be large. Also, it is known that sensors located close to each other
may contain highly correlated information. Therefore, dimension reduction tech-
niques are usually used in conjunction with GP models or ANNs to build surrogate
models for the physics-based FE model. A commonly used technique is singular
value decomposition (SVD). SVD is a linear algebra technique used to transform
high-dimensional matrices into a reduced dimensional space preserving most of the
original information. Figure 3 presents a generalize procedure of surrogate modeling
based on a GP model and dimension reduction methods for a model with inputs x and
Fig. 3 Building surrogate model for FEM model using GP and dimension reduction
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 53
θ . Following that, Fig. 4 shows the prediction process using the trained Kriging model
for given FEM input parameters and features obtained from dimension reduction.
The proposed framework explained in this work is based on a surrogate model
using a Kriging model combined with SVD to develop a fast emulator of the FEM.
Figure 5 shows the testing accuracy obtained at 46 sensors installed in the Greenup
miter gate at a particular point in time, which shows how close ML-based surrogate
model emulates the original FE model.
Given the SHM data and a ML-based surrogate model of the physics-based FE model,
damage diagnosis can be performed using Bayesian estimation of parameters, an , that
directly relate to a damage mode (e.g., cracks, gap, thickness loss due to corrosion, etc.
See Fig. 2 for examples). Bayesian estimation of this parameters can be performed
based on the following state-space equation
where h(an−1 ) is the state equation that describes the evolution of an over time, εh
is the process noise, ĝ(an , xn ) is the ML surrogate model built in Sect. 3.1 where xn
represents the known/measurable input variables to the physics-based model (e.g.,
loads, geometry, material properties, etc.), and sn is the observations from SHM
system as discussed in Sect. 2.4.
Note that the state equation given in Eq. (8) may be unknown due to a lack
of understanding of the damage evolution mechanism. In that situation, a random
walk type equation can be used as the state equation with large process noise as
discussed in [23]. Assume there are NS strain sensors installed in a structure such as
the miter gate and the damage parameter, an , to infer at time step tn is the extent of the
loss of bearing contacting (i.e., gap length). Then, the posterior probability density
function of the gap length an at time step tn conditioned on strain measurements s1:n
is estimated using Bayesian inference method recursively as follows
in which f (sn |an ) is the likelihood function, obtained from the measurement equa-
tion, of observing sn for given an at time step tn , and the term f (an |an−1 ) represents
the probability distribution of an for a given an−1 obtained using the state equation,
which describes the damage evolution over time.
The recursive Bayesian updating of Eqs. (9) and (10) is analytically intractable.
In practical application, various filtering methods, such as particle filtering [30],
extended Kalman filter [31], and unscented Kalman filter [32], have been developed
to approximate recursive updating process.
Figure 6 presents an illustrative example of strain measurement data sn of 10
sensors (NS = 10) over a certain time period of interest. After that, Fig. 7 shows the
damage diagnostics results of gap length of miter gate (i.e., the gap damage in the
miter gate as shown in Fig. 2) over time using recursive Bayesian updating and the
ML-based surrogate models. It shows that the recursive Bayesian inference method
can effectively perform damage diagnostics by fusing the information from physics-
based simulation model and SHM data. More details of the integration of ML-based
surrogate model and SHM observation data using Bayesian recursive updating for
damage diagnostics are available in [23].
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 55
Fig. 7 Gap length diagnostics result using Bayesian method and ML surrogate model (note that
gap here refers to the damage mode shown in Fig. 2)
Sensor placement optimization (SPO) plays a critical role in improving the effec-
tiveness of the SHM system for damage diagnostics (see Eqs. 9 and 10). Designing
an optimal sensor network is very challenging in practice since the observations are
not available in the design stage. In that case, the physics-based simulation model
needs to be employed to provide information on how to properly allocate the sensors.
56 M. A. Vega et al.
d∗ = argmax{(d)},
d∈d (11)
s.t. C(d) ≤ Ce ,
where d is a sensor network design, d is the sensor design domain, (d) is a cost
function, C(d) is the total cost of the sensor network, and Ce is the allowable budget.
The cost function (d) may be formulated from different perspectives. For
example, probability of detection [33], Bayes risk [34], and information gain [35]
have been used as cost functions in sensor placement design optimization. Taking the
information gain measured by the Kullback–Leibler (KL) divergence as an example,
(d) is formulated as
(d) = f θ (θ ) f s|θ,d (s|θ, d)DKL (d, s)dsdθ , (12)
θ s|θ
where DKL (d, s) = θ f θ |s,d (θ|s, d)log f θ|s,d (θ|s, d)/ f θ (θ) dθ is the KL divergence
(e.g., relative entropy) for given observations s and sensor placement design d,
which measures the difference (i.e., information gain) between the prior and posterior
distributions of damage state variables θ .
Solving the sensor placement optimization model given in Eq. (11) is extremely
challenging due to the high computational effort required in the repeated evaluations
of Eq. (12). Bayesian data analytics and machine learning techniques are essential
to overcome the challenge. First, ML models and Bayesian inference methods as
discussed in Sects. 3.1 and 3.2 enable for the efficient estimation of the posterior
distributions of the damage states variables. More importantly, ML-based optimiza-
tion methods make it possible to solve the model given in Eq. (11) when a large
number of sensors need to be allocated to a large civil infrastructure.
For example, when sensors need to be placed on the miter gate to detect the
“gap” of the gate due to damage, the dimension of the design variables given in
Eq. (11) will be very large. Assuming that 20 sensors need to be placed, the number
of design variables will be 60, if the three-dimensional coordinates of each sensor
are considered as design variables. In that case, directly solving the optimization
model of Eq. (11) will be computationally prohibitive. Alternatively, a greedy-based
framework can be employed to place the sensor one-by-one. In order to identify
the optimal placement of the ith sensor, the optimization model given in Eq. (11) is
re-formulated as
di∗ = argmax di , d1:i−1 ∗
,
di ∈d
∗
(13)
s.t. C(d) ≤ Ce and d = di ∪ d1:i−1 ,
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 57
∗
where d1:i−1 are the coordinates of the previous i − 1 sensors and di are the
coordinates of the ith sensor.
Through the formulation of the model in Eq. (13), the dimension of the design
variables is reduced to 3 in each iteration of the greedy optimization scheme. Even
for the three-dimensional optimization model, the global optimization of Eq. (13)
is still computationally challenging. Bayesian optimization method, which is also
known as the efficient global optimization method, can be employed to efficiently
solve the optimization model by leveraging the prediction capability of the Gaussian
process model [36, 37]. GP-based Bayesian optimization is a process of adaptively
training a GP surrogate model for the objective function of an optimization model. In
each iteration of the adaptive training of the GP, training data are identified as those
design locations which have the highest probability of being the maximum/minimum
design point. The key to the GP-based optimization is the definition of the expected
improvement function (EIF)
μ(d) − ϕ ∗ μ(d) − ϕ ∗
EIF(d) = μ(d) − ϕ ∗ + σ (d)φ , (14)
σ (d) σ (d)
where φ(·) and (·) are, respectively, the probability density function and cumulative
distribution function of a standard normal random variable, ϕ ∗ is the current best
values in the training dataset, μ(d) and σ (d) are the mean and standard deviation of
the GP surrogate model prediction.
By maximizing the EIF in Eq. (14), new training points may be identified to
adaptively refine the GP surrogate model to approach the optimal value. Figure 8
shows an illustrative example of Bayesian optimization using GP.
As shown in Fig. 8, an initial GP surrogate model is trained first. Based on the
trained GP surrogate model, the EIF values over the design space are computed as
shown in Fig. 8a. By maximizing the EIF, a new training point is identified, and the GP
model is retrained in the second iteration as shown in Fig. 8b. After a few of iterations,
as shown in Fig. 9, the maximum point can be identified. These iterations show that
the GP-based optimization needs very few evaluations of the objective function to
identify the global optimization, which is much more efficient than the other global
optimization algorithms, such as genetic algorithm and simulated annealing.
The ML-based optimization method allows us to effectively allocate the optimal
sensors and thereby increases the effectiveness of damage diagnostics as discussed
in Sect. 3.2. Note that the greedy algorithm-based sensor placement design is an
approximation of the original model. It may not find the “true” globally optimized
sensor network. This limitation can also be mitigated using other ML methods, such
as the reinforcement learning method, which may be better suited for dynamic opti-
mization problems. Figure 10 presents an example of sensor placement optimization
results of miter gate obtained using the method presented in this section. More details
are available in [38].
58 M. A. Vega et al.
Failure prognostics is a process of predicting the end of life (EOL) of civil infrastruc-
tures to inform life-cycle management. Based on the state estimation from failure
diagnostics as discussed in Sect. 3, the state of the system from the current time to a
future time is obtained to predict the potential failure time or estimate the remaining
useful life (RUL) of the system. Figure 11 shows a schematic of how to use the
predictions to calculate the EOL and RUL distributions.
An essential part of the failure prognostics or RUL estimation is the degrada-
tion modeling since it is required to perform the projection of the state into future.
To build such degradation models for prognosis purposes, researchers have tried to
model the evolution/degradation of damage using physical degradation models such
as applications in fatigue crack growth [39–41] and corrosion growth [6, 7]. These
physical degradation models are developed based on the understanding of the phys-
ical behavior and are usually validated by experiments. On the other hand, empirical
degradation models are used when the evolution/degradation of damage is not well
understood either due to the limited understanding of the physical phenomenon or
when the damage state cannot be measured continuously but rather occasionally.
Approaches that combine physics-based approaches with data-driven methods have
also been developed in recent years and lead to a group of hybrid approaches. A
60 M. A. Vega et al.
As been discussed in Sect. 2.2, inspection data contains some degradation information
of civil infrastructures even though they are highly abstracted. Since the damage state
estimated from SHM is in continuous form, the continuous state can be converted into
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 61
inspection data in discrete state through certain protocols. Taking the gap damage of
a miter gate given in Fig. 2 as example, as shown in Fig. 12, the estimated gap length
can be converted into gap state, which can then be used for failure prognostics using
the gap state transition matrix given in Sect. 2.2.
More specifically, a certain protocol is usually needed to map the continuous
damage state to the abstracted discrete state of inspection data. For instance, the gap
length at of a miter gate (see Fig. 2) can be mapped into a gap state X i,t based on an
engineering protocol as follows
⎧
⎪
⎪ X 1,t = A, at ∈ [0, β1 ]
⎪
⎪
⎪
⎪ X 2,t = B, at ∈ [β1 , β2 ]
⎪
⎨
X 3,t = C, at ∈ [β2 , β3 ]
R = h OCA (at , β) = , (15)
⎪
⎪ X 4,t = D, at ∈ [β3 , β4 ]
⎪
⎪
⎪
⎪ X = F, at ∈ [β4 , β5 ]
⎪
⎩
5,t
X 6,t = CF, at ∈ [β5 , ∞)
where the PDF f (an |s1:n ) of damage parameter an , is estimated using the Bayesian
updating method presented in Sect. 3.2.
Once the gap state at current time step is estimated, the damage state of the failure
m time steps into the future is obtained through the transition matrix given in Sect. 2.2
as
62 M. A. Vega et al.
in which PReport and P(Xn |s1:n ) are, respectively, given in Eqs. (1) and (16).
Using Eq. (17), the RUL for the system based on the current damage state and the
future failure state can be estimated. Details on such predictions can be found in [23].
This approach assumes that the transition matrix PReport can accurately represent the
underlying degradation pattern of the structure. However, this assumption is usually
not true since the abstracted ratings are often biased by human errors as discussed
in Sect. 2.3. To tackle this issue, an approach has been proposed in [10] to map
the reported transition matrix to a more useful transition matrix that has eliminated
some of the effects of the human errors. In order to map the reported rating transition
matrix PReport to the underlying “true” transition matrix PTrue , the underlying true
transition rating is defined at time t as X ttr and that at t + 1 as X t+1
tr
. Similarly, the
obs
reported ratings from field engineers are defined at time t as X t and that at time t
+ 1 as X t+1
obs
. Based on these definitions and using the human error matrix given in
Eq. (3), PReport can then be mapped into PTrue by following the procedure shown in
Fig. 13. More details of Fig. 13 are available in [1]. Once PTrue is obtained, it can be
used to substitute PReport in Eq. (17) to get more accurate failure prognostics results.
In addition, a Bayesian method has also been developed in [23] to update the errors
in the transition matrix using SHM data. The advantage of the approach presented
here is that it requires minimal information for failure prognostics. The disadvan-
tage is that there is very large uncertainty in the obtained RUL estimation results.
Alternatively, a stochastic continuous degradation model can be built to improve the
confidence of such predictions of the damage parameters which leads to better failure
prognostics.
Using the gap growth of miter gate as an example, as shown in Fig. 14, an alternative
approach to perform failure prognostics is to map the abstracted inspection data into
a degradation model in continuous space and then perform failure prognostics in the
continuous space. It is expected that prognostics in continuous space can increase
the confidence of RUL estimation.
For a transition matrix PReport given in Sect. 2.2 or PTrue obtained in Sect. 4.1 after
mitigating human error bias, a continuous degradation model can be obtained through
Bayesian or optimization-based calibration. More details of the mapping from inspec-
tion data to a degradation model are available in [10]. Here, a brief summary of
the major steps is presented. When optimization-based method is employed, the
optimization model is given by
gopt (θ ; PTrue ) = P̂(θ) − PTrue2 , (18)
where P̂(θ ) is the simulated transition matrix for a given degradation model with
parameters θ.
The most critical part is how to estimate θ for given PTrue . To do that, P̂(θ )
is calibrated for any given θ . The degradation model is modeled as a multi-stage
degradation model as follows
da(t)
= exp σ j (t)U (t) Q j (t)(a(t))w j (t) , ∀ j = 1, . . . , Nd , (19)
dt
where a(t) is the gap length at time t, Q i and wi are stage-dependent degradation
parameters, σi is a standard deviation variable of degradation stage i, and U (t) is a
stationary standard Gaussian process with autocorrelation function given by
Fig. 14 Failure prognostics in continuous space by mapping inspection data into a degradation
model
64 M. A. Vega et al.
1 , θ 2 , θ 3 , θ 4 , θ 5 , e1 , e2 , e3 , e4 , σ), where
θ (θ
(22)
θ j σ j , ζ j , Q j , w j , j = 1, 2, . . . 5 .
Then, the degradation model can be estimated as follows using the optimization
model given in Eq. (18). Figure 15 summarizes the overall procedure of estimating
P̂(θ ) for a given degradation model and model parameters θ .
Once the degradation model is available, it can be used for failure prognostics in
the continuous space as illustrated in Fig. 11. This procedure, however, is not limited
to the model given in Eq. (19). It is also applicable to other degradation models
and can be integrated into a Bayesian framework. More detailed discussions of this
approach can be found in [10].
Figure 16 shows a comparison of the RUL estimates obtained using the approaches
presented in Sect. 4.1 (denoted as TM mean prediction, TM Conf. limit) and Sect. 4.2,
respectively. It shows that mapping inspection data into a continuous degradation
model can significantly increase the confidence of the failure prognostics results.
Failure diagnostics and prognostics of civil infrastructure usually require the usage of
multiple models including degradation model [39–41] and physics-based model [12–
14], as discussed earlier. In addition to various analysis models, heterogeneous data
and uncertainty sources are involved in the process of diagnostics and prognostics. A
flexible tool that can be used to tackle the challenges in diagnostics and prognostics
caused by the heterogeneity of model and data sources is Bayesian networks (BN).
A BN, which is also called probabilistic graphic model, is a directed acyclic
graph that connects different variables in a probabilistic way. It allows for the flexible
integration of multi-type of models and information sources in a systematic Bayesian
framework, and thereby enable decision makers to update information and reduce
uncertainty in a holistic manner [43]. Due to its capability of fusing information and
data sources, BN plays a vital role in building digital twins for SHM in various assets,
from aerospace engineering to civil and mechanical engineering [40, 44].
66 M. A. Vega et al.
Fig. 17 An example of
dynamic Bayesian network
n
p(X) = p(X 1 , X 2 , . . . , X n ) = p(X i |πi ), (23)
i=1
where πi is a set of parent nodes of X i , p(X i |πi ) is the conditional probability mass
(CPM) function or conditional probability density (CPD) function, and nodes without
parent nodes are called root nodes. For root nodes, assume p(X i |πi ) = p(X i ).
A type of widely used BN in failure diagnostics and prognostics is the dynamic
Bayesian network (DBN). Figure 17 shows a simple example of DBN. As shown
in this figure, the DBN consists of state variable denoted by at and measurement
variable represented by st . The CPD function f (at |at−1 ) describes the transition of
the state variable at over time and h(st−1 |at−1 ) models the probabilistic relation-
ship between the state variable and measurement variable. When the state variables
are variables related to the failure modes or degradation stages of the civil infras-
tructure, the measurements collected from measurement variable st can be used for
failure diagnostics using Bayesian inference methods. Based on the failure diagnos-
tics or estimation of damage related state variables, failure prognostics may then be
performed according the transition of state variables over time, which is governed
by the CPD function f (at |at−1 ).
In practical engineering applications, the node in the DBN can be a mathematical
model, a finite element model, or a data-driven machine learning model. The DBN
used for failure diagnostics and prognostics can be constructed using physics-based
method [12–14], Bayesian network learning method [44], or a hybrid of physics and
data-driven methods [19, 20].
Figure 18 shows a schematic dynamic Bayesian network at one time instant for
the failure diagnostics and prognostics of a miter gate (see Fig. 2), which is used an
example to explain the presented approaches in this chapter. Note that, for the sake
of simplification, the transient BN is not depicted. As shown in this figure, the DBN
connects variables of a degradation model with variables of a strain analysis model
of the miter gate. For example, the parent nodes of node e1 in Fig. 18 are nodes σe
and μ1 . The CPDs in the DBN can be derived according to the approaches discussed
in Sects. 3.2 and 4.2 of this chapter. Using the strain measurement collected from
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 67
Fig. 18 A schematic Bayesian network of a miter gate at one time instant for diagnostics and
prognostics
SHM system, the uncertain variables of the degradation model can be updated. The
updated degradation model can then be used to estimate RUL of the gate dynamically
over time.
Figure 19 shows an illustrative example of the posterior distribution updating of
parameter w1 over time using the strain measurements of the miter gate. Following
that, Fig. 20 shows the RUL estimation of the miter gate over time. It shows that the
confidence of RUL estimation increases over time as more and more measurements
are collected. Additionally, it is worth mentioning that the degradation model and
strain analysis model are updated in an integrated manner in the DBN-based frame-
work. This example illustrates the flexibility of DBN in connecting multiple models
68 M. A. Vega et al.
for failure diagnostics and prognostics. A good application of DBN for the failure
diagnostics and prognostics is available in [45].
Moreover, the degradation model parameters obtained in Sect. 4.2 using optimiza-
tion or Bayesian-based method can be used as prior information for the DBN-based
method.
In summary, this section presents three different approaches for failure diagnostics
using Bayesian data analytics and machine learning. The advantage of the approach
presented in Sect. 4.1 requires minimal information for failure prognostics. The
advantages of the approaches presented in Sects. 4.2 and 4.3 are that they can provide
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 69
prediction results with less uncertainty. Next, the maintenance planning based on
failure diagnostics and prognostics is briefly introduced.
continuous state deterioration (such as in Sects. 4.2 and 4.3). A most extensive list
of CBM application can be found in here [50–53]. Most of the CBM applications
available in the literature are for mechanical systems, aerospace systems, or manu-
facturing systems. For large civil engineering infrastructure, most of the applications
have been applied to bridge engineering [54–56]. In CBM, maintenance schedules
are predicted based on the results from diagnosis and prognosis, as discussed in
Sects. 3 and 4. For diagnosis and prognosis, CBM approaches can be classified into
physics-based approach [12–14], data-driven approach [15–18], and hybrid approach
[19, 20]. The approaches presented in this book chapter can be classified as hybrid
approaches since they combine physics-based approach with data-driven approach
to improve CBM predictive capabilities.
6 Conclusion
References
1. Vega MA (2020) Diagnosis, prognosis, and maintenance decision making for civil infrastruc-
ture. UC San Diego
2. Moaveni B, Conte JP, Hemez FM (2009) Uncertainty and sensitivity analysis of damage iden-
tification results obtained using finite element model updating. Comput Civ Infrastruct Eng
24:320–334. https://doi.org/10.1111/j.1467-8667.2008.00589.x
3. Jang S, Li J, Spencer BF (2013) Corrosion estimation of a historic truss bridge using model
updating. J Bridge Eng 18:678–689. https://doi.org/10.1061/(ASCE)BE.1943-5592.0000403
4. Vega MA, Ramancha MR, Conte JP, Todd MD (2020) Efficient Bayesian inference of miter
gates using high-fidelity models. In: Model validation and uncertainty quantification, volume
3: proceedings of the 38th IMAC, a conference and exposition on structural dynamics 2020,
Houston, TX. Springer
5. Kim N-H, An D, Choi J-H (2017) Prognostics and health management of engineering systems.
Springer International Publishing, Cham. https://doi.org/10.1007/978-3-319-44742-1
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 71
6. Guedes Soares C, Garbatov Y, Zayed A (2011) Effect of environmental factors on steel plate
corrosion under marine immersion conditions. Corros Eng Sci Technol 46:524–541. https://
doi.org/10.1179/147842209X12559428167841
7. Wang C, Elsayed EA (2020) Stochastic modeling of corrosion growth. Reliab Eng Syst Saf
204:107120. https://doi.org/10.1016/j.ress.2020.107120
8. Si X-S, Wang W, Hu C-H, Zhou D-H (2011) Remaining useful life estimation—a review on
the statistical data driven approaches. Eur J Oper Res 213:1–14. https://doi.org/10.1016/j.ejor.
2010.11.018
9. Wang Z, Shafieezadeh A (2020) Real-time high-fidelity reliability updating with equality infor-
mation using adaptive Kriging. Reliab Eng Syst Saf 195:106735. https://doi.org/10.1016/j.ress.
2019.106735
10. Vega MA, Hu Z, Fillmore TB, Smith MD, Todd MD (2021) A novel framework for integration
of abstracted inspection data and structural health monitoring for damage prognosis of miter
gates. Reliab Eng Syst Saf 211:107561
11. Mitchell JS (2007) From vibration measurements to condition based maintenance seventy years
of continuous progress. Sound Vib 41:62–78
12. Orchard ME, Vachtsevanos GJ (2007) A particle filtering approach for on-line failure prognosis
in a planetary carrier plate. Int J Fuzzy Log Intell Syst 7:221–227. https://doi.org/10.5391/IJFIS.
2007.7.4.221
13. Daigle M, Goebel K (2011) Multiple damage progression paths in model-based prognostics.
In: 2011 aerospace conference. IEEE, pp 1–10. https://doi.org/10.1109/AERO.2011.5747574
14. An D, Choi J-H, Schmitz TL, Kim NH (2011) In situ monitoring and prediction of progressive
joint wear using Bayesian statistics. Wear 270:828–838. https://doi.org/10.1016/j.wear.2011.
02.010
15. Zio E, Di Maio F (2010) A data-driven fuzzy approach for predicting the remaining useful life
in dynamic failure scenarios of a nuclear system. Reliab Eng Syst Saf 95:49–57. https://doi.
org/10.1016/j.ress.2009.08.001
16. Mohanty S, Das S, Chattopadhyay A, Peralta P (2009) Gaussian process time series model for
life prognosis of metallic structures. J Intell Mater Syst Struct 20:887–896. https://doi.org/10.
1177/1045389X08099602
17. Galar D, Kumar U, Lee J, Zhao W (2012) Remaining useful life estimation using time trajectory
tracking and support vector machines. J Phys Conf Ser 364:012063. https://doi.org/10.1088/
1742-6596/364/1/012063
18. Ye Z-S, Xie M (2015) Stochastic modelling and analysis of degradation for highly reliable
products. Appl Stoch Model Bus Ind 31:16–32. https://doi.org/10.1002/asmb.2063
19. Xu J, Xu L (2011) Health management based on fusion prognostics for avionics systems. J
Syst Eng Electron 22:428–436. https://doi.org/10.3969/j.issn.1004-4132.2011.03.010
20. Liao L, Kottig F (2014) Review of hybrid prognostics approaches for remaining useful life
prediction of engineered systems, and an application to battery life prediction. IEEE Trans
Reliab 63:191–207. https://doi.org/10.1109/TR.2014.2299152
21. Spencer BF, Hoskere V, Narazaki Y (2019) Advances in computer vision-based civil infrastruc-
ture inspection and monitoring. Engineering 5:199–222. https://doi.org/10.1016/j.eng.2018.
11.030
22. Gibb S, La HM, Le T, Nguyen L, Schmid R, Pham H (2018) Nondestructive evaluation sensor
fusion with autonomous robotic system for civil infrastructure inspection. J Field Robot 35:988–
1004. https://doi.org/10.1002/rob.21791
23. Vega MA, Hu Z, Todd MD (2020) Optimal maintenance decisions for deteriorating quoin
blocks in miter gates subject to uncertainty in the condition rating protocol. Reliab Eng Syst
Saf 204:107147. https://doi.org/10.1016/j.ress.2020.107147
24. Campbell LE, Connor RJ, Whitehead JM, Washer GA (2020) Benchmark for evaluating perfor-
mance in visual inspection of fatigue cracking in steel bridges. J Bridge Eng 25:04019128.
https://doi.org/10.1061/(ASCE)BE.1943-5592.0001507
25. Yin T, Zhu H (2018) Probabilistic damage detection of a steel truss bridge model by optimally
designed bayesian neural network. Sensors 18:3371. https://doi.org/10.3390/s18103371
72 M. A. Vega et al.
26. Chua CG, Goh ATC (2005) Estimating wall deflections in deep excavations using Bayesian
neural networks. Tunn Undergr Space Technol 20:400–409. https://doi.org/10.1016/j.tust.2005.
02.001
27. Arangio S, Bontempi F (2015) Structural health monitoring of a cable-stayed bridge with
Bayesian neural networks. Struct Infrastruct Eng 11:575–587. https://doi.org/10.1080/157
32479.2014.951867
28. Vega MA, Todd MD (2020) A variational Bayesian neural network for structural health moni-
toring and cost-informed decision-making in miter gates. Struct Health Monit. https://doi.org/
10.1177/1475921720904543
29. Parno M, O’Connor D, Smith M (2018) High dimensional inference for the structural health
monitoring of lock gates, 1–29
30. Doucet A, de Freitas N, Gordon N (2001) Sequential Monte Carlo methods in practice. Springer
New York, New York, NY. https://doi.org/10.1007/978-1-4757-3437-9
31. Mochnac J, Marchevsky S, Kocan P (2009) Bayesian filtering techniques: Kalman and extended
Kalman filter basics. In: 2009 19th international conference radioelektronika. IEEE, pp 119–
122. https://doi.org/10.1109/RADIOELEK.2009.5158765
32. Julier SJ, Uhlmann JK (2004) Unscented filtering and nonlinear estimation. Proc IEEE 92:401–
422. https://doi.org/10.1109/JPROC.2003.823141
33. Tenney R, Sandell N (1981) Detection with distributed sensors. IEEE Trans Aerosp Electron
Syst AES-17:501–510. https://doi.org/10.1109/TAES.1981.309178
34. Flynn EB, Todd MD (2010) A Bayesian approach to optimal sensor placement for structural
health monitoring with application to active sensing. Mech Syst Signal Process 24:891–903.
https://doi.org/10.1016/j.ymssp.2009.09.003
35. Nath P, Hu Z, Mahadevan S (2017) Sensor placement for calibration of spatially varying model
parameters. J Comput Phys 343:150–169. https://doi.org/10.1016/j.jcp.2017.04.033
36. Jones DR, Schonlau M, Welch WJ (1998) Efficient global optimization of expensive black-box
functions. J Glob Optim 13:455–492
37. Hu Z, Du X (2015) Mixed efficient global optimization for time-dependent reliability analysis.
J Mech Des 137. https://doi.org/10.1115/1.4029520
38. Yang Y, Chadha M, Hu Z, Vega MA, Parno MD, Todd MD (2021) A probabilistic optimal sensor
design approach for structural health monitoring using risk-weighted f-divergence. Mech Syst
Signal Process 161:107920
39. An D, Choi J-H, Kim NH (2012) Identification of correlated damage parameters under noise
and bias using Bayesian inference. Struct Health Monit 11:293–303. https://doi.org/10.1177/
1475921711424520
40. Li C, Mahadevan S, Ling Y, Choze S, Wang L (2017) Dynamic Bayesian network for aircraft
wing health monitoring digital twin. AIAA J 55:930–941. https://doi.org/10.2514/1.J055201
41. Leung MSH, Corcoran J, Cawley P, Todd MD (2019) Evaluating the use of rate-based moni-
toring for improved fatigue remnant life predictions. Int J Fatigue 120:162–174. https://doi.
org/10.1016/j.ijfatigue.2018.11.012
42. Cubillo A, Perinpanayagam S, Esperon-Miguez M (2016) A review of physics-based models
in prognostics: application to gears and bearings of rotating machinery. Adv Mech Eng
8:168781401666466. https://doi.org/10.1177/1687814016664660
43. Sankararaman S, Ling Y, Mahadevan S (2011) Uncertainty quantification and model validation
of fatigue crack growth prediction. Eng Fract Mech 78:1487–1504. https://doi.org/10.1016/j.
engfracmech.2011.02.017
44. Hu Z, Mahadevan S (2018) Bayesian network learning for data-driven design. ASCE-ASME
J Risk Uncertainty Eng Syst Part B Mech Eng 4:1–12. https://doi.org/10.1115/1.4039149
45. Rafiq MI, Chryssanthopoulos MK, Sathananthan S (2015) Bridge condition modelling and
prediction using dynamic Bayesian belief networks. Struct Infrastruct Eng 11:38–50. https://
doi.org/10.1080/15732479.2013.879319
46. Yam RCM, Tse PW, Li L, Tu P (2001) Intelligent predictive decision support system for
condition-based maintenance. Int J Adv Manuf Technol 17:383–391. https://doi.org/10.1007/
s001700170173
Diagnosis, Prognosis, and Maintenance Decision Making for Civil … 73
47. Weibull W (1951) A statistical distribution function of wide applicability. J Appl Mech
103:293–297
48. Barlow R, Hunter L (1960) Optimum preventive maintenance policies. Oper Res 8:90–100.
https://doi.org/10.1287/opre.8.1.90
49. Ahmad R, Kamaruddin S (2012) An overview of time-based and condition-based mainte-
nance in industrial application. Comput Ind Eng 63:135–149. https://doi.org/10.1016/j.cie.
2012.02.002
50. Zhu Y, Elsayed EA, Liao H, Chan LY (2010) Availability optimization of systems subject to
competing risk. Eur J Oper Res 202:781–788. https://doi.org/10.1016/j.ejor.2009.06.008
51. Tian Z, Jin T, Wu B, Ding F (2011) Condition based maintenance optimization for wind power
generation systems under continuous monitoring. Renew Energy 36:1502–1509. https://doi.
org/10.1016/j.renene.2010.10.028
52. Tian Z, Liao H (2011) Condition based maintenance optimization for multi-component systems
using proportional hazards model. Reliab Eng Syst Saf 96:581–589. https://doi.org/10.1016/j.
ress.2010.12.023
53. Alaswad S, Xiang Y (2017) A review on condition-based maintenance optimization models for
stochastically deteriorating system. Reliab Eng Syst Saf 157:54–63. https://doi.org/10.1016/j.
ress.2016.08.009
54. Petcherdchoo A, Neves LA, Frangopol DM (2008) Optimizing lifetime condition and reliability
of deteriorating structures with emphasis on bridges. J Struct Eng 134:544–552. https://doi.
org/10.1061/(ASCE)0733-9445(2008)134:4(544)
55. Saydam D, Frangopol DM (2015) Risk-based maintenance optimization of deteriorating
bridges. J Struct Eng 141:04014120. https://doi.org/10.1061/(ASCE)ST.1943-541X.0001038
56. Gong C, Frangopol DM (2020) Condition-based multiobjective maintenance decision making
for highway bridges considering risk perceptions. J Struct Eng 146:04020051. https://doi.org/
10.1061/(ASCE)ST.1943-541X.0002570
Real-Time Machine Learning for
High-Rate Structural Health Monitoring
S. Laflamme (B)
Department of Civil, Construction, and Environmental Engineering,
Iowa State University, Ames, IA 50021, USA
e-mail: [email protected]
C. Hu
Department of Mechanical Engineering, Iowa State University, Ames, IA 50021, USA
J. Dodson
Air Force Research Laboratory, Munitions Directorate, Eglin AFB, FL 32542, USA
1 Introduction
Fig. 1 Example of high-rate system: a experimental setup; and b time series signals from five
consecutive tests
For example, an MRAS was applied to the DROPBEAR to identify the position of a
moving cart mimicking some high-rate features. The algorithm identified the posi-
tion of the cart online through the estimation of the system’s fundamental frequency
using a sliding mode observer [14]. The authors in [25] used MRAS to estimate the
stiffness of the same moving cart in real time, also leveraging sliding mode theory. An
average computing time of 93 µs was reported. DROPBEAR was used in [8], where
the authors proposed to estimate the cart position by matching extracted frequen-
cies to pre-generated finite element models. The authors implemented the real-time
algorithm in a field-programmable gate array and evaluated experimentally, with a
reported 4.04 ms computing time with minimum error and an improved iteration
time minimizing errors and computation time of 0.83 ms to match the pre-generated
models.
Other work was conducted on data generated from the droptower tests shown
in Fig. 1. The authors in [12] proposed a wavelet neural network to conduct signal
estimations. A particularity of the algorithm was its on-the-edge learning capability,
without pre-training, provided by a self-organizing input mechanism that permitted
adaptation of the neuro inputs to local stationarities. While the algorithm showed
promise for online applications, its computing time was not suitable for high-rate
applications. Inspired by this work, a fast deep learning algorithm was proposed in
[1], but with real-time HRSHM applicability. An average computing time of 25 µs
was reported.
In this chapter, opportunities and limitations in using real-time learning for
HRSHM are discussed, with an emphasis on physics-informed methods that care-
fully craft the input of the representation enabling time series feature extraction.
The discussion starts by reviewing opportunities in crafting the input space [10] for
accelerating convergence speed in estimating non-stationary time series. After, the
deep learning strategy developed in [1] for fast real-time applications is presented.
Lastly, a path in conducting rapid real-time state estimation is discussed.
On a simplistic perspective, one can initially view the problem of HRSHM as one of
non-stationary time series estimation and forecast. A critical challenge in building
a black-box representation for a non-stationary system is in the time-varying nature
of dynamic characteristics, whereas representative features are expected to evolve
with time. Some machine learning techniques lend themselves to characterizing time
series signals, in particular recurrent neural networks (RNNs) that can learn tempo-
ral dependencies, and their infamous long short-term memory (LSTM) architecture
capable of preserving long-term knowledge. However, machine learning techniques
typically heavily rely on pre-training, and curated training data related to high-rate
system are highly limited due to the important costs involved in experimentation and
highly uncertain dynamic environment. A successful machine learning algorithm for
HRSHM is one that can be trained on these very limited datasets, and with strong
Real-Time Machine Learning for High-Rate Structural … 79
Fig. 2 Pure on-the-edge machine learning algorithm (“smooth trans VIO”) versus hybrid algorithm
(“hybrid”): a overall time series; and b zoom on the first section of the time series
where y(t) is a measurement taken at time t, τ the time delay, and d the embedding
dimension. In other words, there exists a delayed vector that preserves the essential
dynamics of the system. While originally developed for autonomous systems, the
theorem has been extended to non-autonomous systems with deterministic forcing
[20], state-dependent forcing [5], and stochastic forcing [21].
The embedding theorem does not apply to non-stationary systems. It was demon-
strated in [10] that, while using ν constructed based on the embedding theorem to
estimate a non-stationary system did yield adequate performance, the method showed
greatly unstable with respect to the choice of τ and d, where a mere over-embedding
by one dimension could result in close to 100% underperformance in accuracy and
convergence. Yet, the authors in [17] proposed one such application for neurocontrol.
The idea was to use a sliding window through the time series measurement to extract
the delayed feature vectors (i.e., τ and d), assuming that each time series window
was stationary, and use these features as the input to the representation. Thus, for
each new window, new features were extracted and the input space adapted accord-
ingly. This algorithm was termed “self-organizing input space.” Results yielded a
leaner neuro-representation, computationally faster, and of better convergence com-
pared with the same representation built with the best non-variable input space (i.e.,
ν constant) found through brute force.
The method was used by the authors in [12] to evaluate its promise for HRSHM
applications. This was done by constructing an observer equipped with such self-
organizing input space, termed “variable input observer” or VIO. The architecture of
the VIO is illustrated in Fig. 3, where the τ and d parameters are sequentially selected
by analyzing the dynamics of the local time series and are used to construct ν that
serves as the input delay vector to the wavelet network. Here, the wavelet network
has self-adaptive capabilities (i.e., nodal weights, bandwidths, and centers), and the
adaptation is conducted using the estimation error as the cost function.
Numerical simulations using the VIO were conducted on the droptower dataset
(Fig. 1), and results also compared against those obtained from a non-variable input
space found through brute force. Results showed that the use of the VIO yielded a
neuro-representation that was approximately 50% leaner (i.e., fewer nodes), showed
a reduction of approximately 50% in the root mean square error (RMSE), and an
approximate 20% improvement in convergence speed. Conclusions of the study
pointed toward important opportunities in the input space, similar to conclusions
drew in [10] that also demonstrated that the selection of the wrong input space for
a high-rate system could have strong negative consequence on the quality of the
representation. Nevertheless, the VIO had two important limitations impeding its
application to HRSHM. Firstly, the architecture of the representation was incapable
of convergence given the constantly changing input space. Secondly, the computing
time used in finding the optimal input space features was quite significant. Based on
these findings, the authors developed a deep learning strategy specifically crafted for
HRSHM applications, presented in what follows.
Opportunities and limitations identified in the research on the VIO led to a deep
learning algorithm to build a stable predictor for non-stationary time series [1]. The
strategy consists of utilizing n RNNs organized in parallel, termed ensemble of
RNNs, each ith RNN associated with an input space νi constructed with a unique
combination of τ and d (i.e., τi and di ). The architecture of the proposed ensemble
of RNNs is illustrated in Fig. 4. The selection of the n delay vectors will be discussed
later in this section. The outputs of each RNNs are assembled through an attention
layer that assigns weights to each output and passed through an linear neural to yield
the prediction. The combination of the attention layer and linear neuron forms the
function that maps the feature vectors h to the prediction. The RNNs are constructed
with short-sequence LSTM cells to better cope with non-stationarities and empower
faster computing. Here, the pure on-the-edge capabilities constraint assigned to the
VIO is relaxed. Yet, it is assumed that only limited training data is available, and
that the available training data does not represent most of the dynamics that could
be experienced by the system. Thus, transfer learning is utilized to train the LSTM
cells in real-time, using the one-step ahead prediction error for the cost function.
Using a discrete time k notation for simplicity, the ith delay vector at time k is
written
where τ is a positive integer. The role of each LSTM cell is to recursively update a
hidden state vector h. Here, these hidden states represent features extracted from the
time series, and these extracted features are mapped to the prediction. This process
is illustrated in Fig. 5 for the ith RNN that maps the ith delay vector νki to the
step ahead prediction ŷ i + k + 1 as a function f of the last extracted feature vector
hid . The role of the attention layer is to combine all the individual predictions to
compute the ensemble prediction ŷk+1 . Remark that, as implicitly shown in Eq. 2, the
algorithm predicts the k + 1 output ŷk+1 , consisting of the one-step ahead prediction.
For multiple steps ahead, the one-step ahead prediction is iterated using predicted
values as synthetic measurements. Formal equations characterizing LSTM networks
can be found in [9]. The physics-informed input space ν i is the corner stone of the
algorithm, enabling more targeted feature extraction and thus improved prediction
accuracy and horizon for non-stationary time series. Its construction is discussed in
what follows.
The construction of the input space is based on physical knowledge and is conducted
in two steps. The first step is the extraction of data structure from the training time
series through principal component analysis (PCA). The strategy is to decompose
the signal into principal components (PCs), and to keep the first PCs that represent
most of the signal. For a time series dataset, PCA is typically conducted by a singular
value decomposition of the signal’s autocorrelation matrix. See [2] for a mathematical
description of the process. Here, the first n PCs that account for 95% of the signal are
taken, and n different input space are constructed, given raise to n RNNs running in
parallel. As an example, consider test data generated from the droptower shown in
Fig. 1a taken from accelerometer 1 (test 1—TS1 ). Figure 6 shows 95% of the signal
reconstructed with only the first five PCs that are individually plotted in Fig. 7.
Real-Time Machine Learning for High-Rate Structural … 83
Fig. 6 Reconstructed time series TS1 using the first five PCs
Once the data structure is extracted through PCA, the topology of the ith PC (Fig. 7
for instance) is investigated to select parameters τi and di that will be used to construct
ν i . This is done through well-established numerical techniques, in particular the
mutual information (MI) test [3] to select τ , and the false nearest neighbors (FNN)
test [12] to select d. The MI tests is based on information theory and is a measure of
the nonlinear dependence of measurements for different sampling periods and selects
a delay value that adds most information to the sequence. The FNN test searches
for the embedding dimension d that minimizes the number of false neighbors when
altering the dimension.
Once constructed, the delay vectors are used to train the RNNs. The RNNs are
trained individually, on the full training dataset, using the associated delay vector
(Fig. 5). After, the trained LSTM cell is extracted and becomes the feature extractor
84 S. Laflamme et al.
that will map νki to hi (Fig. 4). During the real-time prediction phase, the neuro-
paremeters are sequentially updated based on a cost function based on the one-step
ahead prediction error.
Table 1 Input space hyper-parameters selected using the “GS inputs” and “PCA inputs” techniques
LSTM GS inputs PCA inputs
τ d τ d
1 14 8 25 5
2 11 10 15 6
3 8 12 11 8
4 5 14 7 12
5 4 15 5 15
Real-Time Machine Learning for High-Rate Structural … 85
horizon, where the PCA strategy shows to outperform the LOI up to 34 steps ahead
with respect to the RMSE, and 29 steps head with respect to the MAE.
Figure 9 plots the naive prediction length and DTW metrics as a function of
the prediction horizon q. A study of the naive prediction length shows that the
prediction by the PCA strategy is more stable and richer in information compared
with the GS strategy. This can also be observed through the DTW plot where the
PCA strategy significantly outperforms the GS strategy over mid- and long-term
prediction horizons, demonstrating that the proposed physics-informed technique
generates features (h) that are well representative of the non-stationary time series,
thus explaining the better quality of predictions over long prediction horizons.
Overall, through this numerical demonstration, we demonstrated that a deep learn-
ing tool with a physics-informed input space could be used to quickly learn and
predict a non-stationary time series. Hand-coded in Python and without leveraging
a parallel processing architecture for the RNNs, the average computing time of the
algorithm was 25 µs per time step. With the 1 MHz sampling rate used in acquiring
data on the droptower test, this is equivalent to 25 steps (q = 25). This would govern
what should be the minimum data sampling rate, unless the algorithm is coded in
a way that step ahead predictions could be processed in parallel. Remark that it is
anticipated that hardware implementation of the algorithm along with parallel pro-
cessing of the RNNs themselves will significantly increase computing speed, thus
improving performance with respect to the minimal sampling rate and/or required
prediction horizon for HRSHM applications.
follows that a complete HRSHM algorithm is one that yields information that can be
used within a feedback system to make rapid decisions. A simple application of time
series prediction to decision making is the detection of prediction errors to detect
changes in dynamics, although this method is limited to a binary output “change”
versus “no change” that is difficult to integrate in a decision system. Often, state
estimation rather than measurement prediction is preferred, whereas physical prop-
erties such as stiffness and damping are estimated and can be used as a direct way
to detect, localize, and quantify changes in dynamics (e.g., damage). For example,
work presented in [6] proposed an algorithm for online estimation of bridge stiffness
over a series of discrete elements in order to localize and quantify damage, where
the estimated stiffness values can be used in a decision loop to detect the presence
Real-Time Machine Learning for High-Rate Structural … 87
of important damage post-seismic events and decide upon urgent inspections and/or
bridge closure.
State estimation inherently relies on physical representations, which complicates
the task in the high-rate dynamics realm given the lack of physical knowledge com-
bined with the complexity of the dynamics. As discussed in the introduction, attempts
have been made in HRSHM using MRAS approaches, in particular in [25] where
the authors showed that the position of a moving cart could be identified with an
average computing time of 93 µs. The algorithm is schematized in Fig. 10. It con-
sists of an adaptive model running in parallel with a reference model (i.e., “real”
system). The adaptive model is used to generate synthetic measurements, and the
errors on measurements are used to update the uncertain parameters of the adaptive
model. A critical contribution of the work is that it integrates the notion of concurrent
learning to cope with the lack of persistent excitation, as it is often the case for high-
rate events consisting of impulse loads. Concurrent learning consists of concurrently
using real-time measurements augmented with a temporally limited historical data
set to guarantee exponential convergence.
In this section, we summarize the demonstration of the algorithm on experimen-
tal data acquired on the DROPBEAR testbed found in [25]. After introducing the
experimental methodology that includes the construction the MRAS algorithm, we
present and discuss key results that show rapid state estimation capabilities and con-
clude by discussing how this method could be integrated with our proposed deep
learning approach.
4.1 Methodology
The DROPBEAR testbest, shown in Fig. 11, was designed to validate online param-
eter estimation algorithms for high-rate dynamic systems. It is discussed in detail in
[14]. Briefly, it consists of a 505 mm cantilever steel beam equipped with a sliding
cart to produce a gradual or sudden change in stiffness, along with a mass attached
88 S. Laflamme et al.
Fig. 12 a Frequency response functions (FRFs) for the beam with mass at various fixed cart
positions; and b wavelet transfer for the beam without mass with the moving cart
Fig. 13 Estimation versus reference with and without CL: a displacement z 1 ; and b displacement
error z̃ 1
one sensor is utilized in the experimental setup, the reduced order representation only
has one degree-of-freedom. After, the adaptive dynamic parameters of the state-space
representation, which consist of the equivalent mass, stiffness, and damping, are
updated sequentially during the acquisition of measurements. This is done through
sliding mode theory. As discussed above, a key issue in using impact loads is the lack
of persistent excitation, where the adaptive dynamic parameters are not expected to
converge. Concurrent learning is used as a solution, where only a short sequence
representative of the input, termed history stack, is used in the adaptive mechanism.
At each time step, the algorithm investigates whether the new input adds information
to the history stack, and if it does, the new input replaces the less informative value
in the sequence. Figure 13 demonstrates the use of concurrent learning on a toy
single degree-of-freedom example, plotting the estimation of displacement z 1 and
its estimation error z̃ 1 over time. Results show that without concurrent learning, the
estimation does not converge to the reference value.
Results from the fixed cart experiment are first described, where five tests were
conducted under each cart location. Under these tests, the mass was taken as known,
and both the stiffness and damping properties estimated, with both values taken
initially as null. Figure 14a, b plot typical results for the estimation of stiffness and
damping, respectively, under each cart location, with the blue vertical line showing
the convergence point. Results show that both the stiffness and damping converge
with the stiffness converging to its real value and damping to an adequate value
(the real value of damping was unknown). Average results are tabulated in Fig. 14c,
where the estimated frequency computed from the estimated stiffness shows a great
match. It can also be observed that the convergence time decreases significantly with
the increasing frequency of the system. Also, while the convergence time appears
90 S. Laflamme et al.
Fig. 14 Key results from fixed cart experiments: a typical results for stiffness estimation; b typical
results for damping estimation; and c average frequency estimate per series of tests and convergence
time
Fig. 15 a Displacement z 1 estimation versus reference during cart movement; b error on displace-
ment estimation z̃ 1 ; c impulse load inputs; and d cart position estimation (red-dashed line) versus
real (black line)
4.3 Discussion
State estimation (i.e., cart position) of the moving cart was conducted with an average
computing time per step of 95 µs using MATLAB. This compares well with results
obtained from Downey et al. [8], where the authors reported an average computing
time per step of 4.04 ms using a Fourier transform to identify the fundamental
frequency and conduct model matching. Analogous to the deep learning algorithm,
we anticipate that hardware implementation would significantly decrease computing
time and thus increase HRSHM applicability.
While both this work and the one on the ensemble of RNNs discussed in the
previous section are yet to be combined, we view the use of the MRAS-based tech-
nique as a path to providing rapid state estimation for HRSHM. An obvious method
to combine both methods would be to produce time series predictions and train
the MRAS algorithm on the predictions, thus enabling predictive capabilities on
the cart position. Alternatively, the MRAS algorithm could be integrated in parallel
with the ensemble of RNNs to yield a physics-informed machine learning approach,
where the deep learning algorithm would learn and leverage a physical representation
to improve predictive performance. As compared with purely data-driven machine
learning models, the resulting physics-informed machine learning models would
likely generalize better to scenarios unseen during model training, especially those
that are not well represented by the training data.
Obviously, there exists a vast number of different algorithms and hybrid permuta-
tions that could be used to enable rapid state estimation for HRSHM. Regardless of
the selected algorithms, computing speed is a critical metric enabling real-time appli-
cations, and convergence speed will define the accuracy of the algorithm. As briefly
92 S. Laflamme et al.
Table 2 Comparison of convergence time after second, third, and fourth impulse under J = 10,
30, and 50, along with average computation time per sample
J Convergence time (ms) Computation
Second impulse Third impulse Fourth impulse time/sample (µs)
10 161 432 327 64
30 144 289 264 95
50 131 249 136 112
discussed in the introduction, computing time and convergence speed will often be
conflicting goals. As an example, Table 2 lists the performance of the MRAS algo-
rithm on the moving cart experiments as a function of the length of data kept in the
history stack J¯. While the convergence time after each impulse decreases with J¯
increasing, the average computing time per sample increases, showing the important
trade-off between computing time and accuracy. It follows that the minimum com-
puting time in an HRSHM application will be governed by the required accuracy,
which itself will depend on the required performance of the feedback system through
the quality of closed-loop decisions.
5 Conclusion
In this chapter, we have introduced the problem of high-rate structural health mon-
itoring (HRSHM) and discussed the use of real-time machine learning to empower
rapid state estimation, in the sub-millisecond realm. We have started by defining the
HRSHM challenge, where the dynamics of interest are importantly characterized by
(1) large uncertainties in the external loads; (2) high non-stationarities and heavy
disturbance; and (3) unmodeled dynamics generated from changes in system config-
urations, and argued that a successful HRSHM algorithm is one capable of leading to
closed-loop decisions under 1 ms. Given the complexity of the dynamics, we found
that adaptive mechanisms are well suited for the problem, as long as they are capable
of fast convergence and short computation time. This can be done through the proper
design of machine learning algorithms.
Among solutions, we reviewed opportunities in crafting the input space to yield
lean and effective representations. In particular, it was discussed that being capable
of adapting the inputs to the local dynamics could dramatically improve performance
of the algorithm. We presented a machine learning algorithm based on an ensemble
of recurrent neural networks (RNNs) that leveraged the idea of adaptive input spaces
and demonstrated that a physics-informed selection of inputs yielded significant
improvement in predictive performance. Importantly, the algorithm was capable of
time series predictions in with an average computation time of 25 µs.
Real-Time Machine Learning for High-Rate Structural … 93
References
15. Kantz H, Schreiber T (2004) Nonlinear time series analysis. Cambridge University Press
16. Kettle RA, Dick AJ, Dodson JC, Foley JR, Anton SR (2016) Real-time state detection in
highly dynamic systems. In: Rotating machinery, hybrid test methods, vibro-acoustics & laser
vibrometry, vol 8. Springer, Berlin, pp 27–34
17. Laflamme S, Slotine JE, Connor J (2012) Self-organizing input space for control of structures.
Smart Mater Struct 21(11):115015
18. Lee SJ, Jang MS, Kim YG, Park GT (2011) Stereovision-based real-time occupant classification
system for advanced airbag systems. Int J Automot Technol 12(3):425–432
19. Qu ZQ (2013) Model order reduction techniques with applications in finite element analysis.
Springer Science & Business Media
20. Stark J (1999) Delay embeddings for forced systems. I. Deterministic forcing. J Nonlinear Sci
9(3):255–332
21. Stark J, Broomhead D, Davies M, Huke J (2003) Delay embeddings for forced systems. II.
Stochastic forcing. J Nonlinear Sci 13(6):519–577
22. Takens F (1980) Detecting strange attractors in turbulence. In: Dynamical systems and turbu-
lence, Warwick, pp 366–381
23. Wadley HN, Dharmasena KP, He M, McMeeking RM, Evans AG, Bui-Thanh T, Radovitzky R
(2010) An active concept for limiting injuries caused by air blasts. Int J Impact Eng 37(3):317–
323
24. Xu Z, Rahman F, Xu D (2007) Comparative study of an adaptive sliding observer and an
ekf for speed sensor-less dtc ipm synchronous motor drives. In: Power electronics specialists
conference, IEEE, pp 2586–2592
25. Yan J, Laflamme S, Hong J, Dodson J (2021) Online parameter estimation under non-persistent
excitations for high-rate dynamic systems. MSSP (under review)
26. Zhang Y, Zhao Z, Lu T, Yuan L, Xu W, Zhu J (2009) A comparative study of luenberger
observer, sliding mode observer and extended kalman filter for sensorless vector control of
induction motor drives. In: Energy conversion congress and exposition, IEEE, pp 2466–2473
Development and Validation
of a Data-Based SHM Method
for Railway Bridges
1 Introduction
The concerns about maintenance and monitoring of structures have become a big
challenge for engineers, researchers, and the civil engineering community. There
has never been a more appropriate time to develop robust and reliable structural
damage detection systems as aging civil engineering structures, specifically bridges,
continue to be used past their life expectancy and well beyond their original design
load capacity. The increasing monetary pressure on bridge authorities to extend
the life span of the existing structures as far as possible is mainly driven by the
considerable expenses associated with building a completely new bridge. In addition
to this are also the costs associated with the physical removal of the old bridge and
consequent costs from withdrawing the bridge from service, which affects passengers
and freight transportation in several different ways. The reasonable thing to do is to
keep the present structures in operation while ensuring public and structural safety at
minimum cost. Concurrently, this requires managers and decision makers to prioritize
rehabilitation and replacement programs. The problem is that when a significant
damage to the structure is discovered, the deterioration has often already progressed
far, and required repair is substantial and costly. One way to tackle this issue is
by defining clever maintenance strategies that make use of structural monitoring in
real time, thus enabling the detection of damage in its earliest stage and providing
accurate remaining life predictions.
Structural health monitoring (SHM) aims at providing support for these strategies,
through the collection of reliable data on the real condition of a bridge, the observation
of its evolution over time, and characterization of the degradation. By permanently
installing a number and variety of sensors, which continuously measure structural
and environmental parameters, it is possible to obtain a real-time representation of
the structure’s current state. However, this information is only useful to the decision-
making process if it is not misleading. This can be avoided by assuring that reliable
SHM systems, methods for data analysis, and statistics tools are put into place.
element (FE) model of the target structure. These models are often of gradually
increasing complexity during the development stage, so as to make sure that the
measured responses can be reproduced as accurately as possible by the FE model.
The well-known vibration-based damage identification (VBDI) methods consist of
measuring and evaluating the dynamic behavior of the structure often by comparing
it to the behavior simulated by numerical models, for instance FE models. Natural
frequencies are the most fundamental vibration parameter, and methods directly
measuring shifts in natural frequency can be used for identifying damage. Since
many algorithms of damage detection are based on the difference between the models
before and after occurrence of damage (Axiom II of SHM [1]), problems such as
parameter identification and damage detection are closely related to model updating.
Over the recent years, a second approach has emerged sparked by the massive amount
of acquired experimental SHM data and breakthroughs in technology, computing
capability, and storage solutions. The data-based approach is free of geometrical
and material information, allowing to circumvent the burden of having to develop
a detailed FE model of the target structure. With this approach, time series anal-
ysis, often along with signal processing techniques, is employed to extract damage-
sensitive features from measured signals. In these regards, choosing an appropriate
damage feature is crucial for the success of damage detection as these features are
used to establish baseline statistics and to monitor changes in the normal structural
behavior. This procedure makes the data-based approach well suited for permanent
automated long-term monitoring.
Machine learning (ML) is an application of AI that provides systems the ability
to automatically learn and improve from experience without explicitly programming
them to do so. The idea behind using the data-based approach for SHM, and more
specifically for damage detection, is to use the data sets of signals obtained from a
structure over time and to use soft computing methods to warn about damage and
its characteristics. Pattern recognition is a particularly useful branch, wherein labels,
such as “healthy” or “damaged,” are assigned to a given input value with the help of
an algorithm. The type of learning performed by the computer usually falls into two
major categories—supervised and unsupervised learning—depending on what type
of data is used. In the context of damage detection, supervised learning implies that
both data from the undamaged (reference) and damaged (novel) conditions of the
structure are provided to the algorithm. Since prior knowledge about possible damage
scenarios is available, classification and regression analysis can be carried out. This
approach thus allows to characterize damage up to the level 3 [2] of damage iden-
tification (i.e., quantification of damage). On the other hand, unsupervised learning
implies that only data from the undamaged (reference) condition of the structure is
provided to the algorithm. This means that the data is not labeled, and therefore, it is
only possible to group and interpret data based on input data. It is at most possible
to reach level 2 [2] of damage identification (i.e., localization of damage) with this
approach, the main goal being to carry out novelty detection and eventually also
clustering and dimensionality reduction. In this manner, a reference condition of the
structure is first established by the algorithm based on appropriate features extracted
from data, which can be either measured or obtained from a numerical model of the
98 A. C. Neves et al.
Fig. 1 Seven-step closed loop cycle of the proposed data-based approach to bridge SHM
ANNs are among the most common ML techniques used for novelty and damage
detection, as well as to solve classification and regression problems. The ANN has
the capability of modeling and predicting [4] the behavior of intricate linear and
Development and Validation of a Data-Based SHM Method … 99
nonlinear systems exclusively based on data. For this reason, even the most basic
feedforward ANN has shown to be a suitable method for predicting the behavior of
civil engineering structures, as demonstrated by plenty of research works [3, 5–10].
When creating the ANN model, the total amount of train data is divided into three
data subsets (Fig. 2), each to be used at different stages of the training process. The
first data subset corresponds to the training subset, in which a number of examples
are used to fit the parameters (primarily weights of connections between neurons)
of the ANN model. The second data subset corresponds to the validation subset,
which provides an unbiased evaluation of the model fit to the training subset while
tuning the hyperparameters of the model. The last data subset corresponds to the
testing subset, which provides an unbiased evaluation of the final model fit to the
training dataset. In the context of damage detection, a portion of the recorded healthy
acceleration data sets during train passages are used as train data for the ANN. Since
the train data stems from one single state of the structure, learning can be said to be
carried out in an unsupervised manner. After the ANN is trained, new acceleration
data relating to healthy and damaged states can be used for testing the performance
of the network.
A portion of the available healthy acceleration data sets recorded during train
passages will be used as the damage-sensitive feature to train an ANN. For simpli-
fication, only part of the recorded acceleration time history of a train passage with
duration T is plotted in Figs. 3 and 4.
During the training process, inputs (i) and matching targets (t) are assigned within
each sample (train passage) of the training subset. After the training process is
completed, the ANN will be asked to make predictions in the form of outputs (o). The
inputs, targets, and outputs correspond to the discrete acceleration values registered
by one sensor at any given time instant n. Training can be conducted, for example,
by giving four accelerations {i n−4 , i n−3 , i n−2 , i n−1 } ( ) as input accelerations to
Fig. 2 Flowchart representing the data partition for the ANN model designed for damage detection
100 A. C. Neves et al.
where Nw are the very first four accelerations (in this example) for which the predic-
tions cannot be calculated, and hence, the average is made for N − Nw . According
to a detection threshold that is defined for the system, the damage detection process
consists in verifying Eq. (2),
The information coming from each sensor individually may be more or less reli-
able depending on the sensor characteristics and their location on the structure, but
when the information coming from all sensors is combined, the reliability of the
detection method is expected to increase. A damage index (DI) provides a tool to
combine the information obtained from multiple sensors into one single number,
which works as a detection threshold set for the system as a whole. One possible
way to carry out this combination is through the Mahalanobis distance, a covariance-
weighted distance metric shown to successfully allow for condition discrimination in
several damage detection studies [11]. The distance between a point P and a distribu-
tion D is measured in standard deviations. Each newly obtained DI point is compared
to the cloud of DI points that is considered as the reference. For each train passage
p, the damage index DI p is formulated according to Eq. (3) as
T
DI p = RMSEprediction − μ .C −1 . RMSEprediction − μ , (3)
where RMSEprediction (obtained according to Eq. (1)) is the prediction error in any
state, μ is the mean prediction error of the distribution within the reference state, and
C is the covariance matrix of the distribution within the reference state.
Naturally, since reality is much more complex, this binary discrimination reliant on
a hard threshold can lead to instances of false diagnostics (Table 1). In a way, damage
detection can be said to be based on hypothesis testing where the null hypothesis
states that the structure is healthy. Incorrect diagnosis comprises False Positives (FPs)
and False Negatives (FNs), where a healthy structure is perceived to be damaged and
where existing damage in a structure goes undetected, respectively. On the other
hand, correct diagnosis comprises True Positives (TPs) and True Negatives (TNs),
where a damaged structure and a healthy structure are properly assessed, respectively.
FPs and FNs are also commonly recognized as Type I and Type II errors and are to be
avoided: FP could be linked to extra inspections or unnecessary repairs. FN could be
linked to particularly undesired consequences due to the continuous accumulation of
undetected damage and ensuing failure of the structure, with life-threatening safety
implications and casualties to society. Figure 5 shows the information contained in
Table 1 in a graphical form. Given a chosen detection threshold ( ) for the
system, the obtained data sample errors from the ANN in light of healthy ( ) and
damaged ( ) data according to Eq. (1) can be classified into FP, TP, TN, and FN.
The receiver operating characteristic (ROC) curve is an easily interpretable two-
dimensional graphic that illustrates the performance of a binary classifier system as a
function of an adjustable threshold. The area under the curve (AUC), which measures
the entire two-dimensional area underneath the entire ROC curve, is routinely used
to compare the classifiers’ accuracy and hence its usefulness. For instance, in Fig. 6,
system is the ideal system with AUC = 1.0, whereas system is the worst with
AUC = 0.5. The green hatched area ( ) represents the AUC for an average ROC
curve that is more likely to occur for real systems . In a ROC curve, the TP rate is
plotted against the FP rate for all possible values of detection threshold, the points
along which the curve is drawn. It is understandable that a very high threshold (strict
threshold) will result in 0% of FPs and TPs. On the contrary, a very low threshold
(lenient threshold) will result in 100% of FPs and TPs. It is important to define
the detection threshold that finds a balance between the two types of error, as it is
impossible to minimize them simultaneously. It is important to stress that the ROC
curves cannot generally be obtained in practice. Firstly because it requires the data
from damaged condition to also be available. Second because the true condition of the
structure at any moment has to be known, which is precisely the point of employing
the monitoring system. Nonetheless, ROC curves can become very useful in the
developmental stage of methods for damage detection as their performances can be
compared on the same grounds.
• 49 input neurons: The number of neurons equals the number of features, which
n
are the 30 accelerations at−i registered by the 6 sensors n ∈ {1, . . . , 6} in the last
5 samplings i ∈ {1, . . . , 5}, the 18 axle loads and 1 axle position relative to a
reference point;
• 30 neurons for the hidden layer based on empirical rules-of-thumb [14];
• 1 output neuron: the current acceleration atn at time t predicted by sensor n.
All the input parameters were assembled into one single-input matrix. The only
input variable that remained present in all the trials of ANN configurations was the
acceleration. Any other deemed relevant parameter can also be given as input. What
the network does is to predict a new acceleration at a certain instant in time based on
previous accelerations ( ) as shown in Figs. 3 and 4. The number of delays has to
be chosen before the training phase. Perceptibly, it cannot be a small number as that
will give little information for the training of the network, and it cannot be a large
number since that leads to the increase in computation time. In this case study, the
five previous accelerations {i n−5 , i n−4 , i n−3 , i n−2 , i n−1 } preceding the acceleration
to be predicted on were chosen as the time window for the input matrix.
One way of evaluating the performance of the trained network is by determining
the deviation in the predictions. The use of the root mean squared error (RMSE) is
very common and makes an outstanding general purpose error metric for numerical
predictions. For each sensor and for each train passage (or each speed), in a similar
way for both healthy and damaged scenarios, one can estimate the R M S E pr ediction
as expressed in Eq. (1). Figure 7 illustrates the RMSE of the predicted accelerations
by the six sensors, in the presence of an undamaged structure ( ) and for a damaged
Fig. 7 RMSE against increasing speed of the train. DC1: damage extension of 0.9 × 0.4m2 . Data
from healthy structural condition; Data from damaged structural condition
106 A. C. Neves et al.
Fig. 8 Gaussian Process fitted by prediction errors against increasing train speed. Here, a log-
normal distribution of the error is considered. Mean; Standard deviation; Data to fit the GP;
Data from healthy condition; Data from damaged condition, considering DC1 with a 0.9×0.4m2
section
yielding slightly different results every time a ROC is regenerated. In any case, the
desired virtues of a good model such as reliability and robustness are proven by the
above-presented ROC curves. The model is considered reliable if damage is early
detected with a high probability of detection, i.e., if small damage can be identified.
The model is considered robust if two conditions are satisfied: the probability of
detection increases with increasing damage severity, and changing an input parameter
by a small amount does not lead to failure or unacceptable variation of the outcome
but rather to proportional small changes.
108 A. C. Neves et al.
The railway bridge KW51 (Fig. 10) is a steel bowstring bridge on railway line
L36N located in Leuven, Belgium. The 115 m long bridge enables the crossing
of the Leuven Mechelen canal and is only used by passenger trains with an imposed
maximum speed of 160 km/h. The railway line consists of two tracks, hereafter
referred to as track A at the north-side and track B at the south-side. At the crossing
of the bridge, both tracks are curved. The monitoring campaign started on October 2,
2018, and under the period May 15 to September 14, 2019, the bridge was retrofitted
(Fig. 11) to solve a construction error noticed during inspection. The retrofitting
consisted of strengthening the connections between the diagonals and the arches and
the bridge deck.
The sampling frequency is 1651 Hz. Regarding the data processing, the strain
signals are low-pass filtered with an eighth-order Chebyshev Type I filter with 0.1 dB
of passband ripple and a cut-off frequency of 16 Hz. The acceleration signals are high-
pass filtered with a fourth-order Butterworth filter with cut-off frequency of 0.5 Hz.
The measurements include acceleration and strain measurements on the bridge, strain
measurements on the rails, and temperature and relative humidity measurements.
Detailed information about the measurement setup can be found in [16]. Although
not all sensors are used for this study, they are here mentioned for completeness.
The optimal selection of the ANN training and architectural parameters can be an
exhaustive process. Judging which input (damage sensitive) features are relevant is
an open problem as well, and finding the optimal ANN model turns out to be a trial–
error process. Nonetheless, some empirical rules exist, and these can be followed
as a starting point for suggesting reasonable models and ad-hoc parameters. Table
2 [17] lists relevant ANN key parameters to be defined, such as input features, the
number of neurons in the hidden layer, the training algorithm, et cetera.
Several ANN configurations were tested, and each was obtained from different
combinations of the parameters as presented in Table 2. The ideal hyperparam-
eter search procedure would consist in searching all possible combinations of influ-
encing features. Realistically, this would be impractical due to major computational
resources and the curse of dimensionality [18]. This being said that the objec-
tive of this work is to provide some sort of check list for the implementation of
the proposed method for SHM of railway bridges. As such, a limited number of
different manually selected configurations are studied where, for each, the effects
of making changes within one hyperparameter/feature while keeping the remaining
constant are analyzed. Two aspects are kept invariable throughout all configurations:
only one hidden layer is used and the training function is the Levenberg–Marquardt
backpropagation algorithm.
The idea is to start simple by defining an ANN that has, for example, 10 neurons
in the hidden layer and that uses the 20 previous raw accelerations from sensor A
as input (e.g., a1A to a20A in m/s2 ) to predict the current acceleration (e.g., a21A in
m/s2 ). In other words, the dataset is broken into windows of size 20. Progressively,
more complex ANNs (Fig. 12) are then defined and tested. For example, increasing
the amount of training data so that the ANN uses the 30 previous (e.g., a1A to a30A
110 A. C. Neves et al.
Fig. 12 A more complex input matrix and output vector for ANN
in m/s2 ) but also the 30 posterior (e.g., a32A to a61A in m/s2 ) raw accelerations from
sensor A as input to predict the current acceleration (a31A in m/s2 ). Additionally,
the ANN can also consider strains from sensor B (e.g., s1B to s60B ), temperature
from sensor C (t31C in °C), and average train speed (v in m/s) as inputs. It is to be
noted that the matrix and vector presented in Fig. 12 are related with one input data
corresponding to one train passage, but training of the ANN is carried out by looking
at many train passages.
In the following figures in this and in next section, the represented prediction errors
for each train passage p are normalized relative to the RMS of the target response
signal. The chronological dates of the train passages are presented in Table 3. The
labels “before-,” “during-,” and “after-” that appear in this and following sections
concern the structural states in reference to retrofitting. The data sets collected under
the different states are labeled according to Table 4. If optimization is the goal, the
authors recommend the use of validation techniques (e.g., K-fold) to evaluate which
data sets from the “before-retrofitting” period are better used as training and test data
sets; however, this process was left out of the scope of this case study.
In this case study, two criteria are used to evaluate the performance of the algo-
rithm: accuracy and sensitivity. The first concerns comparing the magnitude of the
prediction errors between train and test data in the reference state (before-retrofitting);
the second concerns comparing the magnitude of the prediction errors between test
data in the reference state (before-retrofitting) and test data in other states (during-
and after-retrofitting). The accuracy criterion is a reflection of how good of an approx-
imator the ANN. But in the end, it is all the more important that a novel state of the
structure (e.g., damage) is distinguishable from the reference one, which is expressed
by the sensitivity criterion. Even if the prediction errors in the reference state are high,
it is sufficient that the prediction errors in the novel state result even higher.
The following results focus on the prediction errors related to accelerometer 3.
Furthermore, the prediction errors represented in Figs. 13, 14, 15, and 16 concern the
average normalized prediction errors [19] for all the considered train passages within
each of the structural states (Table 3). The errors in the just mentioned figures are
represented in relation to the train samples (before-retrofitting) errors: for example,
a small ε means that the taken test sample yields prediction errors with the same
order of magnitude of the ones obtained for the train samples. The general thinking
adopted when visualizing these figures is that the smaller the error for before-test
and the larger the error for during-test and after-test the better the performance of
the system.
Even though the monitoring spans over periods of ambient, free, and forced vibra-
tions, only the forced vibration part of the signal in which the train is on the bridge
was used. Not only because the bridge vibration activated by the crossing vehicle can
produce higher response amplitudes that can ease the interpretation of results but also
because the use of higher amplitude loads may be required in order to detect some
types of damage (e.g., cracks in concrete). Moreover, the ANN is trained and tested
for a specific loading that is frequent for this bridge, a 52-axle train. The information
regarding the train configuration (e.g., number of axles) was derived from the strain
measurements on the rails. The axle loads for this type of train are accepted to vary
within a relatively limited range of values. Therefore, including them as an input for
the feature is thought to yield little to no improvement of the ANN performance, as
shown in [8]. Furthermore, it was assumed that there is only one train on the bridge
at a time. Giving data collected with deck accelerometers 1–6 as input to train the
ANN and then asking it to predict for accelerometer 3 seems to lead to similar results
if only the input of accelerometer 3 is given to train the ANN. This suggests that
only the data collected by accelerometer 3 seems to be necessary as input to train
the ANN to predict to said sensor. The same reasoning is followed for the remaining
sensors, and accordingly, six different ANNs, one for each sensor, were trained.
Besides acceleration data, which alone was used as the original input for the ANN,
strain, temperature, and average speed of the train were also added to see if these
Development and Validation of a Data-Based SHM Method … 113
improved the ANN. Nearly no improvements are perceived (Fig. 13) if strain data
is added. Strain gauges are not normally used in high frequency applications, and
it may be necessary to further down sample the original signal (1651 Hz) since the
strains only exceed the noise floor below 10 Hz. Adding temperature or train speed
data, however, seems to increase the sensitivity of the ANN. However, it may not
be worth the effort required to have the extra sensors in the bridge, especially given
that the ANN is performing well without that extra information. Therefore, the input
matrix used for training the ANNs can comprise solely accelerations.
Measuring, processing, and storing of massive amounts of data can provide valu-
able information in order to manage structures [20]. At the same time, the sampling
frequency with which the input data is given to the ANN (i.e., the amount of data)
affects the ANN’s performance. As such, different versions of ANNs were trained
by having the input data sampled with the original sampling frequency and also by
down sampling it by a factor of 1.5, 2, and 3. Resampling is performed by using
a polyphaser anti-aliasing filter. The advantage with down sampling is the reduced
amount of resultant input data and thus faster training process of the ANN. On the
other hand, by doing so, less information is fed into the ANN during the training
process. That seems to reflect in higher prediction errors within test data collected
under the reference state but also in the reduced ability of the ANN to distinguish
between the different states (Fig. 14). Therefore, input data collected at the original
frequency was preferred.
Different configurations of ANNs were tried where the minimum number of given
previous acceleration data points to predict the current one starts at 10. Increasing
the number of total time steps from 10 to 20 and from 20 to 40 seems to significantly
improve the results (Fig. 15). Increasing the number of time steps from 40 to 60
marginally improves the results and over that the improvement is negligible. There-
fore, to save in computational effort spent on training, 60 time steps are deemed
as the most suitable for this case study. The 60 chosen time steps correspond to
giving 30 previous and 30 posterior acceleration data points to calculate the current
acceleration. Providing not only earlier accelerations as input but also the following
accelerations to the time step to be predicted seems to yield both a more accurate
and discriminative ANN.
A compromise must be reached for the number of neurons in the hidden layers.
Two main problems can arise from having too few and too many neurons in the hidden
layer(s): underfitting and overfitting, respectively. The former occurs when there are
too few neurons in the hidden layers to adequately detect the signals in a complex data
set. The latter occurs when the ANN has so much information processing capacity
that the limited amount of information enclosed in the training set is not enough
to train all of the hidden neurons. In light of the present case study, increasing the
number of hidden neurons from 10 to 20 seems to significantly improve the results,
whereas increasing it from 20 to 30 or even 30 to 40 produces no changes in the
results (Fig. 16). Therefore, 20 hidden neurons are reasoned as adequate for this case
study.
Figure 17 depicts the distribution of normalized prediction errors ε p (Eq. 4) asso-
ciated with accelerometer 3 against each train passage p ordered chronologically
114 A. C. Neves et al.
(Table 3). The data sets collected under the different states of the structure are labeled
according to Table 4. It is possible to see with the naked eye that most of the lowest
prediction errors correspond to the before-retrofitting period. That conclusion is
more obvious by looking at the mean of the prediction errors (μ) for each period (
) and how they compare to each other. The dashed lines ( )
represent the boundaries obtained by summing the mean with the two standard devi-
ations of the prediction errors for each data set, i.e., [μ − 2σ, μ + 2σ ], which delimit
the colored areas , , and . Additionally, it is possible to perceive the progres-
sive retrofitting taking place with the gradual increase of the prediction errors within
the duration of the retrofitting period ( ). That also explains the larger variance of
errors within the period. In the after-retrofitting period, any interventions come to
an end and the scaffolding and necessary equipment are removed from the bridge.
The new condition of the bridge is noticeably different from the reference one as the
ANN predicts consistently worse.
The ROC curves associated with a system made up of accelerometer 3 alone (
) are depicted in Fig. 18. It is obvious that such a simple damage detection
system robustly identifies the new states of the structure. The same outstanding
performance may not persist for other sensors, for example if accelerometer 6 alone
constitutes the damage detection system ( ). The ROC curves ( ) that are
obtained when the information from all the 6 accelerometers placed under the bridge
deck is combined, according to Eq. (3).
6 Conclusions
Acknowledgements The authors are grateful for the data provided by Dr. Kristof Maes from KU
Leuven.
References
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 117
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_6
118 A. Meixedo et al.
1 Introduction
Improved
Safety
Fc
Fb
Fa
Undamaged Damaged
DI
SHM
Increased Reduced
Longevity Maintenance
systems do not include the measurement of EOVs, latent variable methods can be
employed. These methods can suppress independent actions using only structural
measurements.
Data fusion focuses on reducing the volume of data while preserving its most
relevant information. The fusion process may combine features from a single sensor,
and features from spatially distributed sensors or even heterogeneous data types. The
Mahalanobis distance has been thoroughly used in this context due to its capacity to
describe the variability in multivariate datasets [11].
Feature classification aims at discriminating the features into healthy or damaged.
It can be divided into supervised or unsupervised learning algorithms [12]. When
training data is available from both undamaged and damaged structures, supervised
learning algorithms can be used, such as statistical process control [6] or MLP neural
networks [13]. Since data obtained from damaged structures is rare or inexistent,
unsupervised learning algorithms have been increasingly observed in the literature.
Novelty detection methods are the primary class of algorithms used in this situation.
This type of algorithm is a two-class problem that indicates if the acquired data comes
from normal operating conditions or not. Due to its simplicity and effectiveness,
outlier analysis is a broadly implemented damage detection technique [14]. In spite
of the SHM feature classification resorting to clustering methods has been reported
mainly following the supervised strategy of pre-defining cluster partitions to describe
one or more known structural behaviours, and subsequently compare them with
new ones, and this type of techniques have an unsupervised nature [15]. The major
advantage of cluster-based strategies, over those previously described, consists of the
greater sensitivity exhibited by these algorithms, which is related to their capacity to
analyses data compactness and separation instead of defining boundaries between or
around data objects. The works describing cluster-based classification for damage
detection refer its high sensitivity to structural changes and associate it with the
ability of these methods to analyses compactness and separation within feature sets.
While most of SHM works rely on responses derived from ambient vibrations or
static responses, recent works have also been using the structural responses generated
by traffic on bridges to take advantage of the repeatability of these actions, their
known behaviour, and their large magnitude, which imposes a greater excitation of
the bridge in a short time, when compared with ambient or static loads [16, 17].
However, robust and effective implementations of SHM in bridges based on traffic-
induced dynamic responses are still scarce. In most damage detection methodologies
that have been proposed, the EOVs in the structural response is often disregarded, the
type of damages is limited, or the loading scenarios are very specific, which limits
their usability in real and complex bridges.
In this context, the present research work aims at implementing and validating
a real-time unsupervised data-driven SHM strategy for early damage detection in
railway bridges using traffic-induced dynamic responses.
Real-Time Unsupervised Detection of Early Damage in Railway … 121
Fig. 2 Schematic representation of the SHM strategy for early damage detection
122 A. Meixedo et al.
A bowstring-arch railway bridge over the Sado River was selected as the case study
used throughout this research work. It is located on the southern line of the Portuguese
railway network that establishes the connection between Lisbon and Algarve (Fig. 3).
The bridge is prepared for conventional and tilting passenger trains with speeds up
to 250 km/h, as well as for freight trains with a maximum axle load of 25 t. Even
though the bridge accommodates two rail tracks, only the upstream track is currently
in operation. The bridge has a total length of 480 m and is divided into 3 continuous
spans of 160 m each. The bridge deck is suspended by three arches connected to
each span of the deck by 18 hangers distributed over a single plane on the axis of the
structure. The superstructure is composed of a steel–concrete composite deck, while
the substructure, which includes the piers, the abutments, and the pile foundations,
is built with reinforced concrete. The deck is fixed on pier P1, whereas on piers
P2, P3, and P4 only the transverse movements of the deck are restrained, while the
longitudinal movements are constrained by seismic dampers.
The structural health condition of the railway bridge over the Sado River has been
controlled with a comprehensive autonomous online monitoring system, as detailed
in Fig. 4a, since the beginning of its life cycle. This monitoring system was defined
based on an operational evaluation and allowed the acquisition of data necessary to
LISBON ALGARVE
160 160 160
SA 2m
SA 2sn SA 2ss
OT-P1 Ac1 S2s OT-P4
AL-P1 DLU-P1 AsV1 AsV2 AL-P2 DLU-P2 AsT1 AsV3 AsT2 AL-P3 DLU-P3 AsV4 AsV5 AL-P4 DLU-P4
DLD-P1 AsV6 AsV7 DLD-P2 AsV8 AsV9 AsV10 DLD-P3 AsV11 AsV12 DLD-P4
P1 P2 P3 P4
1/3 north (tn) a)
1/2 (m)
OT-P(e) OT-P(r)
Upstream Downstream
AL-P OT-P(r)
OT-P(e)
DLU-P DLD-P
AL-P AL-P
DLU-P b)
Upstream Downstream
Ac 1
Ac1
AsV3 AsV9
S2m
Upstream Downstream
As V3
As T1
AsT1 AsV8
S2tn c)
Ac: Vertical accelerometer in the concrete slab Al-P: Longitudinal accelerometer between bearings
AsV: Vertical accelerometer in the steel box DLU-P: Upstream longitudinal displacement transducers
AsT: Triaxial accelerometer in the steel box DLD-P: Downstream longitudinal displacement transducers
OT-P: Optical sensors Ts: NTC thermistors in the s teel box
SA: Arch section with NTC thermistors Tc: PT100 sensors in the concrete slab
S2s: Deck section with NTC thermistors Ta: arch NTC thermistors
Fig. 4 SHM system installed in the railway bridge over the Sado River: a overview, b longitudinal
accelerometer, displacement transducer, and optical sensor, c vertical and triaxial accelerometers,
and d NTC thermistors and PT100 sensors
124 A. Meixedo et al.
implement the strategy for early damage detection (Fig. 2). To identify each train that
crosses the bridge and compute its speed, two pairs of optical sensors are installed
at both ends of the bridge (Fig. 4b). The structural temperature action is measured
using PT100 thermometers and NTC thermistors. Three sections of the arch were
instrumented with twelve NTC thermistors. Additionally, four NTC thermistors are
fixed to the steel box girder and three PT100 thermometers are embedded in the
concrete slab (Fig. 4d). To control the behaviour of the bearing devices, the responses
from longitudinal displacement transducers are obtained from eight sensors, each
adjacent to a bearing device (Fig. 4b). The set of sensors also includes one vertical
piezoelectric accelerometer fixed at the mid-span of the concrete slab, two triaxial
force balance accelerometers at the thirds of the mid-span steel box girder, and twelve
vertical force balance accelerometers fixed along each span of the steel box girder
(Fig. 4c). Four longitudinal MEMS DC accelerometers are also installed at the top
of each pier (Fig. 4b). Data acquisition is carried out continuously, at a sampling rate
of 2000 Hz, by a locally deployed industrial computer to save the time history during
the passage of the trains.
A realistic simulation of healthy and damage scenarios was conducted to test and
validate the strategies proposed herein since it was not possible to simulate damage
scenarios experimentally. After a successful validation of the methodology, it can
be directly applied to experimental data from different types of bridges, where a
baseline scenario is defined, and further experimental data can be tested to detect the
occurrence of eventual structural changes.
For this purpose, a 3D finite element (FE) numerical model of the bridge is devel-
oped in ANSYS software [19] and fully validated with experimental data (Fig. 5).
Among the modelled structural elements, those defined as beam finite elements
consist of piers, sleepers, ballast-containing beams, rails, arches, hangers, transverse
stiffeners, diaphragms, and diagonals. Shell elements were used to model the concrete
slab and the steel box girder, while the pads, the ballast layer, and the foundations
were modelled using linear spring-dashpot assemblies. The mass of the non-structural
elements and the ballast layer was distributed along the concrete slab. Concentrated
mass elements were used to reproduce the mass of the arches’ diaphragms and the
mass of the sleepers, which were simply positioned at their extremities. The connec-
tion between the concrete slab and the upper flanges of the steel box girder, as well
as the connection between the deck and the track, were performed using rigid links.
Special attention was paid to the bearings supports, as they can strongly influence the
performance of the bridge. Hence, in order to simulate the sliding behaviour of the
bearings, nonlinear contact elements were applied. Moreover, constraint elements
located between the bearings were used to restrict the transversal movement in each
pier, and the longitudinal and transversal movements in the case of the first pier.
Real-Time Unsupervised Detection of Early Damage in Railway … 125
Z
X Y
a)
Numerical Frequency (Hz)
7 80
b2 6
ec 15
n2 5
ar 6
r 6
0 7
ov 6
1 2 3 4 5 6
ly 6
16
ay 6
g 6
pt 6
n2 6
ct 6
Fe 01
Ja 1
M 01
Ap 01
N 01
Ju 01
M 201
Au 0 1
Se 201
Ju 0 1
O 01
20
20
20
2
2
2
2
2
ov
Experimental Frequency (Hz)
N
D
b) c)
0.03
Vertical Acceleration (m/s2)
AL-P2 Experimental Ac1 Experimental
Long. Acceleration (m/s2)
0
0
-0.01
-0.02 -0.05
-0.03
0 2 4 6 8 10 0 2 4 6 8 10
Time (s) Time(s)
d) e)
Fig. 5 Numerical modelling and validation: a 3D FE numerical model of the bridge over the Sado
River, b agreement between numerical and experimental modal frequencies, c static validation of
the displacements measured on pier P4, d dynamic validation of longitudinal accelerations at pier
P2 with the AP at 216 km/h, and e dynamic validation of vertical accelerations at the concrete slab
(Ac1) with the AP at 216 km/h
To ensure that the numerical model accurately simulates the structural behaviour
of the bridge, the responses obtained from modal, static, and dynamic analyses were
compared with those measured by the SHM system [20]. The numerical natural
frequencies were compared with those obtained experimentally during an ambient
vibration test [21]. Figure 5b shows a very high coefficient of determination (R2 =
0.9993) between the numerical and the experimental results. To validate the static
behaviour of the numerical model, the response of the structure to the action of
temperature was studied. The structural static behaviour of the bridge was simu-
lated in the FE model by running a time-history analysis using experimental data
as input. The simulation procedure consisted of using the temperatures acquired
every hour on site over the course of one year. Figure 5c presents a very good
agreement between the numerical and experimental displacements of pier P4 for
the temperature measured on site between November 2015 and November 2016.
Regarding the dynamic behaviour, numerical simulations were conducted consid-
ering the Portuguese Alfa Pendular (AP) train as a set of moving loads crossing the
bridge over the Sado River at a speed of 216 km/h. Figure 5d, e shows a very good
126 A. Meixedo et al.
agreement between the experimental and numerical responses, in terms of the longi-
tudinal accelerations measured on pier P2, and the vertical accelerations acquired on
the concrete slab at the second mid-span (Ac1), respectively. Before the comparison,
the time-series were filtered based on a low-pass digital filter with a cut-off frequency
equal to 15 Hz. A detailed description of the numerical model and its validation com
be found in Meixedo et al. [20].
The dynamic numerical simulations implemented in the present research work
aimed at replicating the structural quantities measured in the exact locations of the
23 accelerometers installed on site (Fig. 4) during the passage of a train in the bridge.
To correctly reproduce these structural responses, the temperature action measured
precisely during the passage of each train was introduced as input in the numerical
model. The measurements of the optical sensors’ setup were used to obtain the train
speed and axle configuration, as well as the type of train [22]. The dynamic analyses
mentioned hereafter were carried out for two of the passenger trains that typically
cross the bridge over the Sado River, namely the AP train and the Intercity (IC) train.
Their frequent speeds on the bridge are 220 km/h for the AP train and 190 km/h for
the IC train. The nonlinear problem was solved based on the full Newton–Raphson
method, and the dynamic analyses were performed by the Newmark direct integration
method, using a methodology of moving loads [9]. The integration time step (t)
used in the analyses was 0.005 s.
Figure 6a summarizes the 100 simulations of the baseline (undamaged) condi-
tion that aim at reproducing the responses of the bridge taking into account the
variability of temperature, speed, loading schemes (LS), and type of train. These
baseline scenarios compose the training dictionary and do not include any damage
on any location. During each simulation, real temperatures measured by the SHM
system were introduced in the elements of the bridge. The average values for each
season were 21 °C for spring, 30 °C for summer, 16 °C for autumn, and 10 °C for
winter, but the dispersion across the structure was considered by measuring and using
temperature values in all elements of the bridge. The simulations included the AP
and IC trains crossing the bridge with ten different loading schemes, according to
the experimental observations previously made by Pimentel et al. [23]. Three train
speeds were considered for each type of train, as observed in Fig. 6, thus resulting in
100 time-history simulations for the baseline condition, each taking approximately
10 h on a 4.2 GHz Quad-Core desktop with 32.0 GB of RAM.
On the other hand, the damage scenarios were chosen based on possible vulner-
abilities identified for the type of structural system, taken into account its materials,
behaviour, loadings, and connections. As shown in Fig. 6b, damage scenarios were
simulated according to different groups: (i) damage in the bearing devices (type
D1), (ii) damage in the concrete slab (type D2), (iii) damage in the diaphragms
(type D3), and (iv) damage in the arches (type D4). Each scenario was simulated
considering only one damage location. Nevertheless, if, by any chance, two or more
damage scenarios in different locations are observed at the same time, the effects
from multiple damage locations are expected to superimpose, and the influence on
the features extracted from the data will be greater. Therefore, the multiple damage
scenario will be more observable than the scenarios tested here. Regarding the group
Real-Time Unsupervised Detection of Early Damage in Railway … 127
AP Train IC Train
3.85 2.7 16.3 2.7 4.2 2.7 16.3 2.7 4.2 2.7 16.3 2.7 3.0 7.5 3.0 6.4 2.6 15.4 2.6 5.8 2.6 15.4 2.6
2.7 16.3 2.7 4.2 2.7 16.3 2.7 4.2 2.7 16.3 2.7 2.6 15.4 2.6 5.8 2.6 15.4 2.6 5.8 2.6 15.4 2.6
LS 1 LS 2 LS 3 LS 4 LS 5 LS 6 LS 7 LS 1 LS 2 LS 3 a)
NORTH D4 D4 D4 D4 D4 D4 SOUTH
D4
D2 D4 D2 D2 D2 D4D2D4 D2 D2 D2 D4D2D4 D2 D2 D2 D2
D3 D3 D3 D3 D3 D3 D3 D3 D3
D1 D1 D1
P1 P2 1/6 P3 P4
1/6 1/6
1/4 1/4 1/4
1/3 1/3 1/3
1/2 1/2 1/2
Fig. 6 Baseline and damage scenarios: a combination of 100 simulations for the baseline condition,
b types of damages and their location on the bridge over the Sado River
of type D1, four severities of damage were included, namely increases of the fric-
tion coefficient from a reference value of 1.5% to 1.8%, 2.4%, 3.0%, as well as
to a full restrain of the movements between the pier and the deck. The remaining
damage scenarios consisted of 5%, 10%, and 20% stiffness reductions in the chosen
sections of the bridge (Fig. 6b) on the concrete slab (D2), the diaphragms (D3),
and arches (D4). These structural changes were simulated by reducing the modulus
of elasticity of concrete (type D2) and of steel (types D3 and D4). A total of 114
damage scenarios were simulated for AP train crossings at 220 km/h and adding as
input the temperatures measured on site during a summer day. Additional damage
scenarios could have been simulated for different combinations of EOVs. However,
as observed in Sect. 4.2, the proposed methodology is effective in removing these
effects and keeping only those generated by structural changes.
The time-series illustrated in Fig. 7 are examples of simulated responses for
baseline and damage conditions, acquired from the accelerometer Ac1.
To obtain the most reliable reproduction of the real SHM data, the noise measured
on site by each accelerometer was added to the corresponding numerical output.
These noise distributions were acquired while no trains were travelling over the bridge
and under different ambient conditions. Each simulation was corrupted with different
128 A. Meixedo et al.
Vertical Acceleration(m/s 2 )
Vertical Acceleration(m/s 2 )
Ac1 Ac1
0.05
0.05
0
0
-0.05
-0.05
IC:Load 1 IC:Load 2 IC:Load 3 AP 220km/h AP 215km/h AP 210km/h
-0.1
0 2 4 6 8 10 12 0 2 4 6 8 10
Time(s) a1) Time(s) a2)
0.1 0.1
Vertical Acceleration(m/s 2 )
Vertical Acceleration(m/s 2 )
0.05 0.05
0 0
-0.05 -0.05
Fig. 7 Numerical simulations of sensor Ac1: a1 baseline time-series using different LS of the
IC train at 190 km/h, a2 baseline time-series using different speeds of the AP train with LS5, b1
damage time-series considering friction increase D1 (P2), and b2 damage time-series considering
stiffness reduction D2 (m2)
noise signals acquired at different days, thus ensuring the most representative
validation for the techniques developed herein.
The variations associated with different train types, loading schemes and train
speeds are shown in Fig. 7a1, a2. A clear distinction between the bridge responses
for the IC (Fig. 7a1) train and the AP train (Fig. 7a2) passages can be observed,
thus displaying the necessity of considering different train types for implementing
damage detection strategies. Contrariwise, Fig. 7a1 allows observing that different
LS generate smaller changes in the dynamic responses. The train speed also has an
important influence in the structural response induced by trains crossing the bridge,
as shown in Fig. 7a2.
The influence of damage scenarios in the signal obtained for the train crossings
appears to be much smaller than that observed for EOVs, even when regarding
sensors adjacent to the damages and for the biggest magnitudes considered (20%
stiffness reductions). This conclusion can be easily observed in Fig. 7b1, b2, where
the bridge responses considering friction increments in the bearing devices of pier
P2 and stiffness reductions in the concrete slab are, respectively, presented.
Real-Time Unsupervised Detection of Early Damage in Railway … 129
The first step of the strategy is to extract damage-sensitive features based on the
hybrid combination of Continuous Wavelet Transform (CWT) and principal compo-
nent analysis (PCA). After extracting the wavelets coefficients by applying a CWT
to the acceleration measurements, a PCA is performed to significantly compress
information.
Wavelet functions are composed of a family of basic functions that can describe
a signal in localized time (or space) and frequency (or scale) domain. The main
advantage achieved by using wavelets is the ability to perform local analysis of a
signal, i.e. zooming on any interval of time or space. Signal-based damage detection
techniques that involve wavelet analysis take advantage of this to be capable of
revealing some hidden aspects of measured signals [24].
The CWT is a well-established method of implementing multiscale signal anal-
ysis. This technique will only be introduced here, and the reader is referred to Cohen
and Ryan [25], amongst others, for a more detailed mathematical explanation. The
CWT decomposes the analysed signal into a set of coefficients in two dimensions,
shift and scale, where scale is approximately inversely proportional to frequency. A
basis function is translated (shift) and stretched (scale), and compared against the
signal. High coefficients represent a good match between signal and wavelet at a
particular instant in time and related frequency [26]. Hereupon, the CWT provides
variable resolution and delivers a map of the energy content of the signal in time and
frequency.
Let f (t) be the acceleration response of the system, where t denotes time. The
wavelet coefficients are described as the inner product of the function f and the
wavelet ψa,b corresponding to parameters a (scale) and b (shift) [26]:
+∞
1 ∗ t −b
Wψ f (a, b) = f, ψa,b =√ f (t)ψ dt for a > 0 (1)
a a
−∞
In the previous equation, ψ(t) is the mother wavelet, in which the superscript
asterisk indicates complex conjugation. The wavelet functions ψa,b (t) are constructed
by a translated and dilated version of the mother wavelet, using the two parameters
a and b (Eq. 2). The parameter b localizes the basis function at t = b and its neigh-
bourhood by windowing over a certain temporal stretch depending on the parameter,
a.
1 t −b
ψa,b (t) = √ ψ for a, ba, b ∈ R + (2)
a a
130 A. Meixedo et al.
fc
f = (4)
at
Considering a vector of 2112 acceleration measurement points, the present anal-
ysis used the Morlet mother wavelet [28] to extract matrices of 2112-by-82 features
(wavelets coefficients) for each of the 23 sensors and for each of the 214 structural
conditions. To illustrate the feature extraction procedure, Fig. 8 shows these matrices,
plotted as images with scaled colours, for two different sensors: (1) AL-P2—longi-
tudinal accelerometer located in pier P2, (2) Ac1—vertical accelerometer in the mid-
span section of the concrete slab; and for four structural conditions: (a) undamaged
scenario AP|220 km/h|AUT—the AP train crossing the bridge at 220 km/h during
an autumn day, (b) undamaged scenario IC|190 km/h|SPR—the IC train crossing the
bridge at 190 km/h in a spring day, (c) damaged scenario D1 (P2: restrained)—the
full restraint of the bearing devices in pier P2, and (d) damaged scenario D2 (e2:
20%)—the 20% stiffness reduction in a section of the concrete slab aligned with pier
P2.
In Fig. 8 is possible to clearly observe different energy concentrations depending
on the sensor. Sensor AL-P2 seems to be more sensitive to the different structural
conditions analysed since the CWT coefficients from this sensor (Fig. 8a1, b1, c1,
d1) provide a much clearer image of the evolution of the energy content from the
undamaged scenarios to the damaged scenarios. On the contrary, the energy concen-
tration on the images from sensor Ac1 seems to be very similar between the undam-
aged scenario AP|220 km/h|AUT (Fig. 8a2) and both damage scenarios D1 and D2
(Fig. 8c2, d2). The main variation in frequency values is observed for the undamaged
scenario IC|190 km/h|SPR (Fig. 8b2), which indicates that the influence of varying
environmental and operational conditions (such as different temperature, train, and
speed) may be greater than the damage occurrence. These results also show that
different sensors can store different information about the bridge structural condition,
thus combining this information may enhance the features sensitiveness.
Real-Time Unsupervised Detection of Early Damage in Railway … 131
2000 0.05
0.04
AP|220km/h|AUT
0.04
Wavelet coeff.
1500
0.03
0.03
1000
0.02 0.02
a1) a2)
10-3 0.03
2000
8 0.025
IC|190km/h|SPR
Wavelet coeff.
1500 0.02
6
0.015
1000
4
0.01
500 2
0.005
b1) b2)
10-3 0.05
2000
15
D1 (P2:restrained)
Wavelet coeff.
0.04
1500
10 0.03
1000
0.02
5
500 0.01
c1) c2)
10-3
2000 7
0.05
Wavelet coeff.
D2 (e2:20%)
0.04
1500
5
0.03
1000
3 0.02
500 0.01
1
0 20 40 60 80 0 20 40 60 80
Wavelet coeff. d1) Wavelet coeff. d2)
D2 D2
D2
D1 AL-P2
AL-P2 * D1 Ac1 D1* Ac1
Fig. 8 Wavelet coefficients extracted from the acceleration responses according to the structural
condition of the bridge and the sensor location
132 A. Meixedo et al.
Y = X ·T (5)
C = T · · TT (6)
WI M
M
AUM
WM
AUM
R
T
T
SUR
T
R
SP T
T
T
R
AUIN
N
P2N
q3n
q21n
s1sn
e2n
e3sn
q1n
q s
m1s
q2s
s n
s32ns
e2s
s3s
t2n
AU
t1n
m2
s13
m2
t3n
AU
AU
t1s
t2s
SU
m1
t3s
m3
SU
SU
SU
SP
SU
SP
SP
WI
WI
P3
P4
q3
e43
e1
e3
e2
s2
e4
e1
m
e
6
Feature 12
4
AL-P2
2
0
-2
-4
5
Feature 47
0
-5
AL-P2
-10
4
Feature 116
2
0
-2
-4 AL-P2
-6
6
Feature 323
4 AL-P2
2
0
-2
-4
0 15 30 45 60 75 90 105 120 135 150 165 180 195 210
Structural Condition a)
Feature 12
2
Ac1
0
-2
4
Feature 47
2 Ac1
0
-2
6
Feature 116
4 Ac1
2
0
-2
-4
10
Feature 323
8 Ac1
6
4
2
0
-2
0 15 30 45 60 75 90 105 120 135 150 165 180 195 210
Structural Condition b)
AP 220km/h AP 215km/h AP 210km/h D1 D3
IC 195km/h IC 190km/h IC 185km/h D2 D4 AL-P2 Ac1
Fig. 9 Amplitude of four of the 328 statistical features extracted based on CWT and PCA, for all
214 structural conditions and from two of the 23 sensors: a longitudinal accelerometer located on
pier P2 (AL-P2), b vertical accelerometer located in the mid-span section of the concrete slab (Ac1)
134 A. Meixedo et al.
The analysis of the features is presented in Fig. 9 that revealed the necessity to
adequately model these statistical parameters to remove the changes generated by
EOVs and highlight those generated by damage. Assuming that environmental condi-
tions has a linear effect on the identified features, the implementation of a latent vari-
able method as PCA to the extracted features may efficiently remove environmental
and operational effects, without need to measure these actions [11, 30].
Considering now an n-by-m matrix X with the features extracted from the dynamic
responses, where n is the number of structural conditions and m is the number
of features from all the sensors (i.e. 328), a transformation to another set of m
parameters, Y, can be achieved by applying Eq. (5).
As demonstrated by Santos et al. [10], the PCA is able to cluster meaningful
information related to EOVs in the first components, while variations related to other
small-magnitude effects, such as early damage, may be retained in latter components.
Since the purpose of the present research work is to detect damage, which has gener-
ally a local character, the feature modelling operation consists of eliminating the most
important principal components (PCs) from the features and retaining the rest for
subsequent statistical analysis. Bearing this in mind, the matrix from Eq. (6) can be
divided into a matrix with the first e eigenvalues and a matrix with the remaining m-e
eigenvalues. Defining the number of e components remains an open question with
regard to the representation of the multivariate data; although several approaches
have been proposed, there is still no definitive answer. In this work, the value of e
(or the number of PCs to discard) is determined based on a rule of thumb in which
the cumulative percentage of the variance reaches 80% [31]. After choosing e, the
m-e components of the matrix Y can be calculated using Eq. (5) and a transformation
matrix T̂ built with the remaining m-e columns of T. Those m-e components can be
remapped to the original space using the following:
FPCA = X · T̂ · T̂ T (7)
WI M
M
AUM
WM
AUM
R
T
SUR
T
R
SP T
T
T
R
AUIN
N
P2N
q3n
q21n
s1sn
e2n
e3sn
q13n
q s
q3s
q2s
s21n
s3ns
e1 s
s
t2n
AU
m21
m3
m2
t1n
t3n
AU
AU
t1s
t2s
m41
SU
t3s
SU
SU
SU
SU
SP
SP
SP
WI
WI
P3
P4
e3
s2
s3
e2
e3
e2
e1
e4
m
m
e
s
6
Feature 12
4
AL-P2
2
0
-2
-4
5
Feature 47
-5
AL-P2
-10
4
Feature 116
2
0
-2
-4 AL-P2
-6
6
Feature 323
4 AL-P2
2
0
-2
-4
0 15 30 45 60 75 90 105 120 135 150 165 180 195 210
Structural Condition a)
Feature 12
2 Ac1
0
-2
4
Feature 47
Ac1
2
0
-2
6
Feature 116
4 Ac1
2
0
-2
-4
10
Feature 323
8 Ac1
6
4
2
0
-2
0 15 30 45 60 75 90 105 120 135 150 165 180 195 210
Structural Condition b)
AP 220km/h AP 215km/h AP 210km/h D1 D3
IC 195km/h IC 190km/h IC 185km/h D2 D4 AL-P2 Ac1
Fig. 10 Amplitude of four of the 328 features modelled based on PCA, for all 214 structural
conditions and from two of the 23 sensors: a longitudinal accelerometer located on pier P2 (AL-P2),
b vertical accelerometer located in the mid-span section of the concrete slab (Ac1)
136 A. Meixedo et al.
SUM
SUM
SUM
SUM
SUM
M
AUT
AUT
AUT
AUT
SPRT
SP T
SPR
SPR
SUR
WIN
WIN
WIN
WIN
q 1n
q 2n
q 3n
e2n
q 1s
q 2s
q 3s
s1n
s2n
s3n
e3n
s3s
e2s
s1s
s2s
e3s
t1n
t2n
t3n
AU
AU
m1
m2
m3
m1
m2
m3
t1s
t2s
t3s
P2
P3
P4
e4
e1
e1
e2
e3
e4
60
50
Mahalanobis Distance
AL-P2
40
30
20
10
0
0 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 210
Structural condition a)
SUM
SUM
SUM
SUM
SUM
M
AUT
AUT
AUT
AUT
SPRT
SP T
SPR
SPR
SUR
WIN
WIN
WIN
WIN
q1 n
q2n
q3n
q1 s
q2s
q3s
e2n
s1n
s2n
e3n
s1s
s3s
s3s
e2s
e3s
t1 n
t2 n
t3 n
AU
m1
m2
m3
m1
m2
AU
m3
t1 s
t2 s
t3 s
P2
P3
P4
e1
e2
e3
e4
e4
s2
e1
60
50 Ac1
Mahalanobis Distance
40
30
20
10
0
0 10 20 30 40 50 60 70 80 90 100 110 120 130 140 150 160 170 180 190 200 210
Structural condition b)
Fig. 11 Features fusion based on the Mahalanobis distance for all 214 structural conditions,
considering the responses from accelerometers: a AL-P2, b Ac1
Time-series analysis and distance measures can help perform data analysis and
suggest the existence of different structural behaviours within a dataset, as shown in
the previous sections. However, the development of real-time SHM strategies should
resort to machine learning algorithms that can autonomously decide whether one or
more distinct structural behaviours are being observed from patterns in the features.
Hence, feature discrimination is addressed herein using unsupervised classification
algorithms.
Cluster analysis was the data mining technique chosen to address feature classi-
fication. The aim of the clustering process is to divide a dataset into groups, which
must be as compact and separate as possible. This can be mathematically posed
as an attempt to minimize the dissimilarity between features assigned to the same
cluster (within-cluster distance), which, consequently, maximizes the dissimilarity
between the features assigned to different clusters (between-cluster distance) [15].
Considering a given partition containing K clusters, Pk = {C1 , . . . , Ck }, the overall
138 A. Meixedo et al.
N N
1
OD = di j (10)
2 i=1 j=1
60 60 60
Cluster 1
Cluster 2
MD (AL-P2)
MD (AL-P2)
MD (AL-P2)
40 Centroids 40 40
20 20 20
0 0 0
0 30 60 90 120 150 180 210 0 10 20 30 40 0 10 20 30 40
Structural Condition MD (Ac1) MD (AsV4)
40 40
Cluster 1
Cluster 2
AL-P2
MD (Ac1)
MD (Ac1)
Centroids
20 20
AL-P2
0 0
0 30 60 90 120 150 180 210 0 10 20 30 40
Structural Condition MD (AsV4)
40
Cluster 1
Ac1 Cluster 2
MD (AsV4)
Centroids
20
Ac1
0
1 0 30 60 90 120 150 180 210
Structural Condition
0.8
0.6
SIL
0.4
0.2 AsV4
0
2 3 4 5 6
Number of clusters a) AsV4 b)
Fig. 12 Allocation of damage-sensitive features into clusters: a silhouette index (SIL), b clusters
defined for all structural conditions and their centroids for three of the 23 sensors
by the k-means method and are shown in Fig. 12b for three of the 23 sensors: (i)
AL-P2, (ii) Ac1, and (iii) AsV4. These plots demonstrate that the clustering method
can divide the features without any human interaction or input. In Fig. 12b can be
observed the dissimilarity between the two centroids of three different combinations
of sensors, while the plots in the diagonal of this figure show that the two clusters
found for each couple of sensors are compact over time and separated when the
simulated damages start. This result undeniably shows that the k-means method is
capable of analysing the feature set and, in a fully automated manner, separating it
according to the structural conditions observed on site. Also, it is demonstrated that
the clusters have de advantage of allowing a multidimensional representation of the
features, which, in this case study, led to a classification without false detections.
After the definition of the baseline, which can be promptly achieved after one day
of trains crossings, each new train crossing can be used to test the bridge structural
condition based on the proposed strategy.
140 A. Meixedo et al.
5 Conclusions
This research presents a data-driven AI-based SHM strategy for conducting real-
time unsupervised early damage detection in railway bridge vibration response
from traffic-induced excitation, applying time-series analysis and machine learning
techniques. The strategy consists of fusing sets of acceleration measurements to
improve sensitivity and combines: (i) CWT and PCA for feature extraction, (ii)
PCA for feature modelling, (iii) Mahalanobis distance for feature fusion, and (iv)
clustering algorithms for feature classification. A comprehensive dataset of baseline
and damaged scenarios was simulated using a highly reliable digital twin of the
Sado Bridge tuned with experimentally obtained actions as input, namely tempera-
ture, train loadings, and speeds. Damage severities of 5%, 10%, and 20% stiffness
reductions in the concrete slab, diaphragm, and arches were simulated, as well as
friction increases in the movements of the bearing. The damage-sensitive features
were extracted from the bridge accelerations induced by train crossings by combining
CWT with PCA. The wavelets coefficients were first extracted from the time-series.
Afterwards, PCA was implemented to the wavelets coefficients and statistical param-
eters were extracted from the PCs to allow data compression. The study of the
wavelet-PCA-based features extracted from different structural conditions allowed
drawing conclusions about the supremacy of the EOVs when compared with damage,
proving the importance of feature modelling. Moreover, the information obtained
from each feature was different depending on the sensor location and the statis-
tical parameter. PCA was once again implemented to modelling the features. This
latent variable method proved its importance and effectiveness in removing observ-
able changes induced by variations in train speed or temperature without the need
to measure them and without losing sensitivity to damage. To describe the vari-
ability present in the modelled features, a Mahalanobis distance was implemented
to the 328 features extracted from each sensor signal. This implementation allowed
corroborated that different sensors have greater or lesser sensitivity, depending on
the location of the damage. Moreover, this step proved to be crucial to achieve the
highest possible level of information fusion and to obtain a clear distinction between
undamaged and damaged conditions. In order to automatically detect the presence of
damage, a clustering-based classification was performed. The robustness and effec-
tiveness of the proposed strategy were demonstrated by automatically detecting the
damage scenarios as different from those belonging to undamaged structural condi-
tions. Using features modelled based only on structural responses, no false detections
occurred. An additional important conclusion obtained from this work is that, even
with an SHM system not capable of measuring EOVs, it is possible to successfully
detect different types of damage using the bridge’s responses to train crossings. This
achievement renders the strategy the ability to be less dependent on spatial actions
very difficult to characterize, thus contributing for the normalization of SHM proce-
dures. This strategy also has the advantages of minimizing the number of sensors
that need to be installed and, consequently, the cost of the SHM system, as well as
allowing for a more automatic and straightforward implementation.
Real-Time Unsupervised Detection of Early Damage in Railway … 141
Acknowledgements This work was financially supported by the Portuguese Foundation for
Science and Technology (FCT) through the PhD scholarship SFRH/BD/93201/2013. The authors
would like to acknowledge the support of the R&D project RISEN through the H2020|ES|MSC—
H2020|Excellence Science|Marie Curie programme, the Portuguese Road and Railway Infras-
tructure Manager (I.P), the Portuguese National Laboratory for Civil Engineering (LNEC), and
the Base Funding—UIDB/04708/2020 of the CONSTRUCT—Instituto de I&D em Estruturas e
Construções—financed by national funds through the FCT/MCTES (PIDDAC).
References
15. Cardoso R, Cury A, Barbosa F (2019) Automated real-time damage detection strategy using
raw dynamic measurements. Eng Struct 196(109364). https://doi.org/10.1016/j.engstruct.2019.
109364
16. Azim R, Gül M (2019) Damage detection of steel girder railway bridges utilizing operational
vibration response. Struct Control Health Monit 26(e2447):1–15. https://doi.org/10.1002/stc.
2447
17. Nie Z, Lin J, Li J, Hao H, Ma H (2019) Bridge condition monitoring under moving loads
using two sensor measurements. Struct Health Monit 19(3):917–937. https://doi.org/10.1177/
1475921719868930
18. Farrar CR, Doebling SW, Nix DA (2001) Vibration–based structural damage identification.
Philos Trans R Soc London A: Math Phys Eng Sci 359(1778):131–149. https://doi.org/10.
1098/rsta.2000.0717
19. ANSYS. Academic Research. Release 17.1 2016
20. Meixedo A, Ribeiro D, Santos J, Calçada R, Todd M (2021) Progressive numerical model
validation of a bowstring-arch railway bridge based on a structural health monitoring system.
J Civ Struct Heal Monit 11(2):421–449. https://doi.org/10.1007/s13349-020-00461-w
21. Min X, Santos L (2011) Ensaios dinâmicos da ponte ferroviária sobre o rio sado na variante
de alcácer. Lisboa [Portuguese]
22. Meixedo A, Gonçalves A, Calçada R, Gabriel J, Fonseca H, Martins R (2016) On-line moni-
toring system for tracks. In: exp.at 2015—3rd experiment international conference, Sao Miguel
Island, Azores. https://doi.org/10.1109/EXPAT.2015.7463240
23. Pimentel R, Ribeiro D, Matos L, Mosleh A, Calçada R (2020) Bridge weigh-in-motion system
for the identification of train loads using fiber-optic technology. Structures 2021(30):1056–
1070. https://doi.org/10.1016/j.istruc.2021.01.070
24. Ren WX, Sun ZS (2008) Structural damage identification by using wavelet entropy. Eng Struct
30:2840–2849. https://doi.org/10.1016/j.engstruct.2008.03.013
25. Cohen A, Ryan RD (1995) Wavelets and multiscale signal processing. Chapman & Hall,
Boundary Row, London
26. Cantero D, Ülker-kaustell M, Karoumi R (2016) Time–frequency analysis of railway bridge
response in forced vibration. Mech Syst Signal Process 76–77:518–530
27. Ülker-kaustell M, Karoumi R (2012) Influence of non-linear stiffness and damping on the
train-bridge resonance of a simply supported railway bridge. Eng Struct 41:350–355. https://
doi.org/10.1016/j.engstruct.2012.03.060
28. Teolis A (1998) Computational signal processing with wavelets. Birkhauser
29. Ribeiro D, Leite J, Meixedo A, Pinto N, Calçada R, Todd M (2021) Statistical methodologies for
removing the operational effects from the dynamic responses of a high-rise telecommunications
tower. Struct Control Health Monit 28(4):e2700. https://doi.org/10.1002/stc.2700
30. Yan A, Kerschen G, De BP, Golinval J (2005) Structural damage diagnosis under varying
environmental conditions—Part I: a linear analysis. Mech Syst Signal Process 19(4):847–864.
https://doi.org/10.1016/j.ymssp.2004.12.002
31. Jolliffe IT (2002) Principal component analysis, 2nd edn. Springer, New York, pp 112–147
32. Hastie T, Tibshirani R, Friedman J (2011) The elements of statistical learning, data mining
inference, and prediction, 2nd edn. Springer, Stanford, pp 460–462
33. Santos J, Crémona C, Calado L (2016) Real-time damage detection based on pattern
recognition. Struct Concrete 17(3):338–354. https://doi.org/10.1002/suco.201500092
Fault Diagnosis in Structural Health
Monitoring Systems Using Signal
Processing and Machine Learning
Techniques
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 143
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_7
144 H. Fritz et al.
SHM system in operation at a railway bridge. The results demonstrate the suitability
of the FD approach for ensuring reliable SHM systems.
1 Introduction
In SHM systems, generally, seven basic sensor fault types occur, which are depicted
in Fig. 1 together with formal descriptions as functions over time. Bias is a deviation
by a constant value between sensor data and actual structural response. Drift is the
incrementing deviation between sensor data and actual structural response over time.
In the case of a gain, sensor data is scaled by a constant value; while in precision
degradation, sensor data is contaminated with white noise. For a complete failure,
sensor data consists of a constant value or noise regardless of changes in the actual
structural response. Outliers are isolated, non-continuous changes (“dropouts”) of
the signal at individual points in time, with signals returning to their real values after
each single dropout.
Table 1 presents an overview of studies available in literature analyzing reasons
for the occurrence of individual sensor fault types in SHM systems focusing only on
failures in the sensing mechanism and not on failures coming from synchronization
faults. In addition to the causes in Table 1, it should be noted that the mechanism
and recurrence frequency of individual sensor fault types are highly dependent on
the physical design of the sensors as well as the environmental conditions.
Fault identification of sensors deployed in SHM systems is a crucial step toward
FD, as fault types may allow inferring causes of sensor faults in SHM systems. For
autonomous fault identification, ML provides promising classification algorithms.
ML generally represents learning processes of computer systems, often described as
the conversion of experience into expertise or knowledge [37]. ML algorithms may
help analyze large amounts of data, recognize data patterns, and adapt to the patterns
autonomously. Compared to traditional algorithms (such as expert systems), ML
algorithms offer two advantages, (i) ML algorithms work with previously unknown
(i.e., new) data for which the system has not been trained, and (ii) ML algorithms
can adapt to environmental conditions and resulting changes in the data.
Theoretically, automatic identification of sensors faults can be performed by using
ML classification algorithms with raw sensor data as input. However, because raw
sensor data typically contains noise and random components that may introduce
aleatory uncertainties, patterns in the data indicative of sensor fault types (“fault
Fault Diagnosis in Structural Health Monitoring Systems … 147
Sensor output
0 100 200 300 400 500 600 700 800 900 1000 0 100 200 300 400 500 600 700 800 900 1000
Time Time
Sensor output
0 100 200 300 400 500 600 700 800 900 1000 0 100 200 300 400 500 600 700 800 900 1000
Time Time
Sensor output
0 100 200 300 400 500 600 700 800 900 1000 0 100 200 300 400 500 600 700 800 900 1000
Time Time
Drift Outliers
Sensor output
Sensor output
..
0 100 200 300 400 500 600 700 800 900 1000 0 100 200 300 400 500 600 700 800 900 1000
Time Time
Table 1 (continued)
Fault type Cause Source
Outliers • Mechanisms such as heavy-tailed distributions or [34]
data that are coming from different kinds of
distributions
• Environmental variations (e.g., cold temperature [35]
influence)
• Low battery supply, loose electrical contact, and [36]
sensor saturations
[38]. The wavelet transform uses a series of functions, the so-called wavelets, each
with a different scale. Since wavelets are localized in time, the wavelet transform
can extract coupled time–frequency information from a signal.
Definition 1 Let u, ψ ∈ L 2 (R, R). For b ∈ R and a > 0, the wavelet transform of
u with ψ is defined by
1 t −b
L ψ u(a, b) := u(t) √ ψ dt, (1)
a a
R
When transforming the fault types of Fig. 1 through wavelets, all sensor fault
types, except outliers, are recognizable because of unique patterns the fault types
follow. The reason outliers are not identified is that they correspond to point singu-
larities in a continuous signal, which are, in general, not detectable by standard
integral transforms. However, outliers can be easily detected and removed from the
original signal by thresholding, and therefore, outlier identification is not critical to
the complete FD approach. Thus, outliers will not be considered in this work.
After pre-processing faulty sensor data, ML algorithms may be used to allow
automatic identification of sensor faults. To integrate fault identification into the
overall FD process, built upon the wavelet transform, a generalized FD approach is
proposed in the following section.
150 H. Fritz et al.
Frequency Frequency
Fig. 2 Plots of wavelet transform of artificial signal (2) with different fault types injected into the
signal
This section presents the methodology for the autonomous FD using wavelet trans-
form and machine learning. The methodology for autonomous FD presented in this
chapter is realized via two ML algorithms, (i) a ML regression algorithm for fault
detection, fault isolation, and fault accommodation, and (ii) a ML classification algo-
rithm for fault identification. The ML approach for fault detection and isolation has
already been introduced in previous work [3] and is generalized herein. For training
the ML regression algorithm, correlations among sensor data at non-faulty sensor
operation are exploited. Correlations between sensor data may occur as a result of
sensor placement on the structure and may be readily visible in the raw data or
exposed through data-pre-processing (e.g., for acceleration data through using the
Fourier amplitudes of correlated sensor data in the frequency domain). Fault isolation
is accomplished by designing a separate instance of the ML regression algorithm for
each sensor. For the ML classification algorithm of the FD approach, patterns indica-
tive of faulty sensor data, as described in the previous section, is used to automate the
Fault Diagnosis in Structural Health Monitoring Systems … 151
Fig. 3 Implementation
Sensor i
framework for the FD
approach Measuring
Measurement f t
Pre-processing
ML regression
Approximation t
algorithm for sensor i
Comparison
Fault detection
Fault isolation
no
t f t a
yes
Sensor alert
ML classification
algorithm Fault classification
Wavelet transform of
Fault accommodation
f t
152 H. Fritz et al.
4 Validation Tests
In this section, validation tests are performed to showcase the ability of the FD
approach to diagnose sensor faults reliably and accurately on real-world data recorded
by an SHM system in operation on a railway bridge. First, the ML algorithms used
in the validation tests are briefly described. Next, the railway bridge is presented.
Then, the data pre-processing and the determination of ML algorithms parameters
are explained. Finally, the validation test results of the FD approach are shown and
discussed.
The validation tests are performed using an ANN as a ML regression algorithm and a
CNN as a ML classification algorithm. In this subsection, the ANN and the CNN are
briefly described, with emphasis on how each algorithm is applied to the respective
FD tasks.
Fault detection, fault isolation, and fault accommodation using artificial neural
networks. The ANN employed for the validation tests of this study performs fault
detection, fault isolation, and fault accommodation. The ANN concept is based on
the function of biological neurons, which “fire” (become activated) upon receiving
stimuli. Mathematically, the activation (i.e., output) of neuron M is described as
follows [39]
K
N M := f wi x i + b , (3)
i=1
which takes the input x i (components of some input vector x), makes the summation
with weights wi , adds a bias b, and passes it with a transfer function f . Typically,
“sigmoid” functions (e.g., hyperbolic tangent) are used as activation functions. The
overall ANN output is then obtained as a weighted combination of activations of
individual neurons organized in layers, constituting an input layer, one (or more)
hidden layer(s) and an output layer, see [39] for details. Further information on the
concept of artificial neural networks for fault detection and fault isolation used in
this study may be found in [3].
Figure 4 shows the topology of an ANN for fault detection and isolation in sensor
i. As mentioned above, to achieve fault isolation, one ANN model is created for
each sensor. As for fault detection, the ANN model for sensor i goes through three
phases, the training phase, the testing phase, and the application phase. In the training
phase, a training data set is formed in a SHM system state designated as “non-faulty,”
comprising both data of sensor i (target output) and corresponding (i.e., at the same
time points) sensor data correlated with the data of sensor i (input data), collected
Fault Diagnosis in Structural Health Monitoring Systems … 153
Virtual sensor
Sensor data of
data tj of
sensor 1, … , k
sensor i
tj tj
j ( f (t j ) − fˆ(t j ))2
R2 = 1 − (4)
j ( f (t j ) − f )2
154 H. Fritz et al.
where f denotes the mean value of f t j for t j ∈ T . Typically, values of R 2 lie
between 0 and 1, with values close to 1 indicating a good approximation quality.
Values less than zero are possible when the approximation function fits the data
worse than the mean value of the data. A sensor fault detection alert is issued if the R 2
value differs more than a tolerance ε from the threshold already defined at the testing
phase. For differentiating between sensor faults and structural damage, the virtual
data of ANN instances of several sensors need to be analyzed collaboratively. Fault
detection alerts issued by the majority of sensors may indicate structural damage.
If a sensor fault alert is issued, fault identification and fault accommodation are
initiated. In case of a faulty sensor, for fault accommodation, the faulty sensor data
f Si (t) is replaced by the virtual data fˆSi (t).
Fault identification using convolutional neural networks. In what follows,
autonomous fault identification is implemented through a CNN, representing the
supervised ML classification algorithm in this study, for the sensor fault types
presented in Sect. 2. Convolutional neural networks represent a class of ML algo-
rithms for classification problems, commonly applied to image and pattern recog-
nition [41]. In contrast to artificial neural networks, convolutional neural networks
combine three characteristics, making CNN algorithms suitable for classification:
• Local receptive fields, allowing the first layer to extract features from segments
of the input data by means of filters. The subsequent layers combine the features
extracted, detecting higher-order abstractions.
• Shared weights, which derive from the idea that filters can be used across the entire
input data to extract features irrespective of the section of the input data. The exact
location of a feature is not important, as only its relative position to other features
is relevant. Thus, by applying the weights of a filter to different sections of the
input data, a feature map is generated. Input data of the same type may by roughly
different; but, if the relative position of the features is approximately the same, a
CNN is capable of classifying the input data as belonging to the same type.
• Subsampling, reducing the spatial resolution of a filter and the sensitivity of the
output to variations, shifts, and noise.
The concept of using a CNN for fault identification is presented in Fig. 5. A CNN
takes as input a signal x(u), which is here an image obtained after a wavelet transform
of the faulty sensor data. Neuron values for x j for layer j are computed from values of
...
.. Bias; Gain, Drift, Precision
degradation; Complete
failure (constant, and noise)
the previous layer x j−1 by applying a linear operator W j and a pointwise nonlinearity
rectifier ρ, see [42] for details:
x j = ρW j x j−1 . (5)
Details on the parameters of the CNN used in the validation tests are given below.
By help of (5), a higher abstraction of the wavelet transforms is obtained (“abstracted
wavelet transforms”), generalizing the main features of the different fault types, as
indicated in Fig. 5. The complete CNN is then obtained by constructing a cascade of
several convolutions of the form (5). Additionally, the abstracted wavelet transforms
are flattened, generating a one-dimensional vector that is interpretable by the final
output layer of the CNN. The output layer returns a percentage for every possible
fault type, representing the probability of the abstracted wavelet transform to belong
to each fault type. Thus, the CNN is able to classify the abstracted wavelet transforms
into each fault type, based on the highest percentage obtained from the activation
function.
The six fault types that, as previously mentioned, are considered in this study
constitute six classes for the ML classification. It is important to note that a seventh
class for non-faulty data is not required, because non-faulty data is sieved out by the
ANN on the fault detection step of the FD approach proposed in this paper. If the
accuracy of the ANN in the FD process is not sufficient for application, the seventh
class of “non-faulty” data may be added to the CNN, for avoiding error accumulation.
The validation is performed using sensor data recorded by an SHM system installed
on a reinforced concrete railway bridge that includes monitoring data of one year,
collected as part of a previous research project [43]. The bridge is a double-track
railway overpass. The bridge deck consists of ten spans each resting on four piers
with circular cross-sections, which are monolithically connected to the deck. The
piers are founded on bored piles, whose heads are connected with rigid beams (pile
head beams), and the sensors of the SHM system are installed in the pile head beam.
The data set contains sensor data from one data acquisition unit of one pile head
beam of the bridge comprising 11 sensors, whose types and labels are listed in Table
2. The top view of the pile head beam and the sensor positions of sensor S1–S11 are
illustrated in Fig. 6. Further figures and information of the validation structure and
the monitoring data may be found in [43].
The data set contains data from eight strain sensors and three temperature sensors
from the year 2017, each data point being collected every 10 min, i.e., at a sampling
frequency of 1.7 MHz, corresponding to n = 52,560 data points per sensor. The
validation of the FD approach proposed is shown for faults artificially injected into
the sensor data of strain gauge S1.
156 H. Fritz et al.
Prior to training the ANN, the Pearson correlation coefficients, as shown in Fig. 7,
are calculated to unveil correlations between sensor data, in particular between the
temperature sensors and the strain gauges on the steel reinforcement. Not all sensor
data in the data acquisition unit is correlated. If sufficient training data is available,
a lack of correlation in the training data between the input layer and the output layer
does not degrade the precision of the ANN. However, when having limited training
data, only correlated sensor data should be used.
Before training, the sensor data serving as input to the ANN (i.e., sensor data from
sensors S2–S11) is normalized because the magnitude of the values of the sensor
data is different from sensor to sensor due to the different types (e.g., temperature
values are much higher than strain values). The ANN topology as well as the neuron
activation functions are defined through trial-and-error, as listed in Table 3. The input
layer of the ANN consists of ten neurons, representing data points at time step t j for
sensors S2–S11. The output layer of the ANN has one neuron, which yields virtual
data at time step t j for sensor S1. The activation function selected for the input layer
and the hidden layers is the rectified linear unit (ReLU). For the output layer, the
Fault Diagnosis in Structural Health Monitoring Systems … 157
a) 0.992
Sensor S1
Strains [mV/V]
0.99
0.988
00:00 04:48 09:36 14:24 19:12 00:00
Time [hh:mm]
b)
Complete failure (constant) Bias Gain
Scale
setting parameter σ = 40. Exemplarily, a one-day sensor data set used for training
the CNN is presented in Fig. 8a. The strain curve of the one-day-sensor data set
exhibits fluctuations due to the changes in ambient temperature.
S3
Top view pile head beam
S7
S2 S6 S4 S8
S9-11
S1
Pile
S5
Bored pile
Each input for the CNN in the training process is one 2D image of wavelet-
transformed one-day sensor data of faulty signals, as exemplarily shown in Fig. 8b.
In total, a number of n = 5000 training images per fault type are fed to the CNN.
The hyperparameters and properties of the CNN, selected through trial-and-error,
are listed in Table 3, namely the network topology (the number of neurons per layer)
and the neuron behavior (the activation functions). The ReLU activation function is
used for the input and intermediate layers; while for the output layer, the “softmax”
activation function is selected. “Pool size” refers to the factor by which pooling
layers reduce the feature map, and “strides” represent the frame size with which the
filters of convolution layers “convolve” around their input. Finally, “dropout” is a
probability value below which neurons of a layer are ignored (“dropped”).
It should be noted that since the ANN is able to adapt to the condition of
the monitored structure, the ANN properties are case-specifically defined for the
validation tests presented herein, i.e., the ANN properties depend on the char-
acteristics of the monitored structure and on the SHM system setup. The CNN,
however, due to identifying artificial faults within a wide range of fault magnitudes
is measurement-independent.
For the validation tests, the fault types previously mentioned (bias, drift, gain, preci-
sion degradation, and complete failure) are simulated and injected into the sensor
data of sensor S1. Faulty data is characterized by function f˜ defined by the faults
compromising non-faulty data f . Tests are performed by varying the magnitude of
the faults and observing the change in the coefficient of determination. The results
of the ANN performance with increasing fault magnitudes are shown in Table 4. The
ANN for sensor S1 is trained with n = 20,000 data points as training data. In the
testing phase, using a testing set of n = 2000 data points, the high precision of the
training is showcased through the coefficient of determination equal to R 2 = 0.992.
The magnitude of sensor faults b is based on the mean value x̄ of the data set to be
tested. Since the training precision of the ANN is high, the threshold for indicating
a fault is accordingly set high to a = 0.98.
Furthermore, tests have been performed to identify above which percentage of a
faulty signal the ANN is capable of detecting faults. Here, three percentage levels
of faulty data are considered, 1%, 3%, and 5%. For all fault types, except drift, it is
possible to detect the fault within under 1% of faulty data. Since drift requires time
to fully manifest, detecting drift with 5% of faulty data is not possible. With 30% of
faulty data in the sensor data, a drift is detectable (marked with * in Table 5).
For validating the CNN, a “confusion matrix” is calculated, as presented in Table
6. The confusion matrix presents the classification outputs of the CNN for the test
data. The CNN classifies correctly most of the fault types, achieving an accuracy of
99.39%. Furthermore, Table 7 presents the evaluation of the CNN during testing. For
each fault type, precision, recall, and F1-score are presented. Precision represents the
160 H. Fritz et al.
ratio between correctly predicted fault types and the total predictions of the fault type.
Recall represents the ratio between correctly predicted fault types and the actual total
observations of the fault type. Finally, the F1-score represents the weighted average
of precision and recall. For example, the bias fault has 394 correctly predicted fault
types, eight incorrectly predicted fault types as drift, and one observation incorrectly
Fault Diagnosis in Structural Health Monitoring Systems … 161
classified as bias. Therefore, bias has a precision of 99.74%, a recall of 98.00%, and
a F1-score of 98.86%. It can be observed that the evaluation of CNN gives results
close to 100% in all scores, thus denoting reliable results.
Acknowledgements The authors gratefully acknowledge the support offered by the German
Research Foundation (DFG) under grants SM 281/9-1, SM 281/12-1, SM 281/14-1, SM 281/15-1,
and LE 3955/4-1. This work is also partially sponsored by the German Federal Ministry of Transport
and Digital Infrastructure (BMVI) under grant VB18F1022A. Any opinions, findings, conclusions,
or recommendations expressed in this chapter are those of the authors and do not necessarily reflect
those of DFG or BMVI.
References
1. Glisic B, Inaudi D, Casanova N (2010) SHM process as perceived through 350 projects. In:
SPIE 7648, smart sensor phenomena, technology, networks, and systems. San Diego, CA,
USA, 04/08/2010, pp 76480P
2. Zhang Z, Mehmood A, Shu L, Huo Z, Zhang Y, Mukherjee M (2018) A survey on fault diagnosis
in wireless sensor networks. IEEE Access 6:11349–11364
3. Dragos K, Smarsly K (2016) Distributed adaptive diagnosis of sensor faults using structural
response data. Smart Mater Struct 25(10):105019
4. Patton RJ (1991) Fault detection and diagnosis in aerospace systems using analytical
redundancy. Comput Control Eng J 2(3):127–136
5. Willsky AS (1976) A survey of design methods for failure detection systems. Automatica
12:601–611
6. Kraemer P, Fritzen C-P (2007) Sensor fault identification using autoregressive models and the
mutual information concept. Key Eng Mater 347:387–392
7. Smarsly K, Law KH (2014) Decentralized fault detection and isolation in wireless structural
health monitoring systems using analytical redundancy. Adv Eng Softw 73(2014):1–10
8. Luckey D, Fritz H, Legatiuk D, Dragos K, Smarsly K (2020) Artificial intelligence techniques
for smart city applications. In: Proceedings of the international ICCCBE and CIB W78 joint
conference on computing in civil and building engineering 2020. Sao Paolo, Brazil, 08/18/2020
9. Dai X, Gao Z (2013) From model, signal to knowledge: a data-driven perspective of fault
detection and diagnosis. IEEE Trans Industr Inf 9(4):2226–2238
10. Zaher A, McArthur SDJ, Infield DG, Patel Y (2009) Online wind turbine fault detection through
automated SCADA data analysis. Wind Energy 12(6):574–593
11. Tolani DK, Yasar M, Ray A, Yang V (2006) Anomaly detection in aircraft gas turbine engines.
J Aerosp Comput Inf Commun 3(2):44–51
12. Xu X, Hines JW, Uhrig RE (1999) Sensor validation and fault detection using neural
networks. In: Proceedings of the maintenance and reliability conference, Gatlinburg, TN, USA,
05/10/1999
Fault Diagnosis in Structural Health Monitoring Systems … 163
13. Campa G, Fravolini ML, Seanor B, Napolitano MR, Del Gobbo D, Yu G, Gururajan S (2002)
Online learning neural networks for sensor validation for the flight control system of a B777
research scale model. Int J Robust Nonlinear Control 12(11):987–1007
14. Cork L R, Walker R, Dunn S (2005) Fault Detection, Identification and Accommodation
Techniques for Unmanned Airborne Vehicle. In: Proceedings of the Australian International
Aerospace Congress. Melbourne, Australia, 03/15/2005.
15. Vemuri AT, Polycarpou MM (1997) Neural network-based robust fault diagnosis in robotic
systems. IEEE Trans Neural Networks 8(6):1410–1420
16. Eski I, Erkaya S, Savas S, Yildirim S (2011) Fault detection on robot manipulators using
artificial neural networks. Robot Comput-Integr Manuf 27(1):115–123
17. Venkatasubramanian V, Vaidyanathan R, Yamamoto Y (1990) Process fault detection and
diagnosis using neural networks—1. Steady-state processes. Comput Chem Eng 14(7):699–712
18. Obst O (2009) Distributed fault detection using a recurrent neural network. In: 2009 Inter-
national conference on information processing in sensor networks. San Francisco, CA, USA,
04/13/2009
19. Jeong S, Ferguson M, Hou R, Lynch JP, Sohn H, Law KH (2019) Sensor data reconstruction
using bidirectional recurrent neural network with application to bridge monitoring. Adv Eng
Inform 42:100991
20. Steiner M, Legatiuk D, Smarsly K (2019) A support vector regression-based approach
towards decentralized fault diagnosis in wireless structural health monitoring systems. In: The
12th international workshop on structural health monitoring (IWSHM). Stanford, CA, USA,
10/09/2019
21. Zhao C, Sun X, Sun S, Jiang T (2011) Fault diagnosis of sensor by chaos particle swarm
optimization algorithm and support vector machine. Expert Syst Appl 38(8):9908–9912
22. Yu CB, Hu JJ, Li R, Deng SH, Yang RM (2014) Node fault diagnosis in WSN based on RS and
SVM. In: Proceedings 2014 international conference on wireless communication and sensor
network (WCSN). Wuhan, China, 12/13/2014
23. Balaban E, Saxena A, Bansal P, Goebel KF, Curran S (2009) Modeling, detection, and
disambiguation of sensor faults for aerospace applications. IEEE Sens J 9(12):1907–1917
24. Huang H-B, Yi T-H, Li H-N (2017) Bayesian combination of weighted principal-component
analysis for diagnosing sensor faults in structural monitoring systems. J Eng Mech
143(9):04017088
25. Dragos K, Jahr K, Smarsly K (2016) Nonlinear sensor fault diagnosis in wireless sensor
networks using structural response data. In: Proceedings of the 23rd international workshop
of the European group for intelligent computing in engineering (EG-ICE). Krakow, Poland,
06/29/2016
26. Li L, Liu G, Zhang L, Li Q (2019) Sensor fault detection with generalized likelihood ratio and
correlation coefficient for bridge SHM. J Sound Vib 442:445–458
27. Webster JG (1999) The measurement, instrumentation, and sensors handbook. CRC/IEEE
Press, Boca Raton
28. Teh HY, Kempa-Liehr AW, Wang KI-K (2020) Sensor data quality: a systematic review. J Big
Data 7(11):1–49
29. Ternan JG (1983) Thermoelectric drift of thermocouples due to inhomogeneous changes in
composition. J Appl Phys 55:199–209
30. Hamada T, Suyama Y (2004) EMF drift and inhomogeneity of type K thermocouples: In: SICE
2004 annual conference. Sapporo, Japan, 08/06/2004
31. Wang D, Fotinich Y, Carman GP (1998) Influence of temperature on the electromechanical
and fatigue behavior of piezoelectric ceramics. J Appl Phys 83(10):5342–5350
32. Tong L, Sun D, Atluri SN (2001) Sensing and actuating behaviours of piezoelectric layers with
debonding in smart beams. Smart Mater Struct 10:713–723
33. Li C, Weng GJ (2002) Antiplane crack problem in functionally graded piezoelectric materials.
J Appl Mech 69(4):481–488
34. Hawkins DM (1980) Identification of outliers. Springer, Dordrecht
164 H. Fritz et al.
35. Dervilis N, Antoniadou I, Barthorpe RJ, Cross EJ, Worden E (2015) Robust methods for outlier
detection and regression for SHM applications. Int J Sustain Mater Struct Syst 2(1/2):78354
36. Fu Y, Peng C, Gomez F, Narazaki Y, Spencer BF (2019) Sensor fault management techniques
for wireless smart sensor networks in structural health monitoring. Struct Control Health Monit
26(7):e2362
37. Russel SJ, Norvig P (2014) Artificial intelligence: a modern approach, 3rd edn. Pearson
Education, New Jersey
38. Najmi AH, Sadowsky J (1997) The continuous wavelet transform and variable resolution
time-frequency analysis. J Hopkins APL Tech Dig 18(1):134–140
39. Kanevski M, Pozdnoukhov A, Timonin V (2009) Machine learning for spatial environmental
data. Theory, applications and software. EPFL Press, Lausanne
40. Netter J, Kutner MH, Nachtsheim CJ, Wasserman W (1996) Applied linear statistical models.
Irwin Professional Publishing, Burr Ridge
41. LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document
recognition. Proc IEEE 86(11):2278–2324
42. Mallat S (2016) Understanding deep convolutional networks. Phil Trans R Soc A 374:20150203
43. Bauhaus-Universität Weimar (2020) Messdaten eines Monitoringsystems einer Eisenbahn-
brücke (online). Available at: https://www.mcloud.de/zh/web/guest/suche/-/results/detail/F9B
FABA3-29A9-48A2-973B-E1C29E8555B0
44. Kingma DP, Ba J (2014) Adam: a method for stochastic optimization. In: 3rd International
conference for learning representations. San Diego, CA, USA, 05/07/2015
A Self-adaptive Hybrid
Model/data-Driven Approach to SHM
Based on Model Order Reduction
and Deep Learning
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 165
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_8
166 L. Rosafalco et al.
1 Introduction
Civil infrastructures are crucial in our digital and smart society. However, they are
exposed to aging and to a progressive accumulation of damage, also caused by
exceptional loading conditions. As a prompt detection of structural damage can
prevent catastrophic events, SHM is becoming more and more an active field of
research [8]. Global damage detection methods are looked for, to be robust against
operational (e.g., in terms of the load amplitude) and environmental (e.g., due to
varying thermal and hygrometric conditions) variability.
Customary vibration-based monitoring techniques rely on the fact that modal
parameters, and therefore, the overall dynamic response of a structure depends on
its mass and stiffness properties. Any change in their values due to the inception of
a structural damage accordingly affects the dynamic response of the structure [15].
Within such a perspective, through the SHM system the observations in time (e.g.,
acceleration recordings shaped as multivariate time series) recorded by a sensor net-
work are analyzed to extract some damage-sensitive features able to discriminate
virgin and damaged states. Signal processing is to be carried out via effective pro-
cedures, able to handle the raw vibration signals, and retrieve the aforementioned
damage-sensitive features (from now on simply referred to as features), to allow the
SHM procedure to detect damage.
Local damage detection methods, based on visual inspections or non-destructive
testings, are widespread in civil engineering; however, they require some a-priori
knowledge on the position of damage, so a proper expertise [12]. Thanks to the recent
advancements in sensor technology, global monitoring techniques based on contin-
uous vibration measurements are now more often used. Among the vibration-based
techniques, two main approaches can be distinguished: the model-based approach
and the data-driven approach. Within a model-based approach, the discrepancy
between data and model output is minimized via a model update procedure, e.g., by
means of Kalman filters [9, 10]. Exploiting the measurements from real structures,
damage localization, and quantification can be achieved. However, a model-based
approach may be affected in its accuracy and efficacy when a large amount of noisy
data has to be processed. Moreover, the associated inverse problem to be solved for
damage identification is usually ill-posed. On the other hand, a data-driven approach
[11] does not typically rely upon any physics-informed model. Because of its ability
to also handle big data acquired on-the-fly through deployed sensors, this approach is
increasingly attracting interest. Within a standard implementation, damage-sensitive
features are extracted from raw vibration signals by exploiting their statistical char-
acteristics. According to the statistical pattern recognition paradigm [12], the SHM
task can be arranged into four sequential steps: operational evaluation, data acquisi-
A Self-adaptive Hybrid Model/Data-Driven Approach … 167
tion, features extraction, and statistical modeling for features discrimination. Since
civil structures are always subject to varying operational and environmental condi-
tions, potentially hampering the monitoring task, the SHM strategy should account
for such variability within the dataset or exploit data normalization techniques [22]
to distinguish the effects of the aforementioned variability from those due to damage.
The pattern recognition paradigm is well suited for implementation through
machine learning algorithms. Indeed, machine learning allows to statistically handle
data by learning the functional link between the damage-sensitive features and the
structural states. Recently, DL algorithms have been allowed to further empower the
approach; thanks to DL, it is possible to accomplish features extraction and discrim-
ination in one single shot. DL algorithms can deal with high dimensional data to
automatically catch temporal and spatial correlations, within and across time record-
ings. This allows to extract and exploit damage-sensitive features in an end-to-end
efficient way.
According to Rytter [21], damage identification consists of four levels of increas-
ing complexity: damage detection, localization, quantification, and prognosis. Data-
driven approaches can work within a supervised [25, 27] or unsupervised [14, 17]
setting. Unsupervised methods exploit unlabeled data relevant to a reference condi-
tion, often the damage-free baseline, to test the current state and basically accomplish
early damage detection. Vice versa, supervised methods exploit labeled data related to
both undamaged and damage scenarios, to be possibly undergone by the structure, and
are therefore well suited to accomplish also damage localization and quantification.
When dealing with civil structures, data related to damage conditions are often
unavailable. To solve this drawback, a commonly adopted paradigm is the simulation-
based classification [1, 20, 24], which aims to fuse the model-based and the data-
driven approaches into a hybrid one. With this novel paradigm, real data are replaced
by the output of numerical experiments where the effect of damage on the structural
response is simulated through a physics-based model of the structure. Accordingly,
the inception and subsequent growth of damage can be framed in a way similar to
model update, wherein the tuning/training of the model itself is not carried out to
allow for epistemic uncertainties, but to track the time-varying health of the structure.
In this work, we propose a NN-based, mixed offline-online SHM procedure aiming
to first detect and localize damage, and then quantify it in a dynamic environment,
so as to promptly react in case of detection of any variation of the structural state.
The remainder of this chapter is organized as follows. In Sect. 2, the proposed
procedure is detailed to highlight the features of the mixed offline–online strategy,
together with the potentiality offered by transfer learning (TL). In Sect. 3, dataset
assembling is addressed together with the adopted MOR strategy. In Sect. 4, two
NN architectures are discussed for, respectively, simultaneous damage detection and
localization, and damage quantification. The capability of the proposed method to
track damage evolution is assessed in Sect. 5. In Sect. 6, conclusions on the present
work and future developments are finally discussed.
168 L. Rosafalco et al.
2 Monitoring Procedure
Model update
FOM analysis
Online phase
ROM construction
Experimental data acquistion
NN training
No Damage
inception/growth
Yes
NN testing
Damage quantification
The two NNs share the same fully convolutional network (FCN) architecture,
already successfully adopted, e.g., in [18, 20, 25]. This class of NNs has been
reported to be powerful in detecting hierarchical patterns in the data, being also
computationally efficient thanks to their shared-weights architecture. By adopting
the same architecture for the classification and the regression tasks, TL [4, 7, 13] can
be exploited too. TL takes advantage of the knowledge gained during the training
of a NN, to train another NN aimed at solving a different task. The present scenario
is known as inductive transfer learning, which assumes that labeled data for related
but different tasks are available. This framework allows reducing both the training
time and the amount of training data, without compromising the performance of the
procedure.
3 Numerical Modeling
In the following, the procedure to build a FOM of civil structures or buildings is first
described; then, we specify the features of the dataset used to train the NNs; finally,
we discuss how to speedup the dataset construction, exploiting a MOR technique.
170 L. Rosafalco et al.
where t ∈ (0, Tf ) denotes time; v = v(t) ∈ RM is the nodal displacement vector, with
M denoting the number of degrees of freedom (dofs); v̈ is the vector of nodal accel-
erations; M ∈ RM ×M is the mass matrix; K(g, δ) ∈ RM ×M is the stiffness matrix,
with g and δ the parameters providing its dependence on damage as specified below;
f(η) ∈ RM is the vector of external loads; η is the vector of parameters ruling the
operational conditions; v0 and v̇0 are the initial conditions, in terms of nodal dis-
placements and velocities at t = 0. Due to the small relevance of damping in the
identification of continuously excited systems [6], dissipation effects have been dis-
regarded.
Damage is modeled as a local stiffness reduction, assumed frozen within the time
interval (0, Tf ) of interest. This assumption involves a timescale separation between
damage growth and health assessment; the considered scenarios are thus character-
ized by a low damage evolution rate and allow keeping the linear structural response
valid. According with the classification framework, only a finite set of damage states
are considered: those have to be preliminary determined in order to allow covering
relevant structural failure modes. For the problem at hand, G damage conditions
g ∈ {1, . . . , G} have been then accounted for, in addition to the undamaged one
g = 0. Each condition is characterized by a damage in the corresponding subdomain
g , wherein the stiffness reduction has a time-invariant magnitude taking value in a
specified range for δ.
Synthetic vibration recordings can be built starting from the dynamic response of the
model through nodal displacements mimicking the SHM sensing network output.
Regarding the optimal location of sensors, even in the present pseudo-experimental
setting, to understand how to maximize the sensitivity of measurements to the damage
scenarios to be detected, readers are referred to [5]. The monitoring system is assumed
to consist of Nu sensors recording the structural displacements un (t), n = 1, . . . , Nu .
Each recording un (t) consists on its own of LM data points, which are gathered
in the matrix U(η, g, δ) = [u1 , . . . , uNu ] ∈ RLM ×Nu . Next, data are corrupted with a
A Self-adaptive Hybrid Model/Data-Driven Approach … 171
Gaussian noise featuring a specific signal-to-noise ratio, in order to mimic the sensor
self-noise that inevitably affects the experimental data.
The dataset, denoted by D, is built by assembling I instances according to:
⎧ ⎫
⎨U1 (η1 , g1 , δ1 ) , UI (ηI , gI , δI )⎬
D= g1 , ... gI . (2)
⎩ ⎭
δ1 , δI
To speedup the generation of D, the number of dofs in Eq. (1) can be reduced through
MOR techniques for parametrized problems [1, 20]. The reduced basis method is
here adopted to also control the approximation error.
The FOM solution v = v(t, η, g, δ) is approximated as v ≈ WvR , by linearly
combining W M basis functions ww ∈ RM , w = 1, . . . , W collected into W =
[w1 , . . . , wW ] ∈ RM ×W , where W has to be set to attain the target accuracy. vR =
vR (t, η, g, δ) thus becomes the vector gathering the ROM dofs.
To build W, the proper orthogonal decomposition (POD), see, e.g., [16], is
adopted. POD provides the projection bases via a singular value decomposition of
the snapshot matrix S = [v1 , . . . , vS ] ∈ RM ×S , collecting S snapshots of the FOM,
according to:
S = PZ , (3)
By retaining the first W ≤ r left singular vectors, the POD bases W = [p1 , . . . , pW ]
are obtained. The dimension W can be set by prescribing the reconstruction error in
reducing the order of the problem from M to W , to be smaller than a tolerance
according to:
W
l=1 (σl )
2
r ≥ . (4)
l=1 (σl )
2
where
Once integrated in time, the reduced order solution VR = [v1R , . . . , vLRM ] ∈ RW ×LM ,
can be back-projected to recover the FOM solution at each time instant.
4 Deep Learning
In this section, we aim at detailing the two NNs used to perform the damage detection
and localization first, and the damage quantification next. We describe how TL can be
exploited to speedup and also improve the training of the NN adopted in the second
stage of the procedure.
The two NNs aim at approximating a target function, in our case linked to the
structural response to the external loading, respectively, mapping U onto the discrete
set g of damage patterns for the joint detection and localization task (working as a
classification model), and onto the damage level δ for the quantification task (working
as a regression model).
between ϕbg ∈ {0, 1}, stating if the g-th damage class has to be associated with the b-
th instance Ub of the processed mini-batch, and ψbg ∈ [0, 1], providing the confidence
by which the b-th instance Ub is associated with g. The values of ϕbg and ψg are
collected in the matrices ∈ {0, 1}B×(G+1) and ∈ [0, 1]B×(G+1) , respectively.
In Fig. 2, a schematic representation of the classification model is reported. To
compute , B input instances are processed by three convolutional units (CUs),
featuring, respectively, N1 = 32, N2 = 64 and N3 = 32 kernel filters, followed by a
global average pooling (GAP) layer and by a linear projection whose output is the
score matrix = (ϑ 0 , . . . , ϑ G ) ∈ RB×(G+1) . is next handled as the argument of
a softmax function in the following form:
eϑbg
ψbg = (8)
G ϑbg
g=0 e
to finally obtain the matrix associated with the input mini-batch. Each CU consists
of a convolutional layer L, a rectified linear unit (ReLU) activation layer, and batch
normalization (BN) layer. Readers are referred to, e.g., [18, 20, 25], for further details
on how CUs and GAP perform signal processing and feature extraction.
174 L. Rosafalco et al.
B
1
Cr (δ r , δ) = (δrb − δb )2 , (9)
B
b=1
where δ = [δ1 , . . . , δB ] ∈ RB are the target damage levels associated with the pro-
cessed mini-batch instances.
It is to note that the information regarding damage localization is not allowed for
by the regression model. Running several tests, we have figured out that this choice
has no impact on the procedure performance in the present case; anyhow, we cannot
exclude that, for more complex applications, allowing for the outcome of damage
localization as a further input to the damage quantification stage may be beneficial.
A Self-adaptive Hybrid Model/Data-Driven Approach … 175
The weights of the regression model can be randomly initialized or, alternatively,
their initialization can account for the weights of the trained classification model. In
this way, the regression model can exploit the signal processing and feature extraction
capacity of the classification model. Indeed, features are synthetic descriptions of
data obtained for a specific task, but they may be also adopted for other tasks if there
is redundancy across the learnt and new tasks, see [26].
Two benefits are expected by the use of TL: speeding up of training; reduction
of the data required for training. All these outcomes will be investigated next with
reference to a specific case test.
5 Results
To assess the capability and the performance of the proposed SHM procedure, we
apply here our strategy to the case study as shown in Fig. 4 and referred to a two-
dimensional portal frame.
The adopted FE mesh shown in Fig. 4a contains 1884 dofs. The frame is perfectly
clamped at the bases and has a thickness of 0.1 m. It is assumed to be made of con-
crete, whose mechanical properties are: Young’s modulus E = 30 GPa, Poisson’s
ratio ν = 0.2, and density ρ = 2500 kg/m3 . The structure is excited by seven differ-
ent loading conditions C ∈ {1, . . . , 7}, obtained by combining the three distributed
loads qk , k = 1, 2, 3 as shown in Fig. 4b. Each load qk varies in time according to
qk (t) = Qk sin (2π fk t), where Qk and fk are the relevant amplitude and frequency
that vary in the ranges [10−50 kPa] and [50−95 Hz], respectively. These parame-
ters, which embody the variability of the operational conditions, are collected in the
vector η = {C, Q1 , Q2 , Q3 , f1 , f2 , f3 } .
The G = 4 damageable regions g are shaded in Fig. 4b. The considered dam-
age scenarios are built by assuming that stiffness reduction can vary in each domain
within the range δ ∈ [10−25%]. In the same Fig. 4b, also the Nu = 7 sensed dis-
placements un (t) are shown. The sensors are scheduled to gather the measurements
with a sampling frequency of 200 Hz; this allows avoiding aliasing regarding the first
7 vibration modes of the structure. The signal-to-noise ratio of such measurements
is assumed to be equal to 80.
Y = 400 samples of {η, g, δ} have been adopted to assemble S; for each sample,
the solutions at X = 121 time instants have been collected. Therefore, a total of S =
X × Y = 48,400 snapshots are handled. By prescribing an error tolerance = 10−3 ,
W = 70 POD bases get selected, to replace the original 1884 dofs of the FOM: the
first ten of such bases are sketched in Fig. 5.
As far as damage detection and localization are concerned, the evolution of the loss
function Cc and the prediction accuracy during the training of the NN are reported
in Fig. 6 against the number of epochs, both for training and validation. The training
176 L. Rosafalco et al.
is stopped either when a maximum number of epochs has been attained, or when the
loss Cc does not decrease during 15 epochs in a row. The loss function plot shows
that the minimization of Cc is successfully carried out; this assures the increasing
accuracy of classification, which represents the percentage of instances for which
the damage is correctly detected and localized. The training has been early stopped
and overfitting is shown to be avoided, since the outcomes regarding the training and
the validation datasets are very similar. Once trained, the classification model has
been tested by employing another dataset collecting pseudo-experimental instances
simulated through the FOM. The relevant classification outcome is reported in the
confusion matrix of Fig. 7, which shows an overall accuracy of around 99%, with
just a few misclassification errors. The mentioned misclassification errors affect the
damage localization task, while damage detection, to distinguish the undamaged
state g = 0 from all the other scenarios, is perfectly accomplished.
As far as damage quantification is instead concerned, Fig. 8 provides the evolution
of the loss Cr against the training epochs. The training has been early stopped after
5 epochs without a decrease of the loss on the validation set. The plot highlights a
consistent reduction of such loss, especially in the first stage of training. By com-
paring the evolutions relevant to the training and validation sets, it can be stated that
overfitting is avoided again.
The trained regression model has been next adopted to perform damage quantifi-
cation on a test set provided as before by the FOM. The relevant outcomes are shown
in Fig. 9, in terms of a parity plot to compare the NN predictions (reported along
the vertical axis) and the target damage levels (reported instead along the horizontal
axis), and in terms of the corresponding prediction errors. The reported values of the
damage level are distributed in a rather narrow range around the line bisecting the
quadrant, to testify the NN generalization capacity for unseen instances generated
by the FOM. In the bar chart, the counts are proportional to the number of instances
1.00 m
q2(t) q3(t) 0.75 m u3(t) 0.75 m
0.50 m
1.00 m
Ω2 u6(t) u7(t) Ω3
q1(t)
1.00 m 6.00 m
u1(t) u2(t)
u4(t) u5(t)
Ω1 Ω4
1.00 m
0.25 m
5.50 m
(a) (b)
Fig. 4 Portal frame. a Geometry, FE discretization and applied loads; b subdomains {1 , . . . , 4 }
where damage may occur, and sensed displacement components.
A Self-adaptive Hybrid Model/Data-Driven Approach … 177
(a)
(b)
178 L. Rosafalco et al.
Output Class
1 0 77 0 0 0
0.0% 0.0% 97.5% 0.0% 0.0%
2 0 0 78 0 0
0.0% 2.5% 2.5% 100.0% 0.0%
3 0 2 2 80 0
0.0% 1.2% 0.0% 0.0% 100.0%
4 0 1 0 0 80
0 1 2 3 4
Target Class
for which a certain prediction error has been observed a-posteriori; a Gaussian-like
distribution of the mentioned prediction error can be seen. This outcome is actually
connected to the definition of Cr : when the mse is employed as the loss function,
the regression model tries to predict the expected value E [δb |Ub ] of the sequence δb ,
with b = 1, . . . , B, conditioned on Ub , see [3]; due to the central limit theorem, the
distribution asymptotically tends to a Gaussian one if the number of samples grows
to infinity. Accordingly, the prediction errors tend to inherit a Gaussian distribution
too.
Regarding the exploitation of TL, from the results reported in Fig. 10 it is possible
to appreciate the induced speedup of the learning process: the number of training
epochs turns out to be more than halved than that necessary for the case without
TL, as shown in Fig. 8. The reduced number of epochs is obtained together with a
reduction of the training time, as reported in Fig. 11. In the same plot, the training
time required when the training dataset is halved is also reported; by halving the
dimension of D, we aimed at assessing if TL may enable the use of a smaller dataset,
assuring the same or similar performances of the NN. TL is expected to enhance the
NN performance because it eases the training procedure, via a better initialization
A Self-adaptive Hybrid Model/Data-Driven Approach … 179
(a) (b)
Fig. 9 Regression outcomes in terms of damage level: a Parity plot comparing the predicted values
to the ground truth ones; b histogram of the relevant prediction errors
of the weights, and therefore allows obtaining optimal results even if D is narrowed.
The performance improvement can be qualitatively assessed through Fig. 12, and
specifically comparing the prediction error distribution with that reported in Fig. 9:
the error results to be smaller if TL is employed.
A quantitative assessment of the effect of TL on the regression performance is
reported in Table1, by adopting the mean absolute error (mae) on the test set as the
evaluation metric. TL leads to an improvement of the regression performance when
the entire dataset D is considered. The same improvement has not been obtained when
the dimension of D has been halved, as an almost identical mae value is obtained for
the analyses with and without TL. On the other hand, in all the cases TL provides
a considerable speedup of training, with a reduction of the number of epochs from
46 to 21 for the entire dataset, and from 79 to 25 for the halved one. In this table,
the mean μ and standard deviation ς of the prediction error, as evaluated on the
180 L. Rosafalco et al.
Fig. 11 Comparison among the training times of the three adopted models, at varying number of
instances in the processed dataset
(a) (b)
Fig. 12 Regression outcomes when TL is exploited, in terms of damage level: a Parity plot com-
paring the predicted values to the ground truth ones; b histogram of the relevant prediction errors
test set, are also reported. Having adopted the mse as loss function, these statistic
features are employed to assess the performance of the regression model, as they
fully characterize a Gaussian distribution. The error mean μ shows a small bias in
the prediction of δ, probably because the solution has got entrapped into a local
minimum during the optimization procedure. The greater accuracy of the regression
model trained by exploiting TL is highlighted by the smaller values of ς in both the
analyses.
The proposed classification and regression models are finally adopted to track
the evolution of the damage level δ, under the loading conditions detailed in Table2.
In the same Table, also the time-evolving value of δ to be identified on-the-fly is
reported. During the analysis, each time damage is detected and localized by the
classification model, and the regression model is plugged in to predict its magnitude.
A Self-adaptive Hybrid Model/Data-Driven Approach … 181
Table 1 Regression model: comparison of the performances if TL is exploited or not, and if dataset
is reduced in size, in terms of training time, generalization capability, and mean μ and standard
deviation ς of the prediction error
Model Dataset (%) Training Mae (test set) μ ς
epochs
Regression 100 46 0.80 0.26 0.98
Regression 50 79 0.79 0.14 1.06
Regression (TL) 100 21 0.70 0.22 0.92
Regression (TL) 50 25 0.81 −0.27 1.02
Table 2 Data regarding the damage level and damage class, and output of the evolutionary clas-
sification model. The considered test case features: g = 1, Q1 = Q3 = 0, Q2 = 20 kPa, f2 = 60
Hz
Damage level (%) Output class Target class
5 0 1
6 0 1
7 0 1
8 0 1
9 1 1
10 1 1
.. .. ..
. . .
25 1 1
At this point, the dataset D is updated by dropping the instances featuring a value
of δ in the domains where damage has been localized, smaller than the identified
one, and by adding new instances featuring a value of δ larger than or equal to the
previously mentioned one. Within the present approach, those instances are provided
by the FOM to keep the accuracy of the online identification task the highest possible
and avoid a bias in the estimations.
Regarding the time necessary to update the baseline, generate new data and re-
train the NNs every time damage grows, some results are provided next as obtained
with a laptop featuring an Intel Core i5 CPU @ 2.6 GHz and 8 GB RAM. The
CPU time required by each ROM analysis is 0.2 s, and by each FOM analysis is
1.3 s, respectively. The duration of the subsequent training of the classifier amounts
to around 10 min, while that of the regressor amounts to 4 min if TL is exploited.
Overall, the update stage requires less than 1 h of outage for the SHM procedure. If
the FOM analyses become too time-taking for baseline update to be handled close
to online, techniques for the dynamic update of the ROM bases may be exploited.
This duration can be obviously (by far) reduced if a more powerful computer is used
to run the procedure.
182 L. Rosafalco et al.
As shown in Fig. 13, the proposed procedure is able to self-adapt and to track
damage evolution. For the sake of brevity, results are shown for this damage case
only, but the present framework can be successfully applied also to other cases
characterized by different damage patterns.
6 Conclusions
the accuracy of the NN-based approach and allow reducing the amount of data
required to attain the sought accuracy. The environmental variability is going to be
also allowed for. The procedure will be then applied to real experimental settings.
References
16. Quarteroni A, Manzoni A, Negri F (2015) Reduced basis methods for partial differential equa-
tions: an introduction. Unitext, vol 92. Springer, Berlin
17. Rafiei MH, Adeli H (2018) A novel unsupervised deep learning model for global and local
health condition assessment of structures. Eng Struct 156:598–607. https://doi.org/10.1016/j.
engstruct.2017.10.070
18. Rosafalco L, Manzoni A, Mariani S, Corigliano A (2020) Fully convolutional networks for
structural health monitoring through multivariate time series classification. Adv Model Simul
Eng Scie 7:38. https://doi.org/10.1186/s40323-020-00174-1
19. Rosafalco L, Manzoni A, Mariani S, Corigliano A (2021) An autoencoder-based deep learning
approach for load identification in structural dynamics. Submitted
20. Rosafalco L, Torzoni M, Manzoni A, Mariani S, Corigliano A (2021) Online structural health
monitoring by model order reduction and deep learning algorithms. Submitted
21. Rytter A (1993) Vibrational based inspection of civil engineering structures. Ph.D. thesis,
University of Aalborg, Denmark
22. Sohn H, Worden K, Farrar CR (2002) Statistical damage classification under changing envi-
ronmental and operational conditions. J Intell Mater Syst Struct 13(9):561–574. https://doi.
org/10.1106/104538902030904
23. Sudret B, Defaux G, Pendola M (2007) Stochastic evaluation of the damage length in rc beams
submitted to corrosion of reinforcing steel. Civ Eng Environ Syst 24(2):165–178. https://doi.
org/10.1080/10286600601159305
24. Taddei T, Penn J, Yano M, Patera A (2018) Simulation-based classification; a model-order-
reduction approach for structural health monitoring. Arch Comput Methods Eng 25(1):23–45
25. Torzoni M, Rosafalco L, Manzoni A (2020) A combined model-order reduction and deep
learning approach for structural health monitoring under varying operational and environmental
conditions. Eng Proc 2(1). https://doi.org/10.3390/ecsa-7-08258
26. Zamir AR, Sax A, Shen W, Guibas LJ, Malik J, Savarese S (2018) Taskonomy: disentangling
task transfer learning. In: 2018 IEEE/CVF conference on computer vision and pattern recog-
nition, 18 June–23 June, pp 3712–3722. Salt Lake City, UT. https://doi.org/10.1109/CVPR.
2018.00391
27. Zhang T, Biswal S, Wang Y (2020) Shmnet: condition assessment of bolted connection with
beyond human-level performance. Struct Health Monit 19(4):1188–1201. https://doi.org/10.
1177/1475921719881237
Predictive Monitoring of Large-Scale
Engineering Assets Using Machine
Learning Techniques and
Reduced-Order Modeling
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 185
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_9
186 C. Bigoni et al.
1 Introduction
Many existing private and public assets, such as civil engineering infrastructures,
buildings, or aircraft, require reliable damage detection techniques to be safely used,
especially during their inevitable aging. When monitoring a structure over its lifecy-
cle, its deterioration and damages represent a great concern and the early detection of
critical decay might prevent failures that can cause sudden shutdowns or even catas-
trophes with severe life-safety and economic repercussions [1]. To prevent these
critical failures from happening, techniques of structural health monitoring (SHM)
have been developed in recent studies with applications to civil and aerospace engi-
neering, as well as to the conservation of cultural heritage structures. SHM refers
to automated monitoring procedures that seek to provide reliable information on the
performance and integrity of a structure in real time. In the context of SHM, the
combination of sensor measurements, numerical models simulating the underlying
behavior of a structure of interest under different environmental and operational con-
ditions, and machine learning techniques has led to the design of structural digital
twins [2].
The focus of this work is on wave propagation approaches to data-driven, predic-
tive SHM, which aim to detect damages by examining the distortions in propagating
elastic waves as a result of reflections and amplitude attenuations when intersecting
the damage boundary. Featuring a data-driven nature, these approaches, sometimes
called simulation-based SHM [3–7], are decomposed into an offline phase and an
online phase. In the former, a database of synthetic signals is built to represent the
structural behavior under different conditions, while in the latter real experimen-
tal time signals, collected from sensors placed on a structure, are compared with
those simulated offline using a classifier that discriminates between damaged and
undamaged states.
Toward an efficient and robust scheme of data-driven predictive monitoring,
reduced-order modeling techniques are integrated into a wave propagation approach
with cutting-edge machine learning tools. The data-driven SHM setting corresponds
to a multiquery problem, where one has to solve high-dimensional, time-dependent,
parametric equations, which results in great demands for computational resources.
To overcome such a computational burden, model order reduction is employed to
project the original full-order system onto a reduced space with a significantly lower
dimensionality. In this way, a robust dataset of approximated sensor measurements
is generated. In SHM, the task of damage detection is typically reduced to a super-
vised learning process relying on a fully labeled dataset, obtained from both healthy
and damaged structures (either generated with computer-aided procedures or col-
lected experimentally). However, gathering an exhaustive collection of configuration
classes anticipating all types of damages is typically unrealistic and the number of
different classification labels may grow rapidly. Instead, this work relies on semi-
supervised learning techniques, also called one-class classification methods, which
learn the common features among labeled data belonging to the normal class in the
training phase. Unlabeled data from both classes are then used in the test phase to
Predictive Monitoring of Large-Scale Engineering Assets … 187
identify abnormal data which deviate from the normal mode. One-class algorithms
allow to locate the damage and estimate its severity by training a different model
for each sensor location. Finally, guided by an appropriate indicator of the damage
detection performance, a modified sparse Gaussian process method is applied to the
synthetic dataset of healthy configurations to systematically place a fixed number of
sensors on a structure of interest.
Following the introduction, a model problem for sensor measurements governed
by the acoustic-elastic equation is briefly reviewed in Sect. 2. Basic techniques of
reduced-order modeling, which can be used for the multiquery simulations in SHM,
are introduced in Sect. 3. A local semi-supervised method is used for automatic
anomaly detection in Sect. 4, and the variational sparse Gaussian process model is
utilized for optimal sensor placement in Sect. 5. In Sect. 6, the all-round methodology
is demonstrated by a numerical example. Finally, conclusions are drawn in Sect. 7.
This section introduces the model problem for sensor measurements, including its
governing equations and parametric discrete formulations.
Let ⊂ Rd be a polygonal physical domain with piece-wise smooth boundary ∂,
where d = 2, 3 is the spatial dimension, and let [0, T ], with T ∈ R+ , be a suitable
time interval. Here, represents a structure of interest and [0, T ] a suitable time
window to observe the response of a structure undergoing a predefined excitation (i.e.,
the effect of an active or passive source) through sensor measurements. Moreover,
let P ⊂ Rdμ be a suitable parameter domain, where dμ indicates the number of input
parameters required to describe the healthy variations that a structure may undergo
during its life time, and let μ = (μ1 , . . . , μdμ ) be a parameter vector representing one
possible healthy variation of the environmental and operational conditions, i.e., μi
may relate to the material properties, the boundary conditions, the initial conditions,
or the source function for 1 ≤ i ≤ dμ .
Let u = u(x, t; μ) : × [0, T ] × P → Rd be a vector-valued displacement
field, solution of the acoustic-elastic equation equipped with suitable boundary and
initial conditions:
⎧
⎪
⎨ρ(ü + η u̇) − ∇ · σ = h(t; μ)b(x; μ) in
u · n = 0, (σ · n) · τ = t N on ∂ , (1)
⎪
⎩
u|t=0 = u0 , u̇|t=0 = v 0 in
188 C. Bigoni et al.
where ü = ∂ 2 u/∂t 2 and u̇ = ∂ u/∂t are the acceleration and velocity fields. Here, ρ is
the density coefficient, η is a dimensionless damping coefficient,
h : [0, T ] × P → R and b : × P → Rd are two source functions depending on
time and space, respectively, σ = σ (u; μ) is the stress tensor, n and τ are the outward
normal and tangential (unit) vectors to ∂, respectively, t N = t N (x, t; μ) is the trac-
tion vector used in the definition of the free-slip boundary conditions, u0 = u0 (x; μ)
and v 0 = v 0 (x; μ) describe the initial displacement and velocity in space, respec-
tively.
The ultimate goal is to emulate the real sensor response at m given sensor locations
x i ∈ for 1 ≤ i ≤ m. To do so, let : Rd × P → Rdg be an input–output function
and gi : [0, T ] × P → Rdg a (parametric) output of interest, i.e., an approximation
of the sensor response at time t and location x i :
Before proceeding with classic discretization techniques such as the finite ele-
ment method, consider the vector space V = {w ∈ [H 1 ()]d : w · n = 0 on ∂},
equipped with a suitable inner product ·, ·V and the corresponding induced norm
· V . Moreover, consider a parametrized linear form f : V × P → R where the
linearity is with respect to the first argument, and the parametrized bilinear forms
m : V × V → R and a : V × V × P → R, where the bilinearity is with respect to
the first two arguments. Then, the acoustic-elastic problem in abstract form reads:
given t ∈ [0, T ] and μ ∈ P, find u = u(t; μ) ∈ V such that
ρ m(ü, ψ) + η m(u̇, ψ) + a(u, ψ; μ) = h(t; μ) f (ψ; μ) ∀ψ ∈ V,
(3)
u(t = 0; μ) = 0, u̇(t = 0; μ) = 0,
with
m(u, ψ) = u · ψ , f (ψ; μ) = b(μ) · ψ ,
(4)
a(u, ψ; μ) = [H(μ) : ε(u)] : ε(ψ) ,
in which H is the Hooke’s stiffness tensor and ε(·) = 21 [∇ · +(∇·)T ] is the operator
of Cauchy strain. Note that, for the sake of simplicity, homogeneous boundary and
initial conditions in (3) are considered, i.e., u0 = 0, v 0 = 0, and t N = 0.
where M h ∈ R Nh ×Nh is the mass matrix, Ah (μ) ∈ R Nh ×Nh the parametrized stiffness
matrix, and f h (μ) ∈ R Nh the parametrized right-hand side vector with entries
A classic Newmark method is then applied to the temporal discretization and the
governing equation becomes: given μ ∈ P, find the acceleration vector anh (μ) ∈ R Nh
for n = 1, . . . , Nt such that
Finally, the displacement solution vector unh (μ) is obtained using the updating rule
of the implicit Newmark method, introduced in [10] and defined as:
190 C. Bigoni et al.
unh (μ) = un−1
h (μ) + tv h (μ) β a h (μ) + 2 a h (μ)
n−1 n 1−2β n−1
(9)
n
v nh (μ) = v n−1
h (μ) + t ζ a h (μ) + (1 − ζ )a h (μ) ,
n−1
(10)
Problem (7) is denoted as the truth problem and unh (μ) as the truth solution at n-th
time step, which, in principle, can be achieved with as high accuracy as desired.
However, many degrees of freedom may be involved in the problem, thus leading to
a computationally expensive method due to inversion of the Nh -dimensional matrix
in the left-hand side of (7). In addition, to fully represent the healthy variations of
the structure, one needs to estimate an approximation to the output of interest (2)
for many input parameter values {μ1 , μ2 , . . . } over the whole discrete time window
0 = t 0 , t 1 , . . . , t Nt = T , i.e.,
g ik = gi (t 0 ; μk ), gi (t 1 ; μk ), . . . , gi (t Nt ; μk ) , k = 1, 2, . . . , (11)
evaluated at all the sensor locations x i , 1 ≤ i ≤ m. For each input parameter value μk ,
the total computational cost involves the resolution of Nt linear systems of dimension
Nh .
randomized SVD algorithms [15]. Recent efforts have been made in the direction
of space–time approaches, where projection in space and time is performed simul-
taneously, see, e.g., [16]. A different effective strategy is to replace the time domain
formulation with a frequency domain formulation and to apply a ROM method to
replace the full-order problem in frequency domain. In this way, the number of time
instances Nt where one expects to solve a linear system equivalent to (7) is reduced
to a number of principal frequencies Nz , with Nz Nt . In addition, recurring to
a ROM strategy reduces the number of degrees of freedom of each linear system
to the size of reduced basis, i.e., from Nh to r , with r Nh . Without going in too
much detail, the reader are referred to [5, 17], where the authors, motivated by the
interest of studying the transient response of damaged structures under the effect
active sources, construct a reduced-order model of the acoustic-elastic equation in
the Laplace domain.
The goal here is to provide a brief introduction to several basic elements of the RB
method, which lay the foundation for more advanced techniques. In particular, after
introducing a general formulation of the proper orthogonal decomposition (POD)
in Sect. 3.1, the construction of RB from full-order snapshots using the POD is also
described. The technique to retrieve the RB solution, an approximation of the high-
fidelity solution, is ultimately presented in Sect. 3.2.
C i j = pi , p j X , 1 ≤ i, j ≤ Ns . (12)
Ns
√
φi = p j z (i) j / λ(i) , 1 ≤ i ≤ r with N ≤ Ns , (14)
j=1
r
Pr [ f ] = arg min f − ξ 2
X = f, φi X φi , f ∈ X. (15)
ξ ∈Xr
i=1
It can be shown that the projection error of the snapshots only depends on the trun-
cated eigenvalues, written as
Ns
Ns
pi − Pr [ pi ] 2
X = λ(i) . (16)
i=1 i=r +1
This section presents a data-driven technique to detect, localize, and estimate the
severity of structural anomalies by observing healthy configurations only. One-class
classification learning methods offer the possibility of training a set of samples all
belonging to the same class and test if a new sample is abnormal, i.e., it belongs
to a different class with respect to the training data. Typical one-class classifica-
tion methods, sometimes called semi-supervised methods, include one-class support
194 C. Bigoni et al.
vector machines (oc-SVMs), isolation forests, and local outlier factors. During the
offline phase, these methods learn a description of the salient features that the train-
ing data have in common to ultimately detect if a previously unseen object reflects
this description by means of an online anomaly (or novelty) score. If the new unseen
sample is associated with an anomaly score close to the ones observed in the training
phase, the new object is classified as healthy, otherwise it is classified as damaged.
The crucial part is to define what close to means from a mathematical standpoint.
Let x be an unseen object, then the outcome of all one-class classification methods
can be summarized as follows:
score(x) ≥ θ damaged/outlier
, (20)
score(x) < θ healthy/inlier
where score(x) is the anomaly score associated with x and θ is an ad hoc threshold
to be estimated by observing the anomaly score value of healthy data only. From
a practical perspective, in the semi-supervised context, θ is heuristically chosen by
observing the highest anomaly score value in the training data, i.e., θ should be equal
to the anomaly score of the most outlier sample among all the inliers. Consider D
the dataset of healthy measurements and ε ∈ R, then the threshold value is fixed as
where a positive value of ε indicates the user accepts a higher false alarm rate, while
a negative value implies a higher miss detection rate. The trade-off between false
positive and false negative errors should guide the choice of ε and ultimately of θ .
It becomes clear that to choose an effective threshold value, the training set has to
be the most comprehensive as possible, covering several healthy environmental and
operational scenarios.
An alternative approach to detect anomalies is to include sensor measurements
belonging to damaged structures in the training set, which leads to using traditional
two-class supervised learning methods to distinguish healthy scenarios from dam-
aged ones. In this approach, the choice of the threshold value benefits form the
availability of two (or more) classes in the training phase. However, an increasing
trend toward the assumption that it would be unreasonable to describe all types of
damages is observed in the literature; as a consequence, representing only some
damaged configurations would lead to a bias toward certain types and therefore to
possible misdetections with high probability (see, e.g., [5, 24]). For this reason, in
this chapter, one-class classification methods are used instead.
The general one-class approach is introduced in Sect. 4.1 and explain the need for
feature selection in subsection 4.2. For a detailed description of the classic one-class
models, the reader is referred to [25, 26] for the oc-SVMs, [27] for the isolation forest
and [28] for the local outlier factor. A Python implementation of these methods can
be found for example in scikit-learn library [29].
Predictive Monitoring of Large-Scale Engineering Assets … 195
Considering that the time signals, i.e., the output of interests (2), are collected at mul-
tiple sensor locations, one has to decide on how to best aggregate these data. There
exists two typical approaches in the literature to combine sensor data: decision-level
fusion or feature-level fusion. The latter combines data after feature extraction and
considers one global classifier (sensor independent), thus exploiting the correlations
across sensors. On the other hand, in decision-level fusion, the signals are classified
for each sensor location by a local classifier (sensor dependent) and the results are
then combined into a decision output. The two strategies are summarized in Fig. 1.
While the superiority of one method over the other one depends strongly on the
problem at hand, to exploit the local aspect of the data the authors propose to use
the decision-level fusion approach which facilitates the use of a hierarchical classi-
fication approach where increasing levels of damage identification can be defined to
ultimately gain information on the existence, localization, and severity of the damage.
In a decision-level fusion approach, one has to train as many one-class algorithms
as the number of sensor locations. Thus, the global classification model (20) is
replaced with m local detection models, where m is the number of sensors:
scorei (x) ≥ θi damage in the proximity of the i th sensor,
, (22)
scorei (x) < θi health in the proximity of the i th sensor,
for 1 ≤ i ≤ m. From a computational cost point of view, note that the process can be
run in parallel since the local models are independent. Moreover, in the feature-based
fusion approach, aggregating the local features leads to high-dimensional input data,
while the dimensionality of the input data for the classifiers in the decision-based
Fig. 1 Flowchart to compare the feature-level (top) and the decision-level (bottom) fusion
approaches for the semi-supervised damage detection strategy with multidimensional training data
captured by m sensors
196 C. Bigoni et al.
strategy remains small. As further explained in Sect. 4.2, high-dimensional data may
lead to overfitting, a well-known problem in machine learning.
Overfitting refers to the phenomenon observed when the model performs well on
training data, but fails to generalize well to new observation. An overfitted model
is described by more parameters than can be justified by the data and is typically
associated with high-dimensional input data. While sometimes adding more training
samples may be a remedy to overfitting, a more effective solution to prevent this
behavior is to reduce the dimensionality of the data, i.e., to perform what is called
feature selection.
Recall that the input data correspond to Nt -dimensional sensor signals, with Nt the
number of time steps, which may be very large. Therefore, one needs to express these
high-dimensional time signals by means of few variables, extracted from the signals
themselves. Ideal features should be sensitive to damage and, at the same time, robust
toward noise and healthy variations. Common choices for the engineering-based,
damage-sensitive features can be found for example in [24, 30]. When studying
the acoustic-elastic equation, i.e., in the context of guided-waves, relevant features
are the crest factor, which indicates how extreme the peaks are in a waveform, the
maximum and minimum values of the time response, the corresponding arrival times,
i.e., the onset, and the number of peaks and troughs in the signals. Without further
detail, the reader is referred to [5] and references therein for a thorough description
of damage-sensitive features for a guided-wave monitoring approach.
Finally, note that autoencoders, a particular type of neural networks, trained to
attempt to copy their inputs to their outputs, have gained particular interest in the
framework of anomaly detection, see, e.g., [31–33]. The main advantage of using
autoencoders for anomaly detection is that specific engineering-based damage indi-
cator features do not need to be specified by the user, different from classic one-
class methods. Instead, by learning the features which suffice to describe and recon-
struct the input, autoencoders provide a purely data-driven feature extraction method.
Hence, raw measurements such as sensor time signals can be used directly.
Until this point, this work has covered the problem of how to best detect damages
given a fixed network of sensors. The specular research question is how to choose the
location (and the number) of sensors in order to best detect defaults. Sensor placement
strategies are extremely important to optimally equip structures, whose monitoring
Predictive Monitoring of Large-Scale Engineering Assets … 197
1 Here f denotes a generic regression function and should not be confused with the linear functional
f (·; μ) in Sect. 2.
198 C. Bigoni et al.
the former for a stationary kernel with isotropic lengthscale , while the latter for an
automatic relevance determination (ARD) kernel that considers an individual corre-
lated lengthscale k for each input dimension and allows for differentiated relevances
of input features to the regression.
Given a finite number of training input locations in the domain , a prior joint
Gaussian is defined for the corresponding outputs:
2Here r denotes a (scaled) radius and should not be confused with the reduced dimensionality in
Sect. 3.
Predictive Monitoring of Large-Scale Engineering Assets … 199
points, to facilitate the information gain of the whole dataset, thus allowing for a com-
plexity reduction, i.e., O(Mm 2 ). An overview of well-known sparse GP regression
methods can be found for example in [36], where each sparse method is described
as an exact inference with a specific approximated prior, different from the true GP
prior (23). A different approach is presented in [37], where both the m inducing
points, indicated as D = [d 1 |d 1 | · · · |d m ]T , and the hyperparameters θ are consid-
ered as variational parameters to be estimated by minimizing the Kullback–Leibler
(KL) divergence between the true posterior (26) and a variational posterior. This is
equivalent to maximize the following variational lower bound:
( D∗ , θ ∗ ) = arg max L( D, θ )
D,θ
1
= arg max log N ( y|0, Q( D, θ ) + χ 2 I M ) − Tr(κ(X, X) − Q( D, θ)) ,
D,θ 2χ 2
(28)
where Q = κ(X, D)(κ( D, D))−1 κ( D, X) is the Nystrom approximation of the true
prior covariance. Note that the trace term in (28) acts as a regularization term of the
marginal log likelihood, which can be viewed as an accuracy indicator of how well
the inducing points summarize the overall statistics.
where s ⊂ indicates the admissible domain for sensor locations and, with a
slight abuse of notation, D ∈ s means that each inducing point d i , 1 ≤ i ≤ m is
constrained to belong to s . For real-world problems, the complexity of the domain
may be such that the boundaries of s cannot be easily specified analytically and, in
such cases, it may be worth to replace s with a discrete counterpart. If that is the
case, instead of gradient-based optimization techniques, one could opt for discrete
optimization methods such as the genetic algorithm [38].
A second point to notice is that the output of interest y is in general parame-
ter dependent, i.e., y = y(μ). Hence, choosing the sensor locations as the optimal
inducing points obtained for one specific input configuration may not be optimal
for another context, described by a different parameter. To overcome this, this work
proposes to apply the variational sparse GP approach to Nμ outputs of interest y(μi ),
with μi ∈ P for 1 ≤ i ≤ Nμ . To summarize the information from the so-obtained
Nμ m inducing points, the K-medoids algorithm, a well-known unsupervised cluster-
ing technique, is employed to find m clusters and their corresponding centers, called
centroids. As a last step, the desired sensor locations will be chosen as the clusters’
centers.
To quantify the quality of the placements, the simplest choice is to compare
the relative reconstruction error of the high-fidelity quantity of interest at unsensed
locations with respect to the mean of the posterior distribution of the sparse model
based on the estimated variational parameters. Alternatively, the pointwise relative
variance reduction, defined as
6 Numerical Example
Fig. 2 Summary of the one-class SVMs classification results for one healthy geometry (a) and 5
damaged ones (b–f)
a total of Nh = 30 912 degrees of freedom, while for the RB solver the model
relies on 267 basis. The natural variations are described by dμ = 3 parameters, i.e.,
μ = (E, ν, k) ∈ P = [0.999, 1.001] × [0.329, 0.331] × [1.9, 2.1], where E is the
Young’s modulus and ν the Poisson’s ratio, defining the stress tensor σ , while k is a
parameter representing the number of cycles before attenuation of the source impulse.
The position of the active source as well as the density and damping coefficients are
fixed, i.e., ρ = 1, η = 0.1, respectively.
For each sensor location, one should follow the pipeline introduced in Sect. 4 and
generate a training dataset of Ntrain = 1000 samples, obtained by extracting a few
damage-sensitive features from the discrete time signal (11), in their turn obtained
with a suitable reduced-order modeling approach, as described in Sect. 3. Then,
one has to train m separate one-class SVM models to learn the common traits of
the local healthy features and test the results on sensor measurements belonging to
both healthy and damaged geometries. In this example, the test signals are generated
synthetically using the high-fidelity model. Different from the training set, to approx-
imate experimental measurements, a Gaussian noise term εi ∼ N (0, γi2 ) is added to
the test time signal (11), with γ being the 0.01% of the maximum amplitude of 30
randomly chosen training healthy signals in the training set. Damages are obtained
by modifying the geometry of the structure to include discontinuities, as shown in
Fig. 2b–f. The crack-modeling approach is a common choice in the literature, see,
e.g., [39] where artificial damages on the blade of a wind turbine are implemented
via a trailing edge opening. Anomaly detection results are shown in Fig. 2, where for
202 C. Bigoni et al.
each geometry the average outcome of the oc-SVM for 10 simulations (i.e., for 10
different input parameters) is presented. For each damaged scenario, at least one sen-
sor is classified as damaged, while for the healthy scenario all sensors are classified
as healthy. Moreover, for most of the damaged scenarios, one can observe a certain
level of proximity between the cracks and the sensors classified as damaged, thus
guiding the localization of damages. An exception corresponds to the geometry in
Fig. 2c, where almost all sensors are classified as damaged, thus preventing localiza-
tion. This issue is attributable to the relative position of the source and the crack, i.e.,
to localize damage (c) the source should be placed differently. A reasonable solu-
tion is to consider different locations for the active source, for example by defining
an additional input parameter in the model. This approach is already employed in
SHM with Lamb wave propagation where piezoelectric transducers are used as both
sensors and actuators (see e.g., [40]).
Sensor placement
To identify the optimal sensor locations, one has to create a new database of
Nμ = 100 synthetic observations sampled from a Sobol’s sequence [41]. Keep the
same parameters as in the previous paragraph, but, instead of computing the syn-
thetic time signals at few predefined points in space (the sensor locations), collect the
measurements at all the nodes of a new coarse mesh of , for a total of n dof = 360
degrees of freedom. Following the description given in Sect. 5.2, for each one of the
Nμ time signals, one has to perform the constrained variational approximation (29)
with an ARD-Exponential kernel over s = \ ∂ to obtain a set of m inducing
points D∗ (μi ) for 1 ≤ i ≤ Nμ . Figure 3 sketches the clustering results obtained for
different values of Nμ . Observe that the centroids tend to stabilize already after con-
sidering a cluster of 10 samples, especially when the sensor budget is high, i.e., for
large m values. This can be explained by noticing that when m is small, the algorithm
is trying to reconstruct a non-trivial quantity of interest over a complex domain with
only a few points, which may lead the sparse model to get stuck in a local mini-
mum without reaching convergence. Finally, the relative variance reduction (30) is
used to evaluate the quality of the estimated locations whose results are shown in
Fig. 4, where one can observe a variance reduction almost equal to 1 near the sensor
locations and a general reduction above 0.7 for all the unsensed locations even for
m = 4, thus indicating a good sensor placement.
7 Conclusion
This chapter presents how a model-based numerical approach can be integrated with
different data-driven techniques in the context of predictive maintenance. A peculiar-
ity of this work is that the authors make no assumption on the type of damages that a
structure may undergo during its lifetime, while modeling many environmental and
operational healthy scenarios. From a technical point of view, this work describes
how reduced-order modeling techniques can be leveraged to generate large and robust
Predictive Monitoring of Large-Scale Engineering Assets … 203
m=4 m =9 m = 16 m = 25
N N N N
Fig. 3 Comparison of the centroids obtained with the K-medoids clustering algorithm for different
sizes of clusters Nμ . Each plot shows a fixed number m of inducing points, which increases from
left to right
Fig. 4 Relative variance reduction obtained using m centroids and averaged over Nμ = 100 sam-
ples. Each plot shows a fixed number m of inducing points, which increases from left to right.
Values close to 1 indicate a good placement quality
datasets of synthetic sensor measurements and it explains how such datasets can be
used to learn the salient features that healthy scenarios have in common to ultimately
detect damages. The same damage-sensitive features are also used to guide an auto-
mated data-driven sensor placement strategy to increase the detection accuracy for
a given budget of sensors.
Although a simple 2D example is used to validate the proposed method, the gen-
eralization to more complex, possibly nonlinear problems is possible. Note however
that for real-world engineering problems the parameter space describing healthy vari-
ations is expected to be high dimensional, thus requiring a high computational effort
to elaborate the synthetic time signals. To this end, sensitivity analysis techniques
such as the variance-based global sensitivity indices [42] or the derivative-based
global sensitivity measures (DGSM) [43] are two popular choices to identify a few
parameters that influence the output of interest the most. Finally, note that while
the proposed method can be used for real-time predictions thanks to the offline and
online decomposition of tasks, a filtering technique to integrate the evolution of the
structure and update the model would be a valuable addition.
204 C. Bigoni et al.
References
1. Farrar CR, Worden K (2012) Structural health monitoring: a machine learning perspective.
Wiley, New York
2. Wagg DJ, Worden K, Barthorpe RJ et al (2020) Digital twins: state-of-the-art and future
directions for modeling and simulation in engineering dynamics applications. ASCE-ASME
J Risk Uncertainty Eng Syst Part B: Mech Eng 6(3):030901
3. Lecer M, Allaire D, Willcox K (2015) Methodology for dynamic data-driven online flight
capability estimation. AIAA J 53(10):3073–3087
4. Taddei T, Penn JD, Yano M, Patera AT (2018) Simulation-based classification; a model-order-
reduction approach for structural health monitoring. Arch Comput Methods Eng 25(1):23–45
5. Bigoni C, Hesthaven JS (2020) Simulation-based anomaly detection and damage localization:
an application to structural health monitoring. Comput Methods Appl Mech Eng 363:112896
6. Kapteyn MG, Knezevic DJ, Willcox K (2020) Toward predictive digital twins via component-
based reduced-order models and interpretable machine learning. In: AIAA Scitech 2020
Forum, p 0418
7. Rosafalco L, Manzoni A, Mariani S et al (2020) Fully convolutional networks for structural
health monitoring through multivariate time series classification. Adv Model Simul Eng Sci
7(1):1–31
8. Quarteroni A, Saleri F, Gervasio P (2006) Scientific computing with MATLAB and Octave.
Springer, Berlin
9. Zienkiewicz OC, Taylor RL (2005) The finite element method for solid and structural mechan-
ics. Elsevier, Amsterdam
10. Newmark NM (1959) A method of computation for structural dynamics. J Eng Mech Div
85(3):67–94
11. Hesthaven JS, Rozza G, Stamm B (2016) Certified reduced basis methods for parametrized
partial differential equations. Springer, Berlin
12. Quarteroni A, Manzoni A, Negri F (2015) Reduced basis methods for partial differential
equations: an introduction. Springer, Berlin
13. Benner P, Gugercin S, Willcox K (2015) A survey of projection-based model reduction methods
for parametric dynamical systems. SIAM Rev 57(4):483–531
14. Haasdonk B, Ohlberger M (2008) Reduced basis method for finite volume approximations of
parametrized linear evolution equations. ESAIM Math Model Numer Anal 42(2):277–302
15. Wang Q, Ripamonti N, Hesthaven JS (2020) Recurrent neural network closure of parametric
POD-Galerkin reduced-order models based on the Mori-Zwanzig formalism. J Comput Phys
410:109402
16. Choi Y, Carlberg K (2019) Space-time least-squares Petrov-Galerkin projection for nonlinear
model reduction. SIAM J Sci Comput 41(7):A26–A58
17. Bigoni C (2020) Numerical methods for structural anomaly detection using model order reduc-
tion and data-driven techniques. Ph.D. thesis No. 7734, EPFL
18. Barrault M, Maday Y, Nguyen NC, Patera AT (2004) An ‘empirical interpolation’ method:
application to efficient reduced-basis discretization of partial differential equations. Comptes
Rendus Mathematique 339(9):667–672
19. Chaturantabut S, Sorensen DC (2010) Nonlinear model reduction via discrete empirical inter-
polation. SIAM J Sci Comput 32(5):2737–2764
20. Kast M, Guo M, Hesthaven JS (2020) A non-intrusive multifidelity method for the reduced
order modeling of nonlinear problems. Comput Methods Appl Mech Eng 364:112947
21. Zhang Z, Guo M, Hesthaven JS (2019) Model order reduction for large-scale structures with
local nonlinearities. Comput Methods Appl Mech Eng 353:491–515
22. Guo M, Hesthaven JS (2019) Data-driven reduced order modeling for time-dependent prob-
lems. Comput Methods Appl Mech Eng 345:75–99
23. Guo M, Hesthaven JS (2018) Reduced order modeling for nonlinear structural analysis using
Gaussian process regression. Comput Methods Appl Mech Eng 341:807–826
Predictive Monitoring of Large-Scale Engineering Assets … 205
24. Long J, Buyukozturk O (2014) Automated structural damage detection using one-class
machine learning. In: Catbas FN (ed) Dynamics of civil structures, vol 4. Springer, Berlin, pp
117–128
25. Schölkopf B, Williamson RC, Smola AJ et al (2020) Support vector method for novelty
detection. Adv Neural Inf Process Syst 12:582–588
26. Cristianini N, Schölkopf B (2002) Support vector machines and kernel methods: the new
generation of learning machines. AI Mag 23(3):31
27. Liu FT, Ting KM, Zhou Z-H (2008) Isolation forest. In: 2008 Eighth IEEE international
conference on data mining. IEEE, pp 413–422
28. Breunig MM, Kriegel H-P, Ng RT et al. (2000) LOF: identifying density-based local outliers.
In: ACM sigmod record, pp 93–104
29. Pedregosa et al (2011) Scikit-learn: machine learning in Python. J Mach Learn Res 12:2825–
2830
30. Liu SW, Huang JH, Sung JC (2002) Detection of cracks using neural networks and computa-
tional mechanics. Comput Methods Appl Mech Eng 191(25–26):2831–2845
31. Japkowicz N, Myers C, Gluck M (1995) A novelty detection approach to classification. In:
International conference on artificial intelligence, vol 1, pp 518–523
32. Marchi E, Vesperini F, Eyben F et al (2015) A novel approach for automatic acoustic novelty
detection using a denoising autoencoder with bidirectional LSTM neural networks. In: 2015
IEEE international conference on acoustics, speech and signal processing (ICASSP). IEEE,
pp 1996–2000
33. Pathirage CSN, Li J, Li L et al (2018) Application of deep autoencoder model for structural
condition monitoring. J Syst Eng Electron 29(4):873–880
34. Ostachowicz W, Soman R, Malinowski P (2019) Optimization of sensor placement for struc-
tural health monitoring: a review. Struct Health Monit 18(3):963–988
35. Bigoni C, Zhang Z, Hesthaven JS (2020) Systematic sensor placement for structural anomaly
detection in the absence of damaged states. Comput Methods Appl Mech Eng 371:113315
36. Quiñonero-Candela J, Rasmussen CE (2005) A unifying view of sparse approximate Gaussian
process regression. J Mach Learn Res 6:1939–1959
37. Titsias M (2009) Variational learning of inducing variables in sparse Gaussian processes. In:
Proceedings of the twelfth international conference on artificial intelligence and statistics,
PMLR, vol 5, pp 567–574
38. Davis L (1991) Handbook of genetic algorithms. CumInCAD
39. Avendaño-Valencia LD, Chatzi EN, Tcherniak D (2020) Gaussian process models for mitiga-
tion of operational variability in the structural health monitoring of wind turbines. Mech Syst
Sig Process 142:106686
40. Swartz RA, Flynn E, Backman D et al (2006) Active piezoelectric sensing for damage identi-
fication in honeycomb aluminum panels. In: Proceedings of 24th international modal analysis
conference
41. Joe S, Kuo FY (2008) Constructing Sobol sequences with better two-dimensional projections.
SIAM J Sci Comput 30(5):2635–2654
42. Sobol IM (2001) Global sensitivity indices for nonlinear mathematical models and their Monte
Carlo estimates. Math Comput Simul 55(1–3):271–280
43. Kucherenko S, Song S (2016) Derivative-based global sensitivity measures and their link with
Sobol’sensitivity indices. In: Owen AB, Glynn PW (eds) Monte Carlo and Quasi-Monte Carlo
Methods. Springer, Cham, pp 455–469
Unsupervised Data-Driven Methods
for Damage Identification
in Discontinuous Media
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 207
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_10
208 R. Napolitano et al.
1 Introduction
Fig. 1 Schematic of the digital twin constructed with hybrid physics-based and data-driven
modeling
Current crack diagnostic techniques are limited as they are based on manual synthesis
of data, structural intuition, and manual crack mapping. Combining methods for
acoustic emissions, flat jack tests, and finite element modeling, Anzani et al. [13]
assessed the existing damage on a history masonry tower. In that study, quantitative
comparisons were made between the results of acoustic monitoring and the time of
seismic events; only qualitative comparison was used to compare the finite element
model with existing damage; furthermore, only one potential cause of the damage
was hypothesized and therefore simulated [11]. Continuing this trend, Alessandri
et al. [14] combined structural health monitoring and numerical methods, but again
only qualitatively compared cracks on the existing structure to the results of a single
hypothesized simulation. Similarly, when Milani et al. [15] used numerical methods
to evaluate the origins of crack patterns on the narthex of the Church of the Nativity
in Bethlehem, the existing conditions were only qualitatively compared to a single
load case simulation.
The concept of using qualitative methods to compare existing crack patterns to
simulations based on singular load cases is ubiquitous in the literature [7, 16–20] even
though assuming only one specific load is culpable increases bias in the results [11].
Maps of cracks on an existing structure do not record underlying problems, rather
they capture symptoms of these problems. This can pose issues for diagnostics and
rehabilitation as damage may have already been stabilized and not be related to
current risks [21]. Therefore, understanding how cracks on a structure could have
evolved is vital to understanding the structures potential for future stability [21].
210 R. Napolitano et al.
The aim of this work is to create workflow for inferring the cause of deformed (i.e.,
damaged) regions of masonry structures using automated methods. As a proof of
concept, consider only settlement-induced damage from a known initial state and seek
to determine a parametric representation of the settlement condition quantitatively.
Thus, the proposed workflow may require additional expansions to be successful on
a wide range of damage or loading conditions which differ qualitatively. Here the
earth mover’s distance (EMD) is used to quantify similarity between a library of
known damage states and Gaussian process regression (GPR) to accurately interpo-
late between them to identify the most likely cause of deformation. Manifold learning
strategies are also considered to handle extending the method to higher-dimensional
spaces, a challenge which must be overcome in order to address real-world struc-
tural health monitoring scenarios where damage states cannot be parameterized in a
simplified manner.
2 Methods
This first step in creating the workflow is to develop a metric for how different
scenarios will be compared. This step is schematically represented in Fig. 2. Since
the proposed workflow involves numerous steps, a conceptual summary of the method
is provided first. First, a point cloud representing an existing structure is acquired.
Based on historical record or expert knowledge about the construction and history
of the structure, a hypothetical undamaged or native state is produced (i.e., manually
using CAD) and supplied to the algorithm in the form of a triangulated mesh (state P
in Fig. 2). Then, a library of damaged conditions (states Ql in Fig. 2) for the structure
Fig. 2 Illustration of the supervised learning concept. A reference model is compared to a library of
deformed structures with known damage conditions to determine the most likely damage mechanism
for the observed case
Unsupervised Data-Driven Methods for Damage Identification … 211
For this work, the individual bricks were assumed to behave as rigid bodies. This
should enable a good approximation to the behavior of the masonry walls since
cracking at the masonry joints frequently dominates any fracturing of the individual
brick elements. A hypothetical undamaged state, U, composed of a triangulated mesh,
and the parameterized damage conditions, c = (x, y, a), were used as input to the
DEM simulation, where x and y define the coordinate of the settlement event and a
defines the amplitude of the settlement. By comparing each damaged state, Q, to the
observed structure, P, the true damage c0 can be identified which describes how the
structure came to be damaged. To understand how well each hypothesis, Q, matches
the true conditions, P, the earth mover’s distance (EMD) is used which measures
the distance d between probability distributions [29–31] (see Fig. 2). To expedite
the EMD calculation over the parameter space, Gaussian process regression (GPR),
a Bayesian method for model selection and updating, to approximate the distance
field, is applied [32]; this approach is commonly found in the literature [33–35], the
scikit-learn implementation [36] is used here. Thus, in this method GPR interpolates
for continuously varying damage conditions from a discrete set of damage conditions
to mitigate the computational burden of EMD. To create a weighting system which
would strongly illustrate regions where the distance between P and Q is low, and
thus the guess is good, L = d −1 is interpolated. By taking this inversion, the GPR
estimated field corresponds to the likelihood (measured in units of m−1 ) that the
guessed damage condition is the true one.
The starting point for the manifold learning strategy (illustrated in Fig. 3) is the stored
image of the displaced structure and a feature vector which conveys information
about how the structure was damaged. For this work, the feature vector consisted of
the length of different settlement regions, the depth of different settlement regions,
earthquake magnitude, earthquake orientation, point load, point load application
range, and the total sum of all settlement depths. The first step in the process is the
hashing of the stored image into a 2-D discrete cosine transform (DCT) [37]. Once
each image has been hashed this way, the Euclidean distance between each image
can be calculated.
Having computed the distances, the coordinates of each simulation in the low-
dimensional manifold can be computed. This is achieved by non-linear manifold
learning using the diffusion maps approach [38]. In short, this approach uses a spectral
decomposition of an affinity matrix (i.e., the reciprocal of the pairwise distances)
to determine collective variables. Then, the dominant eigenvectors can be used to
visualize a low-dimensional space (manifold) which contains the most significant
collective variables within the data.
2.3 Experiments
To test if the proposed workflow, along with presented methods, could be used
to diagnose existing crack patterns in masonry structures, a single-leaf, dry-joint
masonry test wall was constructed, and cracking was induced in a controlled manner,
as shown in Fig. 4 The test wall had a fixed base (cinder blocks) and region of
settlement (jack). Geometry and definition of variables used in experimental setup
are shown in Fig. 5. The dimensions and parameters for each brick are shown in Table
1. The settlement width was slightly smaller than the width of two bricks (190%
brick length), and the depth was 0.021 ± 0.001 m. Before and after the settlement
Fig. 4 Photographs of masonry walls used for experimental case studies. a Undeformed wall and
b Deformed wall, after settlement is applied in lower right corner
Table 1 Measured
Parameter Value
dimensions of bricks used in
experimental walls Length (m) 0.193 m ± 0.001
Width (m) 0.090 ± 0.001
Height (m) 0.055 ± 0.001
ρ (kg/m3 ) 2508.67 ± 10
Fig. 6 Estimate of the likelihood field (c) obtained from the experimental wall. The black cross
indicates ground truth settlement condition
events, photogrammetry models were generated using best practices [39–41] and
orthographic photographs were taken.
Based on the success of the simple experimental case above, a more complex synthetic
structure (Fig. 7) was considered. Synthetic data is useful for demonstrating the
Unsupervised Data-Driven Methods for Damage Identification … 215
Fig. 7 (left) Geometry of synthetic masonry structure including individual brick dimensions. (right)
Displacement magnitude of each brick in the synthetic structure due to two different settlement
events
efficacy of the method. On the right of Fig. 7, the damaged synthetic structure with
9633 distinct elements can be seen having incurred two Gaussian settlement loads
(one (x, y, a) at (1.63, − 1.73, 0.157) and a second at (0.62, 3.22, 0.157). Since two
independent damage events were considered for this synthetic case, this demonstrates
the applicability of the method to more realistic scenarios where multiple damage
states are present. A library of 256 damage states was generated by selecting damage
cases from a uniform random distribution where the domain was x ∈ [−0.39, 4.94], y
∈ [−2.80, 5.13], and a ∈ [0.00, 0.30]. EMD was calculated between the ground truth
model and each simulated damage state without segmentation; then, a continuous
likelihood field was generated by applying GPR to the inverted distances (Fig. 8).
Considering the two bright, yellow areas which indicate a high likelihood of being
the location and depth of settlement, the highest likelihood of 68.7 m−1 was found
at c = (0.51, 3.28, 0.198). When evaluating the likelihood at the true locations of
settlement, it is found that they have values of 58.6 and 64.6 m−1 and an average
Fig. 8 Estimate of the likelihood field (c) obtained from full point cloud P. Field rendered over (x,
y) slices for discrete amplitudes a as listed in each panel. Yellow color indicates greater likelihood
(in units of m−1 ). Black crosses indicate the two components of cz . White lines are centroids of
bricks to illustrate the position of the structure relative to the position of settlement centers
216 R. Napolitano et al.
field value of 33.3 m−1 . This means that when comparing the average to the true
locations, the true locations are twice as likely to be the cause of the damage. So,
while there is good qualitative agreement in Fig. 8, the quantitative specificity of the
result is relatively weak. To bolster the strength of the method, the damaged structure
and the library of potential damages were segmented into eight different point clouds
(shown in Fig. 9) using MBKM where each likelihood field corresponds to a single
point cloud segment. It can be seen in a subsequent section that the method is not
sensitive to the number of clusters selected in this step.
Segment 0, which is blue in Fig. 9, corresponds to the segment which covers
the parts of the structure around one of the settlement events. Figure 10 presents
the likelihood plot at discrete depths to understand what the possible cause of the
damage in this region of the structure could be. While the true settlement occurs
with a likelihood of 558 m−1 at c = (1.63, − 1.73, 0.157), the highest likelihood
is 575 m−1 at (1.79, − 1.65, 0.156). Since the average likelihood is 105 m−1 , this
indicates that the results are more decisive than what was previously shown in Fig. 8.
The likelihood field for Segment 7, which is gray in Fig. 9, is shown in Fig. 11.
This segment lies directly over the second settlement event (and is separated from the
first settlement event). The algorithm predicts the most likely settlement parameters
to be (0.67, 3.28, 0.204), with a likelihood of 679 m−1 and the average of 94 m−1 .
Fig. 10 Estimate of the likelihood field (c) obtained from P0 which is the blue section of the point
cloud in Fig. 9
Unsupervised Data-Driven Methods for Damage Identification … 217
Fig. 11 Estimate of the likelihood field (c) obtained from P7 which is the gray section of the point
cloud in Fig. 9
Compared to the result for Segment 0, the likelihood is more concentrated in the
x − y plane and more diffuse in a. This is probably due to the delocalized nature
of the settlement compared to the wall, such that certain combinations of (x, y, a)
parameters induce equivalent damage states in the wall itself.
To understand how the number of segments could affect the results of this method,
segmentation was performed for the complex, synthetic structure, where the user
selected 7, 8, and 9 clusters. This was repeated five times using random seeds to
understand if relationships were robust. Figure 12 illustrates all 15 of the different
segmentations which were carried out. Based on the broad range of high likelihood
values, there is significant uncertainty in the optimal value for this parameter. This
is not surprising as the peak likelihood values normally are within a concentrated
region of other relatively high likelihood values compared to other locations. Since
all 15 of the segmentations selected damage conditions corresponded well with the
ground truth, the results are assumed to be reasonably insensitive to the number of
segments selected.
Since the complex example was carried out on a synthetic dataset where no noise
was present, this section sought to understand how robust the method would be if the
point clouds were more realistic and noise was included. Thus, noise on the order
of millimeters was introduced to the synthetic dataset; random noise was added to
every point in P from a uniform distribution over [−δ, + δ] in each of (x, y, z). The
analysis was repeated for Segment 7 (i.e., P7 ) as outlined above. For each increase
in noise, the Z-score was recorded, z = (L(cz ) − μ)/σ, where μ and σ are the mean
and standard deviation of L(c) over all c in the investigated space, and error was
218 R. Napolitano et al.
Fig. 12 Illustration of random segmentations from the MBKM segmentation while varying number
of segments and random number seed. Mean and standard deviation are reported for each case.
The computed maximum likelihood condition from the corresponding surrogate model is reported
alongside each image
measured as the Euclidean distance between the location of peak likelihood and true
cause of damage. Calculated Z-score and error are presented in Fig. 13. As one would
expect, by adding noise to the point cloud, the Z-score decreased, and the prediction
error increased. An encouraging result is that the magnitude of noise expected from
typical point cloud acquisition methods (on the order of 10−3 m) does not appear to
impede the method; the prediction error did not increase between 10−4 and 10−3 m.
Unsupervised Data-Driven Methods for Damage Identification … 219
Fig. 13 Effects of noise magnitude δ (logarithmic scale in graphs) on a the Z-score and b error
between peak likelihood and true value
To evaluate how the number of available simulations (hypotheses) affects the quality
of the surrogate model, a sensitivity study to the library size was also performed.
Simulations were randomly selected from the full library of 256 simulations, and
a new model was constructed based on this sub-sample. The accuracy of the peak
likelihood value versus the ground truth value was computed and reported in Fig. 14
as a function of the sample size. The accuracy clearly improves with more simula-
tions available, but the improvement is most noticeable for libraries of less than 100
simulations.
and extended the analysis to include many Gaussians (i.e., representing an arbitrary
surface as a Gaussian mixture model) were added, the system may have hundreds
of parameters instead of only a handful. Each additional dimension in the param-
eter space makes the space sparser and mandates additional simulations to effec-
tively interpolate using the Gaussian process. This makes it unlikely for the GPR
approach to be tractable for complex problems. Here a method for combining numer-
ical methods and manifold learning to reduce the number of physics-based models
needed for diagnosis is outlined.
The starting point for the manifold learning strategy is the stored image of the
displaced structure and a feature vector which conveys information about how the
structure was damaged. For this work, the feature vector consisted of 12 total numer-
ical features including the length of different settlement regions, the depth of different
settlement regions, earthquake magnitude, earthquake orientation, point load, point
load application range, and the total sum of all settlement depths. The first step in the
process is the hashing of the stored image into a 2-D discrete cosine transform [37].
Once each image has been converted to a discrete cosine transform, the Euclidean
distance between each image can be calculated.
For convenience of visualization, a two-dimensional manifold was constructed
by taking the dominant eigenvectors ψ2 and ψ3 (comprised of one scalar value per
observation) as shown in Fig. 15. It is important to note that the algorithm never sees
the raw input feature vectors; collective variables are derived from only the distances
between observed feature vectors. However, knowing what the input features can
help a user to understand physical meanings behind the different eigenvectors and
start to see how the latent space is related to the physical input parameters.
To see how each of the 12 input features was related to the manifold, the points
on the manifold were colored by that feature. For instance, in the case of earthquake
magnitude, feature 7, dark blue indicates that there was no earthquake while yellow
indicates an earthquake with a magnitude of 0.3 g. As shown in Fig. 15, there is a
Fig. 15 Diffusion map colored by specific features. The first is colored by earthquake magnitude
(gs), the second is colored by settlement width (m), and third by settlement depth (m). Third panel
is zoomed in to show detail of bottom cluster only
Unsupervised Data-Driven Methods for Damage Identification … 221
correlation between feature 7 and where the points fall on the manifold. The simula-
tions with lower ψ3 values seem to not have experienced an earthquake at all, so those
could be purely point load or purely settlement. And across ψ2 there is a gradient
from the very high-magnitude earthquakes at a low ψ2 , to the lesser earthquakes at
higher values of ψ2 .
Not all features show these types of correlations with the eigenvectors, however.
For instance, in the case of feature 1 (settlement amplitude on the left side of the wall),
there does not seem to be any robust clustering possible. This indicates that this input
feature did not dominate the manifold space (at least in the first two eigenvectors).
Based on known information about feature 7, the bottom cluster of dark blue points
in Fig. 15 is the point that did not experience any earthquake. If only the overall
settlement experienced by the wall, feature 12, is considered, there are clear clusters
developing. The sections where there is a lot of settlement have the highest ψ3 and
the lowest ψ2 , then the middle settlement cases have middle range ψ2 , and lastly the
cases where there was little to no settlement have the highest values of ψ2 .
Based on a systematic analysis of the manifold topology, the features which exhib-
ited the strongest correlations to the datasets were: (1) feature 7 depicted in red
dominant colors representing earthquake magnitude, (2) feature 8 in green domi-
nant colors representing earthquake orientation, and (3) feature 12 in blue dominant
colors representing the total amount of settlement. These features could then be used
to interpret the manifold; Figure 16 shows the annotated manifold based on these
results. As ψ3 increases, the earthquake magnitude increases, as ψ2 increases the
earthquake orientation changes, and as ψ2 decreases, the overall settlement in the
system gets larger. Thus, it would seem that ψ2 is a latent variable encompassing
information about both settlement and earthquake orientation. The samples of simu-
lation images on the right are colored by the same conventions. Wall 630 experienced
less settlement than wall 749 which is why it has less blue (RGB = 0, 0.5, 0.5) than
749 (RGB = 0, 0.5, 0.8).
Once a user has created a manifold of possible damage mechanisms, the damaged
wall in question can be added to the set of inputs and tracked to see where in the
manifold it ends up. Based on patterns intuited through clustering the feature-colored
manifold, a user can evaluate where their damaged wall fits in with the rest of the
manifold to understand what could have caused existing damages. In future work,
this method should be validated against the experimental testing data such as those
presented in preceding sections.
4.1 Summary
Fig. 16 (left) Annotated manifold with coloration from three features. (right) Sample walls with colors indicating location in the manifold and showing the
crack pattern
R. Napolitano et al.
Unsupervised Data-Driven Methods for Damage Identification … 223
The proof of concept shown here uses a static library of simulations drawn from
a single hypothesized native state. There are several ways in which this could be
extended to make it more robust. First, more sophisticated approaches are possible
using iterative Bayesian optimization rather than a predetermined library of damage
hypotheses. In this case, simulations would be generated one at a time or in
small batches based on an acquisition function computed from prior observations.
This enables intelligent navigation of the parameter space and allocates simulation
224 R. Napolitano et al.
resources to only those hypotheses which provide the maximal amount of new infor-
mation. Second, the hypothesis space could be extended to include different native
states. This would be motivated by the reality that the native state of a structure is
not always known. Instead, several different native states could be postulated, and
the likelihood calculations could determine the most plausible one.
As already described, while the GPR approach provides high accuracy in deter-
mining damage conditions, it is unlikely to scale efficiently to very large param-
eter spaces. Conversely, the manifold learning strategy described here provides a
qualitative understanding of many variables without yielding reliable quantitative
information about the damage conditions. The combination of these approaches
may provide the advantages of each while mitigating their shortcomings. Another
promising approach is topology optimization based on adjoint simulations. While
this has historically been intractable for complex problems such as DEM, modern
hardware and software developed for training deep learning models can provide
gradients for very large and sophisticated transformations of data. Thus, a combina-
tion of data-driven and physics-based approaches will unlock new opportunities in
structural health monitoring.
Acknowledgements This paper is based on work in part supported by the National Science Founda-
tion Graduate Research Fellowship Program under Grant no. DGE-1656466. Any opinions, findings,
and conclusions or recommendations expressed in this material are those of the authors and do not
necessarily reflect the view of the National Science Foundation. Additional support was provided
by the Dean’s Fund for Innovation at Princeton and the Department of Civil and Environmental
Engineering. The present research was completed as part of the Itasca Educational Partnership under
the mentorship of Dr. Jim Hazzard.
References
1. Glisic B, Inaudi D (2008) Fibre optic methods for structural health monitoring. Wiley, Hoboken
2. National Trust for Historic Preservation T.: The Greenest Building (2017)
3. National Trust for Historic Preservation T.: Atlas of ReUrbanism (2017)
4. Olsen MJ (2013) In situ change analysis and monitoring through terrestrial laser scanning. J
Comput Civ Eng 29:4014040
5. ISO (2005) C.S.N.: 13822 Bases for design of structures—Assessment of existing structures.
CEN Brussels
6. Roca P (2004) Considerations on the significance of history for the structural analysis of ancient
constructions. Structural analysis of masonry historical constructions, Balkema, Amsterdam,
pp 63–73
7. Bayraktar A, Altunicsik AC, Sevim B, Türker T (2011) Seismic response of a historical masonry
minaret using a finite element model updated with operational modal testing. J Vib Control
17:129–149
8. Russo S (2013) Testing and modelling of dynamic out-of-plane behaviour of the historic
masonry façade of Palazzo Ducale in Venice, Italy. Eng Struct 46:130–139
9. Anderson DR, Burnham K (2004) Model selection and multi-model inference, 2nd edn.
Springer, NY
10. Watt DS (2009) Building pathology: principles and practice. Wiley, Hoboken
11. Chamberlin TC (1965) The method of multiple working hypotheses. Science 80:754–759
Unsupervised Data-Driven Methods for Damage Identification … 225
12. Harris SY (2001) Building pathology: deterioration, diagnostics, and intervention. Wiley,
Hoboken
13. Anzani A, Binda L, Carpinteri A, Invernizzi S, Lacidogna G (2010) A multilevel approach for
the damage assessment of historic masonry towers. J Cult Herit 11:459–470
14. Alessandri C, Garutti M, Mallardo V, Milani G (2015) Crack patterns induced by foundation
settlements: Integrated analysis on a renaissance masonry palace in Italy. Int J Archit Herit
9:111–129
15. Milani G, Valente M, Alessandri C (2017) The narthex of the Church of the Nativity in
Bethlehem: a non-linear finite element approach to predict the structural damage. Comput
Struct
16. Wood RL, Mohammadi ME, Barbosa AR, Abdulrahman L, Soti R, Kawan CK, Shakya M,
Olsen MJ (2017) Damage assessment and modeling of the five-tiered pagoda-style Nyatapola
temple. Earthq Spectra 33:S377–S384
17. Ramos LF, Lourenço PB (2004) Modeling and vulnerability of historical city centers in seismic
areas: a case study in Lisbon. Eng Struct 26:1295–1310
18. Michiels T, Napolitano R, Adriaenssens S, Glisic B (2017) Comparison of thrust line analysis,
limit state analysis and distinct element modeling to predict the collapse load and collapse
mechanism of a rammed earth arch. Eng Struct 148:145–156
19. Douglas I, Napolitano R, Garlock M, Glisic B (2019) Reconsidering the vaulted forms of Cuba’s
National School of Ballet. In: Structural analysis of historical constructions, pp 2150–2158.
Springer, Berlin
20. Russo V, Lignola GP, Vassallo E, Zinno A (2010) Second World War damages of the architec-
tural heritage: St. Maria Del Popolo Agli Incurabili Church in Naples. In: Advanced Materials
Research. pp 1137–1142
21. Taylor J (2005) An integrated approach to risk assessments and condition surveys. J Am Inst
Conserv 44:127–141
22. Napolitano RK, Glisic B (2018) Minimizing the adverse effects of bias and low repeatability
precision in photogrammetry software through statistical analysis. J Cult Herit 31:46–52
23. Patias P, Kaimaris D, Georgiadis C, Stamnas A, Antoniadis D, Papadimitrakis D (2013) 3D
mapping of cultural heritage: special problems and best practices in extreme case-studies.
ISPRS Ann Photogramm Remote Sens Spat Inf Sci II-5, pp 223–228
24. Barsanti SG, Remondino F, Fenández-Palacios BJ, Visintini D (2014) Critical factors and
guidelines for 3D surveying and modelling in cultural heritage. Int J Herit Digit Era 3:141–158
25. Napolitano R, Glisic B (2019) Methodology for diagnosing crack patterns in masonry structures
using photogrammetry and distinct element modeling (accepted). Eng Struct
26. Binda L, Gambarotta L, Lagomarsino S, Modena C (1999) A multilevel approach to the damage
assessment and seismic improvement of masonry buildings in Italy. Seism. damage to Mason.
Build. Balkema, Rotterdam, pp 170–195
27. Napolitano R, Hess M, Glisic B (2019) Quantifying the differences in documentation and
modeling levels for building pathology and diagnostics. Arch Comput Methods Eng 1–18
28. Béjar Alonso J (2013) K-means vs Mini Batch K-means: a comparison
29. Yilmaz U (2009) The Earth Mover’s Distance
30. Rubner Y, Tomasi C, Guibas LJ (2000) The earth mover’s distance as a metric for image
retrieval. Int J Comput Vis 40:99–121
31. Lupu N, Selios L, Warner Z (2017) A new measure of congruence: the earth mover’s distance.
Polit Anal 25:95–113
32. Williams CKI, Rasmussen CE (1996) Gaussian processes for regression. In: Advances in neural
information processing systems, pp 514–520
33. Zhou L, Yan G, Ou J (2013) Response surface method based on radial basis functions for
modeling large-scale structures in model updating. Comput Civ Infrastruct Eng 28:210–226
34. Xia Z, Tang J (2013) Characterization of dynamic response of structures with uncertainty by
using Gaussian processes. J Vib Acoust 135:51006
35. Wan H-P, Ren W-X (2014) Parameter selection in finite-element-model updating by global
sensitivity analysis using Gaussian process metamodel. J Struct Eng 141:4014164
226 R. Napolitano et al.
1 Introduction
Construction safety is an important issue for the society. There are many hidden
dangers in construction sites, such as various types of workers, large construction
machinery and complex environment. In the process of dynamic construction, the
load of building structure is time-varying, and various support structures interact
among themselves, leading to the stability risk of the entire construction. Buildings
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 227
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_11
228 Y. Zhang and K.-V. Yuen
and structures under construction are prone to collapse, overturning and other acci-
dents due to the influence of natural conditions, construction level, and construction
quality. In addition, workers suffer from injuries caused by construction machinery,
such as blow or collision. Therefore, the construction safety monitoring is necessary.
In order to improve the construction level of engineering projects and reduce
the incidence of accidents, the concept of smart construction site was proposed [1].
Smart construction site is the concrete embodiment of the concept of smart city in the
construction industry. It provides all-round, three-dimensional, and real-time super-
vision of construction sites with intelligent systems of controllability, data, and visu-
alization. At present, the development of smart construction site is still in the initial
stage, i.e., the perception stage. It uses advanced sensing technologies to monitor
workers, machineries, and structures, and then identifies and locates the potential
hazards. Building information modeling (BIM) takes the three-dimensional graphics
of buildings as the carrier to further integrate all types of building information. The
parameterized model can be used to realize construction simulation, collision detec-
tion, and other applications. Hu et al. used network analysis to improve collision
detection. A component network centered on conflicting objects was constructed
to represent the dependencies of components. This method can effectively identify
the irrelevant conflicts and reduce the number of irrelevant conflicts by 17% [2].
Mirzaei et al. developed a novel 4D-BIM dynamic conflict detection and quantifica-
tion system for the identification of spatiotemporal conflicts [3]. 3D laser scanning
technology uses the principle of laser ranging to scan objects and quickly obtain
3D models. There have been many attempts in building inspection, cultural relics
protection and other methods. Yang et al. [4] used terrestrial laser scanning to detect
the deformation of the arch structure. The surface approximation method was used to
cover the blank of measurement area, and the uncertainty of surface of different order
was studied. Valenca et al. [5] proposed an automatic crack assessment method based
on image processing and terrestrial laser scanning (TLS) technology. The geometric
information measured by TLS was used to correct the captured image. It improved
the identification accuracy of structural cracks. The combination of 3D laser scanning
technology and BIM can realize the detection of structural quality and deformation.
Ham et al. [6] proposed a structural safety diagnosis method based on laser scanning
and BIM. The laser scanning data and BIM model were compared and analyzed to
determine the deformation degree of pipe support. Chen et al. [7] proposed a point-
to-point comparison method for deviation detection between automatic scanning and
BIM. When there is a deviation between BIM and point clouds, it will be highlighted
to remind users for further investigation.
Sensor networks can detect the deformation of high formwork support, tower
crane, and bridge by different sensor nodes. Kifouche et al. [8] developed and
deployed a sensor network that could collect data from multiple types of sensors. The
data was transferred to a server for visualization and real-time processing. Kuang
et al. [9] used fiber-optic sensors to monitor the deflection and cracks of beams.
The results showed that it was possible to detect fine crack and final failure crack
by optical fiber. Casciati et al. [10] used GPS to detect the displacement of steel
structures in real time. The accuracy was of the order of subcentimeters. With the
Applications of Deep Learning in Intelligent Construction 229
Convolutional neural network is the cornerstone for deep learning to achieve break-
through achievements in the field of computer vision in recent years. Convolutional
neural network uses convolutional layers to replace fully connected (FC) layers
for effective feature extraction. A convolutional neural network usually consists of
multiple convolutional modules. The front convolution layer has a small receptive
field, which can capture local and detailed information of an image. The receptive
field of the latter convolution layer is gradually enlarged to capture more complex
and abstract information in the image. At first, deep convolutional neural network
230 Y. Zhang and K.-V. Yuen
was used to perform image classification tasks and determine object categories in
the image. After AlexNet was concerned by researchers, some image classification
networks were produced, such as VGGNet, GoogLeNet, and ResNet. The recogni-
tion accuracy of these networks in the image classification task has almost reached the
human level. However, image classification is a very crude task. Objects in an image
need not only to be classified, but also to be positioned. Some algorithms combine
candidate regions with image classification networks to achieve target location based
on bounding boxes, such as region-based convolutional neural networks (R-CNN),
Fast R-CNN, single-shot multibox detector (SSD), you only look once (YOLO),
Faster R-CNN, and so on. The overall architecture of Faster R-CNN is shown in
Fig. 1. Image classification network is used to extract features, and region proposal
network (RPN) is used to generate candidate regions [17]. It further improves the
accuracy and positioning accuracy of object recognition.
Although bounding boxes can locate objects in an image, the bounding box
contains not only the target object, but also the background and other content. In
order to achieve more precise recognition of objects in images, semantic segmenta-
tion and instance segmentation are proposed. These two types of algorithms belong
to the pixel-level object recognition algorithm. Segmentation algorithms mostly
adopt code-decoding structures [18–20], such as fully convolutional networks (FCN),
SegNet, U-Net, and DeepLab. The overall architecture of FCN is shown in Fig. 2.
Multiple convolutional layers are used as encoding structures for down-sampling,
and multiple deconvolutional layers are used as decoding structures for up-sampling.
Human pose recognition is a special object detection task. Key point detection
of human body is very important to describe human posture and predict human
behavior. Action classification, abnormal behavior detection, and other tasks can be
accomplished by estimating the key points of human body. Mask R-CNN, an instance
segmentation algorithm, can not only identify targets at pixel level, but also estimate
the key points of human body. In addition, network structures such as DensePose,
Applications of Deep Learning in Intelligent Construction 231
Fig. 2 FCN
OpenPose, AlphaPose, and DeepPose have also achieved remarkable effects in the
field of human pose estimation.
In the process of dynamic construction, structural stability is poor, which can easily
cause accidents. Therefore, it is necessary to monitor the status of structures under
construction. Steel structures are often used in construction sites, such as steel trusses,
steel columns, and support structures. Bolted joints have the advantages of simple
structure, convenient installation, and strong reliability. These reasons have made
bolts the preferred fastener. Under the influence of dynamic and static loads, bolts
may be loosened and fallen off. In order to avoid interference with construction, non-
contact monitoring methods can be preferred. The combination of machine vision
and deep learning can quickly detect and count the state of bolts in steel structures. A
camera or smartphone can quickly capture images of bolts. Faster R-CNN was used
to identify and locate bolts in images. The image of each bolt is extracted on the basis
of the localization information. Then, the edge lines of bolts in the image are extracted
by using binarization and Hough transform. The detection result is shown in Fig. 3a.
According to the change of edge information, the looseness angle could be recognized
[21]. The head of the shoulder bolt is a regular hexagon, so this method could detect
looseness angle within 60°. Grades of fastener appear on the head of each bolt. The
“bolt” and “num” were identified and located simultaneously by SSD, as shown in
Fig. 3b. According to the localization information, the looseness angle within 180°
could be recognized [22]. When the bolt looseness angle is greater than 180°, the
above method cannot effectively detect bolt looseness angle. A bolt to be loosen
when the loosening is evident in a geometric change, which may even be visible to
naked eyes. Therefore, Faster R-CNN can be used to directly detect large looseness
232 Y. Zhang and K.-V. Yuen
damage of bolts, as shown in Fig. 3c. This method has strong robustness even under
the influence of structural vibration, illumination, and other environmental factors
[23]. Deep learning is a data-driven recognition algorithm, and a large amount of data
is the basic condition to ensure the recognition accuracy and generalization ability of
model. 3D simulation software can quickly produce many bolt images. The detection
model based on the simulation image still can accurately identify and locate bolts in
a real image [24]. Although the simulation software cannot produce bolt images in
a complex scene, this method provides a new idea for data augmentation.
more accurate measurements with less computation time. Subpixel subdivision tech-
nology can make up for the shortage of hardware and improve image resolution.
Luan et al. propose a deep learning approach based on CNN to extract full-field
high-resolution displacements at subpixel levels [29]. The results showed that the
trained network can identify the pixels with sufficient texture contrast as well as their
subpixel motions.
the identification results. A practical detection method should consider not only the
detection accuracy, but also the detection speed. Cha et al. [35] proposed a real-time
crack segmentation network, which has a great improvement in detection accuracy
and speed. The model processes in real-time images at 1025 × 512 pixels, which is
46 times faster than in a recent work. According to the results of semantic segmen-
tation, the parameters such as length, width, and area of damages can be obtained
more conveniently. However, semantic segmentation requires pixel-level annotation
data, and the annotation cost is very large. Meanwhile, semantic segmentation model
is difficult to train and requires high computational performance.
To reduce the detection costs, object detection and traditional methods are
combined to achieve crack refined detection. Object detection can identify and locate
cracks accurately and quickly by bounding box. The positioning method based on the
bounding box cannot quantitatively analyze the parameters of cracks. Some image
processing methods such as binarization and filtering can be used to extract cracks
in the bounding box at the pixel level. The identification result is shown in Fig. 7.
Jiang et al. [36] proposed a real-time crack evaluation system based on deep learning
and wall-climbing UAV. Firstly, SSD was used to identify and locate cracks in the
images transmitted by UAV. Then, the images in bounding box were converted into
binary image. The grayscale images were transformed into binary image through Otsu
threshold segmentation, and morphological processing was applied to ensure conti-
nuity of cracks. These are image-based identification methods, which can only obtain
two-dimensional spatial parameters of damage. Deep convolutional neural networks
have also achieved remarkable results in the field of 3D point cloud segmentation.
Beckman et al. [37] combined convolutional neural network with depth camera to
detect the volume of concrete surface damages.
A large number of workers on construction sites and the crosscutting of jobs are
the main reasons for the higher casualty rate in construction industry than in any
other industry. There are many potential hazards in the construction sites and workers
should be vigilant of their surroundings. Therefore, it is desirable to track the number
and status of workers in the construction site in real time. Object detection algorithms
such as Faster R-CNN can identify and locate workers in different scenes and postures
[38], as shown in Fig. 8. It can quickly count the number and location of workers in
a large scene, providing a new possibility for construction worker detection.
Heavy equipment may lead to serious worker injuries, and equipment operators’
view is easily obscured. Thus, it is necessary to detect workers near the equip-
ment to provide safety warning for operators. In order to improve visibility, heavy
equipment manufacturers install cameras on each side of the equipment (i.e., front,
right, left, and rear) to provide a comprehensive view of the area around the equip-
ment. However, this monitoring system cannot automatically extract information
from images. Son et al. [39] proposed a real-time warning system using visual data
and Faster R-CNN. This system used monocular cameras to estimate worker’s posi-
tion in three dimensions. In addition, some support structures without guardrails
are also one of the important hazards. During the construction of an engineering
structure, workers tend to take shortcuts by crossing supports to perform daily activ-
ities and save time. However, crossing the support is very dangerous and forbidden.
Fang et al. [40] used instance segmentation to detect workers crossing structural
supports during the construction of a project. First, workers and structural supports
are identified at the pixel level by the Mask R-CNN. Then, the positioning relationship
between the structural support and the workers could be used to determine whether the
worker is passing through the support. The proposed method could obtain the distance
between workers and hazard sources, thus providing safety warning for managers
and performing the correct behavior. At construction sites, workers are always in
dynamic walking positions. Predicting workers’ trajectories has great potential to
improve workplace safety. Object identification-based tracking methods only rely on
entity operation information and do not make full use of context information. Cai et al.
[41] proposed a context-augmented long short-term memory approach for worker
trajectory prediction. Compared with the traditional one-step prediction method, the
proposed method could predict multistep trajectory to avoid error accumulation and
effectively reduce final displacement error by 70%.
Workers should be vigilant of potential hazards. However, personal protection of
workers is also very necessary. Safety helmets play an important role in protecting
construction workers from accidents. Nevertheless, workers sometimes do not wear
safety helmets for convenience. For the safety of construction workers, a high-
precision and strong robustness helmet detection algorithm is urgently needed. Object
detection algorithms such as Faster R-CNN, SSD, and YOLO can detect safety helmet
in an image [42–44], as shown in Fig. 9. The recognition accuracy of detection algo-
rithms based on deep learning depends on a large number of sample data. To evaluate
the performance of Faster R-CNN, more than 100,000 image frames of construction
workers were randomly selected from surveillance videos of 25 construction sites
over a period of more than one year. The experimental results showed that this
method has high accuracy, high recall rate, and fast speed, and could effectively
detect non-helmet-use in different construction sites. It is conducive to improving
the safety inspection and supervision level. Besides safety helmet and protective
clothing, safety harness is the main protective equipment for workers working at
height. Workers often forget or deliberately do not wear seat harness. This is a very
dangerous behavior and is one of the main causes of falling from heights. Fang et al.
[45] developed a visual-based automated approach that uses two convolutional neural
networks to determine whether a worker is wearing a safety harness. First, Faster
R-CNN was used to detect workers in an image, and the image in bounding box
was cropped. Then, these cropped images were fed into a multilayer convolutional
neural network to determine whether the workers wear seat harness. The accuracy
rate and recall rate of the Faster R-CNN model were 99 and 95%, respectively, and
the accuracy rate and recall rate of the CNN model were 80 and 98%, respectively.
Human body is very flexible, and posture can be used to identify worker behavior.
Roberts et al. [46] proposed an activity analysis framework based on vision and deep
learning that could estimate and track the posture of worker. Firstly, YOLO was
used to identify and locate workers in an image, and the image in bounding box was
cropped. Then, the cropped images were fed into a 2D skeleton detection network to
estimate each joint coordinates of the worker. Posture estimation of worker is shown
in Fig. 10. Finally, the pose of the same construction worker in different video frames
was tracked. The proposed method was used to identify different worker activities
(bricklaying and plastering), and the results showed that this method had the poten-
tial to evaluate individual worker activities. A 3D skeleton estimation network can
directly determine the spatial position of human. The distance between workers can
be used to estimate severity of crowding in working environment. High crowding may
lead to dangerous working conditions and negative worker behavior. Yan et al. [47]
proposed a vision-based crowd detection technology. First, Faster R-CNN was used
to identify and locate workers in complex environment, and the image in bounding
box was cropped. Then, the cropped images were fed into the 3D skeleton detection
network to estimate each joint coordinates of the worker. Finally, according to the
result of human skeleton estimation, the spatial position of each worker was obtained.
Experimental results showed that this method could estimate the distance between
two workers with an error of 0.45 m in three-dimensional space.
At present, worker status monitoring methods based on the fixed camera have
some limitations, such as the perceptual mixing, occlusion, and illumination. Mean-
while, there is no camera inside the buildings under construction. Some mobile robots
autonomously inspect indoor work sites and find potentially dangerous anomalies.
Lee et al. [48] applied the perception module based on deep learning and simul-
taneous localization and mapping (SLAM) to target recognition and navigation of
mobile robots. The proposed method could identify abnormal behaviors of some
workers, such as not wearing safety helmets, standing on top of ladders, or falling.
Identification of danger is only the first step of construction safety management. It
is necessary to timely and accurately convey risk information to managers. Tang
et al. [49] proposed a language-image framework that aims at understanding and
detecting semantic roles of activities mentioned in safety rules. This framework
includes semantic parsing of safety rules, construction object detectors using SSD
and Faster R-CNN, and semantic role detectors. The experimental results showed
that this framework can preliminarily describe the dangerous scene in an image in
the form of language.
data could be used for data augmentation, which can improve the vehicle detec-
tion performance [52]. The identification and location information of construction
machinery can provide managers with the type, quantity, and distribution of construc-
tion machinery. Identifying and tracking construction machines also can avoid poten-
tial collisions and other accidents. Firstly, convolutional neural network was used to
detect and track vehicles. Then, the tracking trajectory was fed into a hidden Markov
model (HMM) that automatically discovers and assigns activity labels to objects.
Roberts et al. [53] performed activity analysis of machineries based on the object
detection results. The accuracy of activity analysis was found to be 86.8% for exca-
vators and 88.5% for dump trucks. Of all construction machinery, tower cranes are
the largest in size. The tower crane operator is far from the ground and cannot clearly
observe the surrounding environment of lifting objects. High-definition cameras built
into Unmanned Aerial Vehicles (UAVs) could help solve this problem. Roberts et al.
[54] used SSD to identify and locate tower cranes in an image, which is photographed
by UAVs.
All the aforementioned identified bounding box-based construction machinery
recognition methods use horizontal detection algorithms for detection. However,
construction machinery can be in any direction and position, and they are not neces-
sarily horizontal or vertical. When construction machinery is densely parked together,
rotated bounding box can be more accurately fitted to the machinery region in terms
of the orientation. Guo et al. [55] proposed a construction machinery identifica-
tion method based on orientation-aware feature (OAF) fusion convolutional neural
network. The proposed OAF-SSD could be applied not only to the construction
vehicle detection, but also to the detection problem of dense multiple objects in civil
engineering, which can also identify the orientation of target objects (more useful
for motion tracking and estimation).
In the similar fashion as human body, the posture and movements of construc-
tion machinery need to be estimated automatically. This has a significant impact on
construction safety and the use of the machinery itself. Excavator is one of the most
used construction equipment. Its operation is complicated, and its posture changes
more. Hence, it is more difficult to evaluate the full body posture of excavators than
other equipment. Liang et al. [56] proposed a vision-based marker-less pose esti-
mation system for estimating the joints and components of excavators. This system
adopted and improved the advanced convolutional network, namely the stack hour-
glass network for pose estimation. The results showed that this system could estimate
excavator boom, stick, and bucket joint positions but had higher estimation error for
the bucket location due to the occlusion issue. Luo et al. [57] built an integrated model
based on stacked hourglass and stacked pyramid network to estimate the posture of
excavators in an image. This model evaluated the posture of excavators by identi-
fying six key points of excavators, as shown in Fig. 12. Occlusion can significantly
interfere with the detection results of this method. Compared to the activity identi-
fication of construction workers, research on activity identification of construction
machinery is very limited, mainly because of the lack of construction machinery
actions datasets. Zhang et al. [58] produced a comprehensive video dataset of 2064
clips, which included five actions (digging, swinging, dumping, moving forward,
Applications of Deep Learning in Intelligent Construction 241
and moving backward) of excavators and dump trucks. CNN is used to extract image
features, and long short-term memory (LSTM) is used to extract time characteristics
from video frame sequences. These two types of features were used to identify these
five actions.
6 Conclusion
Deep learning has been growing rapidly in the field of image processing. Some
new network structures promote the development of construction safety monitoring
using machine vision. On construction sites, structure, worker, and machinery are
the three potential hazard sources, which are the most likely to cause accidents.
Machine vision-based detection method is a non-contact detection technology, which
can detect hazard sources without affecting the normal construction. At present,
deep learning networks applied in the field of construction safety monitoring can be
roughly divided into three categories: object detection, semantic segmentation, and
pose estimation.
(1) Object detection using neural network is the simplest and most used detec-
tion method. Faster R-CNN, SSD, YOLO, R-FCN, and other object detection
networks were used to identify and locate workers, machineries, and structural
components. This type of detection method mainly relies on bounding box to
identify and locate hazard sources. It can only complete detection tasks with
low positioning accuracy.
242 Y. Zhang and K.-V. Yuen
(2) Compared with object detection, semantic segmentation has higher positioning
accuracy and can achieve pixel-level object recognition. FCN, SegNet, U-Net,
and other networks are used to locate targets, components, and workers with
high precision. However, the disadvantages of this approach are also obvious.
The training cost of semantic segmentation is high, and the detection speed is
slow.
(3) Pose estimation networks recognize human posture through key point esti-
mation. It can be used to detect different types of workers and assess worker
activities. Meanwhile, it can also be used to locate the key points of construction
machinery and identify the working status of machinery.
Many researchers have studied construction safety with deep learning and
proposed many effective detection methods. However, from the perspective of smart
construction site, the research on construction safety monitoring using deep learning
is still in the initial stage. These detection techniques are far from being intelligent,
and only identify and locate multiple targets on construction sites. In addition, the
current detection technology also has some common problems. Deep learning is a
data-driven algorithm. There is a lack of large image datasets for different construc-
tion scenes. Therefore, how to train a robust and high-precision detection model is still
a difficult problem. Most importantly, visual detection method is easily affected by
illumination, background, occlusion, and other factors. Now more and more complex
construction site environment, some targets are often blocked, which is fatal to visual
detection methods. A single type of sensor data often has great limitations, so the
fusion of multitype sensor data may be a more feasible scheme for construction safety
monitoring.
References
27. Ilg E, Mayer N, Saikia T, Keuper M, Dosovitskiy A, Brox T (2017) Flownet 2.0: Evolution
of optical flow estimation with deep networks. In: Proceedings of the IEEE conference on
computer vision and pattern recognition, pp 2462–2470. IEEE, USA
28. Dong CZ, Celik O, Catbas FN, O’Brien EJ, Taylor S (2020) Structural displacement monitoring
using deep learning-based full field optical flow methods. Struct Infrastruct Eng 16(1):51–71
29. Luan L, Wang ML, Yang Y, Sun H (2020) Extracting full-field subpixel structural displacements
from videos via deep learning. arXiv preprint arXiv:2008.13715
30. Cha YJ, Choi W, Büyüköztürk O (2017) Deep learning-based crack damage detection using
convolutional neural networks. Comput-Aided Civil Infrastr Eng 32(5):361–378
31. Deng J, Lu Y, Lee VCS (2020) Imaging-based crack detection on concrete surfaces using You
Only Look Once network. Struct Health Monit 20(2):484–499
32. Zhang X, Rajan D, Story B (2019) Concrete crack detection using context-aware deep semantic
segmentation network. Comput-Aided Civil Infrastruct Eng 34(11):951–971
33. Dung CV (2019) Autonomous concrete crack detection using deep fully convolutional neural
network. Autom Constr 99:52–58
34. Lee JS, Hwang SH, Choi IY, Choi Y (2020) Estimation of crack width based on shape-sensitive
kernels and semantic segmentation. Struct Control Health Monit 27(4):e2504
35. Choi W, Cha YJ (2019) SDDNet: Real-time crack segmentation. IEEE Trans Industr Electron
67(9):8016–8025
36. Jiang S, Zhang J (2020) Real-time crack assessment using deep neural networks with wall-
climbing unmanned aerial system. Comput-Aided Civil Infrastruct Eng 35(6):549–564
37. Beckman GH, Polyzois D, Cha YJ (2019) Deep learning-based automatic volumetric damage
quantification using depth camera. Autom Constr 99:114–124
38. Son H, Choi H, Seong H, Kim C (2019) Detection of construction workers under varying poses
and changing background in image sequences via very deep residual networks. Autom Constr
99:27–38
39. Son H, Seong H, Choi H, Kim C (2019) Real-time vision-based warning system for prevention
of collisions between workers and heavy equipment. J Comput Civ Eng 33(5):04019029
40. Fang W, Zhong B, Zhao N, Love PE, Luo H, Xue J, Xu S (2019) A deep learning-based
approach for mitigating falls from height with computer vision: Convolutional neural network.
Adv Eng Inform 39:170–177
41. Cai J, Zhang Y, Yang L, Cai H, Li S (2020) A context-augmented deep learning approach for
worker trajectory prediction on unstructured and dynamic construction sites. Adv Eng Inform
46:101173
42. Wu J, Cai N, Chen W, Wang H., Wang G (2019) Automatic detection of hardhats worn by
construction personnel: a deep learning approach and benchmark dataset. Autom Construct
106:102894
43. Zhao Y, Chen Q, Cao W, Yang J, Xiong J, Gui G (2019) Deep learning for risk detection and
trajectory tracking at construction sites. IEEE Access 7:30905–30912
44. Fang Q, Li H, Luo X, Ding L, Luo H, Rose TM, An W (2018) Detecting non-hardhat-use by
a deep learning method from far-field surveillance videos. Autom Constr 85:1–9
45. Fang W, Ding L, Luo H, Love PE (2018) Falls from heights: A computer vision-based approach
for safety harness detection. Autom Constr 91:53–61
46. Roberts D, Torres Calderon W, Tang S, Golparvar-Fard M (2020) Vision-based construction
worker activity analysis informed by body posture. J Comput Civ Eng 34(4):04020017
47. Yan X, Zhang H, Li H (2019) Estimating worker-centric 3D spatial crowdedness for
construction safety management using a single 2D camera. J Comput Civ Eng 33(5):04019030
48. Lee MFR, Chien TW (2020) Intelligent robot for worker safety surveillance: deep learning
perception and visual navigation. In: 2020 international conference on advanced robotics and
intelligent systems, pp 1–6. IEEE, UK
49. Tang S, Golparvar-Fard M (2017) Joint reasoning of visual and text data for safety hazard
recognition. In: Computing in civil engineering 2017, pp 450–457. ASCE, USA
50. Fang W, Ding L, Zhong B, Love PE, Luo H (2018) Automated detection of workers and heavy
equipment on construction sites: a convolutional neural network approach. Adv Eng Inform
37:139–149
Applications of Deep Learning in Intelligent Construction 245
51. Kim H, Kim H, Hong YW, Byun H (2018) Detecting construction equipment using a region-
based fully convolutional network and transfer learning. J Comput Civ Eng 32(2):04017082
52. Kim H, Bang S, Jeong H, Ham Y, Kim H (2018) Analyzing context and productivity of tunnel
earthmoving processes using imaging and simulation. Autom Constr 92:188–198
53. Roberts D, Golparvar-Fard M (2019) End-to-end vision-based detection, tracking and activity
analysis of earthmoving equipment filmed at ground level. Autom Constr 105:102811
54. Roberts D, Bretl T, Golparvar-Fard M (2017) Detecting and classifying cranes using camera-
equipped UAVs for monitoring crane-related safety hazards. In: Computing in civil engineering
2017, pp 442–449. ASCE, USA
55. Guo Y, Xu Y, Li S (2020) Dense construction vehicle detection based on orientation-aware
feature fusion convolutional neural network. Autom Construct 112:103124
56. Liang CJ, Lundeen KM, McGee W, Menassa CC, Lee S, Kamat VR (2018) Stacked hourglass
networks for markerless pose estimation of articulated construction robots. In: 35th interna-
tional symposium on automation and robotics in construction, pp 843–849. Curran Associates,
Germany
57. Luo H, Wang M, Wong PKY, Cheng JC (2020) Full body pose estimation of construction
equipment using computer vision and deep learning techniques. Autom Construct 110:103016
58. Zhang J, Zi L, Hou Y, Wang M, Jiang W, Deng D (2020) A deep learning-based approach to
enable action recognition for construction equipment. Adv Civil Eng 1–14
Integrated SHM Systems: Damage
Detection Through Unsupervised
Learning and Data Fusion
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 247
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_12
248 E. García-Macías and F. Ubertini
1 Introduction
the system may operate, and the data acquisition limitations under operating
conditions.
(2) Data acquisition/normalization and cleansing: The data acquisition process
concerns the selection of suitable excitation methods, sensor types and
coverage (number and location), and data acquisition/storage/transmission.
Data normalization regards the ability of separating changes in sensor readings
caused by damage from those caused by varying EOC. Finally, data cleansing
regards the process of phasing out corrupted or uninformative data from the
feature selection process.
(3) Feature extraction and information condensation: This stage relates the iden-
tification of data features that allow one to distinguish between the undamaged
and damaged conditions. Inherent in the feature selection process, the infor-
mation condensation step deals with the minimization of redundant features
and dimensionality reduction.
(4) Statistical model development: This last stage concerns the implementation of
statistical models mapping between the extracted features and the diagnosis,
that is to say, the class or tag assignment. Such models are thus intended to
enable the discrimination between features related to undamaged and damaged
classes. Generally, statistical models are classified in three categories: unsu-
pervised learning, supervised learning and semi-supervised learning as an
intermediate solution. A pattern recognition model can be trained in a super-
vised learning fashion when data are available from both the undamaged
and damaged structures. Conversely, when no information is available on
the damaged structure, the algorithms fall into the category of unsupervised
learning. In semi-supervised learning, a minimum amount of tagged training
data (data labelled as “damaged”) is available, but not sufficient for a full
supervised learning. In this regard, transfer learning is a popular alternative
to supervised learning, where tagged data are generated through a digital twin
of the structure (virtual damage scenarios). In either way, statistical pattern
recognition models analyse statistical distributions of the measured or derived
features with the aim of performing damage identification. The damage identifi-
cation process is commonly organized in a hierarchical structure of increasing
complexity discussed by Rytter [5]: Level I: Detection; Level II: Localiza-
tion; Level III: Classification; Level IV: Extension; and Level IV: Prognosis.
All things considered, the final outcome of the paradigm of statistical pattern
recognition is a set of classes from which decisions can be made, closing the
data to decision (D2D) chain.
Within the paradigm of SHM as a statistical pattern recognition problem, this
chapter presents a journey through the different stages involved in the damage detec-
tion process by means of integrated SHM and unsupervised learning. Motivated by
an application-oriented philosophy, the presented theoretical concepts are illustrated
with real experimental data acquired from three different case studies of Italian
historic constructions: the Consoli Palace in Gubbio, the San Pietro bell-tower in
250 E. García-Macías and F. Ubertini
Perugia and the Basilica of Santa Maria degli Angeli in Assisi. These three construc-
tions are being currently monitored with long-term integrated SHM systems, and
the monitoring data are continuously processed for damage detection purposes in
real time and remotely in the Laboratory of Structural Dynamics of the University
of Perugia. For a detailed description of the monitoring sites, sensor layouts and
monitoring conditions, readers may refer to [6–9].
The reason for focusing on unsupervised learning techniques is their more exten-
sive application and development in the realm of historic constructions. Note that
data from every conceivable damage scenario must be available to effectively apply
supervised learning. Collecting such data is always a challenging task, through either
modelling or experiments. This is particularly critical when dealing with historic
constructions, where making copies of the structure to induce controlled damage is
simply unfeasible. Modelling approaches such as finite element modelling also repre-
sent a formidable problem, since historic constructions are usually characterized by
geometrical and material complexities. The modelling of damage itself is even more
intricate, being possible to find multiple damage mechanisms with highly nonlinear
behaviours, not to mention the frequent existence of previous damage with uncer-
tain origin, extension and evolution. A major drawback of unsupervised learning is
that it usually limits to damage detection, and damage can only be located to some
extent in some particular cases. Nonetheless, level 1 diagnostic often suffices for the
maintenance of historic constructions, whose invaluable architectural and economic
value justifies the execution of in situ inspections every time any fault is detected. In
this light, the statistical pattern recognition paradigm previously depicted in Fig. 1 is
particularized for the integrated SHM of historic constructions in Fig. 2. The general
workflow is organized into hardware and software components:
• Hardware component: The hardware part is generally composed of a hetero-
geneous sensor network, a data transmission system (wired or wireless) and a
data acquisition (DAQ) system that permanently collects the monitoring data.
Computer files containing records of certain time duration are locally processed
or sent through the Internet or another transmission system to a server or the cloud.
• Software component: This component comprises a sequence of pattern recogni-
tion algorithms designed to translate the signals acquired by the sensing interface
into condition classes (damaged or undamaged). The computer implementation
follows the statistical pattern recognition paradigm previously sketched in Fig. 1:
(i) signal processing; (ii) feature Extraction; (iii) statistical pattern recognition;
and (iv) damage detection.
The remaining of the chapter focuses on the software component from feature
extraction to damage detection. Note that the software structure furnished in Fig. 2 can
be readily scripted in a sequential fashion, enabling real-time damage assessment and
minimizing the need for the intervention of qualified technicians. Readers interested
in more specific details on the computational implementation are referred to [10].
There, the authors reported about the architecture and algorithms included in two
proprietary software programs named MOVA and MOSS. These software codes are
currently being used in the Laboratory of Structural Dynamics of the University of
Integrated SHM Systems: Damage Detection … 251
Fig. 2 General workflow of the SHM of historic constructions under the paradigm of statistical
pattern recognition
Perugia for the SHM of several historic constructions, including the case studies
presented herein.
2 Feature Extraction
In the fields of statistics and machine learning, feature selection and extraction are two
approaches for dimension reduction aimed at producing the possible most informa-
tive, distinctive and compact set of features to improve the success of data storage and
processing applications [11]. On the one hand, feature selection relates the process
of selecting the most relevant attributes of a database, overlooking redundant or non-
informative features. On the other hand, feature extraction creates low-dimensional
data sets by functional mapping of the original features. In the realm of SHM, feature
extraction relates the identification process of certain physical signatures or vari-
ables that are sensitive to the appearance of damage from the raw measurements
[12]. The effectiveness of the adopted feature extraction approach largely determines
the success of the damage assessment and the subsequent decision-making process,
since it is a fundamental axiom of SHM that sensors cannot directly measure damage
[13]. The design of the feature extraction strategy is conditioned by the operational
evaluation. This includes aspects such as the employed sensing technology, acces-
sibility limitations, the specific structural typology and engineering materials and,
most importantly, the damage mechanisms of concern in the particular monitored
252 E. García-Macías and F. Ubertini
system. The ideal outcome is a low-dimensional feature set that is highly sensitive
to the condition of the structure. Such a selection is often not a trivial task, being
necessary to combine robust signal processing techniques with a priori statistics and
engineering judgement. A wide variety of damage-sensitive features can be found
in the literature such as simple signal statistics, frequency spectra, time–frequency
analysis features, or modal properties and derived signatures (e.g. mode shape curva-
ture, modal strain energy or flexibility) [12, 14, 15]. Model updating [1, 16] and time
series modelling [17] methods represent other useful approaches for model-based and
data-driven feature extraction, respectively. In essence, these techniques construct (or
learn) a physics- or data-based model of the monitored structure by exploiting moni-
toring data. Then, the model parameters or the residuals between new experimental
data and the model predictions can be used as damage-sensitive features [12].
In the context of vibration-based SHM systems, it is common to employ modal
properties (resonant frequencies, damping ratios and mode shapes) as damage-
sensitive features. To do so, a variety of automated operational modal analysis (OMA)
techniques has been proposed in the literature [18, 19]. These techniques allow
to automatically extract estimates of modal signatures from continuously recorded
ambient vibrations of sufficient time duration. Subsequently, the time series of modal
features are obtained by so-called frequency tracking techniques. A common practice
is to define a reference list or baseline modal properties extracted from a separate
ambient vibration test (AVT). Once defined, the time series of modal features are
traced by exploiting similarities between the reference modes and all the identified
sets of modal characteristics. To illustrate this, Fig. 3a furnishes the natural frequen-
cies identified by automated covariance-driven stochastic subspace identification
(COV-SSI) of the Consoli Palace from July to December 2020. The adopted auto-
mated OMA procedure analyses sets of stabilization diagrams obtained by varying
the order of the underlying state-space model and the number of blocks in the asso-
ciated correlation Toeplitz matrix [10]. Afterwards, a sequential procedure involving
noise mode elimination and hierarchical clustering is applied to discriminate between
spurious and physical modes [18]. On this basis, a frequency tracking approach was
Fig. 3 Identified resonant frequencies by automated OMA of the Consoli Palace (a) and time series
of meaningful resonant frequencies obtained by frequency tracking (b)
Integrated SHM Systems: Damage Detection … 253
adopted to extract the time series of physically meaningful and consistently iden-
tifiable modes as shown in Fig. 3b. Specifically, a set of ten natural frequencies in
the frequency broadband of 2–10 Hz were identified in a separate AVT and selected
as baseline modal characteristics (see [6] for further details). Then, the frequency
tracking method seeks for the most similar modes amongst all the identified ones in
terms of resonant frequencies and modal assurance criterion (MAC) values. To do so,
similarity thresholds in terms of resonant frequencies and MAC values are defined
to pair modes from different data sets. In this particular case, maximum relative
differences of 5% (modes 1–9) and 10% (mode 10) are defined in terms of resonant
frequencies, while minimum MAC values of 0.8 are set up for all the modes.
It may be also useful to utilize mode shapes as damage-sensitive features. These are
usually less affected by environmental conditions compared to resonant frequencies,
although the number of sensors required for their correct characterization is usually
larger. This motivates the use of mode shape components for damage detection,
particularly during the first months of monitoring when acquired data are not enough
for cleansing time series of natural frequencies from environmental effects. Mode
shapes can be characterized using several metrics, although the use of MAC values
between the identified modes and the reference ones is the most common approach.
This statistical indicator appraises
the coherence or similarity between two arbitrary
complex mode shapes ϕ r and ϕ q as:
2
T ∗
ϕr ϕq
M AC(r, q) = T ∗ T ∗ , (1)
ϕr ϕr ϕq ϕq
where the asterisk denotes the complex conjugate. A MAC value of 1 indicates perfect
correlation between the mode shapes, while a value of 0 evidences no correlation.
The assessment of MAC values may be also used as a diagnostic tool to detect faulty
sensors, whose effect is often not so evident in terms of resonant frequencies. To
illustrate this, Fig. 4a furnishes the time series of MAC values of the modes of the
Consoli Palace previously reported in Fig. 3b. It is noted in this figure that a sudden
drop in the MAC values was found between 30 August and 20 October 2020. During
in situ inspections, it was found out that the slot of channel 6 in the acquisition module
of accelerometers had been damaged during a storm event, although it was working
in an apparently normal way. This was confirmed by a statistical analysis of the
mode shapes as shown in Fig. 4b. In this figure, the modal components (normalized
to a maximum amplitude of 1) of the 12 channels of the SHM system are analysed
during three different stages; Stage 1 17 July–29 August 2020 (before fault), Stage
2 29 August–20 October 2020 (faulty period) and Stage 3 20 October–16 November
2020 (after fault). Substantial differences in the modal components of channel 6 are
noted between Sets 1 and 2. Once the slot was repaired in Set 3, the modal component
of channel 6 went back to its correct value. This analysis allowed us to disregard
the acceleration records of channel 6 during the faulty period as a feature selection
procedure for the subsequent pattern recognition steps.
254 E. García-Macías and F. Ubertini
Fig. 4 Time series of MAC values (a) and analysis of the modal components of the fundamental
mode (b) in the Consoli Palace. The shaded areas in (b) represent the 95% confidence level of the
modal amplitudes
Fig. 5 Time series of resonant frequencies and crack displacements of the Basilica of Santa Maria
degli Angeli from 2017 until 2020
E = Y − Y. (2)
The statistical model must be trained using a set of t p feature samples defining
a baseline in-control population, often referred to as the training period. This base-
line data set must statistically represent the healthy state of the structure under
all the possible EOC, being a one-year period often adopted. When the system
remains healthy, matrix Y reproduces the part of the variance of the features driven
by EOC, while E only contains the residual variance stemming from modelling
errors. Conversely, if a certain damage develops, this only affects the data contained
in Y, while matrix Y remains unaltered. Therefore, matrix E concentrates the
damage-induced variance and is apt for being used for damage identification.
Given the widespread use of MLR and PCA as input–output and output only
statistical models for the elimination of EOC in engineering structures, respectively,
a concise overview and some exemplary numerical results are reported hereafter.
Integrated SHM Systems: Damage Detection … 257
Y = Zβ, (3)
As an example, Fig. 8 shows the results of MLR analysis of the resonant frequency
of the fourth mode of vibration (second flexural mode) of the San Pietro bell-tower
from 2018 until 2020. In the analysis, the ambient temperature measurements by
two thermocouples installed at the base of the cusp of the tower (T1 indoor and T2
outdoor) are used as predictors. Note that the model must be linear in its coefficients
contained in matrix β, while arbitrary nonlinear transformations such as exponential
or logarithmic functions can be applied to the predictors. As an example, the results of
Fig. 9 Analysis of residuals of the statistical predictions by MLR of the fundamental frequency
of the San Pietro bell-tower from 2018 until 2020. Models accounting for linear (a) and quadratic
(b) temperature values
the MLR model using both linear and quadratic temperature values are also depicted
in Fig. 8. Ideally, the obtained residuals should be normally distributed; thereby,
statistical tests assessing the Gaussianity of the residuals can be used to evaluate the
quality of a statistical model for data normalization. To this aim, a variety of statistical
metrics can be used from simple statistics to dedicated normality tests such as the
Kolmogorov–Smirnov or the Shapiro–Wilk tests. For illustrative purposes, Fig. 9
shows the analysis of the residuals obtained from the results in Fig. 8. It is noted that,
although the standard deviation of the residuals decreases when quadratic terms are
involved in the MLR, the kurtosis of the distribution of the residuals moves away
from 3 (theoretical value for an ideal Gaussian distribution) compared to the model
with linear terms, which indicates some overfitting degree.
Principal Component Analysis: PCA is a dimensionality reduction technique used
to transform databases into lower-dimensional subspaces without significant losses in
data variance. Mathematically, PCA is defined as an orthogonal linear transformation
that converts the data to a new coordinate system where the greatest variance of the
data lies on the first coordinate (called the first principal component), the second
greatest variance on the second coordinate and so on. Principal components (PCs) are
the eigenvectors of the covariance matrix of the original data; thereby, PCs constitute
an orthogonal basis of uncorrelated components. Denoting by Yn the normalized
version of matrix Y (i.e. the feature time series are transformed to have zero mean
and unit variance) and Y ∈ Rn×n its covariance matrix, the PCs can be obtained by
the eigenvalue–eigenvector decomposition of Y = YnT Yn /(N − 1) as:
Y U = US2 , (5)
Integrated SHM Systems: Damage Detection … 259
where the eigenvectors of Y are the columns of U (loading matrix) and represent
the PCs, and the eigenvalues are the diagonal terms of S2 . The PCs are sorted in
descending order according to the diagonal terms of S2 , which represent the propor-
tion of total variance in Y (i.e. tr( Y ) = n) explained by the PCs. The transformed
data matrix T ∈ R N ×n (scores matrix) is the projection of the original data in Yn
over the space spanned by the PCs in U:
T = Yn U. (6)
In the realm of SHM, PCs providing the largest contributions to the variance are
assumed to encapsulate the effects of EOC on the features matrix Y. In this light,
matrix Y can be estimated by mapping back the reduced subset of PCs onto the
original data space. Specifically, if only the first l columns of matrix U are collected
into a reduced matrix U ∈ Rn×l , matrix Yn (normalized) can be obtained as:
T
Yn = Yn UU . (7)
Fig. 10 Principal component analysis of the resonant frequencies of the San Pietro bell-tower.
Principal components (a) and variance analysis (b). Resonant frequencies are reported in normalized
values
260 E. García-Macías and F. Ubertini
Fig. 11 Comparison between experimental data and predictions by PCA of the first two resonant
frequencies of the San Pietro bell-tower. Filled red circle experimental data and filled blue circle
PCA
K
p(Y) = πk N Y|μk , k . (8)
k=1
Each component of the mixture is defined as a Gaussian distribution N Y|μk , k
with mean and covariance matrix denoted by μk and k , respectively. The parameters
π = [π1 , . . . , π K ]T in Eq. (3) are called the mixing coefficients, and they range
K
between 0 and 1 (0 ≤ πk ≤ 1) and sum to one ( k=1 πk = 1). The model parameters,
μ, and π, are fitted by minimizing the log-likelihood function:
Integrated SHM Systems: Damage Detection … 261
tp K
ln p( Y|π, μ, ) = ln πk N Y|μk , k . (9)
n=1 k=1
The maximum likelihood solution for the parameters (μ, and π) is estimated
using the iterative expectation–maximization (EM) algorithm. In the expectation (E)
step, the parameters (initial guess at the beginning) are held fixed and the posterior
probability of assigning xn to the k’s cluster is given by the so-called responsibilities
γ (z nk ) as:
πk N Yn |μk , k
γ (z nk ) = K , (10)
j=1 π j N Yn |μ j , j
tp tp
T
μnew
k = γ (z nk )Yn , new
k = γ (z nk ) Yn − μnew
k Yn − μnew
k (11)
n=1 n=1
tp
Nk
πknew = , Nk = γ (z nk ). (12)
N n=1
The log-likelihood in Eq. (3) can then be evaluated. Convergence of either the
parameters of the log-likelihood is checked, and if the criteria are not satisfied, the
process is iterated using the updated data values until the criteria are met. Once the K
clusters have been defined in the training period, new data samples can be assigned
to the cluster with the least Mahalanobis distance.
Outliers are always present to a certain degree in every feature set in SHM, being the
result of manifold sources such as noise, identification errors, faulty sensors, imper-
fect mounting of sensors and more. The presence of such outliers in the training data
set has two major pernicious effects. On the one hand, outliers affect the effectiveness
of statistical models for pattern recognition, biasing the model parameters. On the
other hand, the ability of damage classification methods based upon novelty analysis
to detect early-stage damage is also affected by outliers. As it will be discussed in
Sect. 5, novelty analysis techniques heavily rely on the determination of statistical
moments of the damage-sensitive features in the training population. The presence of
outliers will bias such moments, reducing the damage sensitivity of the classification.
262 E. García-Macías and F. Ubertini
Now take H2 another h-subset such that {d1 (i); i ∈ H2 } = {(d1 )1:d , . . . , (d1 )h:d }
where (d1 )1:d ≤ (d1 )2:d ≤ · · · ≤ (d1 )d:d are the ordered distances, and compute
μ2 and 2 based on H2 . Then, det( 2 ) ≤ det( 1 ) holds with equality if and only
if μ2 = μ1 and 2 = 1 . This process, also known as C-step, can be iteratively
repeated. If det( 2 ) = 0 or det( 2 ) = det( 1 ), the algorithm stops; otherwise,
another C-step is run yielding det( 3 ) and so on. The sequence det( 1 ) ≥ det( 3 ) ≥
det( 4 ) ≥ . . . is nonnegative and hence convergent, so there must be an index s
such that det( s ) = 0 or det( s ) = det( s−1 ). An application example is shown
in Fig. 13 for the resonant frequencies of the San Pietro bell-tower between 2019
and 2020. The dimension h of the subsets has been selected as 0.8d, and once
the algorithm converged, one per cent of the samples with the largest Mahalanobis
distances with respect to the converged sample subset is considered as outliers.
Data fusion is the integration of information from multiple sources into a new
database to enhance the observability and identifiability of a system [12]. This
new database is intended to be more informative, reducing uncertainties in the
damage identification by increasing the information completeness and enhancing the
decision-making. In general, data fusion can be achieved in three levels, namely data
level, feature level and decision level [35]. In data-level fusion, the raw measurements
from multiple sensors are directly combined before further processing. Feature-level
fusion directly operates on statistical features or signatures of heterogeneous nature.
Finally, in decision-level fusion, decision-making is performed by integrating the
decisions achieved from different data sources through particular combination rules.
Some common data fusion techniques are data registration, Bayesian probabilistic
Integrated SHM Systems: Damage Detection … 263
=
Considering sample estimations of E and 0 , the upper control limit (UCL) can be
derived as [38]:
n(k + 1)(r − 1)
UCL = F1−α;n,kr −k−n+1 (15)
k(r − 1) + 1 − n
where the term F1−α;n,kr −k−n+1 denotes (1 − α)100% confidence level of the F distri-
bution with n and kr −k−n+1 degrees of freedom. The definition of UCL has a strong
influence on the sensitivity of a control chart to detect damage; a too low value will
lead to an excessive number of false alarms, while a too high value may not be able
to detect damage. In order to optimally select UCL, as well as to compare the effec-
tiveness of different statistical models, the analysis of simulated or synthetic damage
scenarios is particularly helpful. For instance, a common approach is to include shifts
in the mean values of the time series of estimators (e.g. resonant frequency decays
obtained through a nonlinear modal analysis of a finite element model) and simu-
late control charts under different damage scenarios. As an example, Fig. 15a shows
Hotelling’s control charts of the residuals obtained using MLR and PCA for the data
normalization of the first eight resonant frequencies of the Basilica of Santa Maria
degli Angeli. Ambient temperature measurements (outdoor and indoor) are used as
statistical predictions for the MLR model, while 2 PCs explaining more than 90%
of the total variance in the resonant frequencies were used for the PCA model. To
evaluate the effectiveness of the statistical models for damage detection, an artificial
damage scenario was introduced after 1 January 2020. The artificial damage consists
in a constant shift in the time series of the fundamental frequency of 0.5% of its
mean value in the training period. Then, the quality of the damage classification
can be appraised through the analysis of the confusion matrix, including receiver
operating characteristic (ROC) curves, precision–recall curves, Youden’s indexes or
F1-scores. Figure 12 shows an example of a ROC curve and the determination of
the optimal UCL as the cut-off threshold yielding the maximum vertical distance
between the ROC curve and the diagonal curve (line of no discrimination). The area
under the ROC curves (AUC) is a simple metric often used to evaluate the quality
of the classification and compare different statistical pattern recognition approaches.
From the particular examples shown in Fig. 15a, the resulting ROC curves are shown
in Fig. 15b. It is noted that the MLR model slightly outperforms the PCA model,
with AUC values of 0.87 and 0.85, respectively. These differences are also notice-
able in the control charts in Fig. 15a, where two UCLs corresponding to confidence
levels of 99% (danger) and 95% (warning) are represented. Text inserts indicate
the proportion of data points between the 95 and 99% UCLs and above the 99%
UCL. For instance, it is noted that the proportion of outliers in the assessment period
varies from 4 to 9.48% and to 7.65% for the PCA and the MLR models, respec-
tively. Interestingly, although at first glance the PCA model may seem to provide
better damage identification reporting a higher proportion of outliers in the damaged
period, the number of false alarms in the undamaged period is larger compared to
those obtained using MLR. Global metrics such as the AUC of the ROC curves
englobe the performance of a statistical pattern recognition model for minimizing
Integrated SHM Systems: Damage Detection … 265
the number of false alarms while offering high discrimination capabilities of true
out-of-control processes.
Fig. 12 Clustering results using a GMM (a) and statistical predictions using a cluster-wise extension
of the MLR model of the fourth resonant frequency of the San Pietro bell-tower (b)
Fig. 13 Outlier identification in the database of the first six resonant frequencies in the San Pietro
bell-tower between 2019 and 2020 using the MCD method. Filled blue circle experimental data
and filled red circle outliers
266 E. García-Macías and F. Ubertini
Fig. 14 Schematic representation of damage detection using control charts (a) and quality
assessment through ROC curves (b)
Fig. 15 Hotelling’s control charts of the residuals of the first eight resonant frequencies of the
Basilica of Santa Maria degli Angeli considering MLR and PCA (2 PCs) (a), and quality assessment
in terms of ROC curves (b)
This chapter has gone through the main stages in damage detection through unsuper-
vised integrated SHM according to the most recent literature, and the effectiveness
of the presented techniques has been illustrated with real field applications. Overall,
the advanced state of development of damage detection through novelty analysis has
been evidenced, being possible to completely automate the whole process. Clustering
techniques have been presented as a general and effective approach for accommo-
dating nonlinear environmental effects, and the use of outlier removal techniques
along with the analysis of ROC curves has been shown suitable for the optimal defi-
nition of statistical process control charts. Most pressing open problems in the field
Integrated SHM Systems: Damage Detection … 267
References
1. Chen HP, Ni YQ (2018) Structural health monitoring of large civil engineering structures.
Wiley Blackwell, Hoboken
2. Ramos LF, Marques L, Lourenço PB, De Roeck G, Campos-Costa A, Roque J (2010) Moni-
toring historical masonry structures with operational modal analysis: two case studies. Mech
Syst Signal Process 24(5):1291–1305
3. Ottoni F, Blasi C (2015) Results of a 60-year monitoring system for Santa Maria del Fiore
Dome in Florence. Int J Architect Heritage 9(1):7–24
4. Farrar CR, Doebling SW, Nix DA (2001) Vibration–based structural damage identification.
Philos Trans R Soc Lond. Ser A: Math Phys Eng Sci 359(1778):131–149
5. Rytter A (1993) Vibrational based inspection of civil engineering structures, PhD thesis.
Aalborg University
6. Kita A, Venanzi I, Cavalagli N, García-Macías E, Ubertini F (2020) Enhanced continuous
dynamic monitoring of a complex monumental palace through a larger sensor network. In: XI
International conference on structural dynamics, EURODYN 2020
7. Cavalagli N, Gusella V (2015) Dome of the Basilica of Santa Maria Degli Angeli in Assisi:
static and dynamic assessment. Int J Architect Heritage 9(2):157–175
8. Cavalagli N, Botticelli L, Gioffrè M, Gusella V, Ubertini F (2017) Dynamic monitoring and
nonlinear analysis of the dome of the basilica of S. Maria degli Angeli in Assisi. In: XI
COMPDYN 2017, 1, pp 2542–2553
9. Ubertini F, Cavalagli N, Kita A, Comanducci G (2018) Assessment of a monumental masonry
bell-tower after 2016 Central Italy seismic sequence by long-term SHM. Bull Earthq Eng
16(2):775–801
10. García-Macías E, Ubertini F (2020) MOVA/MOSS: two integrated software solutions for
comprehensive structural health monitoring of structures. Mechan Syst Signal Process
143:106830
11. Entezami A, Sarmadi H, Behkamal B, Mariani S (2020) Big data analytics and structural health
monitoring: a statistical pattern recognition-based approach. Sensors 20(8):2328
12. Farrar CR, Worden K (2012) Structural health monitoring: a machine learning perspective.
Wiley, Hoboken
13. Worden K, Farrar CR, Manson G, Park G (2007) The fundamental axioms of structural health
monitoring. Proc R Soc A: Math Phys Eng Sci 463(2082):1639–1664
14. Hou R, Xia Y (2020) Review on the new development of vibration-based damage identification
for civil engineering structures: 2010–2019. J Sound Vib 115741
15. Avci O, Abdeljaber O, Kiranyaz S, Hussein M, Gabbouj M, Inman DJ (2021) A Review
of vibration-based damage detection in civil structures: from traditional methods to machine
learning and deep learning applications. Mechan Syst Signal Process 147:107077
16. García-Macías E, Venanzi I, Ubertini F (2020) Metamodel-based pattern recognition approach
for real-time identification of earthquake-induced damage in historic masonry structures.
Autom Construct 120:103389
17. Fassois SD, Sakellariou JS (2007) Time-series methods for fault detection and identification
in vibrating structures. Philos Trans R Soc A: Math Phys Eng Sci 365(1851):411–448
268 E. García-Macías and F. Ubertini
Carlo Rainieri
1 Introduction
The safety of civil engineering structures may be affected to a large extent by degra-
dation phenomena due to aging, severe environment or fatigue, as well as by damage
induced by hazardous events, such as earthquakes, fire or explosions. Thus, peri-
odic structural assessment and maintenance become critical, above all in the case of
strategic structures, such as bridges and hospitals, or densely occupied structures.
Structural survey as well as destructive and non-destructive investigations are the
conventional approaches usually applied to support the structural safety assessment.
C. Rainieri (B)
Construction Technologies Institute, National Research Council of Italy, 80146 Naples, Italy
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 269
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_13
270 C. Rainieri
However, the local nature of tests, issues related to the subjectivity of the expert
judgment and limited frequency of inspections and the costs of detailed structural
assessment have raised the interest toward alternative approaches that are able to
automatically provide information about the structural health and performance in
near real-time to a remote user. In this context, the modern structural health moni-
toring (SHM) technologies are very appealing because of the opportunity to carry
out timely damage detection. Modal-based SHM, in particular, is currently a very
active field of research even if the original idea dates back to a few decades ago.
SHM basically assumes that any change of the structure that adversely affects its
functionality or load bearing capacity can be referred to as damage [1]. In modal-
based SHM, the relationship between the physical properties of the structure (i.e.,
mass, stiffness, damping) and its modal parameters is exploited for damage detection.
In fact, changes in the structural response associated to variations of stiffness due
to cracking in concrete or masonry structures of external (soil settlements) and/or
internal restraints (loosening of tightening force in bolts of steel structures) or of mass
may suggest an anomalous structural behavior. Taking into account the relationship
between mass and stiffness on one hand, and modal properties on the other hand,
the latter are used as damage-sensitive features in the context of modal-based SHM,
and remote damage detection is based on the analysis of the variations of damage-
sensitive features defined in terms of modal parameters. In spite of the increasing
interest about this technology in civil engineering, practical applications to civil
structures are still quite limited because of some shortcomings affecting the general
applicability and reliability of the technology.
Two major approaches can be identified in the field of modal-based SHM: model-
based and data-driven techniques [1]. Model-based techniques require the solution
of an inverse problem to identify damage by updating a numerical model [2, 3]; as a
consequence, a high degree of engineering knowledge and heavy hardware and soft-
ware resources are usually needed. The main drawback is related to the uncertainties
associated to the setting and refinement of the numerical model, while the main advan-
tage with these techniques is the potential to cover all levels of SHM, from damage
detection to damage prognosis [4]. Data-driven techniques, on the contrary, set data-
driven models to characterize the operational response of the structure under healthy
conditions so that anomalies are identified from changes in the damage-sensitive
features with respect to the reference (baseline) condition [5]. They are character-
ized by lower requirements in terms of engineering knowledge and hardware and
software resources, so they better fit the needs of continuous monitoring and near
real-time damage detection. However, their effectiveness is limited to damage detec-
tion and damage location [1]. Nevertheless, these techniques currently appear as
fairly mature to make the transition from research to engineering practice.
The main obstacles to the wide application of data-driven SHM systems are the
availability of automated procedures for automatic feature extraction and robust
anomaly detectors [4]. The first element still represents a challenge in research, even
if a number of automated data processing procedures have been developed in recent
years to extract the most widely-used features (eigen frequencies and mode shapes)
from vibration measurements (the interested reader can refer to [6, 7] for an extensive
Environmental Influence on Modal Parameters … 271
review). The second element, instead, usually exploits approaches borrowed from
statistics and machine learning for anomaly detection (the interested reader can refer
to [1] for an extensive review).
Different damage-sensitive features based on modal parameters have been defined
and tested over the years [1, 8], but the possibility of getting accurate estimates even in
the presence of a few installed sensors [9] makes natural frequencies very appealing
as damage sensitive features. The most significant drawback with the use of natural
frequencies for damage detection is their high sensitivity to damage as well as to
the influence of environmental and operational variables (EOVs) [10–15]. Since
EOVs yield changes in natural frequency estimates that are often of the same order
of magnitude of those due to damage, an effective compensation of the influence
of EOVs on natural frequency estimates plays a crucial role in the development of
reliable modal-based SHM systems that are able to minimize the occurrence of false
or missed alarms.
The nonlinear relationship with the mechanical properties of materials and the
boundary conditions and the usually large thermal inertia of structures make direct
modeling of the influence of EOVs (and, in particular, temperature) on the modal
properties of a structure very challenging. This is the reason why black-box models
are usually adopted as the preferred alternative. If experimental measurements of the
EOVs influencing the selected damage features are available, a model able to map
the changes of the features with the EOVs can be set [14, 16–18]. As an example, a
linear regression model can be developed to this aim [14], even if more sophisticated
modeling techniques, such as neural networks and support-vector machines [17, 19],
have been also applied to this aim. While these alternative approaches show high
prediction capabilities, they are more computational demanding than the simple
regression models [14]. In any case, it is worth mentioning that since the developed
black-box model does not refer to physical laws, its applicability is restricted to
the structure whose monitoring data have been used for its setting. Another common
drawback with the above methods concerns the identification of the EOVs to measure
and the selection of the positions of the corresponding sensors. The latter might be
very challenging in the case of full-scale structures because the EOVs are often
not uniform and time dependent [9, 20]. If there is a single relevant environmental
variable, it might be not constant along the structure. This is the case, for instance, of
the temperature. In fact, differences occur among temperature of air, indoor surfaces
and outdoor surfaces of the monitored structure as a result of solar radiation and
thermal inertia [5, 9, 20]. When multiple EOVs [9, 14] are responsible for the changes
of the dynamic properties, multicollinearity problems associated to the selection of
regressors may also occur [14].
In order to circumvent all the above-mentioned issues, methods not requiring
EOV measurements [5, 10, 13, 15] are often adopted as an alternative to the previous
ones. The basic idea behind those methods is that the EOV influence on the selected
damage feature lies in a subspace which can be identified from the time histories of
the damage feature itself and that the effect of damage lies in a different subspace.
As a result, the effects of EOVs on one hand and damage on the other hand can
be effectively discriminated as long as the variations of the modal properties due to
272 C. Rainieri
damage are in some way orthogonal or uncorrelated to those caused by the EOVs
[10]. However, the effectiveness of such methods often depends on the dimension
of the vector of monitored features that must be large enough to make possible
the identification of the subspace which the effects of EOVs belong to. In other
words, multivariate data with enough redundancy to remove the unwanted effects
using the data correlation structure must be available, and the training data in the
healthy condition of the structure should include measurements under a wide range
of environmental or operational conditions [21].
Most of the models assume linear correlation between the measured variables or
features. However, the environmental or operational variations often cause nonlinear
effects. For example, as the temperature falls below zero, its influence on the natural
frequencies can change abruptly. This often results also in nonlinear correlation
between the features, especially if the data dimensionality is low [22]. On the other
hand, a linear model may be sufficient with a large data dimensionality, because the
correlation structure may become linear [21]. As an alternative, when the dimen-
sion of the feature vector is insufficient, nonlinear methods can be applied. They
rely on the identification of a nonlinear manifold instead of a linear subspace [23]
and, as such, they are usually more complex and computational demanding. The
potential of nonlinear approaches for the effective compensation of EOV influence
on modal parameters is confirmed by the increasing number of studies on this topic
reported in the literature. Different approaches have been tested to this aim. The
use of autoassociative neural network is described in [24], while the application of
Gaussian mixture model (GMM) for the compensation of nonlinear environmental
or operational effects without the measurement of the underlying variables is illus-
trated in [22]; in a similar way, kernel principal component analysis (kPCA) has
been applied for the compensation of EOV influence on natural frequencies [25] and
mode shapes [26].
The primary role of techniques for compensation of environmental and operational
influence on damage-sensitive features in the context of modal-based SHM motivates
the present paper, aimed at illustrating the applicative perspectives and possible
drawbacks of selected linear as well as nonlinear methods for the compensation of the
influence of EOVs on modal parameters. The theoretical background of the selected
methods is summarized in Sect. 2, while Sect. 3 illustrates and discusses relevant
results obtained from their application to real datasets collected from vibration-based
monitoring systems.
The literature review reported in the previous section remarks how the compen-
sation of the influence of EOVs on damage-sensitive features (and, in particular,
natural frequency estimates) has been carried out by applying a variety of methods,
depending also on the application. While promising results have been obtained in
many cases, large computational efforts characterize some of the applied procedures.
Environmental Influence on Modal Parameters … 273
This might represent a limitation to their extensive application in the context of data-
driven SHM, when large volumes of data become available and need to be processed
in a relatively short time for timely damage detection. As a result, in this paper,
attention is focused on selected linear as well as nonlinear procedures for compen-
sation of EOV effects on modal parameter estimates: only methods characterized
by limited computational burden even in the presence of relatively large amounts of
data (typically, some months of monitoring data) are herein considered.
This section briefly presents the theoretical background of the methods used in
this study, pointing out relevant details specifically related to removal of the EOV
influence on natural frequency estimates. Among the linear methods for compensa-
tion of EOV effects, multiple linear regression (MLR), principal component analysis
(PCA), and second-order blind identification (SOBI) are considered, while kernel
PCA (kPCA) is applied as nonlinear compensation method.
f = RB + E (1)
where f , B, and E hold the natural frequencies of the monitored modes, the coeffi-
cients of the MLR model, and the prediction errors associated to the analyzed modes,
respectively.
The objective of regression analysis is the estimation of the model coefficients
so that the best fit between model predictions and observations is obtained. They
are usually obtained through the least squares method. Regression models can be
also classified as static or dynamic. The former assume that the value of the output
variable at a given time instant depends only on the values of the predictors at the
same time instant. The latter, on the contrary, assume that the output variable at the
current time instant also depends on the values of the predictors at previous time
instants. Dynamic regression is often used to model phenomena characterized by
delayed effects, such as heating and cooling processes in the presence of relevant
thermal inertia.
274 C. Rainieri
z=Tf (2)
where the transformation matrix T is obtained from the singular values decomposi-
tion of the correlation matrix of f . Once the reference transformation matrix T ref is
computed from data referring to the healthy state of the monitored structure, Eq. (2)
can be applied again to obtain the principal components for the current dataset; they
can be remapped afterwards in the original space as follows:
f = T re f T T f (3)
where the superscript T denotes transpose. If only the first principal components,
accounting for most of the variance in the data, are retained, a dimensionality reduc-
tion is achieved by applying Eqs. (2) and (3). As a result, the factors responsible
for most of the variability in the data (such as temperature, and other significant
EOVs) are retained, while secondary effects, such as those due to random errors
in the identification of the natural frequencies, are neglected. On the analogy with
MLR, the prediction error between the observed and the remapped frequencies E
can be computed:
E= f− f (4)
Environmental Influence on Modal Parameters … 275
Such residues are insensitive to the environmental and operational factors modeled
by the retained principal components, so they can vary only as a result of damage
and, as such, they can be used as damage sensitive features.
where both the mixing matrix A and the sources s are obtained from the natural
frequency time series. Some noise e can be also present in the data. Under the
assumption of uncorrelation of the sources among them and with the noise, the
eigenvalue decomposition of the zero-lag covariance matrix allows the estimation of
the whitening matrix from the largest eigenvalues and the corresponding eigenvec-
tors, and of the noise variance from the remaining ones. Multiplying the measured
data by the whitening matrix, the whitened data are obtained and used to compute the
covariance matrix at different time lags. Applying the joint approximate diagonal-
ization [29] to a number of time-shifted matrices, the mixing matrix and the sources
are extracted from the measured data by exploiting their time coherence.
In the context of modal-based SHM, the compensation of environmental and
operational effects by SOBI starts by estimating the reference mixing matrix Aref
from a set of natural frequency estimates collected when the monitored structure
is healthy. Once additional datasets are collected, the corresponding mixing matrix
and sources are estimated; combining the sources associated to the current dataset
with Aref allows predicting how the natural frequencies vary because of the EOVs
represented by the identified sources. However, when setting the model of the envi-
ronmental variability of natural frequencies, multiplying the reference mixing matrix
by the sources estimated from the current dataset is not sufficient. Since the blind
identification problem is solved up to a scaling and a permutation, the variances of the
identified sources remain unknown, and the sources are, by convention, normalized
to have unit variance. This circumstance implies the application of an appropriate
scaling factor to avoid bias. It can be obtained as the ratio between the first singular
value of the reference mixing matrix and that of the mixing matrix estimated from the
276 C. Rainieri
current dataset [27]. Once the model f (t) of the environmental variability of natural
frequencies is set, subtracting the predicted natural frequency time series f (t) from
the measured data f (t) yields residues, which are independent of the environmental
and operational factors represented by the sources and, as such, can vary only as a
result of damage. As a consequence, they can be used as damage sensitive features
in the context of modal-based SHM.
Two relevant aspects in the practical application of SOBI for compensation of envi-
ronmental effects concern the estimation of the number of sources, which cannot be
larger than the number of analyzed natural frequency time series, and the possibility
of estimating also the pattern of the unknown EOVs [27].
The number of sources that are responsible for the variability of the measured
natural frequencies can be estimated as the number of non-zero eigenvalues of the
zero-lag covariance matrix of the measured data [27]. The influence of multiple EOVs
determining the variability of natural frequency estimates can be modeled [27]. In
addition, the estimation of the patterns of the unknown EOVs gives the opportunity
to gain a fundamental insight in the causes of that variability. In fact, the ability
of SOBI to trace the patterns of the EOVs up to a scaling and an offset provides a
fundamental informative support to the analysis of the operational response of the
monitored structure, and it may circumvent the problem of their measure when they
cannot be identified in advance or when the most appropriate sensor location cannot
be predicted. The interested reader can refer to [27] for more details.
The effect of EOVs on the modal properties of structures might be nonlinear and
it can affect the monitored features in different ways. Thus, the use of nonlinear
data processing procedures, such as kPCA, for modal-based SHM under changing
environmental conditions may be advantageous. In the context of the present study,
kPCA has been selected among other nonlinear methods for compensation of the
influence of EOVs on natural frequencies because it is robust and computationally
efficient; moreover, the type of nonlinearity has not to be explicitly defined. However,
it requires the setting of two parameters, which influence the quality of the results,
as explained in the following.
The compensation of environmental and operational effects on natural frequencies
by kPCA follows the approach described in [25]. A nonlinear relationship of the form:
f k = H 0 u k + ek (6)
is assumed between the selected damage features f k (the natural frequency esti-
mates in the present case) and the unknown EOVs uk , where the operator denotes
the nonlinear mapping of damage features onto a high-dimensional feature space,
and the residue vector ek holds the information about the misfit between measured
Environmental Influence on Modal Parameters … 277
data and model predictions; by taking advantage of the Mercer’s theorem [30], the
kernel function k:
T
k f i, f j = f i f j (7)
KV = V (8)
V = [V 1 V 2 ], V 2 ⊂ R n t ×(n t −n u ) (9)
where K is the nt × nt kernel matrix computed on the training dataset, and K test
is the nt × ns kernel matrix computed between the training samples and the entire
dataset. The quantities nt , ns , and nu denote the number of training samples, the total
number of samples, and the number of principal components in the feature space,
respectively. The superscript T denotes transpose.
The Gaussian radial basis function (RBF) is applied as kernel function in agree-
ment with [25]. It is a general-purpose kernel mostly applied in the absence of prior
knowledge [31]. Such a kernel function implicitly defines an infinite-dimensional
mapped feature space by setting a single positive parameter σ:
f i − f j 2
k f i , f j = e− 2σ 2 (11)
The misfit in the feature space can be computed once the parameters nu and σ
have been set. Their values are set in the following in agreement with the approach
reported in [25] and here summarized. The number of principal components nu is
set so that a given fraction of the normal variability (usually, 99%) is accounted for.
This ensures that the resulting value of nu is large enough to account for the normal
variability of the estimates due to EOVs but also small enough to be sensitive to
anomalies. The value of σ is instead selected as the one maximizing the Shannon’s
information entropy [32]. The interested reader can refer to [25] for more details
about the theoretical background of the method.
278 C. Rainieri
3 Applications
The first case study under analysis concerns the compensation of the environmental
influence on the fundamental frequencies of a pre-stressed steel cable: the analyzed
natural frequency time histories were collected by a monitoring system over a period
of about two months [7].
The monitored pre-stressed steel cable connected the heads of two columns
supporting one of the steel arches of the structure of the University Sports Hall
in Campobasso, Southern Italy. Figure 1 shows the steel arches and the monitored
steel cable (on the right in the picture). The cable was equipped with a monitoring
system consisting of four piezoelectric accelerometers (10 V/g sensitivity, ± 0.5 g
full-scale range), a thermocouple for temperature measurements and a 24-bit data
acquisition system.
The analyzed dataset holds the natural frequencies of the three fundamental modes
of the cable. The reduced number of monitored modes is, in principle, unfavorable
to the application of linear PCA and SOBI, because those methods require a number
of monitored modes larger than the number of (unknown) factors responsible for the
variations of the natural frequencies. However, previous studies have shown that the
natural frequency variations were due to temperature only [7]. In addition, while the
Fig. 1 Outdoor view of the University Sports Hall in Campobasso, Southern Italy, and of the
monitored pre-stressed steel cable (on the right in the picture)
Environmental Influence on Modal Parameters … 279
relationship between the natural frequencies and the cable temperature is theoretically
not linear, the experimental measurements have shown that, at least for the range of
temperatures recorded during the monitoring period, this relation is approximately
linear, so linear methods for compensation of EOV influence can be applied [27].
The details of application of MLR, PCA and SOBI to this dataset are reported
elsewhere [27]. The most relevant results are summarized here and elaborated further
to highlight advantages and drawbacks of the different approaches.
Basic statistical analyses of the data have shown that the temperature yields up to
4.5% of variation of natural frequencies with respect to the corresponding average
values [27]. Thus, the large environmental variability of natural frequencies can
prevent early damage detection [11], and its effective compensation is fundamental
to enhance the reliability of SHM.
When setting the models for prediction of the environmental variability of natural
frequencies, the dataset has been divided into two parts of equal length. The first part
has been used to train the model, while the second part has been used to assess the
predictive performance and, as a consequence, the capability of the different methods
of compensating the temperature influence on natural frequency estimates.
MLR requires explicit temperature measurements to be applied. The recorded
temperature over the considered monitoring period is plotted in Fig. 2; this shows
how the temperature range in the testing period was approximately the same as in
the training period, with the only exception of the data collected at the beginning of
the testing period when the temperature went below 0 °C for some days due to an
exceptional snowfall.
25
20
15
Temperature [°C]
10
-5
-10
Dec 05, 2014 Dec 12, 2014 Dec 19, 2014 Dec 26, 2014 Jan 02, 2015 Jan 09, 2015 Jan 16, 2015 Jan 23, 2015 Jan 30, 2015
Time
Fig. 2 Pre-stressed cable: recorded local temperature (in blue, the first half of the time series
representing the measured temperature in the training period)
280 C. Rainieri
Mode I Mode I
0.06 0.03
Data Data
Violation Violation
Center 0.02 Center
0.04
LCL/UCL LCL/UCL
0.01
0.02
0
0
-0.01
I
-0.02 -0.02
-0.03
-0.04
-0.04
-0.06
-0.05
-0.08 -0.06
500 1000 1500 2000 2500 3000 500 1000 1500 2000 2500 3000
Sample number Sample number
(a) (b)
Mode II Mode II
0.1 0.04
Data Data
Violation Violation
Center 0.03 Center
0.05 LCL/UCL LCL/UCL
0.02
0 0.01
0
-0.05
I
-0.01
-0.1 -0.02
-0.03
-0.15
-0.04
-0.2 -0.05
500 1000 1500 2000 2500 3000 500 1000 1500 2000 2500 3000
Sample number Sample number
(c) (d)
Mode III Mode III
0.15 0.04
Data Data
Violation Violation
0.1 Center 0.03 Center
LCL/UCL LCL/UCL
0.05 0.02
0 0.01
-0.05 0
I
-0.1 -0.01
-0.15 -0.02
-0.2 -0.03
-0.25 -0.04
500 1000 1500 2000 2500 3000 500 1000 1500 2000 2500 3000
Sample number Sample number
(e) (f)
Fig. 3 Pre-stressed cable: residues from MLR (left) and SOBI (right)
Environmental Influence on Modal Parameters … 281
The poorer performance of MLR with respect to SOBI (and PCA) can be ascribed
to the well-known difficulty of measuring temperature values representative of the
temperature of the cable along its full extension. In particular, direct sun radiation
effects as well as persistent shadow on portions of the cable jeopardized the repre-
sentativity of the temperature measurements and the predictive performance of the
MLR model. When SOBI (and PCA) have been applied, on the contrary, an effec-
tive compensation of temperature influence was achieved even in the presence of
just three monitored modes. This occurred because only one source (or principal
component) was sufficient to describe about 99% of the variability in the data.
(a)
(b)
Fig. 4 View of the closely spaced monitored hospital’s buildings (source maps.google.it) (a) and
the separation joint (b)
Environmental Influence on Modal Parameters … 283
f 1 [Hz] 2.3
2.2
2.1
2
May Jun Jul
Time 2016
2.4
f 2 [Hz]
2.3
2.2
2.1
May Jun Jul
Time 2016
2.5
f 3 [Hz]
2.4
2.3
2.2
May Jun Jul
Time 2016
3
f 4 [Hz]
2.8
2.6
2.4
May Jun Jul
Time 2016
Fig. 5 Hospital’s buildings—sample monitoring results: natural frequency time series of the first
four modes
buildings—histograms of (a)
0.5
monitored frequencies: mode
I (a), II (b), III (c), IV (d) 0.4
Probability
0.3
0.2
0.1
0
-3 -2 -1 0 1 2 3 4 5
Values
0.5
0.45 (b)
0.4
0.35
Probability
0.3
0.25
0.2
0.15
0.1
0.05
0
-3 -2 -1 0 1 2 3 4
Values
0.6
(c)
0.5
0.4
Probability
0.3
0.2
0.1
0
-3 -2 -1 0 1 2 3 4
Values
0.6
0.5
(d)
0.4
Probability
0.3
0.2
0.1
0
-2 -1 0 1 2 3 4 5
Values
Environmental Influence on Modal Parameters … 285
0.08
0.06
Residue - mode I (Hz)
0.04
0.02
-0.02
-0.04
Apr 2016 Jul 2016 Oct 2016 Jan 2017 Apr 2017 Jul 2017
Time
Fig. 7 Hospital’s buildings: sample residues after compensation of the environmental influence by
SOBI
1.6
1.4
1.2
1
Error norm
0.8
0.6
0.4
0.2
0
01/07/16 01/10/16 01/01/17 01/04/17
time
Fig. 8 Hospital’s buildings: error norm from application of kPCA to the entire dataset
Repeating the analysis on a subset of the entire dataset, the influence of unmea-
sured conditions in the training phase on the quality of predictions can be assessed.
Figure 9 shows the results obtained from application of kPCA to the data collected
in the period April 1st–July 2nd, 2016. The dataset has been divided into two parts of
equal number of samples, of which the first part has been used for training. Figure 9
also shows the evolution of the average local temperature, as measured by a nearby
meteorological station, in the same period.
286 C. Rainieri
1.4 30
1.2
25
Temperature [°C]
20
Error norm
0.8
0.6
15
0.4
10
0.2
0 5
01/05/16 01/06/16 01/07/16
time
Fig. 9 Hospital’s buildings: results from application of kPCA to a subset of the entire dataset;
temperature evolution over time is denoted by the red dash dot line; the horizontal dash dot and
dashed lines represent the average and maximum temperature in the training period, respectively
Comparing the temperature values in the training period with those in the testing
period allows noting that the average daily temperature (red dash dot line in Fig. 9)
is in between 5 and 22 °C—with a mean value of 15.6 °C—in the training period
(from April 1st to May 17th, 2016), and it increases up to 29 °C in the testing period
(from May 18th to July 2nd, 2016). This is probably the reason of the large values of
the misfit that can be observed in the testing period (Fig. 9) in the days characterized
by very high average daily temperature, when this goes out of the range considered
in the training period. The discrete nature of the occurrence of large misfit values,
on the other hand, seems to confirm the assumption that, when the temperature rises,
the dynamic response of the structure abruptly changes as a result of the interlocking
between the two buildings through the narrow separation joint, and if this effect was
not adequately monitored in the training phase, the results in terms of prediction
error are affected also by the occurrence of this phenomenon.
4 Conclusions
Different linear as well as nonlinear methods for the compensation of the influence
of EOVs on natural frequencies in the context of modal-based SHM have been
reviewed in the present paper and applied to real datasets collected from vibration-
based monitoring systems. Based on the achieved results, the applicative perspectives
and possible drawbacks of the selected methods, namely MLR, PCA, SOBI and
kPCA, have been assessed. The following conclusions can be drawn.
Environmental Influence on Modal Parameters … 287
References
1. Farrar CR, Worden K (2012) Structural health monitoring: a machine learning perspective.
Wiley, Chichester
2. Teughels A, De Roeck G (2005) Damage detection and parameter identification by finite
element model updating. Arch Comput Methods Eng 12(2):123–164
3. Simoen E, De Roeck G, Lombaert G (2015) Dealing with uncertainty in model updating for
damage assessment: a review. Mech Syst Signal Process 56–57:123–149
4. Deraemaeker A, Worden K (2018) A comparison of linear approaches to filter out environmental
effects in structural health monitoring. Mech Syst Signal Process 105:1–15
5. Magalhães F, Cunha A, Caetano E (2012) Vibration based structural health monitoring of an
arch bridge: from automated OMA to damage detection. Mech Syst Signal Process 28:212–228
6. Rainieri C, Fabbrocino G (2010) Automated output-only dynamic identification of civil
engineering structures. Mech Syst Signal Process 24(3):678–695
7. Rainieri C, Fabbrocino G (2015) Development and validation of an automated operational
modal analysis algorithm for vibration-based monitoring and tensile load estimation. Mech
Syst Signal Process 60–61:512–534
8. Carden EP, Fanning P (2004) Vibration based condition monitoring: a review. Struct Health
Monit 3:355–377
9. Ubertini F, Comanducci G, Cavalagli N, Pisello AL, Materazzi AL, Cotana F (2017) Environ-
mental effects on natural frequencies of the San Pietro bell tower in Perugia, Italy, and their
removal for structural performance assessment. Mech Syst Signal Process 82:307–322
10. Deraemaeker A, Reynders E, De Roeck G, Kullaa J (2008) Vibration-based structural health
monitoring using output-only measurements under changing environment. Mech Syst Signal
Process 22:34–56
11. Peeters B, De Roeck G (2001) One-year monitoring of the Z24-Bridge: environmental effects
versus damage events. Earthquake Eng Struct Dynam 30:149–171
12. Moser P, Moaveni B (2011) Environmental effects on the identified natural frequencies of the
Dowling Hall Footbridge. Mech Syst Signal Process 25:2336–2357
13. Yan A-M, Kerschen G, De Boe P, Golinval J-C (2005) Structural damage diagnosis under
varying environmental conditions—Part I: A linear analysis. Mech Syst Signal Process
19(4):847–864
14. Cross EJ, Koo KY, Brownjohn JMW, Worden K (2013) Long-term monitoring and data analysis
of the Tamar Bridge. Mech Syst Signal Process 35:16–34
288 C. Rainieri
15. Jin S-S, Cho S, Jung H-J (2015) Adaptive reference updating for vibration-based structural
health monitoring under varying environmental conditions. Comput Struct 158:211–224
16. Peeters B, Maeck J, De Roeck G (2001) Vibration-based damage detection in civil engineering:
excitation sources and temperature effects. Smart Mater Struct 10:518–527
17. Ni Y, Hua X, Fan K, Ko J (2005) Correlating modal properties with temperature using long-term
monitoring data and support vector machine technique. Eng Struct 27:1762–1773
18. Sohn H, Dzwonczyk M, Straser E, Law K, Meng T, Kiremidjian A (1998) Adaptive modeling
of environmental effects in modal parameters for damage detection in civil structures. Proc
SPIE 3325:127–138
19. Ni Y, Zhou H, Ko J (2009) Generalization capability of neural network models for temperature-
frequency correlation using monitoring data. J Struct Eng 135(10):1290–1300
20. Zhou HF, Ni YQ, Ko JM (2010) Constructing input to neural networks for modeling
temperature-caused modal variability: mean temperatures, effective temperatures, and principal
components of temperatures. Eng Struct 32:1747–1759
21. Kullaa J (2011) Distinguishing between sensor fault, structural damage, and environmental or
operational effects in structural health monitoring. Mech Syst Signal Process 25(8):2976–2989
22. Kullaa J (2014) Structural health monitoring under nonlinear environmental or operational
influences. Shock Vibr Article ID 863494
23. Yan A-M, Kerschen G, De Boe P, Golinval J-C (2005) Structural damage diagnosis under
varying environmental conditions—Part II: Local PCA for non-linear cases. Mech Syst Signal
Process 19(4):865–880
24. Sohn H, Worden K, Farrar CR (2002) Statistical damage classification under changing
environmental and operational conditions. J Intell Mater Syst Struct 13(9):561–574
25. Reynders E, Wursten G, De Roeck G (2014) Output-only structural health monitoring in
changing environmental conditions by means of nonlinear system identification. Struct Health
Monit 13(1):82–93
26. Rizzo M, Betti M, Castelli P, Spadaccini O, Vignoli A (2019) Long-term dynamic monitoring of
offshore platform for damage assessment. In: Proceedings of the 9th international conference on
structural health monitoring of intelligent infrastructure: transferring research into practice—
SHMII 2019, vol 1, pp 515–523. St. Louis, USA
27. Rainieri C, Magalhaes F, Gargaro D, Fabbrocino G, Cunha A (2019) Predicting the variability
of natural frequencies and its causes by Second-Order Blind Identification. Struct Health Monit
18(2):486–507
28. Ans B, Hérault J, Jutten C (1985) Adaptive neural architectures: detection of primitives. In:
Proceedings of COGNITIVA ’85, pp 593–597. Paris, France
29. Cardoso JF, Souloumiac A (1996) Jacobi angles for simultaneous diagonalization. SIAM J
Matrix Anal Appl 17:161–164
30. Mercer J (1909) Functions of positive and negative type, and their connection with the theory
of integral equations. Phil Trans R Soc A 209:415–446
31. Awad M, Khanna R (2015) Efficient learning machines: theories, concepts, and applications
for engineers and system designers. Apress Media
32. Shannon CE (1948) A mathematical theory of communication. Bell Syst Tech J 27:379–423
33. Gargaro D, Rainieri C, Fabbrocino G (2017) Structural and seismic monitoring of the
“Cardarelli” Hospital in Campobasso. Procedia Eng 199:936–941
Vibration-Based Damage Feature
for Long-Term Structural Health
Monitoring Under Realistic
Environmental and Operational
Variability
1 Introduction
Civil structures are subject to degradation during their long-term service, mainly due
to materials ageing or deterioration. For this reason, it is important to assess whether
the safety requirement for proper operation is always met. In most cases, human
inspections are still used to monitor the health of a structure: in these situations, a
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 289
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_14
290 F. Lucà et al.
judgement is provided by an expert after visual inspection. This approach comes with
different limitations, such as the evaluation subjectivity and the qualitative content of
the retrieved information. Furthermore, not all the components of a complex structure
can be accessible and thus inspected. Human inspections are also characterized by
being intermittent in time, with the obvious risk that if damage occurs between two
timed inspections, a prompt maintenance intervention cannot always be carried out.
The development of new sensors, networks, information technology, including
computing and storage capabilities, is increasing attention towards condition-based
maintenance, which relies on long-term continuous monitoring. In this context, a key
role is played by structural health monitoring (SHM), which is a multidisciplinary
research field aiming at the definition of automatic damage detection strategies for
civil, mechanical and aerospace structures [1]. All the possible approaches of SHM
can be roughly divided into two families: model-based and data-driven approaches
[2, 3].
In the first case, a physics-based or law-based model of the monitored structure
is required. The inverse problem technique is used to calibrate numerical models,
commonly finite element ones, and damage is detected by relating the measured
data from the structures to the estimated data from the models. However, it is not
always possible to have resources allowing one to create a complete structural model.
Furthermore, an accurate model of a real structure is made difficult by the many
uncertainties about materials, geometries and boundary conditions, making every
structure one-of-a-kind. Moreover, in case of complex structures, it is not possible
to simulate all the possible damage conditions.
For the above-mentioned reasons, the SHM community is more and more focusing
on unsupervised learning data-driven approaches, where damage is detected through
a statistical comparison between data referring to the structure in the current state
(unknown) and a reference scenario, where the structure is assumed to be in a healthy
state.
In this case, the crucial aspect is the definition of a damage-sensitive quantity
or parameter that can be directly related to the health state of the structure. This
quantity is commonly referred to as “damage feature”. A damage feature may be
defined starting from any physical quantity that can be sensitive to changes in any
structural properties. Vibration-based approaches are those most commonly adopted,
starting from the observation that changes in a structural system caused by damage
manifest themselves as changes in mass, stiffness and energy dissipation. These
changes reflect in changes of the dynamic response characteristics: for this reason,
damage features extracted from vibration data are potentially sensitive to damage
[2].
Modal parameters have been largely used in the literature to define damage
features [4]: natural frequencies, mode shapes are those most commonly used, while
modal damping is less used because of the difficulty of its identification [5]. In many
cases, modal parameters can provide a physical interpretation that can help under-
standing the nature of damage. Low-frequency modes are more easily identified,
thus very commonly adopted for damage detection algorithms. Since low-frequency
Vibration-Based Damage Feature for Long-Term Structural … 291
modes capture the global response of the structure, they are less sensitive to local
changes and this can lead to a late damage detection [6].
Another important family of approaches comes from the statistical time series
methods, where time series models are fitted to the measured vibration data and
model parameters are used to define damage features. Autoregressive (AR) [7, 8],
autoregressive with exogenous input (ARX) [9] and autoregressive with moving
average (ARMA) [10, 11] models are commonly used to fit the measured data.
These models are suitable for SHM applications because they inherently account
for uncertainty and do not depend on physical models; thus, they are suitable for
automated damage detection. In the literature, there are several applications, where
either the coefficients or the residuals of the models are used to define effective
damage features [5].
When either modal-based or autoregressive-based damage features are adopted
for unsupervised learning damage detection, one of the main challenges is repre-
sented by the fact that vibration-based approaches are influenced by the effect of
environmental and operational conditions [12]. This happens because the structural
response varies under different environmental conditions, particularly temperature,
causing changes in the material properties and boundary conditions. These variations
determine changes in the structural responses that could be higher than those caused
by damage, at least at an early stage [13].
In the literature, there are not so many applications on real structures, under
realistic environmental conditions. In this chapter, a case study has been selected
to test the effectiveness of a vibration-based damage detection approach based on
AR models, in a challenging realistic scenario. The case study consists of damage
detection of axially loaded beam-type structures (or tie-rods) which are representative
of common structural elements such as braces of truss girders, struts and ties of space
trussed structure, tie-rods of arches and vaults. In operation, both the properties of
the beam and those of the structure vary, due to the changes in the environmental
conditions, mainly temperature. Two nominally identical tie-rods have been installed
in a laboratory and continuously monitored, in a weakly controlled environment,
simulating a real permanent SHM system. The environment was characterized by
temperature variations and a number of working operations, such as the presence of
human activities and the presence of testing machines nearby. It is worth noting that
the experimental set-up allowed for the possibility of measuring the environmental
and operational variables but none of them has been intentionally controlled, to create
a realistic and challenging data set.
This chapter is organized as follows: in Sect. 2, the experimental set-up is
described, along with the chosen instrumentation. Experimental data are presented,
to show the effects of temperature on the tie-rod axial load. The adopted damage
detection approach is described in Sect. 3, where a brief theoretical background
of simple autoregressive models, model order selection and Mahalanobis squared
distance (MSD) is provided. The experimental results are presented and discussed
in Sect. 4. In the end, the conclusions are drawn in Sect. 5.
292 F. Lucà et al.
A brief introduction to the key points of the theory on vibrations in axially loaded
beams is provided to better understand the reasons behind the choice of tie-rods as an
interesting SHM case study. The reader is redirected to [14] for the complete theory of
vibrations on tensioned beams, while the analytical expressions for eigenfrequencies
and mode shapes for different support conditions can be found in [15].
Tie-rods eigenfrequencies and mode shapes are function of the following
parameters:
• Geometrical properties (cross section, momentum of inertia of the section and
length of the beam): while in most cases the properties of the section can be
directly measured, the length of the beam is difficult to be estimated because the
side clamps may create uncertain links.
• Material properties (Young’s modulus of elasticity and density): especially for
ancient structures, the characteristics of the material can be roughly estimated
unless some specific material tests are carried out.
• Boundary conditions (stiffness of the constraints): boundary conditions depend
on different variables such as the clamping systems, the material to which the
tie-rod is fixed and its deterioration.
• Loading conditions (axial load): in most cases, direct in-situ measurements are
difficult and methods for an indirect tension estimation are required. Several
attempts to obtain indirect estimate of the axial load can be found in the literature
[16–18] showing the complexity of the problem in detail.
In a real context, detecting damage through changes in modal parameters is a
complex task, since the above-mentioned parameters are subject to high uncertainty
and may also change with time. A clear example is the effect of temperature, whose
daily and long-term trends cause changes in the material properties that reflect in
daily and long-term trends in the modal parameters. For this reason, an effective
damage detection strategy must be able to separate the effects of damage from those
of environmental variations.
To investigate this topic, an experimental set-up composed of two nominally iden-
tical tie-rods made of aluminium has been designed and installed in the laboratories
of the Department of Mechanical Engineering at Politecnico di Milano. The tie-rods
are characterized by a cross section of 15 × 25 mm2 and a length L = 4000 mm.
The geometrical characteristics are representative of tie-rods that are commonly used
in civil structures to balance lateral forces of arches and vaults. The static scheme
that has been reproduced is a clamped–clamped configuration, where both ends of
the beam are fixed. The constraints have been realized with a structure composed of
three parts (see Fig. 1a).
During the tensioning procedure, where the axial load has been applied to each
of the two tie-rods, the bolted joints of clamp 1 (see Fig. 1b) were tightened-up
while those of clamp 2 (see Fig. 1c) were left loose, so that the beam was not fully
constrained along the axial direction. A load cell has been installed between clamp
Vibration-Based Damage Feature for Long-Term Structural … 293
Fig. 1 Layout of the experimental set-up (a); detail of clamp 1 (b); detail of clamps 2 and 3 (c)
2 and clamp 3, and an axial load of 8 kN has been applied using a tensioner. Each
tie-rod is instrumented with strain gauges composing a full Wheatstone bridge (thus
with temperature compensation, see Fig. 2b) that has been calibrated during this
phase to measure tension all through the monitoring phase.
After the strain gauge bridge calibration, the bolted joints in clamp 2 have been
tightened-up and the load cell between clamp 2 and clamp 3 has been removed. Right
after the tensioning procedure, a preliminary experimental modal analysis has been
carried out to identify the modal parameters. The natural frequencies related to the
first six vibration modes are reported in Table 1.
Table 1 Eigenfrequencies of
Vibration mode Eigenfrequencies [Hz]
the two tie-rods, identified
during the preliminary phase Tie-rod 1 Tie-rod 2
(October 2019) 1 13.89 14.17
2 30.98 31.64
3 53.36 54.46
4 81.82 83.24
5 116.55 118.66
6 157.95 160.40
Each tie-rod has been equipped with four piezoelectric industrial accelerometers
PCB 603C01 (Fig. 2a). These accelerometers have a sensitivity of 100 mV/g, full-
2
scale ±50g and spectral noise up to 10 Hz of 6.2 × 103 μm s2
/Hz. The choice for
industrial accelerometers has been intentionally made to test the proposed procedure
also in case of sensors with expected lower sensitivity and signal-to-noise ratio if
compared to laboratory ones, but also much cheaper, as for a real field application.
The sensor placement has been carried out taking into account the possibility to
detect the first six vibration modes. The autoMAC matrix [19] has been adopted
for many possible layouts, in the end selecting the one represented in Fig. 3, where
dimensions are expressed as a fraction of the entire tie-rod length L (i.e., 1/20 means
that the distance between the first accelerometer starting from left and the clamp 1
is L/20 = 200 mm).
Even if multiple sensors are available on each tie-rod, a strategy that relies only
on one accelerometer has been developed in this work. This choice was made taking
into account a real application, where the adoption of a high number of sensors to
monitor a single tie-rod may not always be possible, for economic reasons.
The acquisition system is composed of three NI9234 modules, with anti-aliasing
filters on-board. The sampling frequency has been set to 256 Hz on a total of 11
channels: eight IEPE channels to acquire the signals coming from the piezoelectric
accelerometers, two channels dedicated to the acquisition of the strain gauge full-
bridges, thus providing a continuous measurement of the axial load, and one channel
to acquire the signal coming from a thermocouple, thus providing a measurement
of the environmental temperature. The choice for the sampling frequency aims at
finding a compromise between the need to detect a sufficient number of modes and
the need to limit the amount of acquired data.
Axial load and temperature have been stored to have an additional information
about the experimental conditions, but, as it will be explained in the following, these
variables are not strictly required by the proposed damage detection strategy. The
trends in time of the axial load for one of the two tie-rods and temperature are reported
in Fig. 4a, for a period of approximately 5 months. As it is possible to see, both daily
and long-term trends can be observed, with daily average axial load reaching values
of 13 kN in the coldest days of the year, where the daily average temperature in the
laboratory is 11 °C.
By zooming in on a shorter time window and comparing the trends of the axial
load with that of temperature, it is possible to notice that the two variables are highly
correlated (see Fig. 4b), with the axial load changes around 700–800 N/°C.
Fig. 4 Axial load and temperature trends in the period between 8th October 2019 and 27th February
2020 (a); axial load and temperature trends in the period 11th–18th October 2019 (b)
296 F. Lucà et al.
Considering the variability of the environmental conditions for the considered case
study, a data-driven approach not relying on modal parameters has been developed.
The damage detection strategy, which combines time series analysis (AR models)
and multivariate outlier detection (Mahalanobis squared distance), is presented in
the following section.
3 Methodology
The first step of the proposed damage detection strategy requires the extraction of
AR coefficients from vibration data. AR models are time series models born in the
field of econometrics to describe time-varying processes where the output variable
linearly depends on its previous values and a stochastic term. These models can be
adopted in a SHM context to represent the dynamic response of structures. In this
work, autoregressive models are exploited to process only the output of a system,
to accurately obtain the dynamic response of the linear system, forced by a random
input.
The first order AR model AR(1) can be defined as it follows [7, 20]:
X t = φ1 X t−1 + at (1)
where X t and X t−1 are two consecutive samples of the system output, φ1 is the
autoregressive coefficient, and at is the residual term. This latter is a Gaussian process
with zero mean and variance σa2 . It can be demonstrated that the coefficient φ1 is
related to the eigenvalues of the first-order linear dynamic system [20].
For systems of order n, with n > 1, the autoregressive moving average model of
order n, or ARMA (n, n − 1), must be used [20]:
where φi are the autoregressive parameters and at is the residual at time t using the
same syntax of Eq. (1). In addition, there is the moving average (MA) part, which is
composed by the residuals at previous times (from t − 1 to t − n + 1), weighted by
the coefficients θi , with i = 1, . . . , n − 1.
Starting from the coefficients of the autoregressive part (φ j with j = 1, . . . , n),
ARMA models may be used to obtain the natural frequencies of the system [20, 21].
A critical point is that to estimate the coefficients θi related to the MA part, a
nonlinear least square approach is required [22] that can be subject to convergence
and local minima problems, and high computational cost [23]. In order to avoid such
problems, an alternative approach may be the adoption of an AR (q), with q n,
that provides and approximation of the ARMA (n, n − 1) model. In this case, the
model is defined by the following equation:
Vibration-Based Damage Feature for Long-Term Structural … 297
q
Xt = φ j X t− j + at (3)
j=1
Before discussing the details about the model order estimate, some pre-processing
choices are briefly discussed. First of all, data continuously acquired by the moni-
toring system are processed every 60s, and thus the q coefficients of the AR(q) model
will be estimated every 60s. This choice starts from the observation that the labo-
ratory environment is frequently characterized by the presence of short time events
(transients, harmonic excitations, impulsive forces). These can result in autoregres-
sive coefficients which are different from those obtained when the structure is excited
only by random noise. By evaluating the coefficients every 60s, a correct clustering
of those records considered outliers is more easily obtained, helping in a better
interpretation of results.
The time histories are initially normalized by removing their mean value and
dividing them by their standard deviation. This step is carried out to allow for a
comparison of AR models referring to different system conditions. Indeed the AR
models are only poles functions that do not depend on the amplitude of the system
input [26].
The BIC has been adopted to select the model order, which is more appropriate
to describe all the possible data acquired during the tie-rods long-time monitoring.
The BIC is defined as:
298 F. Lucà et al.
B I C(q) = q log(N ) − 2 log L φ (4)
where N is the sample size, q is the number of parameters, and φ is the set of all
parameters. L φ is the model likelihood, given the data, evaluated at the maximum
likelihood values of φ. Comparing models with the BIC involves calculating this
parameter for each model and choosing the one with the lowest BIC.
In this work, a set of 200 randomly chosen time records picked from the baseline
set (see Sect. 4) have been considered. For each record, the BIC has been evaluated
for q = 1, . . . , qmax , selecting the first order that satisfied the condition:
With qmax = 100 and toll = 0.05. The results of this process are reported in
Fig. 5.
Finally, a model order q = 45 has been selected so that all the possible considered
records could be correctly described (an example about the fitting quality will be
presented later and discussed in Sect. 4).
Every time a time history is fitted to an AR (45) model, 45 coefficients are evaluated
through a linear least square minimization. Thus, the damage feature can be defined
as it follows:
Vibration-Based Damage Feature for Long-Term Structural … 299
⎡ ⎤
φ1
⎢ φ2 ⎥
φi = ⎢
⎣ ... ⎦
⎥ (6)
φTOT
where the suffix “i” means that the coefficients refer to the time record number i.
In order to carry out unsupervised damage detection, a reference scenario must
be considered. A number Nref of time histories have been considered as a reference,
and thus, Nref observations of the damage feature may be obtained and arranged in
a matrix [φ]ref , defined as it follows:
⎡ ⎤
φT
⎢ 1T ⎥
⎢ φ2 ⎥
⎢ ⎥
⎢ ... ⎥
[φ]ref =⎢
⎢ φT
⎥
⎥ (7)
⎢ i ⎥
⎢ ... ⎥
⎣ ⎦
φ TN
ref Nref ×q
where the symbol “T” means transpose. The matrix [φ]ref is characterized by
multivariate mean vector μref (size q × 1) and covariance matrix [ref ] (size q × q).
Considering a new observation of data for which the health state is unknown, it
is again possible to fit an AR (45) model and arrange the 45 coefficients into a new
vector φ new .
The Mahalanobis squared distance (MSD) can be used to perform an outlier
detection. The MSD between φ new and the reference set [φ]ref can be calculated with
the following expression to define the damage index DI:
T
DI = MSD φ new , [φ]ref = φ new − μref [ref ]−1 φ new − μref (8)
The result of this operation provides a scalar measuring the compatibility between
the new observation φ new and the reference set [φ]ref . Thus, by setting a threshold
level, when DI is above the threshold, the new observation can be labelled as an
outlier with respect to the reference condition, possibly pointing out an eventual
damage. As from the literature, the MSD can filter out environmental effects if the
training set contains a full range of environmental conditions [27].
A threshold can be set by following the procedure described in [28], once the size
of the training set has been defined. The procedure is a Monte Carlo simulation made
up of the following steps:
1. Build a matrix of size Nref ×q, where every element is a random number sampled
from a zero mean and unit standard deviation normal distribution.
2. Calculate the MSD between each row of the matrix and the matrix itself and
store the maximum MSD value.
300 F. Lucà et al.
3. Repeat the process for a high number of trials (i.e. 1000 times), obtaining an
array containing all the largest Mahalanobis squared distances and then organize
them in a decreasing magnitude order. The critical values for 5 and 1% tests
of discordancy for a q-dimensional sample of Nref observations are then given
by the Mahalanobis squared distances, in the array above which 5 and 1% of
the trials occur. This threshold, called t1 , refers to a situation where the baseline
set also contains observations related to the outliers (i.e. damage features related
to the damaged structure are in the baseline set).
4. If the baseline does not contain outliers, the threshold must be set according to
the following expression:
2
N r e f − 1 N r e f + 1 t1
t2 = (9)
Nr e f Nr2e f − Nr e f + 1 t1
In this section, results obtained through the application of the proposed damage
detection strategy are presented. The effect of damage has been simulated on one
of the two tie-rods (named T 1 ) by the addition of concentrated masses, to alter the
dynamic response in a reversible way, according to a strategy very commonly adopted
in literature [29–31]. Two different positions have been selected: one very close to
the clamped ends (L/10) and one at the beam mid span (L/2). Also, two different
masses have been used: 3 and 5% of the total mass of the beam. While tie-rod T 1
was subject to the addition of concentrated masses, the second tie-rod (named T 2 )
was always kept in the same structural condition; i.e. no masses have been applied to
T 2 . In the following, the presented graphs will refer to sensors placed at L/3 on both
tie-rods, while some final remarks regarding results obtained with the other sensors
will be presented in the conclusions.
The data included in the analysis can be divided into 5 sets and labelled according
to the content of Table 2. It is worth recalling that every record contains the infor-
mation related to a 60s vibration data record, acquired with a sampling rate of
256 Hz.
The sets labelled as “Baseline 1” and “Baseline 2” contain data representing the
reference condition, i.e. no mass added to T 1 . Both T 1 and T 2 were only subject to
the effect of environmental and operational variations. The two baseline sets come
from two non-consecutive periods, to provide a wider range of environmental (see
Fig. 6a) and operational conditions (see Fig. 6b). Into details, for every considered
sample, Fig. 6a shows the corresponding average temperature and Fig. 6b shows
the corresponding root mean square (RMS) of the accelerations, normalized to the
maximum observed RMS. Set “V 1 ” is consecutive to “Baseline 1” and set “V 2 ” is
consecutive to “Baseline 2”; “V 1 ” and “V 2 ” are used for validation; i.e. they have
Vibration-Based Damage Feature for Long-Term Structural … 301
been recorded in the same conditions as the baseline but they are not included in
[φ]ref (see Sect. 3.2).
The first step of the analysis consists of fitting an AR (45) model to the data.
Figure 7a shows a good correspondence between the original data and the result
of the autoregressive fitting on a short time window. The point-by-point difference
between the experimental and fitted data is the residual, which is given in Fig. 7b.
The quality of the fitting is also confirmed by the quantile–quantile (QQ) plot of
the residuals in comparison with the standard normal distribution, reported in Fig. 8.
It is possible to observe a good agreement with the standard normal distribution in
the range of ± 2σ.
By fitting the data with the AR (45) model, it is possible to represent the response
of the structure with 45 coefficients every 60s: the time trend of the coefficients for
tie-rod T1 is shown in Fig. 9.
The coefficients show a daily cyclical trend that should be related to the envi-
ronmental and operational conditions [32]. Furthermore, the coefficients show a
different scatter in different portions of the plot. This behaviour is mainly related to
the variability of the excitation sources which are part of the input to the structure:
transients, impulses or harmonic forces are often present during the day, due to the
working activities taking place in the surroundings, while they are reduced or nulled
at night (cyclical trends in Fig. 6b). Consequently, coefficients are more scattered
during working hours rather than during night hours.
At this stage, this representation does not allow for the identification of different
behaviours between data sets “M5”, “M3c” and “M5c” and the baseline. For this
purpose, the damage index DI can be evaluated calculating the MSD between each
sample and the baseline set (composed by both “Baseline 1” and “Baseline 2”), as
explained in 0.
302 F. Lucà et al.
Fig. 6 Average temperature (a) and normalized RMS (b) for considered samples
Fig. 7 Comparison between the experimental data and the AR fitting (a); residuals (b)
By only considering the points referring to T 1 , it is possible to see that when the
same mass is added to the structure, its effect on the damage index is dependent on
the application position. Indeed, red points in data set “M5c” (mass added at the
centre of the tie-rod) are higher than those of data set “m5” (mass added close to
the end clamps). Furthermore, when the same position is taken into account, the
damage index is sensitive to the amount of the simulated damage, as can be seen by
comparing red points in “M5c” and “M3c” zones.
The threshold obtained by following the steps described in 0 is represented by a
black dashed line in Fig. 10. In an automatic damage detection perspective, when
the damage index is above the threshold for a given sample, an outlier is detected.
In this sense, red points referring to conditions “M5”, “M3c” and “M5c” are always
304 F. Lucà et al.
above the threshold; thus, no false negatives (associated with type II error, i.e., the
tie-rod is damaged but the index is below the threshold) are observed.
On the contrary, some false positives (associated with type I error, i.e., the tie-
rod is not damaged but the index is above the threshold) are already detected in
validation data sets “V 1 ” and “V 2 ” for both tie-rods, and in data sets “M5”, “M3c”
and “M5c” for tie-rod T 2 with blue points that are scattered around the black dashed
Vibration-Based Damage Feature for Long-Term Structural … 305
Fig. 10 Damage index calculated on AR coefficients filtered with a moving average of 60 samples
with overlap of 30 samples
Table 3 Percentages of
Correct classifications Type I error Type II error
correct classifications, type I
and type II errors for damage 90.3% 41.1% 0%
detection on T 1
5 Conclusions
Acknowledgements The research presented in this paper has been funded by the Italian National
Research Program, in a project named “Life-long optimized structural assessment and proactive
maintenance with pervasive sensing techniques” PRIN 2017.
Bibliography
1. Farrar CR, Worden K (2010) An introduction to structural health monitoring. CISM Int Cent
Mech Sci Courses Lect 520:1–17. https://doi.org/10.1007/978-3-7091-0399-9_1
2. Farrar CR, Worden K (2013) Structural health monitoring: a machine learning perspective
3. Figueiredo E, Santos A (2018) Machine learning algorithms for damage detection. Vib Tech
Damage Detect Localization Eng Struct 1–40. https://doi.org/10.1142/9781786344977_0001
4. Li H-N, Ren L, Jia Z-G, Yi T-H, Li D-S (2016) State-of-the-art in structural health monitoring
of large and complex civil infrastructures. J Civ Struct Heal Monit 6:3–16. https://doi.org/10.
1007/s13349-015-0108-9
5. Hou R, Xia Y (2021) Review on the new development of vibration-based damage identification
for civil engineering structures: 2010–2019. J Sound Vib 491:115741. https://doi.org/10.1016/
j.jsv.2020.115741
6. Doebling SW, Farrar CR, Prime MB (1998) A summary review of vibration-based damage
identification methods. Shock Vib Dig 30:91–105. https://doi.org/10.1177/058310249803
000201
7. Cox DD, Pandit SM (1993) Modal and spectrum analysis: data dependent systems in state
space. https://doi.org/10.2307/1269677
8. Gul M, Necati Catbas F (2009) Statistical pattern recognition for structural health monitoring
using time series modeling: theory and experimental verifications. Mech. Syst. Signal Process.
23:2192–2204. https://doi.org/10.1016/j.ymssp.2009.02.013
9. Roy K, Bhattacharya B, Ray-Chaudhuri S (2015) ARX model-based damage sensitive features
for structural damage localization using output-only measurements. J Sound Vib 349:99–122.
https://doi.org/10.1016/j.jsv.2015.03.038
10. Carden EP, Brownjohn JMW (2008) ARMA modelled time-series classification for structural
health monitoring of civil infrastructure. Mech Syst Signal Process 22:295–314. https://doi.
org/10.1016/j.ymssp.2007.07.003
11. Entezami A, Sarmadi H, Behkamal B, Mariani S (2020) Big data analytics and structural health
monitoring: a statistical pattern recognition-based approach. Sensors (Switzerland). 20. https://
doi.org/10.3390/s20082328
12. Figueiredo E, Park G, Farrar CR, Worden K, Figueiras J (n.d.) Machine learning algorithms
for damage detection under operational and environmental variability, Struct. Heal. Monit
13. Ni YQ, Hua XG, Fan KQ, Ko JM (2005) Correlating modal properties with temperature using
long-term monitoring data and support vector machine technique. Eng Struct 27:1762–1773.
https://doi.org/10.1016/j.engstruct.2005.02.020
Vibration-Based Damage Feature for Long-Term Structural … 307
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 309
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_15
310 D. García Cava et al.
where different damage scenarios were introduced. Additionally, the best practices of
each procedure are presented through a comprehensive discussion of their potential
advantages and drawbacks.
1 Introduction
In recent years, research on Structural Health Monitoring (SHM) seeks for solutions
to close the loop between designing, manufacturing, building, and maintaining struc-
tures driven by continuous measurements of structural data. Integrated continuous
monitoring systems allow learning from the past to decide in the present about the
structural integrity, and to predict in the future the remaining useful life as well as
improve new designs, all of this backed by experimental evidence [1]. Vibration-
based SHM (VSHM) methods, namely those SHM methods employing measured
structural vibration, have emerged as an attractive alternative with successful devel-
opments in the last decade across different fields of engineering. VSHM methods
operate on the premise that any change in a structure, including damage, introduces
a change in its vibratory behavior. These utilize Damage Sensitive Features (DSFs)
extracted from vibration responses to monitor the status of the structure of interest
[2]. This is done with the help of a baseline model, which provides a reference value
for the DSFs. The baseline model can be derived either from the physics of the struc-
ture, from measured data, or a combination thereof. This chapter is focused on the
data-driven VSHM approach, namely methods based solely on data measured from
the structure. This does not mean that subsequent modeling is necessarily performed
by disregarding the physics of the structure. Instead, physics often drives the selec-
tion of models and may also be used as constraints to build more realistic data-driven
representations [1].
One of the main challenges in the development of VSHM methods relates to the
sensitivity of DSFs to Environmental and Operational Variabilities (EOV), which
hampers the algorithm’s sensitivity to damage [3]. Indeed, damage in the struc-
tural components introduces local changes in its physical properties and/or boundary
conditions. In turn, these changes also modify the structure’s vibrational response,
eventually permeating into the DSFs. However, other types of events unrelated to
damage also influence the properties of the structure, introducing benign changes
in DSFs, which are easily confused with damage. These events can be grouped into
three main categories:
(1) Variations imposed by the operational conditions of the structure, which
modify the way the structure operates over an extended period. Examples
include set-point variables, such as the rotational speed of a rotor system [4],
or the payload of a vehicle [5]. In this case, the structural dynamics exhibit
On Explicit and Implicit Procedures to Mitigate Environmental … 311
includes the results and discussions where both of the procedures are presented.
Section 7 provides a comprehensive discussion and best practices for both methods,
while Sect. 8 summarizes the main conclusions.
2 Problem Statement
Fig. 2 Graphical
representation of different
data normalization methods
in VSHM
which is useful for model selection but also for damage diagnosis [22]. Perhaps the
most recognized method in this group corresponds to the Relevance Vector Machine
(RVM) [23], which may be deemed as a probabilistic interpretation of SVR, while
equally powerful versions of Bayesian nonlinear regressions are also available. In
addition, Gaussian Process Regressions (GPR) comprise a complete probabilistic
perspective into the regression problem, where predictions are calculated on the
basis of their conditional probability given the training data, under the assumption
of a joint multivariate normal probability distribution.
Another important type of stochastic model is formed by ensemble methods,
where predictions are based on a pool of models obtained after adding small variations
in the training data. These variations introduce some sort of randomness into the
obtained models, which ultimately is reflected in the model predictions. For instance,
bagging, or bootstrap aggregating, involves training a pool of models after randomly
sampling different training subsets from the complete training data set. Boosting
and stacking are other methods used to reach a similar effect. Models with small
variations, which produce a variety of different predictions, comprise the resulting
model pool. The ensemble prediction then corresponds to some sort of consensus
achieved among the individual model predictions. In addition, the variations in the
individual predictions also provide a measure of the uncertainty in the ensemble
prediction, in a similar way as the posterior predictive distribution does for Bayesian
regressions.
Implicit procedures rely on capturing the behavior of the DSFs in a reference state
of the structure, by recognizing all its different possible patterns within the target
state under EOV. Once the patterns are learnt, new observations are assessed by their
similarity to the reference patterns. These procedures are performed exclusively on
the basis of DSFs, and thus are referred to as effect-only approaches. In turn, implicit
procedures may follow any of the ensuing philosophies:
Projection methods. In these methods, DSFs are projected into a space where the
influence of EOV might be removed or mitigated to some extent [24]. Factor anal-
ysis, Principal Component Analysis (PCA), Singular Spectrum Analysis and their
non-linear counterparts are examples of these methods [25–29]. Upon transforma-
tion, the obtained variables are sorted according to their sensitivity to EOV. In PCA,
the principal components with larger principal values are presumed to be the most
affected by EOV, and thus are removed. However, there are not specific criteria that
can help guiding the selection of the number of features to remove. More impor-
tantly, since these projections are calculated in an unsupervised or semi-supervised
manner (with only information about the healthy state), it is difficult to guarantee
the separability of EOV from damage effects in the transformed space. Therefore,
it is possible that the variables judged as EOV sensitive may also carry information
316 D. García Cava et al.
on damage characteristics. Hence, while these methods may reduce the sensitivity
to EOV, it may also significantly diminish the damage detection accuracy.
Time-series methods. In these methods, DSFs are treated as time series, whose
structure are modeled to capture non-stationary temporal patterns (trends or season-
ality) taking part of the normal state. Cointegration analysis is one of the best known
method for this task. It is based on the representation of DSFs as a modified vector
autoregressive model referred to as Vector Error Correction Model (VECM). Some
of the VECM properties indicate the presence of trends or seasonality patterns in
the time-series data, and upon detection, the VECM can be used to project out those
patterns [30, 31]. As a result, a stationary residual time series is obtained, which is
expected to be free of the influence of EOV. Extensions of the original cointegra-
tion analysis algorithm have been proposed to deal with heteroscedastic variables
(variables with time dependent variance) [32], and changes of regime in the DSFs
[33].
Manifold learning methods. As explained in the previous section, DSFs hold
functional dependencies with external variables associated with EOV. Explicit
methods attempt at capturing these dependencies with the aid of measurable EOPs.
Manifold learning methods, on the other hand, try to do so without the availability
of EOPs and instead attempt at reconstructing the manifold in which the DSFs are
embedded, solely on their multivariate characteristics. Although numerous methods
can be applied for this purpose, some of the most recognizable in the VSHM context
are non-linear PCA in its different versions (locally linear embeddings [34], kernel
PCA [35], ANN [36], deep NN [37, 38]), and different clustering methods [39–41].
Metric learning methods follow a similar philosophy, but instead of directly learning
the underlying manifold, metric learning methods attempt at capturing the patterns
of the distances measured in the manifold [8, 36].
The essence of explicit procedures consists of capturing the variations in DSFs with
the help of measured EOPs. In this section, the explicit approach is illustrated by
means of a multivariate nonlinear regression, which represents the DSF vector α p as
a deterministic function of EOPs as shown next:
α p = WT · f ξ p + u p , u p ∼ N (0, u ) (1)
f ξ p = f 1 ξ p,1 ⊗ f 2 ξ p,2 ⊗ . . . ⊗ f l ξ p,l ⊗ . . . ⊗ f L ξ p,L (2)
On Explicit and Implicit Procedures to Mitigate Environmental … 317
where W ∈ R L×K is the coefficient matrix of the regression model, f ξ p is a
multivariate functional basis of the EOPs constructed by Kronecker products of
univariate basis functions according to Eq. (2), and u p is a zero mean normal and
independently distributed random vector with covariance u , which accounts for the
ξ p,l represents
variability not attributable to EOVs. In Eq. (2), the individual l-EOP
affecting to the p-observation, so that ξ p = ξ p,1 · · · ξ p,L ∈ R L denotes a vector
with all the EOPs, and f l (·) represents the functional basis associated to the l-EOP
with length determined by product of the number of the basis assigned to each EOP.
Given a set of input EOPs X = ξ 1 ξ 2 . . . ξ p . . . ξ PT ∈ R L×PT and their corre-
sponding DSFs A = α 1 α 2 . . . α p . . . α PT ∈ R K ×PT , a standard least squares
method is used to estimate the coefficient matrix of the regression model and the
innovation covariance matrix, as follows:
−1
Ŵ = F(X)FT (X) F(X)AT (3)
1 T
ˆu =
A − ŴT F(X) A − ŴT F(X) (4)
PT
where F(X) = f ξ 1 f ξ 2 . . . f ξ p . . . f ξ PT ∈ R L×PT is the overall
T
training multivariate functional basis and (·) is the transpose.
To maximize the generalization performance and avoid overfitting, a Leave-One-
Out Cross Validation (LOOCV) method is used to assess different model structures.
The LOOCV method uses a single observation from the training set to calculate the
prediction error (validation), while the remaining ones are used for model estimation
(training). The process is repeated until all the instances in the training set are used
for validation. In the case of the multivariate non-linear regression, the LOOCV
error can be calculated efficiently with the help of the hat matrix, based on a single
calculation on the complete training set, as explained in [20, Sec.7.10].
The compensated DSF vector. An estimate of the DSFs can be obtained based
on the obtained estimates of the coefficient matrix. Then, it is possible to obtain a
corrected DSF α̃ p by subtracting the estimated DSF vector α̂ p = ŴT · f (ξ ) from
the original DSF α p as shown in Eq. (5).
α̃ p = α p − α̂ p = α p − ŴT · f (ξ ) (5)
The corrected DSFs are used to assess the current state of the structure by a novelty
index method obtained with the squared Mahalanobis distance (MD):
318 D. García Cava et al.
d 2 α̃ p , ˆ u−1 α̃ p
ˆ u = α̃ Tp (6)
where α̃ p is the corrected DSF vector observation from Eq. (5) and ˆ u corresponds
to the covariance matrix estimate in Eq. (4). A basic damage diagnosis method can be
made by comparing the above defined MD with a threshold ϑ > 0. A new observation
is then classified according to:
ˆ u ≤ ϑ → Healthy
H1 : d 2 α̃ p , (7a)
ˆ u > ϑ → Damaged
H2 : d 2 α̃ p , (7b)
L R
g2 f p , Z = σ z (2)
pl h zlr(1) frp + zr(1)
0 + z (2)
p0 (9)
l=1 r =1
where g2 f p , Z represents the estimated MD for DSFs f p . The output layer consists
of the weights z (2) (2) (1) (1)
pl and bias z p0 , while z lr and z r 0 indicate the weights and the bias
for the hidden layer. A hyperbolic sigmoid activation function σ (·) is applied in
the hidden layer and a linear identity transfer function h(·) for the output layer. The
weights Z are trained by minimizing the least-square solution for E D which is defined
as follow:
On Explicit and Implicit Procedures to Mitigate Environmental … 319
PT
1 2 2
ED = d f p , F − g2 f p , Z (10)
PT p=1
where E D is the mean squared error between the calculated and the predicted ANN
estimate of the MD-based novelty index.
The ANN is trained on observations from the healthy structure (i.e.PT observations),
subsequently, damage detection is pursued upon calculation of a new novelty index.
The proposed
ANN-based novelty index is calculated
by the relationship between
MD d2 f p , F and the estimated MD g2 f p , Z by the ANN and is defined as
follows:
g2 f p , Z + ε
d̂ p = log 2
2
(11)
d f p, F + ε
H1 : d̂ 2p ≤ θ → Healthy (12a)
The presented study is based on the data collected during the measurement campaign
conducted on a Vestas V27 wind turbine during winter 2014–2015. Though its design
is relatively old, this 225 kW rated power, 27 m rotor diameter, upwind, pitch regu-
lated, horizontal axis wind turbine is representative of many modern wind turbines.
In contrast to modern wind turbines, its blades are relatively stiff, and it has only two
speed-regimes: 32 and 43 rpm.
320 D. García Cava et al.
a) b)
Fig. 3 a) Picture of the V27 wind turbine with the measurement equipment. b) Schematic
representation of the instrumented wind turbine blade
The vibration data was collected from a single instrumented blade, which was
excited by an electro-mechanical actuator. The actuator was mounted on the outer
surface of the blade, about 1 m from the blade root and resembles an automatic
hammer that, due to an electrical pulse, impacts the blade. The actuator was
programmed to impact the blade every 5 min. The vibrational response of the blade
was picked up by an array of 11 monoaxial accelerometers mounted in three rows
along the blade, about 2 m from each other (Fig. 3). The synchronously sampled
vibration signals were collected by the data acquisition system located inside the
rotor spinner. Simultaneously, the data from the pitch sensor and two DC accelerom-
eters were recorded. The latter allowed us to estimate the rotor azimuth angle and
rotor rotational speed. From the data acquisition system, the data was transmitted
wirelessly to a measurement computer located inside the wind turbine tower and
stored in its hard drive as raw data time series.
Simultaneously (but not synchronously), the weather data (temperature, wind
speed, wind direction, etc.) from the nearby weather mast was collected.
The measurement campaign started on Nov. 28, 2014, and finished on Mar. 12,
2015. During the campaign, the wind turbine was governed by its control system,
and it was in one of its three operational modes: idle, 32 rpm, and 43 rpm. During
the 104 days of the campaign, the structure of the instrumented blade was taken to
five different states including: the intact blade, three artificially induced damages in
the form of an opening of the trailing edge of 15, 30 and 45 cm, and the repaired
blade. A more detailed description of the test campaign can be found in [22].
Data set. To illustrate how the selected explicit and implicit methods operate in a
simple and comprehensive form, data corresponding only to the 43 rpm operational
condition is considered. Additionally, the repaired state of the blade is selected as the
reference (healthy) state, due to the longer data availability and further exposition
to EOV. Also, only the accelerometers located on the leading and trailing edges are
considered, leading to a number of M = 8 sensors. The number of observations is
detailed in Table 1. The observations measured on the repaired blade (considered as
healthy) are randomly divided into two groups. Then, 3456 observations are used to
build the reference state (training), while 900 observations are used to evaluate the
model performance (validation). Although the data set is the same in both methods,
On Explicit and Implicit Procedures to Mitigate Environmental … 321
the DSFs extracted from the vibration responses are different. This is explained in
the sections below.
a) b)
Fig. 4 Explicit method. Distribution of the feature variable a) Original α1 and corrected α̃1 against
temperature, b) Original α3 and corrected α̃3 against actuator standard deviation
the DSF vector to all the EOPs. Once the multivariate regression models are defined,
DSF vectors are easily estimated by projecting the EOPs data as shown in Eq. (5).
These new estimated DSFs describe the expected dependencies with their corre-
sponding EOP measurements. To this end, corrected DSF vectors are then calculated
by subtracting the estimated DSF from the actual DSF vectors. Thus, now the new
DSFs are corrected upon consideration of the EOP influence. This can be observed in
Fig. 4a where the corrected feature variable α̃1 seems to be released from its temper-
ature dependency. It can clearly be seen that the trend on α1 previously observed in
Fig. 4a is now removed in the new feature. Additionally, this correction can be also
observed on α̃3 where the effects of the actuator hit is now corrected (see Fig. 4b).
Assessing the damage detectability. A comparison between the damage
detectability obtained when using the original features α p (before compensation)
and the corrected DSF vectors α̃ p is now presented. This is conducted by imple-
menting the outlier analysis, described in Sect. 3.2, on the DSF vectors obtained
before and after compensation. The damage detection threshold is defined as the
value where 98% of the training data (i.e. PT = 3456 observations) lies according to
a chi-squared distribution with degrees of freedom equal to the length of the corre-
sponding DSF vector. Figure 5 presents the obtained control charts in both DSF
vectors (i.e., before and after the compensation) based only on the first PC from each
sensor. Therefore, the dimension of the DSF vectors is K = 8 with (i.e. Nr = 1
and M = 8). In both cases, the damages at different levels are detectable. However,
several observations remain undetected when the uncorrected DSF vectors are used
(see Fig. 5a).
Otherwise, after correction of the DSFs, an increase in the distance between the
healthy and damaged cases is observed, while all the damaged cases appear outside
of the selected threshold. Nonetheless, the false positive rate slightly increases within
the testing set after compensation.
On Explicit and Implicit Procedures to Mitigate Environmental … 323
a) b)
Fig. 5 Explicit method. Damage detection control chart for the wind turbine blade with only the
first principal vector retained (K = 8) for a) non-corrected DSF vectors α p and b) corrected DSF
vectors αp
A similar analysis is conducted using the first three principal vectors on each
sensor (i.e. K = 24 with Nr = 3 and M = 8). Before correction (Fig. 6a), some
of the damaged cases remain undetected, while after correction (Fig. 6b), all the
damaged observations are clearly detected on expense of a small increment of false
positives. Comparing the results obtained with K = 8 (Fig. 5b) and K = 24 (Fig. 6b),
it is observed that all damages can be successfully detected in both cases, while
a slight increment of false positives is observed when more principal vectors are
retained (K = 24). In addition, it is evident that in the present case the first principal
component is critical towards damage detectability.
a) b)
Fig. 6 Explicit method. Damage detection control chart for the wind turbine blade with only the
first principal vector retained (K = 24) for a) non-corrected DSF vectors α p and b) corrected DSF
vectors αp
324 D. García Cava et al.
Moreover, for both analyses, it could be seen that when the DSF vectors were
compensated, some observations from the testing set move above the threshold, thus
increasing the number of false positives. One reason for this could be that the EOVs
experienced for these observations did not exist in the training set and, as such, are not
modelled well in the multivariate regression model. However, as the distance between
the damage and healthy increases, it opens the possibility to raise the threshold and
reduce the number of outliers without affecting the detection rate of the damage.
In this section, the results of the application of an implicit procedure implemented via
the metric learning approach described in Sect. 4 are presented. As in the application
of the explicit method, the data is split according to Table 1. This includes using 80%
of the data for training and defining the threshold for outliers. The DSF vector in this
application example relates the correlation matrix calculated among the considered
accelerometers. The DSF vector f p is then formed by organizing the elements of
the upper diagonal of the acceleration correlation matrix into a single column vector.
For more details about the DSF see [44].
The novelty index obtained by calculating the MD defined in Eq. (8), with mean
and covariance calculated on the elements of the training set (PT = 3456) is observed
in Fig. 7. The threshold corresponding to a 2% false alarm rate is also displayed. As
the DSFs are not pre-processed to mitigate the effects of EOPs, a large number of
instances from the healthy state of the blade, both in the training and validation sets,
are overlapping with the instances from the damaged state.
First, the novelty index when solely using the MD with PT = 3456 is observed
in Fig. 7. As the DSFs are not pre-processed to mitigate the effects of EOPs, the
majority of the 2% allowed false alarms are overlapping with the 15 cm damage.
a) b)
Fig. 8 Implicit method. a) Comparison of the calculated and ANN-predicted MDs of the
correlation-based DSF vector. b) Damage detection control chart for the wind turbine blade based
on the predicted error between the calculated and ANN-predicted MDs
This section is devoted to present the best practices of each procedure, explicit and
implicit, through a comprehensive discussion of their potential advantages and draw-
backs. However, we do not aim at making a direct performance comparison between
326 D. García Cava et al.
the methods discussed in the sections above. Instead, we aim to discuss how these
procedures could be integrated within a VSHM methodology.
Explicit procedures. Additional EOP measurements are required to reconstruct
the functional relationship with corresponding DSFs. Consequently, this type of
procedures requires large datasets of different natures, which in turn calls for extra
computational resources and accurate merging/synchronization of data streams.
Nonetheless, in many structures of interest this information is already available.
Explicit procedures introduce a controlled environment aided by EOV information,
facilitating a better interpretability of the results is achieved. Thereupon, a better
understanding of the monitored structure might be achieved. Also, since DSFs are
associated to EOPs, then it is easy to trace back the cause for any variation in DSFs.
Regardless of their increased complexity, regularized and stochastic regressions
are preferable for the EOV compensation task, due to their increased robustness
towards overfitting and reduced sensitivity to outliers. Moreover, stochastic models
also facilitate the calculation of uncertainty bounds and model comparison [22].
When constructing regressions, it is useful to be able to identify the variables that
effectively influence the output variables, while rejecting those that are irrelevant or
correlated to others. Irrelevant and correlated variables may, in the best case, add
unnecessary complexity to the model, while in the worst case act as confounding
variables, harming the performance of the regression model. In the present context,
this problem has a special significance as the number of potentially influencing vari-
ables can be very large. This is especially true for structures equipped with Super-
visory Control and Data Acquisition (SCADA) systems, which provide hundreds of
different variables and descriptive statistics, many of them of a similar nature (e.g.,
temperatures at different locations). Hence, a good practice is to preprocess the input
data to produce a set of independent variables to be used as inputs in the regression
model. To this end, Principal Component, Factor, or Independent Component Anal-
ysis as well as Singular Spectrum Analysis are amenable methods. Nonetheless, such
transformations come at the price of a reduced physical interpretation of the input
variables. A subsequent step is to remove inputs within the model optimization loop,
aided by sensitivity measures and related techniques [43].
Some types of structures will exhibit large variations in the dynamic characteris-
tics while changing between operational regimes. For instance, wind turbines have
dramatic changes on their dynamics when changing from idling condition to power
production mode. In those cases, independent regression models should be built for
each one of the identified operational modes.
Other aspects to consider relate to the multivariate nature of DSFs. Nonetheless,
most regression methods are designed for a single output, so typically, indepen-
dent regression models are fit to single DSFs. While the predictive accuracy of this
approach is not harmed in any way, the efficiency of the optimization process and
of the obtained model is sub-optimal. Otherwise, considering the multivariate nature
of DSFs will provide a more efficient model, which at the same time, can provide
further understanding into the problem. For instance, collinearity between predicted
On Explicit and Implicit Procedures to Mitigate Environmental … 327
variables or a latent structure in the DSFs can be unveiled. For this purpose, polyno-
mial regressions in a deterministic or Bayesian context can be easily upgraded to the
multiple output case [22], while similar extensions for GPRs are also available [45].
In practice, explicit procedures need data from EOPs and their associated DSFs
from a representative set of environmental and operational conditions to achieve accu-
rate models. This implies large instrumentation equipment, an efficient synchroniza-
tion between different measurement systems (e.g., SCADA and vibration measure-
ment systems), and a sufficiently long initial monitoring period. Overlooking any of
these aspects will most certainly impair the achievable accuracy of the final regression
model.
Implicit procedures. From one hand, implicit procedures are conceptually simpler,
as no extra variables beyond DSFs are required, thus eliminating the concerns for
measuring, synchronization and merging of data streams. On the other hand, this lack
of information is translated either into a reduced performance or higher computational
complexity. Moreover, if extra variables are already available (as is often the case),
why not use this information to potentially enhance the sensitivity to damage?
Implicit procedures learn patterns in DSFs without any contextual or prior knowl-
edge. This implies a potential reduction in damage sensitivity, as the manifolds
obtained in healthy and damages may overlap, especially for early damage stages.
However, the overlapping sections could be resolved with the help of any EOPs.
In practice, implicit procedures need large sets of DSFs to gain as much infor-
mation on the variability of DSFs as possible. In addition, when dimensionality
reduction is involved, it is important to select the correct dimensionality to preserve
maximum sensitivity to damages without adding unwanted confounding features.
These considerations might have negative consequences if overlooked, and they are
generally case-dependent based on the complexity of the problem and quality of the
data measured.
In general. A large number of observations are required to learn the complex
patterns of DSFs. If the dimensionality of DSFs is also large (as in the case of spectra),
then an increased computational complexity is attained. Dimensionality reduction
methods could be integrated to alleviate this issue, but at the same time introduce the
problem of resolving the amount of information to keep/discard without affecting the
sensitivity to damage. It is important to build the reference state(s) by considering a
representative subset of EOVs to the best possible extent. Otherwise, there will be a
higher risk of false positives stemming from unexpected events not accounted for at
the training phase. While often overlooked, the selection of the model structure has
a massive impact on the final performance of the method. This includes the selection
of the model type, function family, structural complexity (or model order), and so
on. A wrong model selection will lead to a bias in the damage indices, as the trends
in the data will be misrepresented.
328 D. García Cava et al.
8 Conclusions
This chapter has been devoted to the contextualization and discussion of different
procedures for mitigation of EOV in damage detection. The core of this discussion has
been focused on two main philosophies: explicit and implicit procedures. A compre-
hensive review has been presented in both cases, aiming to explain the necessity and
scope of each one of them. In short, explicit procedures consider information about
EOV to reconstruct the cause-effect relationship with corresponding DSFs. Mean-
while, implicit procedures do not use a direct information of the EOVs but attempt
at learning the patterns characterizing DSFs or damage indices in a baseline state.
These procedures have been illustrated in the context of a wind turbine in-operation
where different damage scenarios were introduced. The procedures were exemplified
via two machine learning methods: a multivariate nonlinear regression to illustrate
the explicit procedure; and a metric learning method assisted by an artificial neural
network to illustrate the implicit procedure. A detailed discussion of the perfor-
mance of each method has been presented, where their capabilities for accounting
and mitigating EOVs for damage detection in wind turbine blades in-operation are
described. Finally, the advantages and drawbacks of each procedure are also detailed
in order to present a comprehensive best practice guideline. In general, both proce-
dures will require a large number of observations, whether the aim is to learn the
complex patterns of the DSFs or build the relationship with their corresponding
EOVs for constructing a robust SHM methodology. On the other hand, both proce-
dures provide benefits for their implementation. Implicit procedures might be simpler
as they require only information from the DSFs but explicit procedures introduce a
controlled environment aided by EOV information. In terms of future development,
we believe that a holistic approach comprising explicit and implicit compensation
procedures should be considered, as it will facilitate the mitigation of measurable
and unmeasurable effects, while introducing interpretability of the outcomes in a
SHM framework.
Acknowledgements Mr. Callum Roberts and Dr. David García Cava would like to acknowledge
Carnegie Trust for the Caledonian Ph.D. Scholarship (grant reference number: PHD007700).
References
1. Worden K, Cross EJ, Barthorpe RJ, Wagg DJ, Gardner P (2020) On digital twins, mirrors,
and virtualizations: frameworks for model verification and validation. ASCE-ASME J Risk
Uncertainty Eng Syst Part B: Mech Eng 6(3)
2. Farrar CR, Worden K (2007) An introduction to structural health monitoring. Philosophical
Transictions of the Royal society A. Math Phys Eng Sci 365(1851):303–15
3. Sohn H (2007) Effects of environmental and operational variability on structural health moni-
toring. Philosophical Transictions of the Royal society A. Math Phys Eng Sci 365(1851):539–60
4. Avendaño-Valencia LD, Chatzi EN (2020) Multivariate GP-VAR models for robust structural
identification under operational variability. Probab Eng Mech 60
On Explicit and Implicit Procedures to Mitigate Environmental … 329
5. Aravanis TCI, Sakellariou JS, Fassois SD (2020) A stochastic functional model based method
for random vibration based robust fault detection under variable non–measurable operating
conditions with application to railway vehicle suspensions. J Sound and Vib 466
6. Zhu XQ, Law SS (2015) Structural health monitoring based on vehicle-bridge interaction:
accomplishments and challenges. Adv Struct Eng 18 (12)
7. Bogoevska S, Spiridonakos M, Chatzi E, Dumova-Jovanoska E, Höffer R (2017) A data-driven
diagnostic framework for wind turbine structures: a holistic approach. Sensors 17(4)
8. Movsessian A, Qadri BA, Tcherniak D, Garcia Cava D, Ulriksen MD (2021) Mitigation of
environmental variabilities in damage detection: a comparative study of two semi-supervised
approaches. In: Proceedings of the International Conference on Structural Dynamic 1
9. Yuen KV, Kuok SC (2010) Ambient interference in long-term monitoring of buildings. Eng
Struct 32(8)
10. Hansen MH (2007) Aeroelastic instability problems for wind turbines. Wind Energy 10
11. Farrar CR, Worden K (2012) Structural health monitoring: a machine learning perspective
12. Aravanis T-CI, Sakellariou JS, Fassois SD (2020) A stochastic functional model based method
for random vibration based robust fault detection under variable non–measurable operating
conditions with application to railway vehicle suspensions. J Sound Vib 466
13. Sohn H, Worden K, Farrar CR (2002) Statistical damage classification under changing
environmental and operational conditions. J Intell Mater Syst Struct 13(9)
14. Avendaño-Valencia LD, Chatzi EN, Koo KY, Brownjohn JMW (2017) Gaussian process time-
series models for structures under operational variability. Fron Built Environ 3
15. Worden K, Sohn, H. and Farrar, CR (2002) Novelty detection in a changing environment:
regression and interpolation approaches. J Sound Vib 258(4)
16. Xia Y, Hao H, Zanardo G, Deeks (2006) A long term vibration monitoring of an RC slab:
temperature and humidity effect. Eng Struct 28(3)
17. Schlechtingen M, Ferreira Santos I (2011) Comparative analysis of neural network and regres-
sion based condition monitoring approaches for wind turbine fault detection. Mech Syst Signal
Process 25(5):5
18. Kopsaftopoulos F, Nardari R, Li YH, Chang FK (2018) A stochastic global identification
framework for aerospace structures operating under varying flight states. Mech Syst Signal
Process 98
19. Spiridonakos MD, Chatzi EN (2015) Metamodeling of dynamic nonlinear structural systems
through polynomial chaos NARX models. Comput Struct 157
20. Hastie T, Tibshirani R, Friedman J (2009) Elements of Statistical Learning. Springer, Second
21. Laory I, Trinh TN, Smith IFC, Brownjohn JMW (2014) Methodologies for predicting natural
frequency variation of a suspension bridge. Eng Struct 80
22. Avendaño-Valencia LD, Chatzi EN, Tcherniak D (2020) Gaussian process models for mitiga-
tion of operational variability in the structural health monitoring of wind turbines. Mech Syst
Signal Process 142
23. Tipping ME (2001) Sparse bayesian learning and the relevance vector machine. J Mach Learn
Res 1(3)
24. Gómez González A, Fassois SD (2016) A supervised vibration-based statistical methodology
for damage detection under varying environmental conditions & its laboratory assessment with
a scale wind turbine blade. J Sound Vib 366
25. Deraemaeker A, Reynders E, De Roeck G, Kullaa J (2008) Vibration-based structural health
monitoring using output-only measurements under changing environment. Mech Syst Signal
Process 22(1)
26. Deraemaeker A, Worden K (2018) A comparison of linear approaches to filter out environmental
effects in structural health monitoring. Mech Syst Signal Process 105
27. García D, Tcherniak D (2019) An experimental study on the data-driven structural health
monitoring of large wind turbine blades using a single accelerometer and actuator. Mech Syst
Signal Process 127
28. Yan A-M, Kerschen G, De Boe P, Golinval J-C (2005) Structural damage diagnosis under
varying environmental conditions—Part I: a linear analysis. Mech Syst Signal Process 19(4)
330 D. García Cava et al.
29. Movsessian A, Garcia D, Tcherniak D (2020) Adaptive feature selection for enhancing blade
damage diagnosis on an operational wind turbine. In: Proceedings of the 13th International
Conference on Damage Assessment of Structures
30. Harvey DY, Todd MD (2012) Cointegration as a data normalization tool for structural health
monitoring applications. In: Proceedings of SPIE - The International Society for Optical
Engineering, vol 8348
31. Cross EJ, Worden K, Chen Q (2011) Cointegration: a novel approach for the removal of
environmental trends in structural health monitoring data. Philosophical Transictions of the
Royal society A. Math Phys Eng Sci 467(2133)
32. Zolna K, Dao PB, Staszewski WJ, Barszcz T (2015) Nonlinear cointegration approach for
condition monitoring of wind turbines. Math Probl Eng 2015
33. Shi H, Worden K, Cross EJ (2018) A regime-switching cointegration approach for removing
environmental and operational variations in structural health monitoring. Mech Syst Signal
Process 103
34. Yan A-M, Kerschen G, De Boe P, Golinval J-C (2005) Structural damage diagnosis under
varying environmental conditions—part II: local PCA for non-linear cases. Mech Syst Signal
Process (19)4
35. Santos A, Figueiredo E, Silva MFM, Sales CS, Costa JCWA (2016) Machine learning
algorithms for damage detection: Kernel-based approaches. J Sound Vib 363
36. Movsessian A, Garcia Cava D, Tcherniak D (2021) An artificial neural network methodology
for damage detection: demonstration on an operating wind turbine blade. Mech Syst Signal
Process 159
37. Zhao R, Yan R, Chen Z, Mao K, Wang P, Gao RX (2019) Deep learning and its applications
to machine health monitoring. Mech Syst Signal Process 115
38. Tibaduiza D, Torres-Arredondo MÁ, Vitola J, Anaya M, Pozo F (2018) A damage classification
approach for structural health monitoring using machine learning. Complexity
39. Silva M, Santos A, Santos R, Figueiredo E, Sales C, Costa JCWA (2017) Agglomerative
concentric hypersphere clustering applied to structural damage detection. Mech Syst Signal
Process 92
40. Langone R, Reynders E, Mehrkanoon S, Suykens JAK (2017) Automated structural health
monitoring based on adaptive kernel spectral clustering. Mech Syst Signal Process 90
41. Baraldi P, Di Maio F, Rigamonti M, Zio E, Seraoui R (2016) Clustering for unsupervised fault
diagnosis in nuclear turbine shut-down transients. Mech Syst Signal Process 58
42. Kostic B, Gül M (2017) Vibration-based damage detection of bridges under varying temperature
effects using time-series analysis and artificial neural networks, J Bridge Eng 22(10)
43. Roberts C, Cava Garcia D, Avendaño-Valencia LD (2021) Understanding the influence of envi-
ronmental and operational variability on wind turbine blade monitoring. In: Rizzo P, Milazzo
A (eds) European Workshop on Structural Health Monitoring. EWSHM 2020. Lecture notes
in civil engineering, vol 127. Springer, Cham. https://doi.org/10.1007/978-3-030-64594-6_12
44. Movsessian A, Garcia D, Tcherniak D (2020) Adaptive feature selection for enhancing blade
damage diagnosis on an operational wind turbine. In: Proceedings of the 13th International
Conference on Damage Assessment of Structures
45. Alvarez MA, Rosasco L, Neil D, Lawrence ND (2011) Kernels for vector-valued functions: a
review. arXiv preprint arXiv: 1106.6251
Explainable Artificial Intelligence
to Advance Structural Health Monitoring
D. Luckey (B)
Chair of Computing in Civil Engineering, Bauhaus University Weimar, Weimar, Germany
e-mail: [email protected]
H. Fritz · J. J. Peralta Abadía · K. Smarsly
Institute of Digital and Autonomous Construction, Hamburg University of Technology, Hamburg,
Germany
e-mail: [email protected]
J. J. Peralta Abadía
e-mail: [email protected]
K. Smarsly
e-mail: [email protected]
D. Legatiuk
Chair of Applied Mathematics, Bauhaus University Weimar, Weimar, Germany
e-mail: [email protected]
C. Walther
Institute of Structural Mechanics, Bauhaus University Weimar, Weimar, Germany
e-mail: [email protected]
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 331
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_16
332 D. Luckey et al.
such as transparent models and model-specific explanations, are presented and cate-
gorized to identify XAI approaches appropriate for being implemented in SHM appli-
cations. Finally, based on the categorization of ML algorithms and the presentation of
XAI approaches, the conceptual XAI framework is introduced. It is expected that the
proposed conceptual XAI framework will provide a basis for improving ML accep-
tance and transparency and therefore increase trust in ML algorithms implemented
in SHM applications.
1 Introduction
This section provides the basics on AI and presents a categorization of AI and ML.
Next, reviews of ML for SHM are discussed. Finally, current ML algorithms used in
SHM are reviewed and summarized.
Artificial intelligence has been a subject of constant development over the last
decades. As AI covers a broad field of research and industrial applications, it does
not seem possible to formulate a single, generally valid definition. Nonetheless,
334 D. Luckey et al.
general concepts of AI developed in each field of research and industry share simi-
larities, identified as four characteristics: (i) perception of a complex environment,
(ii) acquisition and interpretation of data, (iii) autonomous response to change, and
(iv) automatic achievement of predefined goals [13].
For the purpose of automatic achievement of predefined goals, various AI algo-
rithms have been designed to realize intelligent problem-solving strategies [14].
Furthermore, AI algorithms have been developed for applications in different fields of
engineering and thus cover a wide spectrum of tasks, such as audio-visual perception
of the surrounding environment [15], reproducibility of AI decisions [16], automatic
data interpretation and decision making [17], and evaluation of robustness to changes
in datasets [18]. AI can be broadly categorized in natural language processing, auto-
mated reasoning, ML, knowledge representation, and computer vision as shown in
Fig. 1.
Machine learning is referred to as a process that improves the performance of a
functional unit by new knowledge or skill acquisition or by existing knowledge or
skill reorganization [19]. Designing ML algorithms requires large datasets to learn
how to automate tasks and involves two steps, training and testing. During training,
ML algorithms use datasets, ideally covering the entire possible range of values to
which some meaning can be assigned, to “learn” patterns in the data. Some ML
algorithms also accept incomplete or inconsistent datasets, exhibiting an acceptable
level of robustness. A specific instance of a ML algorithm that is trained to solve
a specific problem is referred to as a ML model. During testing, the ML model
obtained during the training step is given new unknown data, with which the overall
performance of the ML model is tested.
In general, ML can be divided into three different learning processes:
i. supervised learning,
ii. unsupervised learning, and
iii. reinforcement learning [20].
For supervised learning, the datasets used to train a ML model contain not only
the input data from which patterns will be learned but also a target feature, such
Explainable Artificial Intelligence to Advance Structural Health Monitoring 335
Fig. 2 Within supervised learning, the model (e.g. an ANN) is trained and tested for automatic
labeling of data
Several studies have been published reviewing the application of ML to solve SHM
problems. Mitra and Gopalakrishnan [21] have presented a literature review on
guided-wave-based SHM, including statistical methods and ML algorithms, for
damage prediction and estimation. The authors have identified challenges in using
ML for guided-wave-based SHM, such as scarcity of training data and the presence
of noise and outliers. Bao et al. [22] have reviewed the state of the art of data science
engineering in SHM by focusing on deep learning (DL) algorithms and compressive
sampling.
An overview of recent research of DL approaches for SHM of civil infrastructures
has been accomplished by Ye et al. [23], demonstrating history, frameworks, and
datasets for DL, in addition to structural damage detection and condition assessment
applications. Furthermore, the authors have indicated challenges and future trends
of DL-based SHM, concluding that DL-based SHM will witness developments with
new algorithms and enhanced frameworks, sufficient datasets, and computing power.
Azimi et al. [24] have introduced a review of DL methods in SHM. The authors have
covered vibration-based and vision-based SHM through DL, as well as applications
of unmanned aerial vehicles and smartphones for DL-based SHM.
Flah et al. [25] have studied the deployment of AI approaches in SHM for damage
detection of structures. The authors have introduced SHM applications utilizing AI
algorithms, and have elaborated on types of structures being employed in these
applications, advantages, limitations, and recommendations.
have used clustering and support vector machines to classify crack modes from
unlabeled acoustic emission waveform features.
The studies reviewed herein have used supervised learning (SL), unsupervised
learning (UL), and semi-supervised learning (Semi-SL) algorithms. Reinforcement
learning (RL) algorithms have not been used in the studies reviewed. Therefore,
based on the categorization and description of learning processes presented in subsec-
tion 2.1 and omitting RL, Table 1 presents a summary of ML algorithms used in SHM,
grouped by SL, UL, and Semi-SL. Furthermore, since SL algorithms may be used
for classification and regression tasks, subgroups are added in the SL group. It is
observed that mainly SL algorithms are being used in SHM, followed by UL algo-
rithms. Moreover, most algorithms in the SL subgroups are classification algorithms,
revealing a possible focus of XAI approaches towards SL classification algorithms.
Regarding SL algorithms, support vector machine algorithms have been used for
damage detection in built infrastructure [28, 29]. Moreover, Pan et al. [30] have used
support vector machines for structural diagnosis and damage detection in large-scale
cable-stayed bridges. Vitola et al. [31] have used k-nearest neighbors (k-NN) for
damage classification, in combination with principal component analysis (PCA) for
feature extraction from signals representing dynamic responses of structures. Artifi-
cial neural networks have been used for predicting accelerations to be compared with
damage indices, in combination with Gaussian process algorithms, discriminating
between damaged and undamaged structural conditions [32].
As for UL algorithms, a damage detection methodology based on strain field
pattern recognition using clustering and PCA has been presented in [33]. In addi-
tion, Diez et al. [34] have performed damage detection on bridges via clustering for
grouping substructures with similar behavior.
Finally, with respect to the hybrid form of semi-SL algorithms, Sarkar et al. [35]
have used a deep auto-encoder for detection and annotation of cracks on images. At
the end of Table 1, a group of references containing comparative studies of different
ML algorithms has been added for providing a better overview to readers interested
not only in specific applications of ML in SHM, but also in comparative performance
of various ML algorithms.
As shown in Table 1, SL classification algorithms have been used the most in SHM
systems. Since many of these SL classification algorithms are black-box, raising
trust issues regarding their outputs, the explainability and interpretability aspects
of XAI may help increase the transparency and acceptance of the SL classification
algorithms in SHM practice. As a result, the following section presents a conceptual
XAI framework, i.e. a concept towards explainable artificial intelligence, as a solution
to help explain and interpret the reasons behind the predictions of SL classification
algorithms.
338 D. Luckey et al.
Table 1 (continued)
Fuzzy cognitive [59]
map
Semi-SL Deep autoencoder [60, 61]
Manifold adaptive [62]
experimental
design algorithm
Comparative studies [8, 9, 28, 30, 36–38, 40]
According to Adadi et al. [63], AI can be divided into three different categories:
1. Accurate AI, including AI systems primarily exhibiting high accuracy in
decision making,
2. Explainable AI, comprising advanced or newly designed AI systems exhibiting,
in addition to high accuracy, a very high degree of explainability and inter-
pretability in decision making, and
3. Responsible AI, encompassing future AI systems that will act and decide in
a transparent, fair, and comprehensible manner, considering legal, social, and
ethical aspects.
340 D. Luckey et al.
From the point of view of explainability, accurate AI represents the classical use
of AI and ML algorithms in various fields of applications, and is not explainable in
general, except for algorithms being transparent by design, such as decision-trees
and rule-based methods. The other two categories, explainable AI and responsible
AI, support transparency. Before discussing different XAI approaches proposed in
recent years, it is necessary to underline the distinction between interpretation and
explanation, which are defined according to [64] as follows:
algorithms, such as deep learning algorithms, are used. The reason is the nature
of surrogate modeling: Because ML algorithms are used as substitutes of conven-
tional algorithms, the relation between input and output is always clear, despite
being nontransparent from an algorithmic point of view. By contrast, classification
problems require using XAI approaches for the ML algorithms, as explained in the
conceptual XAI framework proposed in the next subsection.
require the use of XAI because classifier decisions are generally not traceable. Given
that tasks in most ML-based SHM applications are formulated as classification prob-
lems, a conceptual framework towards adopting XAI in SHM has been proposed,
including four steps: (i) formulating the problem and confirming that the ML algo-
rithms are used as classifiers, (ii) applying the ML algorithms, (iii) explain the ML
algorithms outputs, and (iv) interpret the ML algorithms outputs within an SHM
context.
From the results of this study, it can be concluded that introducing XAI can
enhance the quality of ML-based SHM applications. Furthermore, it is concluded
that ML algorithms require different levels of explanation, based on the purpose of
using ML and the human individuals to which the explanation is addressed. Future
work may be conducted towards an in-depth analysis of explainability and levels of
explanation for ML algorithms to advance SHM and smart monitoring systems.
Acknowledgements The authors gratefully acknowledge the support offered by the German
Research Foundation (DFG) under grants SM 281/9-1, SM 281/12-1, SM 281/14-1, SM 281/15-1,
and LE 3955/4-1. This research is also partially supported by the Carl Zeiss Foundation. Any opin-
ions, findings, conclusions, or recommendations expressed in this paper are those of the authors
and do not necessarily reflect the views of the sponsors.
References
1. Law KH, Smarsly K, Wang Y (2014) Sensor data management technologies for infrastruc-
ture asset management. In: Wang ML, Lynch JP, Sohn H (eds) Sensor technologies for civil
infrastructures. Woodhead Publishing Ltd., Sawston, UK, pp 3–32
2. Smarsly K, Lehner K, Hartmann D (2007) Structural health monitoring based on artificial intel-
ligence techniques. In: Proceedings of the 2007 ASCE international workshop on computing
in civil engineering. Pittsburgh, PA, USA, July 24, 2007
3. Smarsly K, Law KH, König M (2011).Resource-efficient wireless monitoring based on mobile
agent migration. In: Proceedings of the SPIE (Vol. 7984): health monitoring of structural and
biological systems 2011. San Diego, CA, USA, March 6, 2011
4. Fritz H, Smarsly K (2020) A state-of-the-art review of nature-inspired systems for smart struc-
tures. In: Proceedings of the European workshop on structural health monitoring (EWSHM).
Palermo, Italy, July 6, 2020.
5. Shavlet-Shwartz S, Ben-David S (2014) Understanding machine learning. From theory to
algorithms. Cambridge University Press, New York, NY, USA
6. Li R, Gu H, Hu B, She Z (2019) Multi-feature fusion and damage identification of large
generator stator insulation based on Lamb wave detection and SVM method. Sensors
19(7):3733
7. Abdeljaber O, Avci O, Kiranyaz S, Boashash B, Sodano H, Inman D (2018) 1-D CNNs for
structural damage detection: verification on a structural health monitoring benchmark data.
Neurocomputing 275(2018):1308–1317
8. Nazarian E, Taylor T, Weifeng T, Ansari F (2018) Machine-learning-based approach for post
event assessment of damage in a turn-of-the-century building structure. J Civ Struct Heal Monit
8(2):237–251
9. Sysyn M, Gerber U, Nabochenko O, Li Y, Kovalchuk V (2019) Indicators for common
crossing structural health monitoring with track side inertial measurements. Acta Polytechnica
59(2):170–181
344 D. Luckey et al.
10. Luckey D, Fritz H, Legatiuk D, Dragos K, Smarsly K (2020) Artificial intelligence techniques
for smart city applications. In: Proceedings of the international ICCCBE and CIB W78 joint
conference on computing in civil and building engineering 2020. São Paulo, Brazil, August
18, 2020
11. Kuang C (2017) Can A.I. be taught to explain Itself? The New York Time Magazine
12. Das, A, Rad P (2020) Opportunities and challenges in explainable artificial intelligence (XAI):
a survey. ArXiv abs/2006.11371
13. Samoili S, Lopez Cobo M, Gomez Gutierrez E, De Prato G, Martinez-Plumed F, Delipetrev B
(2020) AI WATCH: Defining artificial intelligence. Publications Office of the European Union,
Luxembourg
14. Swarnkar A, Swarnkar A (2020) Artificial intelligence based optimization techniques: a review,
intelligent computing techniques for smart energy systems. Springer Singapore, 95–103
15. Tian Yh, Chen Xl, Xiong Hk et al (2017) Towards human-like and transhuman perception in
AI 2.0: a review. Frontiers Inf Technol Electronic Eng 18:58–67
16. Shrestha YR, Ben-Menahem SM, von Krogh G (2019) Organizational decision-making
structures in the age of artificial intelligence. Calif Manage Rev 61:66–83
17. Hofmann M, Neukart F, Bäck T (2017) Artificial intelligence and data science in the automotive
industry. ArXiv abs/1709.01989.
18. Hamon R, Junklewitz H, Sanchez I (2020) Robustness and explainability of artificial intel-
ligence—from technical to policy solutions. Publications Office of the European Union,
2020
19. ISO/IEC 2382:2015 Information technology—Vocabulary
20. Russel SJ, Norvig P (2014) Artificial intelligence: a modern approach, 3rd edn. Pearson
Education Ltd., Harlow, Essex, UK
21. Mitra M, Gopalakrishnan S (2016) Guided wave based structural health monitoring: a review.
Smart Mat Struct 25(5):053001
22. Bao Y, Chen Z, Wei S, Xu Y, Tang Z, Li H (2019) The state of the art of data science and
engineering in structural health monitoring. Engineering 5(2):234–242
23. Ye XW, Jin T, Yun CB (2019) A review on deep learning-based structural health monitoring
of civil infrastructures. Smart Struct Syst 24(5):567–585
24. Azimi M, Eslamlou AD, Pekcan G (2020) Data-driven structural health monitoring and damage
detection through deep learning: State-of-the-art review. Sensors 20(10):2778
25. Flah M, Nunez I, Chaabene W, Nehdi M (2020) Machine learning algorithms in civil structural
health monitoring: a systematic review. Arch Comput Methods Eng
26. Joshuva A, Sugumaran V (2018) A study of various blade fault conditions on a wind turbine
using vibration signals through histogram features. J Eng Sci Technol 13(1):102–121
27. Das AK, Suthar D, Leung CK (2019) Machine learning based crack mode classification from
unlabeled acoustic emission waveform features. Cem Concr Res 121:42–57
28. Gui G, Pan H, Lin Z, Li Y, Yuan Z (2017) Data-driven support vector machine with optimization
techniques for structural health monitoring and damage detection. KSCE J Civ Eng 21(2):523–
534
29. Ghiasi R, Torkzadeh P, Noori M (2016) A machine-learning approach for structural damage
detection using least square support vector machine based on a new combinational kernel
function. Struct Health Monit 15(3):302–316
30. Pan H, Azimi M, Yan F, Lin Z (2018) Time-frequency-based data-driven structural diagnosis
and damage detection for cable-stayed bridges. J Bridg Eng 23(6):04018033
31. Vitola J, Pozo F, Tibaduiza DA, Anaya M (2017) A sensor data fusion system based on k-
nearest neighbor pattern classification for structural health monitoring applications. Sensors
17(2):417
32. Neves AC, Gonzalez I, Leander J, Karoumi R (2017) Structural health monitoring of bridges: a
model-free ANN-based approach to damage detection. J Civ Struct Heal Monit 7(5):689–702
33. Sierra-Pérez J, Torres-Arredondo MA, Alvarez-Montoya J (2017) Damage detection method-
ology under variable load conditions based on strain field pattern recognition using FBGs,
nonlinear principal component analysis, and clustering techniques. Smart Mater Struct
27(1):015002
Explainable Artificial Intelligence to Advance Structural Health Monitoring 345
34. Diez A, Khoa NLD, Alamdari MM, Wang Y, Chen F, Runcie P (2016) A clustering approach
for structural health monitoring on bridges. J Civ Struct Heal Monit 6(2016):1–17
35. Sarkar S, Reddy KK, Giering M, Gurvich MR (2016) Deep learning for structural health
monitoring: a damage characterization application. In: Proceedings of the annual conference
of the prognostics and health management society 2016, Denver, CO, USA, October 3, 2016
36. Miorelli R. Kulakovskyi A, Mesnil O, d’Almeida O (2019) Automatic defect localization and
characterization through machine learning based inversion for guided wave imaging in SHM.
AIP Conf Proc 2102(1):050005
37. Finotti RP, Cury AA, Barbosa FDS (2019) An SHM approach using machine learning and
statistical indicators extracted from raw dynamic measurements. Latin American J Solids Struct
16(2):e165
38. Taddei T, Penn JD, Yano M, Patera AT (2018) Simulation-Based Classification; A model-order-
reduction approach for structural health monitoring. Arch Computat Methods Eng 25:23–45
39. Tibaduiza D, Cerón-MH (2017) Damage classification based on machine learning applica-
tions for an unmanned aerial vehicle. The 11th international workshop on structural health
monitoring. Stanford, CA, USA, September 12, 2017
40. Tripathi G, Anowarul H, Agarwal K, Prasad DK (2019) Classification of micro-damage in
piezoelectric ceramics using machine learning of ultrasound signals. Sensors 19(19):4216
41. Salehi, H., Das, S., Chakrabartty, S., Biswas, S. & Burgueño, R. (2018). Damage identification
in aircraft structures with self-powered sensing technology: a machine learning approach.
Structural Control Health Monitoring, 25, e2262.
42. Vitola J, Pozo F, Tibaduiza D, Anaya M (2017) Distributed piezoelectric sensor system for
damage identification in structures subjected to temperature changes. Sensors 17(6):1252
43. Melia T, Cooke A, Grayson S (2016) Machine learning techniques for automatic sensor
fault detection in airborne SHM networks. In: Proceedings of the 8th European workshop
on structural health monitoring (EWSHM), Bilbao, Spain, July 5, 2016
44. Vashisht R, Viji H, Sundararajan T, Mohankumar D, Sarada S (2018) Structural health moni-
toring of cantilever beam, a case study—using Bayesian neural network and deep learning. In:
Proceedings of the 13th international conference on systems. Athens, Greece, April 22, 2018
45. Liu J, Yang X, Zhu M (2019) Neural network with confidence kernel for robust vibration
frequency prediction. J Sensors 2019:6573513
46. Avci O, Abdeljaber O, Kiranyaz S, Boashash B, Sodano H, Inman D (2018) Efficiency valida-
tion of one dimensional convolutional neural networks for structural damage detection using a
SHM benchmark data. In: Proceedings of 25th international congress on sound and vibration
2018 (ICSV 25), Hiroshima, Japan, July 8, 2018
47. Tibaduiza D, Torres-Arredondo MA, Oyaga J, Anaya M, Pozo F (2018) A damage classification
approach for structural health monitoring using machine learning. Complexity 2018:5081283
48. Vitola J, Tibaduiza D, Anaya M, Pozo F (2016) Structural damage detection and classification
based on machine learning algorithms. In: Proceedings of the 8th European workshop on
structural health monitoring (EWSHM), Bilbao, Spain, July 5, 2016
49. Bao Y, Tang Z, Li H (2019) Compressive-sensing data reconstruction for structural health
monitoring: a machine-learning approach. Struct Health Monit 19(1):293–304
50. Cadini F, Sbarufatti C, Corbetta M, Cancelliere F, Giglio M (2019) Particle filtering-based
adaptive training of neural networks for real-time structural damage diagnosis and prognosis.
Struct Control Health Monit (26):e2451
51. Zhao Z, Yua M, Dong S (2019) Damage location detection of the CFRP composite plate based
on neural network regression. In: Proceedings of the 7th asia-pacific workshop on structural
health monitoring. Hong Kong, China, November 12, 2018.
52. Wang H, Yuan S, Chen J, Ren Y (2018) Online updating Gaussian process model for fatigue
crack diagnosis and prognosis. 9th European workshop on structural health monitoring July
10, 2018. Manchester, United Kingdom.
53. Gardner P, Lord C, Barthorpe RJ (2018) A Probabilistic framework for forward model-driven
SHM. 9th European workshop on structural health monitoring. July 10, 2018, Manchester,
United Kingdom
346 D. Luckey et al.
54. Hoang N-D, Liao K-W, Tran X-L (2018) Estimation of scour depth at bridges with complex
pier foundations using support vector regression integrated with feature selection. J Civ Struct
Heal Monit 8(3):431–442
55. Quaranta G, Lopez E, Abisset-Chavanne E, Duval J, Huerta A, Chinesta F (2019) Structural
health monitoring by combining machine learning and dimensionality reduction techniques.
Revista internacional de métodos numéricos para cálculo y diseño en ingenieria. 35(1):20
56. Ye X, Chen X, Lei Y, Fan J, Mei L (2018) An Integrated Machine Learning Algorithm
for Separating the Long-Term Deflection Data of Prestressed Concrete Bridges. Sensors
18(11):4070
57. Rahim SA, Manson G, Worden K (2016) Data visualization approach for operational loading
variations of an aircraft wing box using vibration based damage detection. In: Proceedings of
the 8th European workshop on structural health monitoring (EWSHM). Bilbao, Spain, July 5,
2016
58. Santos A, Figueiredo E, Silva M, Santos R, Sales C, Costa J (2016) Geneticbased EM algorithm
to improve the robustness of Gaussian mixture models for damage detection in bridges. Struct
Control Health Monitor 24(3):e1886
59. Senniappan V, Subramanian J, Papageorgiou E, Mohan S (2016) Application of fuzzy cognitive
maps for crack categorization in columns of reinforced concrete structures. Neural Comput
Appl 28(1):107–117
60. Liu J, Berges M, Bielak J, Garrett J, Kovacevic J, Noh HY (2019) A damage localization and
quantification algorithm for indirect structural health monitoring of bridges using multi-task
learning. 45th annual review of progress in quantitative nondestructive evaluation 2102(1)
61. Karypidis D, Gil Berrocal C, Rempling R, Granath M, Simonsson P (2019) Structural Health
Monitoring of RC structures using optic fiber strain measurements: a deep learning approach.
The Evolving Metropolis IABSE Congress New York City 2019:114
62. Dervilis N, Papatheou E, Antoniadou I, Cross EJ, Worden K (2016) On the usage of active
learning for SHM. In: Proceedings of ISMA2016. ISMA 2016, September 19, 2016, Leuven,
Belgium
63. Adadi A, Berrada M (2018) 2018), Peeking Inside the Black-Box: A Survey on Explainable
Artificial Intelligence (XAI. IEEE Access 6:52138–52160
64. Montavon G, Samek W, Müller K-R (2018) Methods for interpreting and understanding deep
neural networks. Digital Signal Processing 73(2018):1–15
65. Barredo Arrieta A, Díaz-Rodríguez N, Del Ser J, Bennetot A, Tabik S, Barbado A, Garcia
S, Gil-Lopez S, Molina D, Benjamins R, Chatila R, Herrera F (2020) Explainable Artificial
Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible
AI. Information Fusion 58:82–115
Physics-Informed Machine Learning for
Structural Health Monitoring
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 347
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_17
348 E. J. Cross et al.
1 Introduction
As performance and monitoring data from our structures become more abundant, it
is natural for researchers to turn to methods from the machine learning community
to help with analysis and construction of diagnostic/prognostic algorithms. Indeed,
within the SHM research field, use of neural networks, support vector machines and
Gaussian processes for regression and classification problems has become common
place [1]. These methods bring the opportunity to learn complex relationships directly
from data, without a requirement of in-depth knowledge of the system. As an example
from the authors’ own work, in [2] we employed a Gaussian process (GP) regression
to predict strain on a landing gear from measured accelerations across the aircraft.
Use of a suitably trained GP circumvents the need to build complex physics-based
models of the gear for fatigue life calculations. This kind of model is often referred
to as a ‘black-box’ model to reflect the fact the data drives the structure of the model
rather than knowledge of the physics at work.
At the other end of the spectrum the term ‘white-box’ model can be used to
describe a model purely constructed from knowledge of physics, (e.g. differential
equations and finite element models). Physics-based modelling and updating were
common early themes in the structural health monitoring research field [3]. However,
for large or critical engineering structures that operate in (often extreme) dynamic
environments, such as wind turbines, aircraft and gas turbines, predictive modelling
from a white-box perspective presents particularly difficult challenges. Loading is
often unknown and unmeasured, and dynamic behaviour during operation needs to
be fully captured by a computational model, but is sensitive to small changes in (or
disturbances to) the structure. Validation and updating of large complex models bring
their own challenges and remain active research areas [4–7].
Due to the availability of monitoring data, the inherent challenges of the physics-
based approach and the promise of machine learning methods, it is fair to say that the
data-driven approach to SHM has become dominant in the research field. A significant
issue with the use of any machine learning method in an engineering application,
however, is the availability of suitable data with which to train the algorithm. As the
model learns from the data, it is only able to accurately predict behaviour present
in the data on which it was trained. As an example, Fig. 1 shows a black-box model
trained to predict the bending strain on an aircraft wing during different manoeuvres
to inform an in-service fatigue assessment. This data set comprises of 84 flights,
five of which are used for model training. The trained model is able to generalize
well with a very low prediction error for the majority of the flights—Fig. 1a shows a
typical strain prediction for a flight not included in the training set (normalized mean-
squared error1 (nMSE)= 0.29% across the whole flight). However, for the flight shown
in Fig. 1b, the model is unable to predict the strain as accurately (nMSE 4.20% across
1
nMSE = 100
nσ y2
(yi − f i )2 (1)
where yi and f i are the measurements and predictions, respectively, i = 1 . . . n.
Physics-Informed Machine Learning for Structural Health Monitoring 349
the whole flight)—this flight was atypical in terms of operating conditions—it was a
low-altitude sortie over ground, characterized by the turbulent response one can see
in the figure. These conditions are different from those included in the training set
and the model is unable to generalize and predict the strain as well in this case.
In general, but especially because of the inherent flexibility in many of the machine
learning models commonly used, extrapolation should not be attempted in this set-
ting. For an SHM application, this will generally mean that training data are required
from all possible operating conditions that the structure will see. For many applica-
tions, this is currently infeasible, although as data collection becomes more common-
place, the situation will improve somewhat. Where a supervised approach is needed,
this problem is exacerbated by the general lack of access to data from structures in
a damaged state which remains a large barrier to effective diagnosis and prognosis
[10].
Currently a programme of work by the authors is pursuing a physics-informed
machine learning approach to attempt to address some of these issues in a structural
dynamics setting. The aim is to bring together the flexibility and power of state-
of-the-art machine learning techniques with more structured and insightful physics-
based models derived from domain expertise. This reflects a natural wish that any
inferences over our structures will be informed by both our engineering knowledge
and relevant monitoring data available.
The potential means of combining physics-based models and data-driven algo-
rithms are many, ranging from employing ML methods for parameter estimation [11,
12], to using them as surrogates or emulators [13–15]. Of interest here are methods
where the explanatory power of a model is shared between physics-based and data-
driven components. We will often refer to these approaches as ‘grey-box’ models (a
combination of white and black-box components), but the term ‘hybrid modelling’
is equally applicable. The philosophy followed in our work is to embed fundamental
physical insight into a machine learning algorithm. In doing so, our aim is that the
350 E. J. Cross et al.
role of the machine learner is one of augmenting the explanatory power of the model
rather than being employed to correct any potential error or bias in the physical foun-
dation. This chapter will explore this idea in a Bayesian setting, introducing a number
of different approaches and demonstrating their usefulness in an SHM setting.
The term ‘grey-box model’ is perhaps most familiar to those from a control engi-
neering background. Sohlberg [16, 17] provides a useful review and overview of
grey-box models in this context.2 Figure 2 attempts to capture and summarise some
of the currently available modelling approaches relevant for challenges in structural
health monitoring on the white to black spectrum. Note that the ‘degree of greyness’
of the models in the middle region will change according to implementation and
application.
At the whiter end of the spectrum are modelling approaches where data are used
for parameter estimation or model form selection, (with the buoyant field of equation
discovery fitting in here, see [18, 19]). See also [20]. Residual models are those that
use a data-driven approach to account for the observed difference between a physics-
based model and measurements, with general form
where f (x) is the output of the physical model, δ(x) is the model discrepancy and
ε is the process noise (see for example [21–24]). The discrepancy term is often
used to correct a misspecified physical model, giving rise to the term ‘bias correc-
tion’. Residual-based approaches have proven effective across a range of SHM tasks
including damage detection [25] and modal identification [26]. Here, we are inter-
ested in residual modelling in the context of compensation for uncaptured/missing
behaviours in the physics-based model (discussed further in Sect. 3).
The term hybrid architectures reflects the wider possibilities for combinations of
white and black models (which could include the summation form of (2)). Section 5
will demonstrate one such example of combining data-driven and physics-based
models in a state-space setting.
The remainder of the spectrum contains models with structures that are data-
driven/black-box in nature. Sohlberg [16]) describes semi-physical modelling as
when features are subject to a nonlinear transformation before being used as inputs
to a black-box model, we also refer to this as input augmentation, see [8, 27, 28]
2 The use of grey-box models within the control community is undergoing somewhat of a revival and
a good snap-shot of this can be gained by looking at the contributions to the most recent Nonlinear
System Identification Benchmarks Workshop (http://www.nonlinearbenchmark.org).
Physics-Informed Machine Learning for Structural Health Monitoring 351
for more examples. We place these examples under the heading of manipulation of
black-box inputs.
Section 4 of this chapter will discuss constraints for machine learning algorithms—
these are methods that allow one to constrain the predictions of a machine learner
so that they comply with physical assumptions. Excellent examples for Gaussian
process regression are [29–31] and will be discussed in greater detail later.
The final grouping of grey-box approaches mentioned here are physics-guided
black-box learners. These are methods that use physical insight to attempt to improve
model optimisation and include the construction of physics-guided loss functions and
the use of physics-guided initialization. These will not be discussed further in this
chapter but see e.g. [32–35] for more details.
The remainder of the chapter will showcase some of the work of the authors on devel-
oping physics-informed machine learning approaches for SHM tasks. The develop-
ments here fall in the domain of residual and hybrid models (Sects. 3 and 5), and
constrained machine learners (Sect. 4). In order to provide an overview, a variety of
methods and results are presented, however, the implementation details given here are
necessarily very brief and we refer readers to the referenced papers and our webpage3
for specific details and more in-depth analysis. Reflecting the philosophy discussed
in the introduction section, the approaches presented generally incorporate simple
3 https://drg-greybox.github.io/.
352 E. J. Cross et al.
physics-based models or assumptions and rely on the machine learner for enhanced
explanatory power and flexibility (i.e. we are operating towards the blacker end of
the scale).
The machine learning approach used in the work shown here will be Gaussian
process (GP) regression throughout. GPs have been shown to be a powerful tool for
regression tasks [36] and are becoming common in SHM applications (see for exam-
ple [2, 37–40]). Their use here and throughout the work of the authors is due to their
(semi)non-parametric nature, their ability to function with a small number of training
points, and most importantly, the Bayesian framework within which they naturally
work. The Gaussian process formulation provides a predictive distribution rather
than a single prediction point, allowing confidence intervals to be calculated and
uncertainty to be propagated forward into any following analysis (see [9] for exam-
ple). As the use of GPs is now quite common, their fundamental formulation will not
be introduced here, but the mathematical machinery required is briefly summarized
in the Appendix—we refer unfamiliar readers to [36].
In the first examples shown here, the use of priors in the Bayesian framework
is exploited as an appropriate and intuitive means of incorporating physical insight
into a machine learning algorithm. In later sections, we consider the construction of
constraints for GPs and, separately, their incorporation into a state-space formulation
(this latter example relies more heavily on physics-based machinery than the other
examples).
3 Be More Bayes
A Bayesian philosophy is one that employs evidence from data to update prior beliefs
or assumptions, and has been widely adopted across disciplines, including SHM.
However, most commonly, uninformative priors are utilized that do not reflect the
knowledge that we have as engineers of the systems we are interested in modelling.
The formulation of a Gaussian process regression requires the selection of a mean
and covariance function which form the prior process. The process is then conditioned
with training data to provide a posterior mean and covariance as the model prediction.
In the standard approach, no prior knowledge is assumed; a zero mean function is
selected alongside a generic covariance function such as a squared-exponential or
one from the Matérn class which provide a flexible process to fit to most data.
In this section, we will first employ simple physical models as prior mean functions
to a GP and show how they may improve the extrapolative capability of the model.
This simple means of incorporating prior knowledge is equivalent to using a GP with
a zero mean prior to model the difference between the measured data and the physical
model prediction and hence can be classed as a residual approach (see Sect. 2). At
the end of this section, we will show how some knowledge of a system may be used
to derive useful covariance functions in a regression setting.
Physics-Informed Machine Learning for Structural Health Monitoring 353
(a) (b)
(c)
Fig. 3 Model for bridge deck deflections; a shows the training and test datasets for the GPs, b
is a GP prediction with a zero mean function prior and c shows the prediction when a simple
physics-informed mean function is incorporated. See [44] for more details
for health assessment, particularly where one wishes to infer, e.g. fatigue damage
accrued.
The implementation of residual modelling is most effective where the assumptions
and limitations of the white-box model are well understood. As a widely used method
for wave loading prediction, Morison‘s equation [47] is employed here as a physics-
informed mean function. This empirical law is known to simplify the behaviour of
wave loading, not accounting for effects such as vortex shedding or other complex
behaviours [48] and will typically have residual errors in the region of 20%[49].
Here we consider the addition of a data-based GP-NARX to a simplified version
of Morison’s Equation in an attempt to account for these missing phenomena. The
model used is:
Physics-Informed Machine Learning for Structural Health Monitoring 355
Fig. 4 A comparison of wave loading prediction model NMSEs versus test set coverage. Increasing
coverage of the test set by the training and validation sets results in an increased level of model
interpolation. See [50] for more details
yt = Cd Ut |Ut | + Cm U̇t + f ([u t , u t−1 , ..., u t−lu , yt−1 , yt−2 , ..., yt−l y ]) + ε (3)
Morison’s Equation GP-NARX
where yt is the wave force, Cd is the drag coefficient, Cm is the inertia coefficient,
U is the wave velocity, U̇ is the wave acceleration, u t:t−lu are lagged exogeneous
inputs and yt−1:t−lu are the lagged wave force, see [50] for more details (this paper
also shows an example of an input augmentation model, where Morison‘s equation
is used as an additional input to the GP-NARX).
The Christchurch bay dataset is used here as an example to demonstrate the
approach [51]. To explore the generalization capability with and without the physics-
informed mean, different training sets for the GP-NARX are considered with increas-
ing levels of coverage of the input space. A comparison of model errors (nMSE) with
different training datasets is shown in Fig. 4. The coverage level is indicated as a per-
centage of the behaviour observed in the testing set that is also encountered in the
training set [50].
As in the Tamar Bridge example, the model structure offers a significant improve-
ment in extrapolation, where testing conditions are different from those in training
dataset. Following this approach allows predictions to be informed by the prior mean
in the absence of evidence from data. Clearly the prior specification is very important
in this case and a misspecified prior could do more harm than good. Once again we
advocate the use of simple and well founded physics-based models in an attempt to
avoid this issue.
356 E. J. Cross et al.
σ2 ζ ωn
φY (τ ) = E[Y (t1 )Y (t2 )] = e−ζ ωn |τ | (cos(ωd τ ) + sin(ωd |τ |)) (5)
4m 2 ζ ωn 3 ωd
√
where√ standard notation has been used; ωn = k/m, the natural frequency, ζ =
c/2 km, the damping ratio, ωd = ωn 1 − ζ 2 , the damped natural frequency. See
[53] and also [54, 55].
This covariance function can be readily used in the regression context and provides
a useful prior process for oscillatory systems with a response dominated by a single
frequency. This form of covariance function can be described as expressive [56] and
proves useful even when the equation of motion of the system of interest differs from
an SDOF linear assumption.
Figure 5 shows an example of a GP regression for a system with a cubic nonlin-
earity. Here the linear prior provides an appropriate structure for the regression and
is flexible enough to incorporate the nonlinearity in the response. This is a simu-
lated example with the GP training data shown with crosses in the figure (every 8th
point)—here the nMSE is 8.09. For comparison, a GP with a squared-exponential
(SE) covariance function is established with the same training data. The SE process
smooths through the data as expected (nMSE = 66.7), whereas the derived covariance
provides structure through the prior, resulting in good prediction during interpolation.
The hyperparameters in the SDOF covariance function are physically interpretable,
we are, therefore, able to guide their optimisation by providing the likely ranges for
Physics-Informed Machine Learning for Structural Health Monitoring 357
the system of interest. Here the benefit of being able to prescribe the likely frequency
content of the system within the prior is clear and provides much advantage over the
black-box approach—see [53] for more details.
In scenarios where one lacks significant knowledge of the governing equations and
solutions, grey-box methods that tend towards the black end of the spectrum can be
particularly useful. An example of such approaches are constrained machine learners,
which, very generally, aim to embed physical constraints into the learning procedure
such that predictions made by the black-box model then adhere to these constraints.
In the context of Gaussian process regression, there are a number of ways of
constraining predictions, the simplest of which is to include data from boundaries
within the model training. Other methods rely on applying constraints to the covari-
ance function in a multiple output setting (see e.g. [29–31] and [57] where we employ
derivative boundaries for beam deflection predictions).
Here we show an example of building known boundaries (geometry) into a GP
regression via a sparse approximation of the covariance function. The approximation
method relies on an eigendecomposition of the Laplace operator of a fixed domain
[58]:
m
k(x, x ) ≈ S( λi )φi (x)φi (x ), (6)
i
358 E. J. Cross et al.
with φi and λi the eigenfunctions and values, and S the spectral density of the
covariance function. If one chooses the fixed domain to reflect the geometry of the
problem of interest, then inference with this model is appropriately bounded (see
[59, 60] for more details).
As an example here we employ constraints for a crack localisation problem via
measurement of Acoustic emission (AE). The localization approach taken is to use
artificial source excitations and an interpolating GP to provide a map of the differ-
ences in times of arrival ( T ) of AE sources to fixed sensor pairings across the
surface of the structure [61, 62]. Once constructed, the map can be used to assess the
most likely location of any new AE sources. The bounded GP approximation allows
one to build in the geometry of the structure under consideration.
To investigate the predictive capability of the constrained GP, a case study using
a plate with a number of holes in is adopted. The holes, as shown in Fig. 6, provide
complexity to the modelling challenge, introducing several complex phenomena
such as wave mode conversion and signal reflection. Depending on the location of
the source and sensor, the holes may also shield a direct propagation path to the
receiver [61], adding further complication.
Neumann boundaries are imposed here around each hole and at the edge of
the plate. To compare the performance of the standard and bounded GPs, differing
amounts/coverage of artificial source excitations were used for model training. The
initial characterization of a structure via artificial source excitation can be expensive
and time consuming, for structures in operation it may also be infeasible to access
all areas/components. To mimic the scenario where it is not possible to collect arti-
Physics-Informed Machine Learning for Structural Health Monitoring 359
ficial source excitations across a whole structure, here we restrict the training grid
to excitation points in the middle of the plate. Figure 7 compares the performance
of the standard and bounded GPs with training sets of varying grid densities. For
each training set, the prediction error (nMSE) on the test set is averaged across every
sensor pair (there are 8 sensors).
From Fig. 7 one can see that as the training set size reduces, the constrained GP
consistently outperforms the standard full GP. This is particularly encouraging as
the bounded GP remains a sparse approximation. As is consistent with our earlier
observations, the inbuilt physical insight aids inference where training data are fewer.
Figure 8 shows the difference in prediction error across the plate for the standard and
bounded GPs for the 20 mm spacing training case and a single sensor pairing. In
this case, the squared error of the full GP is subtracted from the squared error of
the constrained GP, i.e. positive values indicate a larger error in the full GP, while a
negative value expresses a larger error in the constrained GP.
The figure highlights the locations on the plate where the constrained GP more
accurately predicts the true T values. As expected, the locations at which this
effect is most prominent are those that move further away from the training points,
and particularly towards the extremities of the domain. At these locations, it is clear
that the additional physical insight provided by the constrained GP is able to enhance
the predictive performance in comparison with the pure black-box model.
One of the canonical forms for dynamic models, in structural mechanical systems
and beyond, is the state-space representation of the behaviour of interest.
360 E. J. Cross et al.
Fig. 8 Mapping of the difference between the full GP squared error and the constrained GP squared
error across the test set for sensor pair 4–8
4The notation subscript a : b is used to denote values in that range inclusively, e.g. x0:t is the value
of the states x at all times from t = 0 to t = t.
Physics-Informed Machine Learning for Structural Health Monitoring 361
Fig. 9 Force estimation for MDOF simulation under Christchurch bay loading time history[40]
prior knowledge about the nonlinearity; in the language of grey-box models, there is
a strong white box component (the second-order linear system) but the form of the
nonlinearity is the very flexible nonparametric GP. Contrast this approach with the
purely black-box alternative of the GP-SSM, see [74], which suffers from significant
nonindentifiability and computational challenges.
6 Conclusions
Here we follow the notation used in [36]; k(x p , xq ) defines a covariance matrix K pq ,
with elements evaluated at the points x p and xq , where xi may be multivariate.
Assuming a zero-mean function, the joint Gaussian distribution between mea-
surements/observations y with inputs X and unknown/testing targets y∗ with inputs
X ∗ is
y K (X, X ) + σn2 I K (X, X ∗ )
∼ N 0, (7)
y∗ K (X ∗ , X ) K (X ∗ , X ∗ )
The distribution of the testing targets y∗ conditioned on the training data (which is
what we use for prediction) is also Gaussian:
See [36] for the derivation. The mean and covariance here are that of the posterior
Gaussian process. In this work, covariance function hyperparameters are sought by
maximizing the marginal likelihood of the predictions
1 1 n
log p(y|X, θ ) = − yT K −1 y − log |K | − log 2π (9)
2 2 2
Acknowledgements We would like to thank Keith Worden for his general support and also partic-
ularly for provision of the Christchurch Bay data and support of the state-space work. Additionally,
thanks is offered to James Hensman, Mark Eaton, Robin Mills, Gareth Pierce and Keith Worden for
their work in acquiring the AE data set used here. We would like to thank Ki-Young Koo and James
Brownjohn in the Vibration Engineering Section at the University of Exeter for provision of the
data from the Tamar Bridge. Thanks also to Steve Reed, formerly of DSTL for the provision of the
Tucano data. Finally, the authors would like to acknowledge the support of the EPSRC, particularly
through grant reference number EP/S001565/1, and Ramboll Energy for their support of SG and
DP.
References
1. Farrar CR, Worden K (2010) An introduction to structural health monitoring. In: New trends
in vibration based structural health monitoring, pp 1–17
2. Holmes Geoffrey, Sartor Pia, Reed Stephen, Southern Paul, Worden Keith, Cross Elizabeth
(2016) Prediction of landing gear loads using machine learning techniques. Struct Health
Monitor 15(5):568–582
3. Sohn H, Farrar CR, Hemez FM, Shunk DM, Stinemates DW, Nadler BR, Czarnecki JJ (2003) A
review of structural health monitoring literature: 1996–2001. Los Alamos National Laboratory,
USA, 1
4. Simoen Ellen, De Roeck Guido, Lombaert Geert (2015) Dealing with uncertainty in model
updating for damage assessment: a review. Mech Syst Signal Process 56:123–149
5. Gardner P (2018) On novel approaches to model-based structural health monitoring. PhD thesis,
University of Sheffield
6. Gardner P, Lord C, Barthorpe RJ (2019a) A unifying framework for probabilistic validation
metrics. J Verificat Validation Uncertainty Quantificat 4(3)
7. Gardner P, Lord C, Barthorpe RJ (2019b) Bayesian history matching for forward model-
driven structural health monitoring. In: Model validation and uncertainty quantification, vol 3.
Springer, pp 175–183
8. Fuentes R, Cross E, Halfpenny A, Worden K, Barthorpe RJ (2014) Aircraft parametric structural
load monitoring using gaussian process regression. In: Proceedings of the European workshop
on structural health monitoring 2014. Nantes
9. Gibson S J , Rogers T J, Cross E J (2020) Data-driven strain prediction models and fatigue dam-
age accumulation. In: Proceedings of the 29th international conference on noise and vibration
engineering (ISMA 2020)
10. Barthorpe RJ (2010) On model-and data-based approaches to structural health monitoring.
PhD thesis, University of Sheffield
11. Farrar CR, Worden K, Todd MD, Park G, Nichols J, Adams DE, Bement MT, Farinholt K
(2007) Nonlinear system identification for damage detection. Technical report, Los Alamos
National Laboratory (LANL), Los Alamos, NM
12. Kerschen G, Worden K, Vakakis AF, Golinval J-C (2006) Past, present and future of nonlinear
system identification in structural dynamics. Mech Syst Signal Proces 20(3):505–592
13. Oakley J, O’Hagan A (2002) Bayesian inference for the uncertainty distribution of computer
model outputs. Biometrika 89(4):769–784
14. Leser PE, Hochhalter JD, Warner JE, Newman JA, Leser WP, Wawrzynek PA Yuan F-G (2017)
Probabilistic fatigue damage prognosis using surrogate models trained via three-dimensional
finite element analysis. Struct Health Monitor 16(3):291–308
15. Beucler T, Pritchard M, Rasp S, Ott J, Baldi P, Gentine P (2019) Enforcing analytic constraints
in neural-networks emulating physical systems. arXiv preprint arXiv:1909.00912
16. Sohlberg B, Jacobsen EW (2008) Grey box modelling–branches and experiences. IFAC Proc
Vol 41(2):11415–11420
17. Sohlberg B (2012) Supervision and control for industrial processes: using grey box models,
predictive control and fault detection methods. Springer Science & Business Media
18. Fuentes R, Nayek R, Gardner P, Dervilis N, Rogers T, Worden K, Cross EJ (2021) Equation
discovery for nonlinear dynamical systems: a Bayesian viewpoint. Mech Syst Signal Process
154:107528
19. Brunton SL, Proctor JL, Kutz JN (2016) Discovering governing equations from data by sparse
identification of nonlinear dynamical systems. Proc Natl Acad Sci 113(15):3932–3937
20. Noël J-P, Schoukens J (2018) Grey-box state-space identification of nonlinear mechanical
vibrations. Int J Control 91(5):1118–1139
21. Kennedy MC, O’Hagan A (2001) Bayesian calibration of computer models. J Royal Statist
Soc Series B (Statist Methodol) 63(3):425–464
22. Lei CL, Ghosh S, Whittaker DG, Aboelkassem Y, Beattie KA, Cantwell CD, Delhaas T, Hous-
ton C, Novaes GM, Panfilov AV et al (2020) Considering discrepancy when calibrating a
mechanistic electrophysiology model. Philosoph Trans Royal Soc A 378 (2173):20190349
23. Brynjarsdottir J, Hagan AO (2014) Learning about physical parameters: the importance of
model discrepancy. Inve Probl 30(11):114007
24. Worden K, Wong CX, Parlitz U, Hornstein A, Engster D, Tjahjowidodo T, Al-Bender F, Rizos
DD, Fassois SD (2007) Identification of pre-sliding and sliding friction dynamics: grey box
and black-box models. Mech Syst Signal Process 21(1):514–534
Physics-Informed Machine Learning for Structural Health Monitoring 365
25. Vanli OA, Jung S (2014) Statistical updating of finite element model with lamb wave sensing
data for damage detection problems. Mech Syst Signal Process 42:137–151
26. Wan H-P, Ren W-X (2015) A residual-based Gaussian process model framework for finite
element model updating. Comput Struct 156:149–159
27. Rogers TJ, Holmes GR, Cross EJ, Worden K (2017) On a grey box modelling framework for
nonlinear system identification. In: Special topics in structural dynamics, vol 6. Springer, pp
167–178
28. Worden K, Barthorpe RJ, Cross EJ, Dervilis N, Holmes GR, Manson G, Rogers TJ (2018a)
On evolutionary system identification with applications to nonlinear benchmarks. Mech Syst
Signal Process 112:194–232
29. Solin A, Kok M, Wahlström N, Schön TB, Särkkä S (2018) Modeling and interpolation of the
ambient magnetic field by Gaussian processes. IEEE Trans Robot 34(4):1112–1127
30. Wahlström N, Kok M, Schön TB, Gustafsson F (2013) Modeling magnetic fields using gaussian
processes. In: 2013 IEEE international conference on acoustics, speech and signal processing.
IEEE, pp 3522–3526
31. Jidling C, Hendriks J, Wahlström N, Gregg A, Schön TB, Wensrich C, Wills A (2018) Prob-
abilistic modelling and reconstruction of strain. In: Nuclear Instru Methods Phys Res Sect B:
Beam Interact Mater Atoms 436:141–155
32. Karpatne A, Watkins W, Read J, Kumar V (2017a) Physics-guided neural networks (pgnn):
An application in lake temperature modeling. arXiv preprint arXiv:1710.11431
33. Raissi M, Wang Z, Triantafyllou MS, Karniadakis GE (2019) Deep learning of vortex-induced
vibrations. J Fluid Mech 861:119–137
34. Karpatne A, Atluri G, Faghmous JH, Steinbach M, Banerjee A, Ganguly A, Shekhar S, Sama-
tova N, Kumar V (2017b) Theory-guided data science: A new paradigm for scientific discovery
from data. IEEE Trans Knowl Data Eng 29(10):2318–2331
35. Willard J, Jia X, Xu S, Steinbach M, Kumar V (2020) Integrating physics-based modeling with
machine learning: a survey. arXiv preprint arXiv:2003.04919
36. Rasmussen CE, Williams CKI (2006) Gaussian Processes for Machine Learning, vol 38. The
MIT Press, Cambridge, MA, USA
37. Bull LA, Gardner P, Rogers TJ, Cross EJ, Dervilis N, Worden K (2020) Probabilistic inference
for structural health monitoring: new modes of learning from data. ASCE-ASME J Risk and
Uncert Eng Syst Part A Civil Eng 7(1):03120003
38. Wan H-P, Ni Y-Q (2018) Bayesian modeling approach for forecast of structural stress response
using structural health monitoring data. J Struct Eng 144(9):04018130
39. Kullaa J (2011) Distinguishing between sensor fault, structural damage, and environmental or
operational effects in structural health monitoring. Mech Syst Signal Process 25(8):2976–2989
40. Rogers TJ, Worden K, Cross EJ (2020a) On the application of Gaussian process latent force
models for joint input-state-parameter estimation: with a view to Bayesian operational identi-
fication. Mech Syst Signal Process 140:106580
41. Ki-Young K, Brownjohn JMW, List DI, Cole R (2013) Structural health monitoring of the
Tamar suspension bridge. Struct Control Health Monitor 20(4):609–625
42. Cross E, Worden K, Ki Young K, Brownjohn J (2012) Filtering environmental load effects to
enhance novelty detection on cable-supported bridge performance. In: Bridge maintenance,
safety and management. CRC Press, pages 745–752. 10.1201/b12352-101
43. Elizabeth C (2012) On structural health monitoring in changing environmental and operational
conditions. PhD thesis, University of Sheffield
44. Zhang S, Rogers TJ, Cross EJ (2020) Gaussian process based grey-box modelling for SHM of
structures under fluctuating environmental conditions. Proceedings of 10th European workshop
on structural health monitoring (EWSHM 2020)
45. Robert W (2012) Environmental effects on a suspension bridge’s performance. PhD thesis,
University of Sheffield
46. Worden K, Becker WE, Rogers TJ, Cross EJ (2018b) On the confidence bounds of gaussian
process narx models and their higher-order frequency response functions. Mech Syst Signal
Process 104:188–223
366 E. J. Cross et al.
47. Morison JR, Johnson JW, Schaaf SA (1950) The force exerted by surface waves on piles. Petrol
Trans 189:149–157
48. Guilmineau E, Queutey P (2002) A numerical simulation of vortex shedding from an oscillating
circular cylinder. J Fluids Struct 16:773–794
49. Wood AMM, Fleming CA (1981) Coastal hydraulics. Macmillan Education Limited, London
50. Grey-box models for wave loading prediction. Mech Syst Signal Process 159:107741, 2021.
https://doi.org/10.1016/j.ymssp.2021. 107741 ISSN 0888-3270
51. Najafian G, Tickell RG, Burrows R, Bishop JR (2000) The UK Christchurch bay compliant
cylinder project: analysis and interpretation of Morison wave force and response data. Appl
Ocean Res 22(3):129–153
52. Marcus H-A, Nikolaos D, Keith W, Cross EJ, Mills RS, Rogers TJ (2021) Structured machine
learning tools for modelling characteristics of guided waves. arXiv preprint arXiv:2101.01506
53. Cross EJ, Rogers TJ (2021) Physics-derived covariance functions for machine learning in
structural dynamics. In: Submitted to SYSID 2021
54. Papoulis A (1965) Probability, random variables, and stochastic processes. McGraw-Hill Edu-
cation
55. Caughey TK (1971) Nonlinear theory of random vibrations. In: Advances in applied mechanics,
vol 11. Elsevier, pp 209–253
56. Wilson A, Adams R (2013) Gaussian process kernels for pattern discovery and extrapolation.
In: International conference on machine learning, pp 1067–1075
57. Cross EJ, Rogers TJ, Gibbons TJ (2019) Grey-box modelling for structural health monitoring:
physical constraints on machine learning algorithms. In: Structural health monitoring 2019:
enabling intelligent life-cycle health management for industry internet of things (IIOT)—
proceedings of the 12th international workshop on structural health monitoring, vol 2. pp
2136–2145. ISBN 9781605956015
58. Solin Arno, Särkkä Simo (2020) Hilbert space methods for reduced-rank Gaussian process
regression. Statist Comput 30(2):419–446
59. Jones MR, Rogers TJ, Gardner PA, Cross EJ (2020a) Constraining Gaussian processes for grey-
box acoustic emission source localisation. In: Proceedings of the 29th international conference
on noise and vibration engineering (ISMA 2020)
60. Jones MR, Rogers TJ, Martinez IE, Cross EJ (2021) Bayesian localisation of acoustic emission
sources for wind turbine bearings. In: Health monitoring of structural and biological systems
XV, vol 11593. International Society for Optics and Photonics, page 115932D
61. Hensman J, Mills R, Pierce SG, Worden K, Eaton M (2010) Locating acoustic emission sources
in complex structures using Gaussian processes. Mech Syst Signal Process 24(1):211–223
62. Jones MR, Rogers TJ, Worden K, Cross EJ (2020b) A Bayesian methodology for localising
acoustic emission sources in complex structures. arXiv preprint arXiv:2012.11058
63. Kristensen NR, Madsen H, Jørgensen SB (2004) Parameter estimation in stochastic grey-box
models. Automatica 40(2):225–237
64. Herbert JAF Tulleken (1993) Grey-box modelling and identification using physical knowledge
and Bayesian techniques. Automatica 29(2):285–308
65. Lourens E, Papadimitriou C, Gillijns S, Reynders E, De Roeck G, Lombaert G (2012) Joint
input-response estimation for structural systems based on reduced-order models and vibration
data from a limited number of sensors. Mech Syst Signal Process 29:310–327
66. Azam SE, Chatzi E, Papadimitriou C (2015) A dual Kalman filter approach for state estimation
via output-only acceleration measurements. Mech Syst Signal Process 60:866–886
67. Naets F, Croes J, Desmet W (2015) An online coupled state/input/parameter estimation
approach for structural dynamics. Comput Methods Appl Mech Eng 283:1167–1188
68. Maes K, Karlsson F, Lombaert G (2019) Tracking of inputs, states and parameters of linear
structural dynamic systems. Mech Syst Signal Process 130:755–775
69. Dertimanis VK, Chatzi EN, Azam SE, Papadimitriou C (2019) Input-state-parameter estimation
of structural systems from limited output information. Mech Syst Signal Process 126:711–746
70. Alvarez M, Luengo D, Lawrence N (2009) Latent force models. In: Artificial intelligence and
statistics, pp 9–16
Physics-Informed Machine Learning for Structural Health Monitoring 367
71. Hartikainen J, Sarkka S (2012) Sequential inference for latent force models. arXiv preprint
arXiv:1202.3730
72. Rogers TJ, Worden K, Cross EJ (2020b) Bayesian joint input-state estimation for nonlinear
systems. Vibration 3(3):281–303
73. Friis T, Brincker R, Rogers TJ (2020) On the application of Gaussian process latent force
models for Bayesian identification of the Duffing system. In: Proceedings of ISMA 2020—
international conference on noise and vibration engineering and USD 2020—international
conference on uncertainty in structural dynamics
74. Frigola R (2015) Bayesian time series learning with Gaussian processes. PhD thesis, University
of Cambridge
75. Rogers TJ (2019) Towards Bayesian system identification: with application to SHM of offshore
structures. PhD thesis, University of Sheffield
Interpretable Machine Learning for
Function Approximation in Structural
Health Monitoring
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 369
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_18
370 J.-S. Pei et al.
1 Introduction
traditional, opaque ML ML
nonlinear dynamics models
Physics-
low
Based
Models
Interpretable Machine Learning for Function Approximation … 371
University of Southern
0.2
Normalized velocity
Normalized velocity
0.6
0.4
−0.2
.2
The force-state mapping formulation has proven to be one of the most powerful
methods to model nonlinear dynamical systems for aerospace, mechanical, and civil
engineering, and many bio-mechanical systems, e.g., [11, 21, 26] thus being useful
to SHM. Nonetheless, nonlinear restoring forces can be very complex especially
when involving hysteresis (i.e., history/path-dependency) and when time varying
(e.g., caused by deteriorating system properties due to aging or damage). Further-
more, multi-component systems involving multi-degrees-of-freedom (MDOF) can
be extremely challenging. All this indicates the need of going beyond sigmoidal neu-
ral networks. The first step is to make sigmoidal neural networks interpretable for
force-state mapping, and a small subset of nonlinear static function approximation,
as reviewed herein.
ing the focus in our work. Variations in trained weights and biases from
using different sets of training data sets can hardly be related to changes
of the system [34]. Injecting interpretability into the initial design, we
believe, would help alleviate this challenge for damage detection in
SHM.
Table 2 Damping models from [22] and recommended sigmoidal neural networks for initialization
(with initial values for weights and biases obtained but not presented here)
Type Formula Source Sigmoidal neural
network
Linear viscous c · ẋ Slow fluid 1 layer with 2 nodes
Air α · sgn(ẋ)ẋ 2 Fast fluid 2 layers with (3 + 4)
and 8 nodes
Coulomb β · sgn(ẋ) Sliding friction 1 layer with 3 nodes
Displacement-squared d · sgn(ẋ)x 2 Material damping 2 layers with (3 + 4)
and 8 nodes
Solid, or structural b · sgn(ẋ)|x| Internal damping 2 layers with (3 + 4)
and 8 nodes
The notation (m + n) indicates an ANN layer with two separate components of m and n nodes, from
two separate prototypes for multiplicand and multiplier; the 8 nodes in a layer is for multiplication
(see Table 4)
applications. Preparing prototypes and variants takes time, but can be done in a for-
ward problem fashion by following a clear procedure. The resulting prototypes and
variants are generic; i.e., they can be used for numerous individual training tasks
through proper transformations leading to a high overall efficiency in addition to the
built-in rationality and transparency in this proposed initialization methodology.
To complete sparse learning, we implement user-in-the-loop for feature extraction
from the data, match each dominant feature with a sigmoidal neural network proto-
type before concatenating them within a hidden layer (and sometimes, between two
hidden layers) for the backpropagation to be applied without weight freezing. This
growing technique allows for iterations, especially automated iterations for future
work to put in place, among other major improvements to capture more salient fea-
tures and reveal more inner-workings through theoretical work and visualization.
376 J.-S. Pei et al.
offers one case study for a sigmoidal neural network to illustrate the procedures
covering USC AF2 data, and another for a nonlinear autoregressive exogenous model
(NARX) to demonstrate a possible extension of the proposed IML covering NASA
Jet Propulsion Laboratory (JPL) pyroshock data. Section 5 addresses justifications
for this work in addition to the benefits provided by IML, while Sect. 6 discusses
future work. Section 7 provides concluding remarks.
where we have
+∞
f˜(ω) F ( f ( p)) f ( p)e−iωp d p. (3)
−∞
C f plays an important role to optimize the number of hidden nodes and quantifying
the approximation error bound [2, 3]. However, it has been questioned if C f is
sufficient from a computational point of view [51]. In our work, we question why
C f alone may not be a sufficient complexity measure for constructive methods by
examining the set of target functions
1
f ( p, w, b) = σ (wp + b) = (4)
1+ e−(wp+b)
where z = wp + b and stands for convolution [47]. This means that C f can be w-
dependent. This contradicts our common sense that the actual complexity of approx-
imating this family of target functions is the same: One logistic sigmoidal function
would achieve zero approximation error. For the sake of constructive methods, C f
378 J.-S. Pei et al.
perhaps tells only half of the story. The other half of the story, perhaps, is to know
what is considered complex for sigmoidal neural networks and what is considered
not, the so-called “modeling capabilities” of sigmoidal neural networks.
We have started to systematically explore the capabilities of sigmoidal neural
networks algebraically with Techniques A & B and geometrically with Techniques
C to G. Herein, we demonstrate one technique using algebra.
We utilize Taylor series expansion of a sigmoidal function to approximate poly-
nomials. For example, two sigmoidal functions following Eq. (4) are chosen, f 1 and
f 2 , with w1 = −w2 and b1 = b2 = 0. The Taylor series expansion of both functions
at the origin x = 0 to the second power can be written as
1 w1 1
f1 = + p+ (w1 )2 Q 2 (−w1 ξ1 ) p 2
2 4 2!
1 w1 1
f2 = − p+ (w1 )2 Q 2 (+w1 ξ2 ) p 2 .
2 4 2!
The constant term can be eliminated by taking the difference of the above two
functions. For a non-zero w1 , one then has
2
z= ( f 1 − f 2 ) = p + w1 [Q 2 (−w1 ξ1 ) − Q 2 (+w1 ξ2 )] p 2 . (6)
w1
One may choose z to mimic the first power of input p. It can be seen that the
error bound is determined by values of w1 and p. Using this approach as detailed
elsewhere [39, 44], Table 3 gives the derived (not trained) number of hidden nodes,
weights, and biases for the approximation of p 0 , p 1 , and p 2 terms, while their error
bounds are likewise given elsewhere [44]. See Table 4 for the derived (not trained)
number of hidden nodes, weights, and biases for the four basic arithmetic operations,
with details given elsewhere [48].
A basic tenet of soft computing is to avoid “hard computing” methods, i.e., con-
ventional computing methods, even for items as basic as the four arithmetic oper-
ations. Although the implementation of these basic operations draws attention in
other settings such as hardware implementation required in embedded systems (e.g.,
[7]), the same topic has not been studied sufficiently in neural networks. When the
“mapping” of basic items into neural networks cannot be understood, the acclaimed
superiority of neural networks in approximating unknown functions becomes harder
to accept even given some prior studies on this (e.g., [23, 28, 30, 31]). A systematic
solution (not random trial-and-error) based on the governing nonlinear dynamics of
neural network learning, approximation capabilities of sigmoidal neural networks
and features of target functions would greatly benefit numerous applications in non-
linear system identification in addition to advancing fundamental research in neural
computation.
There are other techniques for other basic target functions. Our effort in approxi-
mating a sinusoidal term [40, 48] and certain wavelet-like functions [48] is scattered
and without error bounds estimated.
Interpretable Machine Learning for Function Approximation … 379
Table 3 Derived weights and biases in approximating polynomials from the the zeroth to second
power—general results. From [44]
Target Function Weights in Input Biases in Input Weights in Output layer, L W
layer, I W layer, b
0
p0 w1[0] = b[0] = w2[0] = 1 1
[0] 0
w1,1
[0]
−w1,1
0
w1[1] = w2[1] = + w[1] − w[1]
2 2
p1 b[1] =
[1] 0 1,1 1,1
w1,1
[1]
−w1,1
⎡
[2] 3
⎤T
1+eb1
⎢ ⎥
⎢ [2] 2 b1
[2] [2] ⎥
⎢ w1,1 e −1+eb1 ⎥
⎢ ⎥
⎢
[2] 3 ⎥
⎡ ⎤ ⎢ 1+eb1 ⎥
b1[2] ⎢ ⎥
⎢ 2 [2] ⎥
⎢ b[2] ⎥ ⎢ w[2] eb1 −1+eb1[2] ⎥
⎢ ⎥ ⎢ ⎥
w1[2] = w2[2] =⎢ ⎥
1,1
p2 b[2] =⎢ 1 ⎥ ⎢
[2] 2 ⎥
⎡ [2] ⎤ ⎣ 0 ⎦ ⎢ 2 1+eb1 ⎥
w1,1 ⎢ ⎥
⎢ −w [2] ⎥ 0 ⎢ − [2] 2 b[2] [2]
⎥
⎢ ⎢ ⎥
1,1 ⎥ ⎢ w1,1 e 1 −1+eb1 ⎥
⎢ [2] ⎥ ⎢ ⎥
⎣ w1,3 ⎦ ⎢ [2] 2 ⎥
[2] ⎢ 2 1+eb1 ⎥
−w1,3 ⎣− ⎦
[2] 2 b[2] [2]
w1,1 e1 −1+eb1
IE 1 What are the behaviors/responses in mechanics that these basic modeling ele-
ments ought to capture/approximate? We introduce an ever-growing list of
useful basis functions and features in our work as in Table 2 and Fig. 3.
IE 2 What are the sigmoidal neural network modeling components as basic/essential
as springs and dampers? Here, we introduce neural network prototypes pro-
posed and constructed by us.
IE 3 What are the steps that make a general initial design procedure to answer the
two key questions? We introduce the following three steps:
Step 1 Decompose an unknown target function into a linear combination of
dominant features or basis functions as in IE 1.
Interpretable Machine Learning for Function Approximation … 381
4 Case Studies
Prior work uses training examples to demonstrate the three steps under IE 3 with two
carefully collected laboratory datasets [41, 47]. Here, we briefly review the dataset
labeled AF2 that was previously given in Fig. 2. Figure 4 illustrates the three steps.
The Lipschitz condition for the ODE in Eq. (1) requires the approximation of r (x)
to be bounded. The specified application has further requirements (or preferences)
382 J.-S. Pei et al.
Fig. 4 A step-by-step
respectively, denoted as
Prototype 2a + 1c. Slightly
adapted from [47]
Prototype #1 Prototype #2 Prototype #3 Prototype #1b + #1c Prototype #1b +
- #2
for restoring force for restoring force
vs. velocity vs. displacement
(d) (e)
Σ 1 1
Prototype #2 Prototype #1
displace- Σ
0.5 0.5
ment Σ
Σ restoring 0 0
Σ
force
Σ -0.5
Variant a
-0.5
Variant a
Σ -1
-1 -0.5 0 0.5 1
-1
-1 -0.5 0 0.5 1
on sigmoidal neural networks: (i) fast training convergence, (ii) reasonable accuracy,
(iii) great generalization capabilities, (iv) small or modest computational resources in
terms of network size and learning time (e.g., for wireless structural health monitoring
and intelligent control systems), and (v) constructive and insightful methods (i.e.,
not treating neural networks as “black-boxes,” a goal that we are striving for here).
Prior work demonstrates all these aspects [41], while Fig. 5 contrasts the proposed
with Nguyen-Widrow initialization in terms of computational stability.
Interpretable Machine Learning for Function Approximation … 383
expt
InputDelay = none
InputDelay = 0.0833 ms throughout
InputDelay = 0.1667 ms throughout
InputDelay = 0.0833 and 0.1667 ms
InputDelay = 0 and 0.1667 ms
InputDelay = 0, 0.0833, and 0.1667 ms
10000
A2-Z-Pole-Top-Ch-6
acceleration (g)
5000
-5000
131 132 133 134 135 136 137 138
time (ms)
Input delay has been witnessed to play a significant role for both performance
improvement and inner-working interpretation. When there is no data measurement,
a very rough estimate for the longest input delay using the nominal sound speed in
aluminum, the specimen material, yields
1 meter
× 103 = 0.1582 ms. (8)
6320 m/s
Estimates using the experimental data are compatible with this number leading
to the adoption of tu = 0.1667 ms in Fig. 6.
6 Future Work
Following the layered representation framework in deep learning [29] and XAI [17],
we may use the neural network prototypes in the layer for basic features/simple
shapes. This layer is treated as pre-trained; the work reviewed here provides solutions
that have been obtained in an interpretable manner. We may introduce additional
layer(s) to aggregate the features. We should also introduce additional layers to
devise and automate feature definition, detection, and selection, eventually leading
to a deep neural network in an interpretable manner.
Neuromanifolds could be a productive path to provide new visualization ideas
thus achieving interpretability in deep learning. When minimizing approximation
error, state-of-the-art optimization techniques often seek insights into the underlying
geometry of an error surface in a parameter space for effective numerical iterative
solutions (e.g., training algorithms). Nonetheless, the geometry in this inverse prob-
lem context can be far more challenging than that in a forward problem defined by
well-behaved functions. As typified by sigmoidal neural networks [1], the geome-
try of an error surface has a singular structure proven to exist mathematically but
is obscure to end users of machine learning, who mine the data for a model but
constantly encounter slow convergence without seeing the singular structure and
knowing the roots for the so-called local minima. For IML, it is proposed to create
a new capability for effectively visualizing the geometry of an error surface as the
approximation error being generated on the fly. This is equivalent to creating and
exploring an imaginary and unknown world of an often high-dimensional parameter
space, whose terrains are determined and altered at every move in the numerical
process of conducting an inverse problem where the data and user interact.
Philosophically, the body of knowledge has been developed, verified, and val-
idated long before AI in the specified domain. Thus, AI could yield a very small
portion of this body of knowledge. With this said, it would be rational to explain why
AI makes sense to the existing body of knowledge, rather than the other way around.
Along the same line, it is important to emphasize that AI should play a secondary
rather than primary role when there is a significant existing body of knowledge in
a specific application domain. Cross validating and merging the knowledge from
these two different sources could eventually become possible in such an application
domain. Physics-informed NN may be one such direction [49].
7 Summary
step back first to study how to replicate what existing/traditional methods can do
but by using sigmoidal neural networks. This backward step allows for the introduc-
tion of interpretability that can then be carried forward. For the specified function
approximation application, we then identify dominant/basic features as constituents
by using either domain knowledge or visualization of data, and devise sigmoidal
neural networks (called prototypes) to approximate these features in predetermined
settings. We finally rely on human-in-the-loop to select from the prototypes for a
complete neural network design solution including the obtained number of hidden
nodes and initial weight and bias values. This constructive method is transparent and
interpretable to users and has the potential of directly benefiting deep learning and
other IML applications.
Acknowledgements The research shown in Fig. 6 was carried out at the Jet Propulsion Labora-
tory, California Institute of Technology, under a contract with the National Aeronautics and Space
Administration (80NM0018D0004). The first and second authors acknowledge Dr. John Antonio
for enabling their collaboration.
References
1. Amari SI, Park H, Ozeki T (2006) Singularities affect dynamics of learning in neuromanifolds.
Neural Comput 18:1007–1065
2. Barron AR (1993) Universal approximation bounds for superpositions of a sigmoidal function.
IEEE Trans Inf Theory 39(3):930–945
3. Barron AR (1994) Approximation and estimation bounds for artificial neural networks. Mach
Learn 14:115–133
4. Blum AL, Rivest RL (1995) Training a 3-node neural network is NP-complete. Neural Netw
5:117–127
5. Brewick PT, Abdelbarr M, Derkevorkian A, Kolaini AR, Masri SF, Pei JS (2018) Fusion of
state-space modeling and data-driven strategies for computational shock response prediction.
The American Institute of Aeronautics and Astronautics (AIAA) Journal Published online on
28 Feb 2018
6. Chassiakos AG, Masri SF (1991) Identification of the internal forces of structural systems using
feedfoward multilayer networks. Comput Syst Eng 2(1):125–134
7. Choi S, Scrofano R, Parasanna V, Jang JW (2003) Energy-efficient signal processing using
FPGAs. In: International symposium on field programmable gate arrays, proceedings of the
2003 ACM/SIGDA eleventh international symposium on field programmable gate arrays, Mon-
terey, CA, pp 225–234
8. Cybenko G (1989) Approximation by superpositions of sigmoidal function. Math Control Sig
Syst 2:303–314
9. Erhan YDB, Courville A, Manzagol PA, Vincent P, Bengio S (2010) Why does unsurpervised
pre-training help deep learning? J Mach Learn Res 11:625–660
10. Farrar C, Worden K (2013) Structural health monitoring a machine learning perspective. Wiley,
New York, http://orcid.org/10.1002/9781118443118
11. Farrar CR, Worden K, Todd MD, Park G, Nichols J, Adams DE, Bement MT, Fairnholt K (2007)
Nonlinear system identification for damage detection. Tech. Rep. LA-14353, Los Alamos
National Laboratory
12. Flood I (2002) Neural networks in civil engineering: A review. In: Topping BHV (ed) Civil
and structural engineering computing: 2001, Saxe-Coburg Publications, pp 185–209
Interpretable Machine Learning for Function Approximation … 387
13. Geman S, Bienenstock E, Doursat R (1992) Neural networks and the bias/variance dilemma.
Neural Comput 4:1–58
14. Ghaboussi J, Wu X (1998) Soft computing with neural networks for engineering applications:
fundamental issues and adaptive approaches. Struct Eng Mech 6(8):955–969
15. Greene R, Yau ST (eds) (1993) Differential geometry: partial differential equations on mani-
folds, vol 54
16. Grossberg S (2013) Adaptive resonance theory. Scholarpedia 8(5):1569. http://orcid.org/10.
4249/scholarpedia.1569, revision #145360
17. Gunning D, Stefik M, Choi J, Miller T, Stumpf S, Yang GZ (2019) XAI—explainable artificial
intelligence. Sci Robot 4(37):eaay7120
18. Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural net-
works. Science 313(5786):504–507
19. Holzinger A, Kieseberg P, Weippl E, Tjoa AM (2018) Current advances, trends and challenges
of machine learning and knowledge extraction: from machine learning to explainable AI.
In: Holzinger A, Kieseberg P, Tjoa AM, Weippl E (eds) Machine learning and knowledge
extraction. Springer, Hamburg
20. Hornik K, Stinchcombe M, White H (1989) Multilayer feedforward networks are universal
approximators. Neural Netw 2:359–366
21. Housner GW, Bergman LA, Caughey TK, Chassiakos AG, Claus RO, Masri SF, Skelton RE,
Soong TT, Spencer BF, Yao JTP (1997) Structural control: past, present, and future. J Eng
Mech Spec Issue 123(9):897–971
22. Inman DJ (1994) Engineering vibration. Prentice Hall
23. Jones LK (1990) Constructive approximations for neural networks by sigmoidal functions.
Proc IEEE 78(10):1586–1589
24. Jones LK (1997) The computational intractability of training sigmoidal neural networks. IEEE
Trans Inf Theory 43(1):167–173
25. Judd JS (1990) Neural network design and the complexity of learning. MIT Press, Cambridge
26. Kerschen G, Worden K, Vakakis AF, Golinval JC (2006) Past, present and future of nonlinear
system identification in structural dynamics. Mech Syst Sig Process 20:505–592
27. Kolaini AR, Nayeri R, Kern DL (2009) Pyroshock simulation systems: are we correctly qual-
ifying flight hardware for pyroshock environments. In: 25th Aerospace testing conference
28. Lapedes A, Farber R (1988) How neural nets work. In: Anderson D (ed) Information neural.
American Institute of Physics, Processing Systems, pp 442–456
29. LeCun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521:436–444
30. Lippmann RP (1987) An introduction to computing with neural nets. IEEE ASSP Mag 4–22
31. Lippmann RP (1989) Pattern classification using neural networks. IEEE Commun Mag 47–64
32. Makhoul J, El-Jaroudi A, Schwartz R (1989) Formation of disconnected decision regions with
a single hidden layer. IJCNN Int Joint Conf Neural Netw I:455–460
33. Masri SF, Caughey TK (1979) A nonparametric identification technique for nonlinear dynamic
problems. J Appl Mech 46:433–447
34. Masri SF, Smyth AW, Chassiakos AG, Caughey TK, Hunter NF (2000) Application of neural
networks for detection of changes in nonlinear systems. ASCE J Eng Mech 126(7):666–676
35. NASA (1999) Pyroshock test criteria NASA technical standard. Tech. Rep. NASA-STD-7003,
National Aeronautics and Space Administration
36. Nelles O (2000) Nonlinear system identification: from classical approaches to neural networks
and fuzzy models. Springer, Berlin
37. Nguyen D, Widrow B (1990) Improving the learning speed of 2-layer neural networks by
choosing initial values of the adaptive weights. Proc IJCNN III:21–26
38. O’Donnell KJ, Crawley EF (1985) Identification of nonlinear system parameters in space
structure joints using the force-state mapping technique. Tech. rep., MIT space systems lab.,
SSL#16-85, p 170
39. Pei JS (2001) Parametric and nonparametric identification of nonlinear systems. Ph.D. disser-
tation, Columbia University
388 J.-S. Pei et al.
40. Pei JS, Mai EC (2008) Constructing multilayer feedforward neural networks to approximate
nonlinear functions in engineering mechanics applications. ASME J Appl Mech 75
41. Pei JS, Masri SF (2015) Demonstration and validation of constructive initialization method
for neural network to approximate nonlinear function in engineering mechanics applications.
Nonlinear Dyn 79(3):2099–2119
42. Pei JS, Smyth AW (2006a) A new approach to design multilayer feedforward neural network
architecture in modeling nonlinear restoring forces: part I—formulation. ASCE J Eng Mech
132(12):1290–1300
43. Pei JS, Smyth AW (2006b) A new approach to design multilayer feedforward neural network
architecture in modeling nonlinear restoring forces: part II—applications. ASCE J Eng Mech
132(12):1310–1312
44. Pei JS, Wright JP, Smyth AW (2005a) Mapping polynomial fitting into feedforward neural
networks for modeling nonlinear dynamic systems and beyond. Comput Methods Appl Mech
Eng 194(42–44):4481–4505
45. Pei JS, Wright JP, Smyth AW (2005) Neural network initialization with prototypes—a case
study in function approximation. In: Proceedings of international joint conference on neural
networks 2005 (IJCNN’05), Montreal, Canada, pp 1377–1382
46. Pei JS, Mai EC, Wright JP, Smyth AW (2007) Neural network initialization with prototypes—
function approximation in engineering mechanics applications. In: Proceedings of international
joint conference on neural networks 2007 (IJCNN’07), Orlando, FL, IEEE Catalog Number
07CH37922C, ISBN 0-4244-1380-X
47. Pei JS, Wright JP, Masri SF, Mai EC, Smyth AW (2011) Toward constructive methods for the
universal approximator—function approximation in engineering mechanics applications. In:
Proceedings of international joint conference on neural networks (2011) IJCNN, San Jose CA
48. Pei JS, Mai EC, Wright JP, Masri SF (2013) Mapping some functions and four arithmetic
operations to multilayer feedforward neural networks. Nonlinear Dyn 71(1–2):371–399
49. Raissi M, Perdikaris P, Karniadakis G (2019) Physics-informed neural networks: a deep learning
framework for solving forward and inverse problems involving nonlienar partial differential
equations. J Comput Phys 378:686–707
50. Rudin C (2019) Stop explaining black box machine learning models for high stakes decisions
and use interpretable models instead. Nat Med Intell 1:206–215
51. Scarselli F, Tsoi AC (1998) Universal approximation using feedforward neural networks: a
survey of some existing methods, and some new results. Neural Netw 11(1):15–37
52. Sontag ED (1992) Feedforward nets or interpolation and classification. J Comput Syst Sci
45:20–48
53. Worden K, Tomlinson GR (2001) Nonlinearity in structural dynamics: detection, identification
and modelling. Institute of Physics Pub., p 680
Partially Supervised Learning for
Data-Driven Structural Health
Monitoring
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 389
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_19
390 L. A. Bull et al.
When monitoring engineering systems via measured signals [23], data labelling to
describe the operational, performance, or health state can be difficult to obtain for a
number of reasons; for example:
• There is some high-cost associated with the data labelling process itself—e.g.
expert inspection via non-destructive testing or evaluation procedures.
• It is infeasible to inspect the structure because of difficulty in access—e.g. for
offshore wind farms.
• Some high cost is associated with downtime—e.g. suspending turning operations
to inspect the cutting tool in a lathe during high-value manufacturing.
In these scenarios, providing descriptive labels for all measured data is economically
infeasible or impractical. This problem renders many conventional data-driven clas-
sification models inapplicable, as they require fully labelled or supervised training
data; in other words, each measurement vector xi ∈ Rd must have its own descriptive
(scalar) label yi ∈ R, to define a training set,
Dl = {xi , yi }ni=1
where n is the number of training samples used to learn the classifier f (xi ) = yi
which monitors the system.
The absence of complete labelling {yi }ni=1 will typically force a dependence on
unsupervised learning, i.e. data-driven models learnt from the unlabelled measure-
ments x̃i only; thus, the training data are
Du = {x̃i }m
i=1
With both labelled and unlabelled data, it would be illogical to learn a classification
algorithm given only Dl , while ignoring information in the larger set Du ; likewise, the
converse is true. Instead, data-driven models should utilise the labelled and unlabelled
data in a combined approach, such that the union of both sets is considered,
D = Du ∪ Dl
Conveniently, there are statistical and machine learning tools designing for learning
from partly labelled data; these are referred to as partially supervised learning [51].
Partially supervised learning is used here as an umbrella term, to refer to methods
of leaning from fewer labelled examples1 [47]. Here, a brief introduction is provided
for:
• Active learning,
• Semi-supervised learning, and
• Transfer learning.
Many related techniques exist, some of which are outlined in Sect. 6. Each method
is applied here to a simulated SHM data set. Results for applications to operational
data (from previous publications) are also provided.
The first family of methods considered here are active learners [53]. The main premise
is to improve the predictive performance of the mapping f (xi ) = yi while querying
(i.e. requesting labels) as little as possible [47]. In most scenarios, queries are taken
from the unlabelled data in Du to automatically extend the labelled set Dl .
In general, there are two main approaches: stream-based and pool-based. In
stream-based methods, the data in Du arrive incrementally (in real-time), and the
active learner must determine whether to query or not. On the other hand, pool-
based methods iteratively select the most informative datum from a static set of
unlabelled examples.
Intuitively, active learning has the potential to manage the inspection budget in
SHM. The learner can automatically suggest the measurements for which inspections
appear necessary, to improve (or maintain) the predictive performance of f . A critical
step, therefore, is determining which data should be investigated and labelled. A brief
review of some probabilistic approaches to active learning is provided.
For the first active learning examples, a simulated (vibration-based) data set is con-
sidered, based on an eight-degree-of-freedom benchmark system designed by the Los
Alamos National Laboratory [23]. Following identification of the system parameters
via modal analysis, time series data were simulated for six conditions. Each health
state represents progressive damage, approximated via reductions in the stiffness of
the system (a common assumption in the literature [23]):
From the simulated time series, eight-second windows were converted to the fre-
quency domain (transmissibilities) to define 100 frequency-domain observations per
class, according to the procedure in [10]. Of these data, two-thirds are set aside for
training (D = Dl ∪ Du ) and one-third are held out as an independent test set D∗ .
To visualise the data set and active learning models, the frequency-domain features
are projected via principal component analysis onto two dimensions, as plotted in
Fig. 1.
Perhaps, the most obvious way to query data is to select instances that appear uncer-
tain, given the current model [7]; this procedure is known as uncertainty sampling
[53]. That is, starting from a small number of labelled data, further points can be
queried according to those that appear ‘uncertain’ based on various statistics.
To demonstrate, a Gaussian mixture model (GMM) is learnt for the data presented
in Fig. 1, estimating the parameters in a Bayesian manner [46]. This involves fitting
a Gaussian distribution to each class-conditional density2 p(xi |yi = k) (where k ∈
{1, . . . , 6}) given some small initial labelled set Dl (i.e. leading to six clusters of
data). The GMM resulting from 3% labelled data is visualised by ellipses, one for
each class-conditional, also in Fig. 1.
Given the current GMM and the unlabelled data, observations in Du that appear
most uncertain can be queried. Two measures of uncertainty are considered in this
first example: high entropy and low-likelihood.
Entropy can be used to query observations whose predicted labels appear to be the
most ‘confused’ or ‘conflicted’ [54]. Specifically, measurements whose classification
into all states is equally likely have the highest predictive entropy [5].
Typically, the Shannon entropy of the posterior predictive distribution over the
unobserved labels ỹi is used3 ,
K
H (ỹi ) = − p(ỹi = k | x̃i , Dl ) log p(ỹi = k | x̃i , Dl ) (1)
k=1
The result of querying labels with maximum entropy is to select those data that appear
at the boundaries between existing classes. To visualise, a heat-map of the predictive
entropy over the feature space4 is presented in Fig. 2; unlabelled data lying in dark
red regions would be queried by the learner.
An alternative measure of uncertainty considers data x̃i that appear unlikely given
the current classification model. These ideas align with that of least confident sam-
pling [53], although least confident samples consider the distribution over yi , as
opposed to xi . In contrast to MES, lowest likelihood samples select data that appear
in the extremities of the model, rather than at class boundaries. Considering the GMM
example, low-likelihood measurements are those with a low (marginal) likelihood,
3 That is, the averaged Shannon information content [42] of the possible label outcomes k ∈
{1, . . . , K}, where K = 6 for the simulated data.
4 Given the model and data from Fig. 1.
394 L. A. Bull et al.
K
p(x̃i |Dl ) = p(x̃|ỹi = k, Dl ) p(ỹi = k|Dl ) (2)
k=1
Low-likelihood regions are represented by dark blue areas in Fig. 3. These queries are
useful in discovering new classes of data, as they sample measurements that appear
novel given the model (rather than confused). As such, these queries are arguably
most useful in novelty detection—when applying active learning to streaming data,
for example [7].
Partially Supervised Learning for Data-Driven Structural . . . 395
U (xi ) = H (p(θ |Dl )) − Ep(ỹi |x̃i ,Dl ) [H(p(θ |D, {x̃i , ỹi }))] (3)
(where EA [p(·)] is the expected value of the distribution p(·) with respect to A). Unlike
expression (1), this utility is focussed on the entropy of the posterior distribution
over the parameters p(θ |D)—rather than the predictive distribution over the labels
p(ỹi |x̃i , Dl ). In practice, however, (3) can also be rewritten in terms of the predictive
distribution, as (3) is usually intractable for complex models in the above form [32].
An advantage of the information-theoretic approach is that it takes both measure-
ment and parameter uncertainty into account; this prevents the learner from select-
ing only ‘the most confused’ labels [46]—particularly for discriminative classifiers.
Variation in the sampling procedure is useful to prevent training data from becoming
unrepresentative of the underlying distribution of data—a recognised issue for active
learning methods, referred to as sampling bias [33, 53] (Fig. 4).
While the reasoning behind active learning is intuitive, the performance of active
learners can sometimes prove to be worse than conventional (passive) learning [20,
58]. Specifically, problems occur if sampling becomes too focussed on specific
queries, such that the training data become poorly representative of the underly-
ing distribution, i.e. sampling bias [33, 53]. As such, queries should not focus too
much on specific regions of the feature space—the cluster boundaries, for example.
To avoid sampling bias, various methods have been proposed: combining different
query methods [33] or more ‘adaptive’ query schemes [32].
The query methods summarised here are applied to the eight-degree-of-freedom data.
Initiating with a small (random) sample of labels, the unlabelled data in D are queried
sequentially, as if each point were the last (i.e. myopic active learning [31]). The three
5 For the GMM, the random vector θ contains cluster centres and covariances.
396 L. A. Bull et al.
Fig. 4 Information-theoretic
active learning: darker
regions represent areas in
which data would be
queried. Similar to Fig. 2,
queries are preferred at the
boundaries between existing
clusters; however, in this
case, the parameter
uncertainty for each cluster
is taken into account
Fig. 5 Averaged
classification accuracy for
active learning methods
compared to conventional
(passive) learning, for 400
repeats
where EVPI(d |yi ) is the expected value of querying an incipient point {x̃i , ỹi } to
obtain the labelled observation {xi , yi }, with perfect information, before making a
future decision d . In Eq. (4), MEU(I) denotes the maximum expected utility of a
decision process I involving the random variables {x̃i , ỹi } and a decision d . Similarly,
MEU(Iyi →d ) denotes the maximum expected utility of a modified decision process
Iyi →d , in which yi is obtained prior to decision d being made. Note that the EVPI
is strictly non-negative, and EVPI(d |yi ) = 0 if and only if the optimal policy for d
in I remains optimal for d in Iyi →d . That is, value of information only arises when
labelling has the potential to result in a change in optimal policy.
For stream-based active learning, the EVPI of an incipient data point can be used
to form a convenient heuristic for triggering inspection. Simply, an inspection is
mandated if the following criterion is satisfied,
where Cinsp can be interpreted as the ‘cost’ of inspection. For pool-based (myopic)
active learning, candidate data can be ranked according to their EVPI and queried
preferentially.
Similar to the previous examples, a GMM is used as the predictive model. Given
a small sample of labelled data, a heat-map of the EVPI over the feature space is
shown in Fig. 6a, with clusters corresponding to increasing damage from left to right.
Figure 6b shows the EVPI once the GMM has been updated via value-of-information
queries. Comparing Fig. 6a, b, the region of high EVPI becomes more pronounced
between the two far-right clusters. This occurs, as queries are preferred in areas
between clusters corresponding to high-severity and lower-severity health states,
for which there are differing optimal decision policies. Thus, Fig. 6 highlights that
the value-of-information method queries data that strengthen more ‘costly’ decision
boundaries.
Figure 7 is provided to demonstrate that the performance of a decision-making
agent can be improved via value-of-information sampling. Again, for details of the
decision process, the reader is directed to [36].
Partially Supervised Learning for Data-Driven Structural . . . 399
Formally, entropy minimisation techniques [30] minimise the following loss function
for the unlabelled data [47],
m
K
L=− p(ỹi = k|x̃i ) log p(ỹi = k|x̃i ) (6)
i=1 k=1
One notices similarities to the entropy expression for active learning (1), and that (6)
is minimised when points are assigned to a single class of data with unit probability.
In simple terms, the parameters of the model θ are adjusted such that the unlabelled
data are classified with the maximum possible confidence.
6This is in contrast to active methods, whereby unlabelled instances are only sampled from to
extend the labelled training set.
400 L. A. Bull et al.
m
K
L(θ | D) = L(θ | Du , Dl ) ∝ log p (x̃i | ỹi = k, θ ) p(ỹi = k | θ ) . . .
i=1 k=1
Du
n
+ log p (xi | yi = k, θ ) p(yi = k | θ ) + log p(θ ) (7)
i=1
Dl
For details of how (7) relates to an entropy minimisation viewpoint, refer to [1, 30].
Expression (7) implies that the full joint log-likelihood of the model is maximised,
considering both the labelled (term one) and unlabelled data (term two).
Fig. 8 Black ellipses represent the supervised and semi-supervised GMMs (the blue ellipse rep-
resents the prior of the class-conditionals). Coloured markers correspond to the labelled set, while
black markers are unlabelled. Improvements to density estimation via semi-supervised learning are
clear
Fig. 9 Averaged
classification error for
semi-supervised learning,
compared to conventional
supervised learning, for 50
repeats [10]
402 L. A. Bull et al.
The final family of methods summarised here are transfer learners [49]. In this
setting, the labelled and unlabelled data are split across two or more domains, rather
being generated by the same underlying distribution. This implies that labelled data
from one domain might be used to aid a task in another unlabelled domain, where
the data are generated by different (but similar) processes.
In the context of health and performance monitoring, the ideas of transfer imply
that labelled data from one structure might help annotate the unsupervised mea-
surements from another. Alternatively, the labelled data generated by some model
or simulation might prove useful to annotate in-the-field measurements. Critically,
in these examples, conventional machine learning cannot be applied, as the data
are generated by different distributions for each case: either different structures or
models.
When modelling data from two or more structures, these ideas align with that of
population-based SHM—an emergent field in the SHM literature [9, 26, 29]. An
example of the population-based framework to match labels between structures is
visualised in Fig. 10.
At this stage, it is interesting to highlight how transfer ideas are somewhat analogous
to semi-supervised learning (when target domain labels are unknown). That is, a lim-
ited set of labelled data in Ds are used to support a task—characterised by f (·)—in an
unlabelled domain Dt . However, there is a critical difference: semi-supervised learn-
ing assumes the labelled and unlabelled data are generated by the same underlying
distribution, whereas transfer learning allows for changes in this distribution.
Fig. 10 Separation of
normal and novel data via
transfer learning. Normal
data are and novel
data are Labelled and
unlabelled data sets are
represented by solid and
dashed lines, respectively
8 Stiffness changes due to temperature effects are a typical assumption in the literature [23].
404 L. A. Bull et al.
Fig. 11 Principal component subspace (X ) for the source and target data (filled circle and open circle
markers, respectively). The labelled source data are associated with structure S1; the unlabelled
target data are associated with structure S2. The source and target data appear to be sampled from
distinct (bi-modal) distributions; (three-sigma) ellipses represent the GMM learnt from the source
data. Test data are shown by light · markers
As the data in each domain have a similar structure, transfer component analysis
(TCA) [50] can be applied to match source and target distributions in a shared latent
space. Distribution matching is achieved using a projection ψ(·) such that the distri-
butions can be considered approximately equivalent,
In brief terms, ψ(·) defines a nonlinear projection9 that minimises the distance
between the source and target distributions in the subspace,
Fig. 12 Transfer component subspace (H ) for the source and target data (filled circle and open circle
markers, respectively). The labelled source data are associated with structure S1; the unlabelled
target data are associated with structure S2. The source and target data appear to be sampled from
the same (bi-modal) distributions; (three-sigma) ellipses represent the GMM learnt from the source
data. Test data are shown by light · markers
Thus, in the latent space, a shared model (f = fs = ft ) can be learnt from the super-
vised data in S1, to classify and label the S2 target data: in some sense, this step can
be considered similar to self-training for semi-supervised learners.
The TCA subspace10 for the data from S1 and S2 is shown in Fig. 12, where the
distributions have been (approximately) matched using ψ(·). In consequence, the
GMM classifier learnt from Ds (represented by ellipses) can be seen to generalise to
Dt ; in other words, a single model can be shared between the structures,
f ≈ f s ≈ ft
Thus, the labels should propagate (more reliably) from the source to the target data.
Fig. 13 A visualisation of
negative transfer. Labelled
and unlabelled data sets are
represented by solid and
dashed lines, respectively.
The normal data ( )
and novel data ( ) are
confused between the
labelled and unlabelled
domains/structures
Increasingly, methods for partially supervised learning are being applied in the per-
formance and health monitoring literature. In particular, transfer learning has been
investigated quite frequently, in a range of settings. Some examples are provided for
each of the topics discussed.
Transfer learning:
The majority of literature concerning transfer learning for SHM focusses on image
classification using convolutional neural networks (CNNs), e.g. [13, 22, 25, 37].
Typically, applications consider crack detection [22, 25, 37] and fine-tuning the
parameters of a CNN trained on a source domain to aid generalisation in a target
domain. Generally, fine-tuning does not aim to transfer label knowledge from source
to target domains; instead, it focusses on repurposing expensive-to-train ‘deep’ neural
networks [28].
In addition to fine-tuning, neural networks have been applied for domain adapta-
tion within SHM, for example [40, 55, 57]. These techniques use a neural network
to define the mapping from the domain data into a shared latent space, where a clas-
sification model is learnt. An interesting study is presented by [45], demonstrating
domain adaptation for condition monitoring, applied to a fleet of power plants.
Besides neural networks, a paper by [14] applies a probabilistic method for trans-
fer learning by defining an objective function such that information in the source
and target domains is considered jointly. This application focusses on the important
issue of sensor coverage (for a single structure) rather than transferring information
between similar systems.
Domain adaptation has been discussed in an SHM context by [27, 28], consider-
ing methods for knowledge transfer between a simulated source and target domain,
as well as a simulated source and experimental target structure. Additionally, transfer
component analysis has been applied to a group of tailplane structures (i.e. experi-
Partially Supervised Learning for Data-Driven Structural . . . 407
mental data) to transfer damage detectors between members of a population for SHM
[11].
Semi-supervised learning:
In the context of bridge monitoring, Chen et al. introduced a graph-based approach
for label propagation [18, 19]. The objective function of a multi-resolution classi-
fier [17, 43] is modified such that the weighting parameters are optimised over the
labelled and the unlabelled data. Another graph-based application considers fault
diagnosis for condition monitoring of bearings and pumps [56]. On the other hand,
label propagation within hierarchical clustering has been investigated with experi-
mental aircraft data [6] and pipe monitoring [52]. Generative mixture models have
also been adapted for probabilistic applications of semi-supervised monitoring with
vibration-based aircraft data [10].
Further examples consider applications of K-means [4], Dirichlet processes, and
fuzzy C-means [34] clustering for semi-supervised heuristics. Huang et al. [34]
use fuzzy C-means within an online SHM strategy; the proposed method becomes
partially supervised during a label-matching step, where the unsupervised clusters
are compared to known classes from the supervised data. Bouzenad et al. [4] define a
similar online heuristic using K-means; in this case, new clusters are created when a
distance-based threshold is broken within the unsupervised algorithm. These methods
can be considered as simple form of clustering with constraints [16], whereby partial
supervision is introduced via constraints on an unsupervised algorithm. (However,
in the examples referenced here, partial supervision is enforced within the SHM
framework, rather than the algorithm.)
Active learning:
Active learning is somewhat less explored in performance and health monitoring
applications. Existing studies include generative mixture models for (batch) online
clustering in tool wear [8] and bridge monitoring regimes [7]. Neural networks
have been applied with uncertainty sampling to classify images of defects in a data
set concerning civil structures [24]. Martinez Arellano and Ratchev [44] propose a
Bayesian convolutional neural network for tool-monitoring, using maximum entropy
sampling. Additionally, an adaptive probabilistic framework is proposed in [15] for
active data selection to aid a particle filter-based damage-progression model.
6 Concluding Remarks
Emerging technologies for learning from fewer labelled data in performance and
health monitoring have been summarised and demonstrated. Collectively, these meth-
ods are termed partially supervised learning. Because of the inherent cost associated
with labelling data in monitoring regimes, it is the authors’ opinion that these algo-
rithms will prove critical in developing structural health monitoring, while ultimately
aiding its transition into standard practice in industry.
408 L. A. Bull et al.
Active learning, semi-supervised learning, and transfer learning have been pro-
posed as examples here:
• Active learning can be used as a technique to guide/automate system inspections.
• Semi-supervised learning can be used to combine these inspection results with
large volumes of unlabelled data.
• Transfer learning is proposed to transfer label information between similar systems
(or monitoring regimes).
As with most data-driven monitoring, care must be taken to ensure that any model
assumptions are appropriate; the caveats for each method have been highlighted
throughout. When implemented correctly, however, partially supervised learning
has the potential to significantly improve performance and health monitoring frame-
works.
Considering the foundations in information theory [42] (particularly active and
semi-supervised learning), there are various related problems, in particular: Bayesian
experimental design [38], Bayesian optimisation, and model updating [3]. There are
also related tools from the machine learning literature, including: meta learning, few
shot learning, and data augmentation [47], all of which indicate interesting avenues
of future work.
Acknowledgements The authors gratefully acknowledge the support of the UK Engineering and
Physical Sciences Research Council (EPSRC) through grant references EP/R003645/1, EP/R00
4900/1, EP/R006768/1.
References
1. Amini MR, Gallinari P (2002) Semi-supervised logistic regression. In: ECAI, pp 390–394
2. Arnold A, Nallapati R, Cohen WW (2007) A comparative study of methods for transduc-
tive transfer learning. In: Seventh IEEE international conference on data mining workshops
(ICDMW 2007), IEEE, pp 77–82
3. Behmanesh I, Moaveni B, Lombaert G, Papadimitriou C (2015) Hierarchical Bayesian model
updating for structural identification. Mechan Syst Sig Proces 64:360–376
4. Bouzenad AE, El Mountassir M, Yaacoubi S, Dahmene F, Koabaz M, Buchheit L, Ke W
(2019) A semi-supervised based k-means algorithm for optimal guided waves structural health
monitoring: a case study. Inventions 4(1)
5. Bull L (2019) Towards probabilistic and partially-supervised structural health monitoring. PhD
thesis, University of Sheffield
6. Bull L, Worden K, Manson G, Dervilis N (2018) Active learning for semi-supervised structural
health monitoring. J Sound Vibr 437:373–388
7. Bull L, Rogers T, Wickramarachchi C, Cross E, Worden K, Dervilis N (2019a) Probabilistic
active learning: an online framework for structural health monitoring. Mechan Syst Signal
Proces 134:106294
8. Bull L, Worden K, Rogers T, Wickramarachchi C, Cross E, McLeay T, Leahy W, Dervilis
N (2019b) A probabilistic framework for online structural health monitoring: active learning
from machining data streams. In: Journal of Physics: Conference Series, IOP Publishing, vol
1264, p 012028
Partially Supervised Learning for Data-Driven Structural . . . 409
57. Zhang W, Peng G, Li C, Chen Y, Zhang Z (2017) A new deep learning model for fault diagnosis
with good anti-noise and domain adaptation ability on raw vibration signals. Sensors 17(2):425
58. Zhu XJ (2005) Semi-supervised learning literature survey. Department of Computer Sciences,
Technical Report, University of Wisconsin-Madison
Population-Based Structural Health
Monitoring
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 413
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_20
414 P. Gardner et al.
via the framework of graph theory. This chapter presents and discusses all of these
very recent developments and provides illustrative examples.
one wind turbine may not be possible for a variety of reasons, such as local differences
in weather, local interactions between structures and different operational patterns.
Despite the fact that some wind turbines may have limited or no labelled health-state
data, the asset manager is still tasked with maintaining and monitoring the complete
population. The asset manager therefore requires a population-based approach to
SHM, where the information across the population is used to create a machine learner
that will both generalise across the population and will allow label information to be
transferred to any wind turbine in the farm, allowing robust health diagnostics for
all wind turbines in the population. Figure 1 demonstrates this process, where data
from across the farm are mapped into some space where a data-based model can be
constructed utilising the population-level information.
The above example describes a scenario involving a homogeneous population
where the structures are nominally identical. A second category of population also
exists, termed a heterogeneous population, where every structure in the population is
different for various reasons, broadly categorised as geometric, material and topolog-
ical differences [4, 8, 9]. Staying with the wind farm illustration, imagine the asset
manager is tasked with overseeing multiple wind farms situated around the world,
with each farm containing wind turbines of different model types. A population-
based approach can be extended to consider a larger population, covering all the
wind turbines in the portfolio. However, more care must be taken in scenarios where
the differences between structures in a population are large. For example, the asset
manager must consider the physics of the structures that are grouped into a popu-
lation; are there common failure types that each turbine will experience such that
labelled data can be shared, are there sub-systems or components that are common
across the population? To answer these questions, it will be necessary to demonstrate
an abstract representation framework for spaces of structures, as this will allow an
engineer to quantify and capture similarities between structures via the framework
of graph theory.
A key component of population-based SHM concerns how learning algorithms are
constructed in a population setting. Clearly, data from one structure cannot naively
be used to classify data from another structure, without some mapping to harmonise
datasets, as the generative distributions from each member of the population will be
different. In the context of population types, this chapter outlines two key technology
types for performing population-based SHM, the concept of a Form and transfer
learning. Each are suited to different problem types in PBSHM, which are discussed
within the chapter. Briefly, a Form seeks to capture the essence of a population,
typically by defining a data-based model that captures the expected normal condition
as well as the variability across the population. By contrast, transfer learning seeks to
leverage information from source structures where label information is known and
transfer this knowledge via some mapping to partially or unlabelled target structures.
These types of learning are outlined in the context of each population type along with
illustrative examples.
416 P. Gardner et al.
Fig. 1 A typical population-based structural health monitoring scenario. The scenario depicts a
population of wind turbines ({S1 , S2 , S3 }) where different types of damage have been observed for
each wind turbine over their operational phase (N —normal condition, Dh —damage to the rotor
hub, Db —damage to a blade, Dt —damage on the tower). Population-based SHM methods seek
to define some model that captures information from across the population, generally via some
mapping φ
Population-Based Structural Health Monitoring 417
2 Homogeneous Populations
For a specific model of turbine, the power curve captures the relationship between
wind speed and power output; the associated function can be used as a indicator of
performance [14, 22]. For a wind farm consisting of identical turbines, this trend
should be relatively consistent across the group. Variations in the power curve can
exist for an array of reasons; the results of operator control, or shadowing and wake
effects from other turbines [2], for example.
Intuitively, the power curve defines a convenient object (i.e. feature) to consider
as the form for a wind farm. This relationship captures the essential nature of power
production, while also the variations across the group. To demonstrate, operational
power-curve data (SCADA [22]) from a wind farm are presented in Fig. 2a. A regres-
sion of these data should generalise to future measurements, given optimal power
generation for turbines within the farm.
418 P. Gardner et al.
K
yi(k) = f (k) (xi ) + i k=1 (1)
i.e. the power yi at each input xi (wind speed) is found by evaluating one of K latent
functions f (k) (xi ) with additive noise i . From the power-curve data, it should be
Population-Based Structural Health Monitoring 419
clear that an appropriate number of components is K = 3: (i) ideal, (ii) 50% limited,
and (iii) zero power.
Labels to assign each observation {xi , yi } to function k are unknown, so a latent
variable is introduced to the model, Z; this is a binary indicator matrix, such that
Z[i, k] = 0 indicates that observation i was generated by function k. There is only
one nonzero entry per row in Z (each observation is found by evaluating one function
only). Therefore, for N data, the likelihood of the model is [11],
K
N ,K
Z[i,k]
p y | f (k) k=1 , Z, x = p yi | f (k) (xi ) (2)
i,k=1
A Gaussian process (GP) is associated with each of the (three) latent functions f (k) .
Briefly, each GP can be described by its mean and kernel function [17], which can be
specified for power-curve modelling; for details, refer
to [5]. Unlike
a conventional
GP, the computation of the posterior distribution p Z, { f (k) } | D is intractable; thus,
methods for approximate inference are implemented to infer the latent variables and
functions—a variational inference and expectation maximisation (EM) approach was
proposed in [1]. Additionally, input-dependent noise is approximated according to
the scheme [10]. The resulting form is shown in Fig. 3.
Within the field of transfer learning, domain adaptation arguably offers one of the
most useful tools for PBSHM and is defined as,
Domain adaptation is appropriate for scenarios where a classifier will not generalise
across domains because of differences in the underlying data distributions, such
as the example outlined in Fig. 1. For this reason, the transfer learning methods
demonstrated in this chapter are therefore all forms of domain adaptation.
To contextualise transfer learning for homogeneous populations, a case study is
provided. The case study considers a special case of the homogeneous population
type, when the source and target structure are exactly equivalent; i.e. they are the
same structure. Even in this context, PBSHM provides a useful framework for over-
coming challenges with data-based SHM; in this particular instance, the problem of
structural repairs and how they change the underlying data distribution of a system.
Structural repairs introduce modifications that change (even if locally) the mass,
stiffness and damping of the structure, causing shifts in the underlying generating
distributions and manifest as drift in the output feature space. This dataset shift means
that a data-based model trained on pre-repair labelled data will not generalise to the
post-repair structure because the distributions of the dataset in training and testing
are not the same. The implication of this dataset shift is that a new labelling campaign
would be required every time structural repairs are made. Instead, a PBSHM view-
point can be taken by treating the two datasets as coming from a population of two
homogeneous structures. The following case study considers two datasets from the
422 P. Gardner et al.
Fig. 5 A representative
schematic of the Gnat • C1
aircraft starboard wing (not
P7
to scale), depicting
inspection panel,
accelerometer and
P8
transmissibility path
CR • •C2
locations. Recreated from
[12]
P9 •C3
BR P5
• •B2
B3 • P6
AR
• P4 •B1
P3 •
A3
P2
P1
A1 • A2 •
Fuselage
same Gnat trainer aircraft, before and after the inspection panels have been removed
and reattached (simulating a repair scenario).
The Gnat aircraft dataset was collected as part of an experimental campaign in
which a network of uni-axial accelerometers were used to obtain transmissibility
features (under white noise excitation) from the starboard wing of a Gnat trainer
aircraft in situ [12]. During the experiments, psuedo-damage was introduced into the
structure by removing individual inspection panels—the locations of which are rep-
resented in Fig. 5. The sensor network was designed such that each transmissibility
path targeted a specific inspection panel (i.e. the transmissibility targeting panel 1
(P1 ), denoted T1 , is computed from the reference accelerometer A R and response
A1), with each transmissibility covering a frequency range of 1024–2048Hz con-
taining 1024 spectral lines, with the magnitudes being utilised as the feature data.
In the following analysis, the feature data are seven stacked transmissibilities (i.e.
R1024×7 ) covering panels {P1, P2, P4, P5, P7, P8, P9} where the label space is
Y ∈ {1, 2, 4, 5, 7, 8, 9} (with only the panels with a large surface area being consid-
ered in this analysis). For more details about the experiments, the reader is referred
to [12].
The dataset simulates a repair scenario, because the experimental sequence
(removing all panels one-by-one and replacing them on the structure with the same
applied torque on the fasteners) was repeated twice. In fact, the maintenance process
on the Gnat aircraft wing typically involves removing and reattaching inspection
panels. In order to visualise the problem caused by the repair action, Figures 6 and 7
present two depictions of the dataset. Figure 6 demonstrates the changes in the first
Population-Based Structural Health Monitoring 423
Fig. 6 A comparison of T1
for the pre- (red) and
post-repair (blue) scenarios;
top and bottom sub-panels
depict the normal condition
and the removal of Panel One
transmissibility path before (show in red) and after (depicted in blue) repairs had
taken place for two health states, the undamaged normal condition (top sub-panel)
and when Panel One has been removed (bottom sub-panel). It is clear from this figure
that there are larger changes due to the repair action than due to damage (e.g. the
large differences around 200–400 spectral lines between pre- and post-repair trans-
missibilities), and hence, a classifier will not generalise from the pre- to post-repair
scenarios. Figure 7 shows the first two principal components of the complete feature
space, demonstrating that the data distributions have changed significantly between
the pre- and post-repair structural states, and hence p(Ys , X s ) = p(Yt , X t ) (where
the pre-repair data has been denoted as the source domain, subscript s , and the post-
repair data is denoted as the target domain, subscript t ), so domain adaptation is
applicable.
In order to harmonise the pre- and post-repair datasets, such that label information
from the pre-repair dataset can be used to diagnose the unlabelled post-repair data,
a domain adaptation algorithm has been applied. The algorithm, metric-informed
joint domain adaptation [7], seeks to find a mapping Z = K W ∈ R(Ns +Nt ×k) , by
learning some weight matrix W ∈ R(Ns +Nt ×k) that projects a kernel matrix, formed
from the joint dataset K ∈ R(Ns +Nt ×Ns +Nt ) , onto a k-dimensional space. The weight
matrix is inferred by minimising the distances between the joint distributions from
the source and target data, formed as an optimisation problem that minimises the
maximum mean discrepancy distance between the marginal and class conditional
distributions (for more information about the algorithm, the reader is referred to [7]).
424 P. Gardner et al.
Fig. 9 Comparison of
classification accuracy given
feature spaces with no
transfer learning (Data and
PCA) and the feature space
after transfer learning
(M-JDA)
3 Heterogeneous Populations
Fig. 11 A visualisation of
the first two natural
frequencies (in Hz) from the
three- (•) and four (+) storey
structures
domain adaptation techniques require that both the source and target feature spaces
to be of the same dimension, such that negative transfer is minimised. In this exam-
ple, the feature spaces are the first three bending natural frequencies such that both
feature spaces are R3 , with Fig. 11 presenting a visualisation of the first two natural
frequencies for each structure.
The approach taken to solve this particular L + 1-problem is outlined next. Firstly,
an unsupervised clustering method (namely a Gaussian mixture model) is used to
identify and group the unlabelled target domain feature data. Once unlabelled target
clusters are identified, each target cluster is removed iteratively from the domain
adaptation training dataset. A mapping, in the form of Z = K W ∈ R(Ns +Nt ×k) , is
subsequently identified from the (complete) source dataset to the particular target
dataset (where one cluster has been removed), where the domain adaptation algorithm
uses the maximum mean discrepancy as a cost function and k = 2. The algorithm then
selects the mapping that produces the smallest distance between the source and target
training datasets. This methodology is based on a naive form of manifold assumption;
i.e. it is expected that the manifold of the source and target clusters is the ‘same’.
Figure 12 presents the ‘optimal’ mapping, where the correct target clusters were used
Population-Based Structural Health Monitoring 427
Fig. 12 Transfer learning results for the three- to four-storey example. Panel a displays the trans-
fer components for the three- (•) and four-storey (+) structures, with panel b demonstrating the
maximum mean discrepancy (MMD) distances in the transformed space when one cluster has been
removed from the target domain
in training, and the distance—in the form of a maximum mean discrepancy distance
(MMD)—between the mappings with different target clusters removed. It can be
seen that the ‘optimal’ mapping is selected by this approach, as the smallest MMD
distance is produced when the target cluster corresponding to Y = 5 (i.e. damage at
the fourth storey) is removed in training. A classifier trained on the source domain
data in the transfer component space in Fig. 12 can be shown to classify the target
structure with 100% accuracy using a semi-supervised Gaussian mixture model. This
case study demonstrates the challenges heterogeneous populations cause for transfer
learning techniques, with care being needed in order to minimise negative transfer.
The second example involves a heterogeneous population comprised of a numeri-
cal physics-based model and an experimental structure, as shown in Fig. 13. The aim
in this example is to transfer damage-extent label information from an unvalidated
(and in this case a deliberately poor-performing) numerical model to unlabelled data
from an experimental structure. This case demonstrates the flexibility of a PBSHM
approach in which a variety of sources of label information can be utilised, and
shows a significant advantage of the PBSHM viewpoint, namely that damage-state
data can be generated in a cost-effective manner from physics-based models, even
when computer model validation is challenging. The numerical model, constructed
using the approach in Fig. 10 as a three degree-of-freedom lumped-mass model, was
formed using the measured dimensions of the experimental structure and with typical
material properties that matched those from the structure.
The SHM problem was to classify the extent of damage, in the form of open
cracks from 0 to 20 mm, where label Y = 1 denotes the undamaged condition, Y = 2
refers to a 5 mm crack, Y = 3 for a 10 mm crack, etc. Again, it is reiterated that the
numerical model has been oversimplified such that the example demonstrates the
effectiveness of domain adaptation in utilising physics-based models in labelling
real world structures. The simplified physics-based model therefore reflects that
physics-based models are challenging to validate in SHM contexts and may not fully
428 P. Gardner et al.
Fig. 13 Experimental
three-storey building
structure [8]
agree with observational data, because of model form-errors. The feature data in this
example were the first three lateral bending natural frequencies of the system.
Domain adaptation was applied such that a mapping in the form of Z = K W ∈
R(Ns +Nt ×k) could be identified (where k = 2), using an MMD-based cost function [8].
The inferred mapping is visualised in Fig. 14 where it can be seen that the numerical
and experimental data have been aligned. A k-nearest neighbour classifier was trained
using the numerical model data, both before and after the transfer mapping. The
classifier trained on the untransformed data produced a testing accuracy in the target
domain of 48%; this is compared to a testing accuracy 88% from the classifier
trained on the transfer components. This result demonstrates the effectiveness and
applicability of utilising both physics-based models and observational datasets within
a PBSHM framework.
Both of these case studies highlight the potential of PBSHM beyond populations
of nominally similar structures. Of course, there are a number of research questions
Population-Based Structural Health Monitoring 429
Negative transfer has been mentioned several times in this chapter because of the risk
it poses to making inferences of health states in a PBSHM approach. As previously
stated, negative transfer occurs when class data from a source domain has been
incorrectly and confidently paired with class data from a target domain, i.e. Class
One in the source domain is mapped onto Class Two in the target domain. Generally,
negative transfer is any scenario where transfer learning reduces the performance
on a classifier when compared to not performing any transfer. Negative transfer is a
particular concern for a PBSHM viewpoint, as typically there are no labelled data
points in the target domain that can be used to validate the inferred transfer mapping.
As such, it is important to quantify how related the source and target structures are
and to only perform transfer when positive transfer is likely. This is a significant
and open research question, with the following section outlining one approach to
identified the similarities between members of a population.
To illustrate the effect of negative transfer, the three- to four-storey example is
reintroduced. The problem in this example was to transfer localisation labels from
the three-storey source structure to the unlabelled four-storey target structure. The
problem is an L + 1-problem, meaning that there is one more class label in the
target domain than in the source domain, and that all other class labels exist in
both domains. Figure 15 shows the identified mapping when data corresponding to
classes Ys ∈ {1, 2, 3, 4} and Yt ∈ {1, 2, 4, 5} were used to train the domain adaptation
mapping—this gave the second smallest distance in Fig. 15 and therefore was not
selected as the optimal mapping, but will be useful for this discussion. Of particular
note, negative transfer has occurred between the source data for class Y = 3 and the
target data for class Y = 4. This is interesting, as source data for class Y = 4 were
used in training, but the domain adaptation algorithm, given no label information
about the target, has inferred a mapping that incorrectly pairs these classes. Negative
transfer has also occurred between class Y = 4 in the source domain and Y = 5 in
the target, which arises as the cost function in the domain adaptation approach only
considers a mapping with the smallest distance between the two domains.
In this example, it should be clear that the risk of negative transfer was high, as
the data from the source and target domains did not correspond to the same classes
in training. However, it is worth noting that negative transfer can occur in scenarios
where both the source and target training data do refer to the same classes, and
where the two datasets are more dissimilar than can be accounted for given the
assumptions and adaptation method (i.e. the type of mapping) used in the transfer
learning algorithm. For PBSHM to be applied with confidence using transfer learning
430 P. Gardner et al.
Fig. 15 An example of negative transfer between the three- and four-storey structures. Panel a
shows the first two natural frequencies of the source domain data (•) used in training, panel b shows
the first two natural frequencies of the training target data (+), and panel c shows the inferred
transfer components after domain adaptation, where negative transfer has occurred
Fig. 16 A simple illustrative example of how a wind turbine (a) may be conceptualised as a series
of elements (b). These elements then form the nodes in a graph representation of the structure (c),
with the physical connections between elements represented by the edges in said graph
pair fully describes the effect of the boundary condition on the behaviour of the
structure.
Isomorphic to the information within the IE model is the attributed graph (AG)
representation of the physical structure; that is, the representations in panel (b) and (c)
within Fig. 16 contain the same physical structural information. One could consider
that the AG provides a more structured form of the data describing an IE model,
and as such facilitates the storage and creation of IE models within a database. The
formal data structure of an AG also facilitates the use of graph-matching algorithms
to find similar physical structures within said database.
In PBSHM, IE models can be used to compare two structures to determine the
overall level of similarity. If two models are sufficiently similar, then transfer between
the two structures should be possible. What ‘sufficiently similar’ means however,
is a complicated question; one way of determining this is to examine the largest
substructure common to two structures. If the largest common substructure is in fact
as large as the two structures in question, then the two structures can be said to
be homogeneous. In which case, as discussed, not only should transfer learning be
possible, but in some cases may not even be required. Of course, for the majority of
comparisons between structures, this situation will not be the case, with the largest
substructure only representing only a small part of the overall structure. In this case,
transfer learning may be possible within the substructure. For example, in Figs. 17 and
18, which show the AGs for two different beam and slab bridges, the intermediate
pier is topologically identical in both. If upon further examination, the attributes
show that this pier is identical in terms of geometrical and material properties, then
it would seem that some form of information transfer should be possible between
432 P. Gardner et al.
Fig. 17 A graph representation of a beam and slab bridge, located near Castledawson in Northern
Ireland. This bridge features a deck supported by four longitudinal beams, as well as columns at
the North and South abutments
the intermediate piers within these two bridges, making allowances for the influence
from the rest of the structure. At the very least, the label spaces would be consistent
between the two.
A more complex problem would be transferring between the deck and parapet
sections for the two bridges, since in Fig. 18 there is an additional support beam.
Here the label spaces are not consistent if one were to attempt a damage localisation
problem. However, in theory, both deck sections should still exhibit some similar
behaviour, since both are plate type elements supported with longitudinal beams for
support and intermediate piers and so some physics should be common to the two
decks. Therefore, there would be an expectation that some information could be
transferred from a classifier trained on one bridge to another.
The largest common substructure found for the two bridges in fact involves not
only the intermediate pier, but also the four longitudinal beams within the deck
structure, the parapets, the deck itself and associated boundary conditions, as well as
the cap beams linking the deck to either the supports or foundations. This situation
creates a problem where there is still a label mismatch between the substructure
and the original substructure, as shown in Fig. 18. How to cope with these issues and
how to perform meaningful similarity comparisons within a transfer learning context
remain open research questions.
A definition of similarity for PBSHM needs to be designed with transfer learning
in mind and will likely vary depending on the SHM problem in question. A set
of criteria informed by cases where negative transfer is likely to occur would be a
Population-Based Structural Health Monitoring 433
Fig. 18 A graph representation of a slightly different beam and slab bridge, located near Randal-
stown in Northern Ireland. This bridge features a deck supported by five longitudinal beams, and
the cap beams sit directly on top of the foundations at either abutment
useful thing to have. Currently, it is hoped that by ensuring material, geometric and
topological similarity, the criteria for transfer learning yielding improved classifier
performance are met.
Clearly, the subject of PBSHM is in its infancy. While there have been scattered
papers on SHM in systems-of-systems or on fleet-based maintenance, etc., the formal
framework presented in this chapter has only the theoretical foundations established
[4, 8, 9, 19, 21]. This situation means that the scope for ‘future work’ is very open
indeed; however, to bring a little focus to the discussion, it will be suggested here
that there are three main areas in which research is needed.
The general framework is based on knowledge transfer in populations, where the
structures of the population are represented in an abstract representation space; this
suggest the three areas for development. In the first case, the abstract representation
of the structures requires research. There will need to be rules for building IE models
with consistency and for developing the theory of the representation space in which
the structures are embedded. Perhaps most important is that the representation space
434 P. Gardner et al.
should have a metric (or metrics), so that structures can be compared; transfer of
knowledge will be contingent on structures being ‘close together’ in some sense, so
that negative transfer is avoided.
The second main area for research is on transfer. As discussed in the main body
of this chapter, transfer learning has only recently been considered for PBSHM; the
problems which arise are difficult. In the worst case scenario, different structures will
have different label spaces and current technology cannot be applied with confidence.
New transfer learning algorithms are needed, perhaps informed by physics, as a great
deal is known about the dynamics of structures.
Finally, in order to transfer knowledge, one must have knowledge. Data for diag-
nostics can come from sensors on the structures or from model predictions; these data
should be optimised for diagnostics across populations, and this produces demands
and constraints that have not been seen before in conventional SHM. In particular,
the role of physics-based models is very interesting in PBSHM. Because the main
algorithms will come from transfer learning, it will not be necessary that models
conform perfectly to their physical counterparts (although digital twin technology
may offer prospects in terms of high-fidelity representation); it will only be required
that transfer is possible. In fact, the way that models of structures will be embedded in
populations means that they are not distinct from real structures. PBSHM therefore
offers the prospect of a final convergence of model-based and data-based SHM.
Acknowledgements The authors gratefully acknowledge the support of the UK Engineering and
Physical Sciences Research Council (EPSRC) via Grant references EP/R003645/1, EP/R004900/1
and EP/R006768/1. They would also like to thank their colleagues David Hester and Andrew Bunce
of Queen’s University Belfast, for their help and advice in building IE models of bridges and for
access to the details of specific bridges.
References
Abstract Damage identification via remotely sensed data amid routine structural
inspections or assessments in the aftermath of extreme events has been the subject
of intensive research in recent years. By analyzing remotely sensed data, these
tasks can be performed rapidly, safely, and economically while maintaining high
accuracy and objectivity. Therefore, many methods have been proposed to detect
damage from remotely sensed point clouds of the civil structures. These methods
use various feature extraction techniques based on geometry and other spectral infor-
mation and classification based on supervised or unsupervised learning algorithms.
However, the proposed solutions are typically optimized to detect particular types of
damage or group various types of damage as a single class, limiting the application of
these methods. This article proposes an advanced workflow to identify damage from
point clouds by extracting semantic information using supervised and unsupervised
learning methods. The supervised learning method is comprised of a deep learning
model developed based on both voxel- and index-based data structures to segment
the point cloud data into various objects semantically. An unsupervised learning
algorithm classifies the segmented scene into instances for damage detection into
various damage states and handles rare cases of damage instances.
1 Introduction
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 437
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_21
438 M. E. Mohammadi and R. L. Wood
2 Background
as anomaly instances, which accounts for the imbalanced nature of the damage. The
GAN model generator architecture consisted of U-net encoder-decoder model, and
the discriminator architecture consisted of a series of convolutional layers [9]. To
test the model, the authors used post-disaster images following a volcanic eruption
took place. Tilon et al. [8] reported the developed model could detect damaged areas
within images with recall and precision of 59 and 97%, respectively.
Similar to images, 3D point clouds have been investigated to address various
tasks within civil infrastructure assessments. One of the popular workflows to detect
damage and changes within the area of interest (ROI) is through change detection
analysis. As one of the early studies, Olsen [10] introduced a workflow to detect
spatial changes for a given ROI through change detection analysis. Within this
study, the point cloud of an ROI at two different epochs are collected and registered
through the georeferencing process. Afterward, the point clouds were segmented
into a series of cell based on selected gird sizes, and the points within the corre-
sponding gird are compared to detect changes and potentially damaged areas. Tran
et al. [11] introduced a machine learning-based change detection analysis of point
clouds collected by aerial laser scanners (ALS). Within this study, initially, each point
cloud dataset collected at different epochs were analyzed to extract four classes of
features, including point distribution, normalized height, ALS-based features per
each dataset, and the localized changes between dataset per each point. Afterward,
these features were analyzed by a learning algorithm to predict the changes. The
developed model based on the designed features could detect changes as accurately
as 90%. While change detection analysis is affected by the imbalance dataset clas-
sification issues, it relies on the availability of a baseline dataset. Its accuracy is the
function of the registration of point clouds collected at two different epochs. In addi-
tion, to change detection-based workflows, a series of studies followed a traditional
classification workflow where a series of features were identified through engineered
features, and then the points are classified based on the extracted features through
a classifier. As one of the early studies, Axia et al. [12] used the normal vector as
the extracted feature and studied the normal vectors variation of a large point cloud
dataset to a global reference vector to identify damaged regions. The dataset utilized
within this study was collected following the 2010 Haiti earthquake and covered a
1.5 km by 1.5 km wide area of Port-au-Prince, Haiti. The damages areas were identi-
fied if the deviation of points’ normal vector exceeds the predefined threshold value.
While the proposed method could be programmed to analyze point clouds rapidly, the
authors reported that this approach could classify partially damaged structures and
undamaged areas. He et al. [13] used a series of 3D shape descriptors to detect roof
damage from aerial point clouds of Haiti’s National Palace, Port-au-Prince, Haiti,
which is collected in the aftermath of 2010 Haiti. Within the proposed method, the
scene’s point cloud representation was initially processed into a digital elevation
model (DEM). Then, the point cloud data and created DEM were further processed
to identify the building locations, and the 3D shape descriptors were computed for
each building. The proposed solution then classified the buildings’ point clouds into
damaged and undamaged classes by thresholding the contour shapes based on the
jaggedness. The authors reported that the proposed workflow resulted in an overall
Machine Learning-Based Structural Damage Identification … 441
accuracy of 89%, while the recall and precision values were 89 and 58%, respec-
tively. Axel and van Aardt [14] improved the damage detection method based on
normal vector variations by introducing preprocessing steps to eliminate the points
representing vegetation and ground surfaces based on above the ground (AGL) height
and surface roughness. To detect damaged structures, Axel and van Aardt [14] used
ALS-based point cloud data of damaged areas following the 2010 Haiti earthquake
and utilize region growing approach to detect and separate structures from the rest of
the scene and then analyzed the structures for damage through evaluating the varia-
tion of normal vectors and height analysis. The proposed method resulted in overall
detection accuracy of 93% and damage classification of 78.9%.
In addition to the studies that focused on detecting damaged areas from point
clouds, multiple studies have also investigated the task of semantic segmentation
from point clouds. Hackel et al. [15] introduced a fast semantic segmentation method
to classify lidar point clouds to a series of semantic objects. Similar to previous
studies, the proposed method started by extracting features from point clouds directly;
however, Hackel et al. [15] relied on the geometric features extracted from a point
and its selected neighboring points at various distances based on principal compo-
nent analysis (PCA) and approximating the three-dimensional shape context features
using histogram-based descriptors. The study has reported that the proposed method
can be predicted and classified points with an overall accuracy of 90.3%.
Besides change detection analysis and traditional methods to detect damage or
semantically classify the point clouds into various objects, various research studies
have focused on investigating the application of deep learning to analyze point
cloud datasets. Here, deep learning corresponds to the learning algorithms that learn
features during the training process. As 3D data can be represented based on various
structures, various deep learning-based solutions based on these 3D representations
are introduced. The first group of studies that are reviewed here developed deep
learning models based on volume element (voxel) grid representations of the 3D
point clouds. Prokhorov [16] was an early study to use voxel representations to
develop a model to classify 3D point cloud data. Prokhorov proposed a model with
an architecture similar to that of CNN to classify various objects’ point clouds.
The developed model had one 3D convolutional layer, one 3D pooling layer, and two
fully connected layers, which is followed by 2-class output later. The training strategy
used within the study includes pretraining the parameters within the convolutional
layer based on lobe component analysis. Maturana and Scherer [17] expanded the
model proposed by Prokhorov [16] for the task of object recognition with a new
training strategy and method to create voxel models, which is known as Voxnet. The
proposed model had two tandem 3D convolutional layers, one max-pooling layer,
and one fully connected layer, which was followed by the output layer. In contrast
to the study conducted by Prokhorov [16], Maturana and Scherer [17] did not pre-
trained the developed network while the model resulted in a performance on par or
better than the network proposed by Prokhorov [16], which could be attributed to
the training strategy and optimization process.
As voxel representation of point clouds requires a larger memory footprint
compared to the raw point cloud representations, multiple studies have investigated
442 M. E. Mohammadi and R. L. Wood
other data structures that require smaller memory to store data. As one of the main
studies, Qi et al. [18] developed a deep learning model using raw point cloud data
as input instances, known as PointNet. Within this model, a multi-linear percep-
tron (MLP) was applied on each input file point, and these MLPs are combined to
extract a signature for the input instance. However, the PointNet architecture did
not account for the local geometric features. To address this matter, Qi et al. [19]
proposed a hierarchical model that applies the PointNet model in a recursive pattern
on point cloud segments that were created by sampling and grouping layers, known
as PointNet++ . Recently, Can et al. [20] utilized PointNet++ to classify the UAS-
SfM-derived point clouds of large urban areas and reported an overall accuracy
and IOU score of 82 and 45.3%. While the PointNet++ took the local geometric
information of point clouds into account, this model’s input size was limited to a
predefined number of points, and the model could not use color or other spectral
information to improve the classification. Kolkov and Lempitsky [21] introduced a
deep learning model based on kd-tree representations of point clouds, which is named
kd-network. The computational graph was created using the kd-tree structure, and
the model shares similarities to CNN models, including sharing learnable parame-
ters and learning the first-order features, and combining the learned features with
higher-order extracted features to analyze point clouds. Kolkov and Lempitsky [21]
compared the developed model’s performance on the Shape-Net-core dataset with
the Voxnet and PointNet and reported that while the kd-network performed 3% better
than Voxnet on average, it demonstrated similar performance to PointNet. Similarly,
Riegler et al. [22] proposed a deep learning model based on the octree representa-
tion of point clouds known as OctNet. The developed model used the octree spatial
partitioning method to create a computational graph, and similar to CNNs, OctNet
utilizes a convolution operation to extract and combine features. Riegler et al. [22]
tested the developed model to semantically segment the façade of a structure and
reported an overall accuracy and IOU score of 81.5 and 59.2%, respectively. Unlink
PointNet, the kd-network, and OctNet models could use the local geometric infor-
mation; however, only OctNet demonstrated the potentials to accept color and other
spectral information along with geometric information. More recently, Wen et al.
[23] introduced a model by combining the graph-based attention layers and CNNs
to analyze raw point cloud representations based on global and local geometric
information, known as GACNN. The developed model contained a graph attention
module and encoder-decoder network. The graph attention module accepted raw
point clouds and identified the kd-tree representation of the input instance. Then, the
model analyzed the data based on a series of MLPs and two attention layers, namely
global and local attention layers, each of which was constructed by a series MLPs. In
the encoder-decoder model with skip connections, the encoder network downsam-
pled the input and used the graph attention module to analyze the parameters. The
decoder interpolated the data and used the graph attention module to compute the
parameters. Wen et al. [23] utilized the model to segment the scene into nine classes
semantically, included powerlines, low vegetation, impervious surfaces, car, fence,
roof, façade, shrub, and tree. The study had concluded that the developed model
could semantically segment the scene with overall precision and recall values of 74
Machine Learning-Based Structural Damage Identification … 443
and 70%, respectively. However, it was noted that the model could only detect some
of the classes (e.g., impervious surfaces that represent road or sidewalks) with high
accuracy compared to models that are developed based on PointNet architecture.
While the previous studies have demonstrated various methods to detect damage
from aerial point clouds, most machine learning-based solutions that were proposed
to analyze the point clouds did not focus on the task of damage detection and damage
state classification. This manuscript aims to introduce a new approach to analyzing
the UAS-SfM-derived point cloud data of large areas after an extreme event (e.g.,
hurricanes or tornadic events) to identify damaged areas semantically segment the
point cloud representations. As a result, a workflow based on 3D fully convolutional
network (3D FCN), k-means clustering algorithm, and 3D CNN model were devel-
oped to semantically segment the UAS-SfM point clouds for the task of damage
identification. The 3D FCN and 3DCNN models within this study are independently
trained to learn the features based on the training instances. The developed workflow
learns each class based on the 3D geometry of input instances and the color informa-
tion collected during the data collection. Ultimately, this method allows the analyst
to perform a damage assessment of built-up areas directly.
3 Dataset
On March 3, 2020, a tornado impacted northern Tennessee and the Nashville metro
area as an EF3 with estimated peak wind speeds of approximately 265 kph (165 mph).
It was reported that the tornado remained on the ground over 60 miles, making it
the longest tornado path officially recorded since 1950 (the earliest documentation
of path length), and resulted in five fatalities and 220 injuries [24]. Following the
event, a team on behalf of NSF-StEER conducted a series of field assessments along
the tornado path to document the distribution and intensity of the damage within the
affected areas. The StEER team reconnaissance focus included door-to-door (d2d)
assessment, aerial surveys via UASs, and vehicle-mounted street view imaging to
document damaged areas and buildings’ performance during the event [2].
One of the surveyed areas included the Lockland Springs neighborhood, as shown
in Fig. 1, (approximately at a latitude of 36.17501 and a longitude of − 86.72643)
[2]. The aerial images were collected using a DJI Mavic Pro 2 platform at an altitude
(i) Waterbodies
paved roads, sidewalks, bridges, parking driveways, and gravel roads or driveways
that are regularly used for vehicle movements. The terrain class contains grasslands,
bushes, utility boxes, postboxes, fences, chairs, and patio furniture. The poles class
consists of transmission poles and wires supporting a variety of services, including
power, telephone, and data.
To classify structures, five different categories are selected based on the StEER
d2d ratings included destroyed structures, severely damaged structures, moderately
damaged structures, minorly damaged structures, and undamaged structures [2].
A structure is considered destroyed if more than 15% or higher percentage of the
structure’s roof or walls is failed, or more than 25% roof deck is damaged. The
severely damaged structure classification comprises structures that sustained 50%
or more damage within the roof covering or walls, lost between 5 and 25% of their
roof sheathing, or lost 15% or less of the roof structure. The moderately damaged
structure classification comprises structures that either 15–50% of the roof covering
446 M. E. Mohammadi and R. L. Wood
or walls sustained damage or less than 5% of their substructure is failed. The minorly
damaged structure classification includes the structures that have been damaged at
15% or less in terms of roof cover or wall damage. Lastly, only structures that do not
exhibit any visible damage are considered undamaged structures. Table 1 summarizes
the number of points that were segmented for the Lockland Spring area dataset.
The classification of UAS-SfM point clouds for the task of damage detection intro-
duces a unique set of challenges. While the color, geometry, and point density vary
between objects of the same class, these variations combined with random and
unique geometric damage patterns existing within the point clouds make the task
of damage detection more difficult than the semantic segmentation of the scene.
Mohammadi et al. [25] and Liao et al. [26] presented deep learning models based on
3DFCN to detect damage from UAS-SfM-derived point clouds. While the 3DFCN
models introduced in those studies were able to learn and predict the geometry input
instances (e.g., terrain class), the overall precision and recall values for damaged and
undamaged instances were limited and only between 10 and 20%, respectively. This
manuscript expands on this earlier work and introduces a new workflow based on the
Machine Learning-Based Structural Damage Identification … 447
fusion of multiple deep learning approaches. This includes 3DFCN, k-means clus-
tering algorithm, and 3DCNN to detect damaged areas from the UAS-SfM derived
point clouds. The main idea behind the introduced workflow is that the damaged
instances can be detected with higher accuracy within the pre-segmented point cloud
data compared to classifying the entire scene for damaged areas and other classes.
The developed workflow initially classifies the point cloud data into seven general
classes including vehicles (includes boats and cars), debris (includes debris and fallen
trees), water bodies, roadways, healthy trees (includes health trees and poles), terrain,
and structures (at various damage states, including destroyed, severely damaged,
moderately damaged, minorly damaged, and undamaged structures). Once the point
cloud data are segmented into these seven classes, only points that are classified
as structures are analyzed by the k-means clustering algorithm to detect each indi-
vidual structure, and lastly, the 3DCNN network evaluates each structure class and
further classify it into five damage states (or classes) of destroyed, severely damaged,
moderately damaged, minorly damaged, and undamaged (or no damage) structures.
This section initially describes the data preparation and transformation process to
convert raw point cloud data into 3D voxels used within this workflow, then presents
the developed network architecture, and finally reviews the training strategy used
to develop the two models. The 3DFCN and 3DCNN presented in this study were
implemented in TensorFlow. The k-means clustering algorithm was implemented
through the Scikit learn library.
The process to create instances within this study is similar to the approach proposed by
Mohammadi et al. [25], which is briefly described here. As stated, the process to create
the training and testing instances are initiated by manually segmenting the UAS-SfM
point clouds into the 14 classes. Afterward, a label corresponding to each segmented
object was assigned to all its points. Then, all the labeled objects are compiled into a
single file. Lastly, the compiled file is divided into 10 m × 10 m (32 ft × 32 ft) parts to
create segments of equal dimensions consisting of multiple objects that can be used
for semantic segmentation. The selected dimensions here permit the user to control
each instance point-to-point spacing or resolution. In the first step, the minimum value
for each instance point coordinates is computed. Then, the values are subtracted by the
identified minimum values for each component (i.e., X, Y, and Z) to transfer the data
into positive ordinates. In the next step, the positive coordinates are downsampled
based on the selected voxel dimensions. Within this study, a voxel dimension of
64 is used, which results in a sampling of roughly 16 cm (6.2 in). The sampling
rate of 16 cm results in an acceptable voxel representation for building damage
assessment following windstorms [27]. In the last step, the voxel coordinate values
are multiplied by the selected dimension for the voxel representations, normalized
based on coordinates computed ranges, and rounded. Lastly, as each voxel model has
448 M. E. Mohammadi and R. L. Wood
a number of empty arrays or cells and occupied arrays, an extra label corresponding
to empty space is added to the training labels known as neutral class.
The 3DFCN developed for this study is inspired by the previous work of Long et al.
[28], which introduced the FCN architecture to analyze 2D images (Fig. 4). The
developed model has an input layer that accepts three 3D voxels with 643 cells
matching the three RGB channels and comprises encoding and decoding parts. The
encoder part has a total of six convolutional layers with linear rectifier units, and
the decoder part has a total of six transpose convolutional layers with linear rectifier
units. The network architecture contains skip connections that allow the convolutional
layers to copy their input into transpose convolutional layers. This allows the latter
layers to learn new information but retain the initial knowledge learned by previous
layers, minimizes the potential gradient vanishing issues [9]. The kernel sizes in the
convolutional and transpose convolutional layers had a size of 3, and in each layer, a
total of 32 kernels existed. Furthermore, the stride parameter was set to a value of 1
for these layers. In the last layer of the network, each kernel also has a self-attention
module that is comprised of two multilayer perceptrons with 32 nodes. These self-
attention layers enabled the model to restore the labels at a higher rate. Lastly, the
output layer is a single voxel with a size similar to input size, each of which cells
represent the label of the input point cloud instance. The model was trained based
on the minibatch size of 64 using an Adam optimizer and an L2 regularization with
the lambda of 0.0005 to reduce the overfitting potential during training.
Once the point cloud data is classified into seven classes by the 3DFCN model, a query
search initially collects all the points classified as structure, and a statistical outlier
removal (SOR) minimizes the number of sparse points within building segments [29].
Afterward, the k-means clustering algorithm categorizes the points that represent
structures into individual buildings. The k-means clustering algorithm here requires
two input parameters, including the number of clusters and a threshold value. The
first parameter, the number of clusters, in this study is determined by double the
number of buildings within the ROI that are being analyzed. The second parameter,
the threshold value, was estimated based on the minimum distance between two
structures. Within this study, the threshold value was set to the value of 0.20 m.
The 3D CNN model developed within this study was inspired by Voxnet [17]. The
developed model, similar to 3DFCN, has an input layer that accepts three 3D voxels
with 643 cells matching to RGB channels, two convolution layers followed by a
pooling layer, and two fully connected layers or multilayer perceptrons. Lastly,
the network has five nodes within its classification layer, as shown in Fig. 5. The
convolutional layers have linear rectifier units with stride parameters of 1 for both
convolutional layers, respectively. Within this model, a max-pooling approach with a
stride parameter of 2 was utilized to reduce the number of parameters and introduce
invariance to translations. The network utilizes the SoftMax function to determine
the class of the inputted voxel instance within its classification layer. Like 3DFCN,
the model was trained based on the minibatch size of 32 using an Adam optimizer
and an L2 regularization with a lambda of 0.001 to reduce the overfitting potential
during training.
The hyperparameter search and the developed 3DFCN and 3DCNN models were
primarily conducted on the GPU resources at the Holland Computing Center, located
at the University of Nebraska-Lincoln. The testing and part of the training process of
3DCNN were conducted locally on a workstation with Nvidia RTX A6000 GPU. To
evaluate the developed models, the confusion matrix and three performance measures
(precision, recall, and IOU) were calculated for both training and testing phases. Both
3DFCN and 3DCNN models were trained on roughly half of the instances and tested
on the rest of the data to investigate the learning transformability of the proposed
workflow. To prepare the training dataset for 3DFCN and to increase the number of
instances, a total of 8677 unique 10 m × 10 m segments were augmented by randomly
rotating each instance two times around its vertical axis. This results in a total of
approximately 17,000 training instances. As for 3DCNN, all the structure classes
were augmented by randomly rotating each structure instance ten times around its
vertical axis resulted in a total of approximately 9000 instances.
The 3DFCN model was trained for a total of approximately 1600 epochs prior to
testing. To identify a set of optimized parameters, multiple networks were initially
trained through a factorial design process, in which the L1 and L2 regularization
lambdas learning rates, and the number of kernels per layer varied. Through this
hyperparameter tuning study, the network with a total of 32 kernels per layer, L1
regularization lambda of 0, L2 regularization lambda of 0.0005, and a learning rate
of 0.0015 proven to be the most successful model. The most successful model was
identified through evaluating model performance measures, including precision and
recall, as well as training mean square error values. As a result, this model was
selected for the extended training process. Table 2 presents the performance measures
for the training and testing sets, where the model was able to detect instances with
average precision, recall, and IOU of 51, 52, and 38%, respectively. Moreover, Fig. 6
demonstrates the confusion matrix for the training and testing phases. As shown,
the model was able to learn all instances with precision and recall value of 40% or
higher except for classes water bodies, vehicles, and roadways. However, it was noted
that these three classes represent the lowest number of instances within the training
dataset, making it difficult for the model to learn. The testing results demonstrate
a trend in performance measures where an average value of 47, 43, and 33% was
observed for precision, recall, and IOU metrics, respectively. Due to a large number
of learnable parameters within 3DFCN and continued performance improvement
of the model on the testing dataset, it is believed that further training will result in
improved testing results. However, the current model was able to learn and predict
Machine Learning-Based Structural Damage Identification … 451
the correct labels of the neutral, terrain, debris, and structure classes which were
subsequently used within the k-means clustering and 3DCNN workflows (Fig. 7).
The 3DCNN model was trained on the created training dataset for only a total of 15
epochs prior to testing, unlike 3DFCN to avoid overfitting informed based on evalua-
tion results of the validation set. Multiple networks were initially trained through the
factorial design process to identify a set of optimized parameters similar to 3DFCN.
The L1 and L2 regularization lambdas, learning rate values, number of kernels in
convolutional layers, and number of nodes in the fully connected layers were varied.
Through the factorial design process, the network with a total of 32 kernels per convo-
lutional layer, L1 regularization lambda of 0, L2 regularization lambda of 0.005, the
452 M. E. Mohammadi and R. L. Wood
learning rate of 0.001, and 256 nodes for fully connected layer proven to be the most
successful model. The best model was identified through assessing the performance
measures, included precision and recall, as well as training-based mean square error
values. Table 3 presents the performance measures for training and testing results
where the model was able to detect instances with average accuracy and recall of 95,
94, and 90%, respectively. In addition, Fig. 8 demonstrates the confusion matrix for
the training and testing processes. As shown, the model was able to learn all instances
with precision, recall, and IOU score values of 68, 67, and 67%, respectively, which
demonstrates significant improvement over previous studies presented by Liao et al.
[25], where a single 3DFCN model was only able predicted damaged areas with
precision and recall value of 34 and 32%, respectively.
The performance of the proposed workflow is further demonstrated by analyzing
a 70 by 70-m point cloud segment. Figure 8 depicts the final semantic segmentation
results for the 3DFCN model. As illustrated, the 3DFCN model learned and classified
the points pertained to the terrain and roadways classification with an acceptable level
of accuracy. In addition, while the points corresponding to the structure class were
detected within the test segment, it was noted that not all the points representing
Machine Learning-Based Structural Damage Identification … 453
(a) RGB colored point cloud (b) Ground truth labels (c) 3DFCN segmentation results
the structures were classified correctly and instead were classified as terrain, which
matches the test results of the confusion matrix presented in Fig. 6. Misclassification
of the points representing the structure class with the terrain class was observed in
both training and testing data. As the training process demonstrates that the model
can learn to distinguish the points within these two classifications, it is expected
that further training improves the classification points that represent these classes.
Once the 3DFCN model segmented the point cloud into seven classes, the workflow
then transforms the voxel representation into the raw point cloud representation
by applying the inverse process described in Sect. 4.1. Afterward, the workflow
runs a query to collect the points with structure classification, as shown in Fig. 9a.
Then, the workflow utilizes the SOR filter to eliminate the sparse points and uses
a k-means clustering algorithm to segment the structures into individual segments.
As shown in Fig. 9b, the algorithm eliminated most of the sparse points using a
neighboring number of 31 and standard deviation factors of 0.25. The clustering
algorithm uses six clusters (double the number of observed structures in the selected
ROI) and a threshold of 0.20 m to segment the structures. The final result of k-
means clustering segmentation results is shown in Fig. 9b. Once the segments are
identified, the workflow converts these segments into the voxel representation, which
the trained 3DCNN model then analyzes. Figure 10 demonstrates the results of the
3DCNN model, where 50% of segments were classified correctly. Figure 11 depicts
the results of the classification.
(a) Points that represent structure class based on (b) Segmented point cloud where each color repre-
3DFCN prediction sents one cluster
(a) RGB colored point cloud (b) Ground truth labels (c) 3DCNN segmentation results
6 Conclusions
This manuscript presented a new workflow that combines three distinct machine
learning algorithms, including 3DFCN, k-means clustering, and 3DCNN. The devel-
oped workflow was developed based on UAS-SfM-derived point clouds to classify
post-tornado scenes for damage assessment semantically. The UAS-SfM point cloud
of built-up areas sustained damage following the March 2020 tornadic windstorm
event in northern Tennessee was used to train and test the workflow. Each point
within this dataset was segmented manually to have a label to match the door-to-
door assessments performed by a ground team of investigators. Afterward, the point
cloud data were divided into two parts for testing and training segments. Initially, the
labeled point cloud was segmented into 10 m × 10 m (32 ft × 32 ft) parts to create
training and testing instances. Then, the 643 voxel representation of the segments
was created, which resulted in a sampling rate of 16 cm (6.2 in). The 3DFCN was
trained based on these voxel models, and the 3DCNN was trained based on structure
voxels.
The developed models were evaluated based on precision, recall, and IOU
that are computed from the confusion matrix. As illustrated by the performance
measures presented for training results, the developed models are optimized to learn
the features; however, the convergence was shown to be slower in the 3DFCN
model, primarily due to the number of learnable parameters. The 3DFCN model
learned and predicted the correct labels of the neutral, terrain, and structure classes
that are essential to the workflow’s success. However, it demonstrated a lower
Machine Learning-Based Structural Damage Identification … 455
precision and recall for objects with much smaller frequency within the dataset
(e.g., vehicle). The 3DCNN model learned all structure classes and demonstrated
significant improvement over the 3DFCN model in detecting damaged structures
[26].
Acknowledgements This work was completed utilizing the Holland Computing Center of the
University of Nebraska, which receives support from the Nebraska Research Initiative. Recon-
naissance data were collected with the administrative and financial support of NSF-StEER, which
included travel funds for the second author. The authors would like to thank Dr. Yijun Liao for her
help and assistance in the field as well as Mr. Awang Ahmad Hashim, Mr. Mario Esquivel, and Ms.
Pooja Rajeev for their assistance in the data preparation and processing.
References
1. He X, Cha EJ (2018) Modeling the damage and recovery of interdependent critical infrastructure
systems from natural hazards. Reliab Eng Syst Safe 177:162–175
2. Wood RL, Roueche D, Cullum K, Davis B, Gutierrez Soto M, Javadinasab Hormozabad S,
Liao Y, Lombardo F, Moravej M, Pilkington S, Prevatt S, Kijewski-Correa D, Djima S, W.
Robertson I (2020) Early access reconnaissance report (EARR). In: StEER - 3 March 2020
Nashville Tornadoes. DesignSafe-CI. https://doi.org/10.17603/ds2-2zs2-r990
3. Vetrivel A, Gerke M, Kerle N, Vosselman G (2015) Identification of damage in buildings
based on gaps in 3D point clouds from very high resolution oblique airborne images. ISPRS J
Photogram 105:61–78
4. Duarte D, Nex F, Kerle N, Vosselman G (2018) Satellite image classification of building
damages using airborne and satellite image samples in a deep learning approach. ISPRS Ann
Photogramm Remote Sens Spatial Inf Sci 2:89–96. https://doi.org/10.5194/isprs-annals-IV-2-
89-2018
5. Spencer Jr BF, Hoskere V, Narazaki Y (2019) Advances in computer vision-based civil infras-
tructure inspection and monitoring. Engineering. 5(2):199–222. https://doi.org/10.1016/j.eng.
2018.11.030
6. Gao Y, Mosalam KM (2018) Deep transfer learning for image-based structural damage
recognition. Comput Aided Civ Infrastruct Eng 33:748–768
7. Simonyan K, Zisserman A (2014) Very deep convolutional networks for large-scale image
recognition. arXiv preprint at https://arxiv.org/pdf/1409.1556.pdf
8. Tilon S, Nex F, Kerle N, Vosselman G (2020) Post-disaster building damage detection from
Earth observation imagery using unsupervised and transferable anomaly detecting generative
adversarial networks. Remote Sens 12(24):4193. https://doi.org/10.3390/rs12244193
9. Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image
segmentation. In: International conference on medical image computing and computer-assisted
intervention, Boston, MA, pp 234–241
10. Olsen MJ (2015) In situ change analysis and monitoring through terrestrial laser scanning. J
Comput Civ Eng 29:04014040. https://doi.org/10.1061/(ASCE)CP.1943-5487.0000328
11. Tran THG, Ressl C, Pfeifer N (2018) Integrated change detection and classification in urban
areas based on airborne laser scanning point clouds. Sensors 18(2):448. https://doi.org/10.3390/
s18020448
12. Aixia D, Zongjin M, Shusong H, Xiaoqing W (2016) Building damage extraction from post-
earthquake airborne LiDAR data. Acta Geologica Sinica 90(4):1481–1489. https://doi.org/10.
1111/1755-6724.12781
456 M. E. Mohammadi and R. L. Wood
Abstract The progress in the world of new sensors is running fast, offering good
performances, and reliable solutions with initial costs orders of magnitude lower
than those faced only a few years ago. The spread of new electronic devices, like
microcontrollers, the increasing power of the networks for data transmission and
management and in the end the availability of the new data-driven approaches have
created a revolution in SHM approaches, not yet fully mastered. The way to design
a SHM system is going to be deeply revised in an industrial perspective, within
a complex framework in which everything has to be planned into details since the
beginning, including the development of a metrological culture, the personnel educa-
tion, the need of spare parts, re-calibration, …. This also means a revolution in data
management: huge data flows not only create hardware problems related to their
transfer; the software too requires a great deal of effort to compress data, also due
to the actual cost of cloud resources. All these facts, accounting for the real metro-
logical performances of the best MEMS sensors available at present, also require
simplified data analyses, as software complexity is now mainly transferred to the
network management. A trade-off must be looked for between the big redundancy
offered by the actual networks and the need of a simple and prompt information,
granting the structure safety: That is why as the data rates increase, the algorithms
to be adopted must be simple, reliable, eventually adapted to edge computing at the
sensor level, where hardware power is now present though at a reduced scale. The
chapter shows such an approach in a real case from the system design, its birth, and
its proper use for damage detection, up to the detection of a structural failure.
© The Author(s), under exclusive license to Springer Nature Switzerland AG 2022 457
A. Cury et al. (eds.), Structural Health Monitoring Based on Data Science Techniques,
Structural Integrity 21, https://doi.org/10.1007/978-3-030-81716-9_22
458 I. Alovisi et al.
1 Introduction
The world of structural health monitoring (SHM) is going through a sort of destruc-
turing, due to the completely new approaches to structural sensing. The new Internet
of Things tools have created a revolution in the way measurements are being carried
out, not just for machines, but also for structures, pushing hard toward the idea of
dense sensing, often translated into that of smart structures.
As often discussed, the general idea of SHM is to mimic the human behavior [1]:
This does not just mean having distributed sensors, but also an efficient network and
an improved data management capability, to provide fast and reliable evaluations. A
measurement system, pretending to act as a human, must be designed with a clear
idea about its final aim, so it has to be deeply understood which functions can be
transferred to the computer through models or data science approaches and which are
still under the specialist’s supervision. In addition, until recent days, the huge effort
on SHM has been mainly carried out on the research side, yielding to a relatively
poor number of routine industrial approaches [2].
The mentioned destructuring of SHM approaches is facing new problems, moving
the most critical implementation phases toward new aspects, different from the past.
If new sensors are considered, they have different performances with respect
to those commonly adopted today for monitoring applications, mainly consisting
in laboratory instrumentation for short-term dynamic testing. In general, the new
cheaper devices have slightly lower output quality, though this gap is narrowing;
a huge difference remains if costs are addressed. The new sensors are often less
expensive than those commonly adopted for SHM: this also means that a trade-
off must be looked for between the available budget and the need for quality in
measurements. Given a fixed budget, the higher number of allowed sensors and
the availability of some redundancy, can help recovering a more than acceptable
information content; at the same time, the new solution allows for a better hope to
be closer to an eventual damage onset, for prompt operation.
Another revolution is offered by the possibility to couple each sensor to a micro-
controller: this unit is a sort of tiny and cheap computer, programmed to work on
simple software-defined operations. Thus, they allow to manage essential functions
like data acquisition or communication to a supervising unit, but they are also capable
of performing simple data analysis like the fast Fourier transform (FFT) or a running
root mean square (RMS) at the sensor level, making the system cleverness distributed
and, therefore, getting closer to the idea of a smart system: however, this implies the
risk of an information jam, which has to be properly managed.
The mentioned destructuring also brings in new problems: in such monitoring
systems with distributed intelligence, each sub-system lives a life on its own within
its time framework, given by its internal clock, unless complex synchronization
tools are applied: This does not just mean abandoning those approaches to damage
detection requiring a single clock shared among all the sensors in the same network
(for example, a loss of synchronization makes traditional modal analysis a harder
task); rather this means changing the strategy to achieve the final result.
New Sensor Nodes, Cloud, and Data Analytics … 459
Even if sensors are now less expensive, it is not possible to assess the same for the
complete network. In case of a wired solution, a more complex layout comes from
the higher number of connections among sensors and supervising units, and this also
implies the need to both provide power supply to all nodes and bring the data back
to a gateway or more generally to a network.
It has to be considered that, in view of a complete description of what’s going on to
get a structure health check, data fusion approaches are often adopted, merging quite
different data kinds, with diverse aims, generated by a large variety of sensors, each
having its own output impedance, sampling rate, output range, etc. Wireless, often
offered as the solution to every kind of problems, is not yet reliable enough when
big data streams have to be managed, or when safety comes into play. The reverse
applies to static data, where wireless can be an interesting option, as the reduced
data flow allows to send a data packet again in case it gets lost over the network. At
this point, the main bottlenecks appear to be those related to the data flux over the
sensor network: this stream is getting bigger and bigger, harder to manage, especially
if in the end a single synthetic piece of information has to be given to the structure
manager, or if an alarm has to be promptly generated.
The new IoT tools for sensor networks are sometimes not really fit for SHM
applications: a powerful protocol like LoRaWAN [3], widely adopted for smart city
projects, due to its power to transmit over long distances, gives each sensor a very
short time to transmit data. This solution is, therefore, not suitable for continuous
data streaming, as in the case of accelerometers, having a higher sampling rate: new
solutions like the narrowband IoT [4, 5] are helping to mitigate these problems. A
high innovation rate is also seen for cabled solution, remaining the majority among
SHM systems: interesting solutions point at the use of industrial protocols for data
transmission, as these have proven to be robust and allow to transfer medium to high
data rates even over complex and busy networks [6].
The preliminary design of the monitoring system becomes the main challenge
[7]; once the aims are properly defined, density of measurements in both time and
space are key issues. The designer has to define beforehand how many sensors are
to be used, their location on the structure, and for how long measurements have to
be taken; this resolves the risks of losing meaningful events in a trade-off against the
need to keep the network not too busy and to store a reasonable though meaningful
amount of data.
In the end, some issues are seldom addressed, and again they belong to the already
mentioned industrial management of SHM. Maintenance of monitoring systems has
costs which are a substantial part of the initial investment; sensors need periodical
checks on their calibration, and this can be a difficult task, especially if they are not
easily reached; a stock of spare parts has to be kept, as electronics systems undergo
a very fast aging, not comparable to the life of the structure. In the end, the technical
staff, taking care of the measurement systems, is asked to follow a proper training,
which has to be strongly multi-disciplinary, also including a plan to take decisions
over the different scenarios which can be depicted by the monitoring system.
460 I. Alovisi et al.
All these aspects make clear that a monitoring system is not just installing sensors
on a structure and getting data, this is only a part in a more complex framework, a
process which requires proper design and awareness [8–10].
If on one hand the recent advances in electronics make new monitoring system
hardware more accessible to everyone, on the other hand it must not be forgotten how
many factors still constitute a problem for these approaches: one above the others
is the required multi-disciplinary approach, needing experts in many different areas,
from civil engineering to data sciences, to electronics, to metrology, to telecom,
mathematics, and so on: all these skills are hardly found in a single professional and
require a trained staff.
Under these principles, the present chapter will go through a list of specifications
to develop a complete industrial project; a path among the others is laid, relying on
the up to date innovation in electronics, networks, communication, data management,
and above all the possibility to give back a prompt and synthetic evaluation about
the structure health.
2 System Design
The approach developed by the authors of this paper aims at a pervasive sensing
of structures. Providing a dense sensing in both time and space means managing
impressive data flows, but it also means designing and realizing a new sensing
infrastructure.
The new approach is shortly summarized in Fig. 1: a traditional SHM system
consists in a number of sensors placed over the structure: their cost usually allows
Fig. 1 Traditional (left) and the new (right) structure of SHM systems
New Sensor Nodes, Cloud, and Data Analytics … 461
Fig. 2 Sensor node (left) and two examples of sensor nodes for measurements in tunnels (centre) and
on bridges (right)
2.1 Hardware
The hardware design turns around the kind of sensors to be adopted for monitoring.
If dense sensing is the main requirement, a deep study about the target performances
and those made available by the chosen sensors is a needed prerequisite.
Micro-Electro-Mechanical-System (MEMS) sensors are considered the most
popular solution when the chosen solution aims at reconstructing a structure kine-
matics due to some external excitation. MEMS sensors are cheap; therefore, we must
assess the proper trade-off between the available budget for measurements and the
desired information. As these systems belong to the world of traditional electronics
and semiconductor industry, they are easily scalable, to include different sensor kinds
and possible network integration.
Fiber optic sensing, the other widely adopted solution today for SHM, has not
been chosen for the monitoring systems described in this chapter, having different
requirements and analysis procedures, and in addition a specific network design and
a completely different economic impact [11, 12].
The hardware analysis presented below is focused on both the sensor level and
the network level.
New Sensor Nodes, Cloud, and Data Analytics … 463
A positive aspect related to the use of cheap sensors is the possibility to implement
new strategies oriented at reducing the effects of noise, such as using redundant or
repeated measurements. Some preliminary attempts in the design phase of the new
units have tried to use four inertial sensors on the same boards, both clinometers and
accelerometers, with different aims. If one considers the same acceleration sensed by
four sensors, then uncorrelated noise gets halved by a classical uncertainty analysis
in case of “in band uncorrelated noise” [15]. In case of clinometers, the difference
of the outputs from a twin configuration in which the two sensors are put close
to each other with opposite sensing axis provides an amplification of the rotation
output, at the same time deleting the effects of temperature; this only works if the
sensitivity to this parameter on each MEMS component is the same. As pointed out
in the introduction, this new strategy is allowed by the low overall cost increment,
due to the increased number of bare sensors, but it has to be framed into an industrial
approach, in which reliability of a new and complex system comes into play as the
main need. A preliminary testing campaign has proven that the sensitivity of most
MEMS sensors to temperature varies between each device, also for integrated circuits
obtained from the same slice; that is why the clinometer temperature compensation
is not always guaranteed or at least it is not always really achieved. At the same
time, halving the accelerometer noise floor, by combining the output of four sensors,
measuring the same quantity, has not been considered enough most structures provide
a good enough S/N ratio, overcoming this problem; all the same some redundancy on
the board design has been preserved. As in an industrial perspective, many devices
have to be deployed at the same time, and an additional issue has been the need for
proper calibration of a meaningful number of sensors together. The possibility to
use static calibration also for accelerometers, ruled by international standards [16,
17], has been the adopted solution, provided a preliminary check against dynamic
New Sensor Nodes, Cloud, and Data Analytics … 465
a b
c d
Fig. 4 Calibration of MEMS accelerometers (a), of MEMS clinometers (b, c), sensitivity to
temperature in a temperature controlled room (d)
tests have been carried out too (see Fig. 4). Of course, a different strategy can also
be implemented, with respect to traditional approaches, as continuous cross checks
among redundant close sensors is a new possible approach to be developed in order
to easily detect any device bad functioning or drift from good operation.
The clinometer calibration proved to be an easier task, as standard back-to-back
approaches against a calibrated reference are an affordable and fast procedure [18].
In this case, careful attention has been paid to define a meaningful resolution limit,
as the number of layers between the structure surface and the sensor has required a
long and burdensome evaluation of any parasitic effect due to temperature. This has
allowed to detect and account for any possible interference on measurements given
by the behavior of glues, resins, plastic, and mechanical fixtures, used to transfer the
rotation information through the mentioned intermediate layers.
Network: Node Synchronization
Each sensor node is also expected to manage the connections with the adjacent nodes
and with all the rest of the network. A first issue has already been mentioned about
the synchronization of nodes, allowing one to define different damage detection
scenarios, if working on a single node or on the comparison among nodes.
Tools are available for at least a rough node synchronization, ranging from GPS
to NTP protocols. These can be considered effective especially for some civil engi-
neering applications, since the dynamics of interest for structures is usually confined
in the low frequency range: even a not perfect synchronization does not alter the vibra-
tion mode detection in a meaningful way, as the percentage phase error is limited. By
466 I. Alovisi et al.
now, it has been stated not to implement such protocols, to avoid added complexity
to the existing firmware governing the microcontroller: the adopted solutions have
been considered in relation to the record lengths to be analyzed, with the possibility
to periodically restart the recording and so realign the clocks over the network.
Network: Data Transfer
Another issue is the meaningful amount of data to be transferred, due to the presence
of many sensors in the network: many solutions have been explored to make this
traffic lighter. A first option is to implement the data elaboration at the sensor level
(edge computing); another solution can be to trigger an immediate data transfer, in
case of significant events, or possibly to store data locally and send them later to the
cloud when the network is not so busy.
In our real case studies, we also tried new approaches. When the data streaming
came mainly from static transducers like clinometers or temperature sensors,
requiring a limited transfer rate, we opted for the use of power line solutions, so
that the same cable bringing the power supply to the peripheral nodes could be also
used to take measurements back (see Fig. 4).
In case of higher data flow rates, the typical case of acceleration measurements,
the need to have robust, and reliable systems has led again to look for industrial
solutions. Among the others, the use of field bus solutions, like the CAN bus, typical
of the automotive industry, has been considered a winning asset, as protocols are
already available, at the same time preventing from the risk of losing data (Fig. 5).
As already pointed out, for such measurements, wireless solutions, though tempted
several times in smaller networks, have not yet been considered reliable enough,
especially when safety issues come into play.
Network: Gateway Tasks
The continuous improvement of electronics is rapidly making available to the market
new devices to collect data in a smart way: Gateways are the heart of the new
monitoring systems, being capable at the same time to collect the big data streaming
rates from the sensor nodes, to order and organize them, in the end to send them to a
cloud server for any further processing. Gateways need to manage both cabled and
wireless connections: in most monitoring systems, we are managing right now, i.e.,
bridges along highways crossing mountain areas, a stable and reliable connection to
send data from the gateway to the cloud can be a challenging hazard. Problems here
are mainly transferred to telecom networks, also requiring a detailed design in terms
of the number of installed gateways to ensure that all data are properly collected,
packed, and sent.
Another very important function, which has been intensively studied with the
progress offered by new gateways, has been the development of “fog computing”
approaches. The need to face SHM as an industrial process has forced to look for
economically sustainable solutions. The cost of cloud resources is still high, espe-
cially if a wide number of sensors, having high data rates, is considered: that is why,
the possibility to store data, at least temporarily, at the gateway level, seems to be
more than attractive, also to guarantee some redundancy at least on the short term.
The gateway can further perform some pre-analysis reducing the amount of data to
be stored, also offering the chance to have a prompt response in case of anomalies:
This is one of the most important trends for the close future.
Data processing at the gateways is an ongoing major technological trend, and
cloud providers are offering solutions to load data processing SW components into
such edge devices. This customer and application driven business logic often comes
in the form of software “containers” that run in an isolated and secured environment.
The gateway software architecture is such that important data are cached locally so
that no information is lost even when Internet connectivity is temporarily unavailable
(Fig. 5).
2.2 Software
tasks must kept under control even after system installation, and it is critical to be
able to retrain a neural network or a machine learning task in field.
Another important aspect of these edge processing tasks is that they must be
documented, replicable, and auditable because they act as the first data filtering
stage, and any misbehavior may result in important or even critical information to
be lost. This is why caution must be taken, and, as soon as an anomaly is suspected,
raw data should always be sent upstream in the architecture up to where they can be
analyzed more carefully and deeply.
As the sampling frequency may differ for each structure, the collected data quan-
tity is variable. In any case, to guarantee a minimum of system redundancy, a larger
amount of data are usually acquired, leading to problems in transmission, storage, and
post-processing analysis, increasing costs and decreasing processing speed. Again,
the challenge is finding the right balance between system capacity and information
needed for structural characterization and health assessment. One possible solution
could be to limit sensors acquisition only to time intervals in which the structure is
really loaded, avoiding periods with no or too low external excitation (accounting
for the sensors noise floor). In this way, selected datasets will always have a suffi-
cient energy content, allowing for the evaluation of both long-term and short-term
processes, the correct estimation of modal parameters and ensuring data acquisition
during extreme events.
To implement this kind of smart acquisition, several issues have to be considered
and fixed. Firstly, features able to drive data selection have to be identified, and these
features should, at the same time, consist in simple functions, so that they can be
performed by the gateway or even by the just mentioned sensor microcontroller. Then,
thresholds have to be calculated for the chosen parameters: the idea is that when a
certain number of sensors simultaneously exceed these values, data acquisition starts.
Groups of sensors designed to be responsible of starting the acquisition are another
key point to be carefully considered. They are essential to give robustness and relia-
bility to the system, avoiding the storage of redundant information, and guaranteeing
data collection for all the significant scenarios.
For this reason, groups are various; their size is different and function of the
expected phenomenon: in case of global events, such as an earthquake, for example,
all the structure will be involved, and almost all sensors will detect it; conversely,
when a local damage occurs, only sensors installed on the interested structural
element or in its proximity will show anomalous values, so allowing, by means
of procedures of modal shape identification, to localize with an acceptable approxi-
mation the region in which the local damage intervened. Furthermore, a robust cloud
architecture should support an effective and reliable alert system (see Sect. 2.3).
Once different alerts statuses have been defined, with related control parameters and
threshold values, false alarms must be prevented, limiting the alerts in case of extreme
events, when groups of sensors exceed the limit values.
470 I. Alovisi et al.
2.3 Cloud
Architecture
The design of a cloud architecture that can scale and support large amounts of
heterogeneous SHM systems has been based on the following principles: (1) data of
different customers need to be kept isolated at all times, (2) real-time alerts should
be handled in the minimum amount of time and clearly notified to interested staff,
(3) cloud should support all the supply chain involved in the design, provisioning,
production, testing, calibration, and installation of full SHM systems, (4) computing
resources necessary to calculate slowly evolving structure trends on massive amounts
of data are allocated dynamically and take advantage of the most advanced algo-
rithm parallelization techniques, to minimize processing time and associated costs,
(5) dependencies to cloud provider specific services has to be minimized.
For each SHM installation, gateway and device provisioning are part of the SHM
design process; their configuration parameters are stored in a database to drive the
production and testing process in the fab. After installation, sensor data are ingested
by using an IoT platform and saved into S3 buckets as parquet files with Hive parti-
tioning to ease later retrieval [19]. Real-time anomalies are detected during data
ingestion, while slowly evolving phenomena and alarm thresholds are checked and
adapted by running parallel Spark jobs in a Kubernetes cluster. This cluster runs pods
for remote gateway control, a LoRaWAN network server, and other housekeeping
tasks that are active 24/7. Once a day, the dimension of the cluster is increased to
launch parallel jobs that query S3 buckets for each structure, run structure dependent
algorithms, and store analysis results that can be visualized for the engineering staff.
Since an important part of sensor data processing needs to take place in the
gateway, a solution derived from OpenStack has been developed; it enables its full
remote control through a combination of web-sockets and MQTT, including injecting
signed Python code, remapping/exporting service ports, and storing one week of full
sensor data locally. If an anomaly is suspected after analysis in the cloud and more
sensor data are needed, it can be obtained by querying the gateway within the one-
week time buffer. At the present stage, no data cryptography has been implemented
yet, though it is considered a fundamental next step.
Data Analytics (Post-processing)
As civil infrastructures are naturally subject to a gradual aging process, induced by
the progressive physical deterioration or accidental actions, there is a specific need
to make the damage evolution controllable and predictable over time through an
effective diagnostic system that allows an adequate asset management and sched-
uled maintenance. Thus, the data collection, selection, and pre-processing must be
followed by a thorough, more detailed, and comprehensive analysis process, in which
the engineering judgment plays a key role in assessing the structural state of health
and conservation. Model updating procedures based on sensors acquisitions, detailed
modal analysis, and reliability evaluation are needed to explore critical scenarios,
New Sensor Nodes, Cloud, and Data Analytics … 471
identify threshold limits, and develop algorithms to perform long-term data interpre-
tation to support subsequent decision-making processes. The SHM system should
be generally designed to display data that have been processed and synthesized to a
simple, meaningful format, to guide the structural engineer and/or the asset owner
to clearly understand and handle potential risks (ref. Sect. 2.4).
attached, so that they can be handled with the appropriate urgency. As stated before,
an alert should always come with raw data attached, so that an in-depth analysis can
be done to confirm or deny the event. Similarly, the frequency of alerts and their
correlation with other critical factors (weather conditions, time, season) should be
kept under control and subject to periodical audit.
The complexity of such monitoring systems coupled to the need to provide a fast
and correct assessment has pushed toward a “back to basics” evaluation of simple
and robust approaches. Given the performances of the new sensors and networks,
cheaper and denser, a careful review of the commonly adopted feature detection
strategies is still running to explore their real effectiveness and capability to provide
satisfactory health evaluations, accounting for the new process uncertainties.
The need of standardization required by the industrial approach is often winning
over the requirements of optimization, also due to the need to guarantee a high
redundancy level in case of sensor failures.
The SHM system developed by Sacertis, including new sensor nodes, cloud, data
analytics, and diagnostics, as outlined in the previous paragraphs, has been designed,
physically realized, and it is currently active on more than 40 bridges and tunnels
within the Italian road network. These infrastructures are equipped to provide contin-
uous assessment of structural integrity, essential to detect the occurrence of structural
changes, or the evolution of damage that could affect the performance and safety of a
structure. In the present chapter, a case study on large-scale SHM system is presented,
to highlight the effectiveness of an innovative, affordable, and minimally invasive
monitoring system in the perspective of damage detection, structural diagnostics,
and proactive asset maintenance.
Data analysis has been carried out following two quite different methods: a data
processing technique related to the industrial approach to SHM, based on simple
features and a complex cross check on the data reliability, which must provide simple
and prompt information, helping the bridge manager with safety issues (Sect. 3.2);
a second, more scientific approach, implements unsupervised learning techniques
to track any eventual damage (Sect. 3.3). Having recognized a real damage during
monitoring, a supervised approach could be implemented too; but, as a single failure
was measured, this event was considered without real statistical relevance, therefore
not suitable for classical classification tools. The two methods have been considered
in a benchmark approach, one as the reference for the other. The first proved to be
more robust and reliable to support an automatic alert and notification system; the
second provided promising results, however, not yet at a technology readyness level
(TRL) allowing it to be delivered as a final product. As mentioned earlier, the sensor
output quality and the will to work with edge computing, to compress the data flow,
has forced to favor the implementation of simplified approaches.
New Sensor Nodes, Cloud, and Data Analytics … 473
Fig. 6 Structural drawings of the bridge: (a) bridge cross-section; (b) bridge images after and
during construction; (c) plan view of the highway bridge; (d) longitudinal section of a span, with
sensor positioning
the bridge could survive even without some tendons, but these had to be substituted.
The application of MEMS accelerometers on the ducts, so as close as possible to any
possible damage, gave the structure owner a prompt and detailed warning about the
real damage position, providing a measure of the fundamental properties necessary
to predict the long-term performance of the bridge and foresee any potential further
damages induced by the corrosion of the strands. Of course the price was a huge data
flow, an important test bed for automatic early warning to detect damage.
The monitoring system includes (Fig. 7):
• the sensor nodes, comprehensive of tri-axial MEMS accelerometers (full scale ±
2.5 g, bandwidth of 50, 100 Hz sampling rate), temperature and humidity sensors,
and a microcontroller used for data sampling. This choice complies with the need
to save space, at the same time keeping some basic information: the band 0–50 Hz
already includes some cable natural frequencies and, thus, fits for the application
of several damage detection approaches.
• The CAN bus network to transfer the data from the end nodes to the local gateways
(as the bridge is long and both the network and the data streaming could not be
managed by a single unit, in this case, we had two at the bridge ends).
• The IoT gateway, where the data are collected, pre-preprocessed, and filtered to
send a selection of significant information to the cloud for further analyses.
• The cloud, where data are stored and more complex data analytics can be
performed.
The SHM system is active since September 20, 2017, and therefore, several years
of data have been stored and analyzed, and the developed damage detection strategy
New Sensor Nodes, Cloud, and Data Analytics … 475
This monitoring strategy reflects the network structure and responds to need of
providing both a quick and reliable near-real-time analysis to detect fragile and
sudden anomalies (local abrupt structural damages, earthquakes vibrations,…) as
well as the identification of long-term damage developed in a longer time-frame (as
corrosion effects).
The choice has been directed toward a system based on two alarm levels:
• Level 1: alarm characterized by a check over the RMS (over 60 s), immediately
estimated and verified at the IoT gateway level, with the aim to highlight the
presence of anomalous vibration levels in the tendon dynamic responses under
external loads.
• Level 2: alarm characterized by a check on a physical parameter, the natural
frequencies of the cables, performed on the cloud platform and activated only
if the first one is exceeded. This approach, among the many possible, has been
considered for its ease of implementation, also due to the general high innovation
and complexity of the whole system.
The Level 1 type of alert highlights the occurrence of anomalous vibrations of
the tendons through the evaluation of the RMS level of each sensor over a sampling
window of 60 s. The release of high energy levels, corresponding to impulses detected
by the accelerometers, may be induced by random actions or to the brittle failure of
476 I. Alovisi et al.
single strands. The RMS is compared to a threshold limit, in real-time, at the gateway
level. The thresholds have been estimated for each sensor over a training dataset repre-
sentative of the standard behavior of the tendon under external loads (traffic, wind,
thermal loads, …). With the increase of the monitoring period, the thresholds have
been recalculated over a longer training set until reaching a convergence (stabiliza-
tion of statistical indicators), obtaining a robust alert level. The thresholds levels for
each sensor and for each of the three measurement directions x, y, and z are shown
in Fig. 8.
In case Level 1 thresholds are exceeded in a group of nearby sensors, a more
complex data processing in the frequency domain is performed at the cloud level,
based on a physical interpretation of the vibration data. The power spectral density is
evaluated by averaging over a window duration of 200 s, (66% of overlap, Hanning
window). If a strand collapses, the cross-section of the tendon is suddenly reduced,
causing a series of complex phenomena in which the overall load is differently
distributed inside each single tendon and among nearby tendons, anyway producing
a natural frequencies drop; to identify a shift in the natural frequency of the tendons,
the frequency domain decomposition (FDD) has been adopted. In case a change
in the natural frequencies is detected (in this case detected over short times, so
hardly depending upon environmental parameters), the second-level alert alarm is
produced. In this section, two main events are described, occurred over the 4-year
monitoring period, respectively, during November and December 2017, showing the
effectiveness of the two-level alert approach. As shown in Fig. 9, on 11/19/2017, the
RMS threshold was exceeded by all the accelerometers of the bridge, pointing at an
anomaly occurring to the entire structure simultaneously. As per the alert procedure,
the second-level analysis was triggered, confirming that the natural frequency of the
tendons did not experience any variation. The sudden increase of the vibration levels
was caused by an earthquake. The SHM system correctly pointed out the anomalous
condition, proving its reliability and robustness, and confirmed the structural integrity
with the second-level check.
In December 2017, instead, the first level threshold was exceeded by a limited
number of sensors, all of them installed in the first span of the bridge. Again, the
second-level analysis was automatically activated for all the sensors that exceeded
the first threshold. As shown in Fig. 10, one of the tendons installed in the first span
New Sensor Nodes, Cloud, and Data Analytics … 477
Fig. 9 (a) Level 1 and (b) Level 2 check on the earthquake excitation on the 19/11/2017
of the monitored bridge also exceeded the second threshold level, experiencing a
frequency reduction caused by the failure of one of the strands, and a partial loss of
prestress force in the structural element. This reduction in the stress distribution is a
function of the severity of the tendon damage (the number of broken strands) and of
the damage location. The energy released when the strand broke was recorded as a
dynamic input also by the nearby sensors.
During the 4 years monitoring period, Sacertis system automatically detected the
failure of strands on other 3 tendons, providing a clear overview of the pre-stressing
status of the structure, and significantly improving the maintenance schedule of the
bridge.
478 I. Alovisi et al.
Fig. 10 (a) Level 1 and (b) Level 2 check on the 12/21/2017 event
The choice adopted for a data-driven approach, to detect damage through synthetic
features, has relied on an unsupervised strategy consisting in a proper mix of autore-
gressive models AR [27] and principal component analysis PCA [28]. The underlying
idea was to separate the effect of environmental quantities out of the dynamic features
pointing at damage.
Given X t , the value of the time record X at time t and X t− j , at time t − j, it is
possible to define the autoregressive model of order p, namely A R( p), thanks to the
well-known expression:
New Sensor Nodes, Cloud, and Data Analytics … 479
p
Xt = φ j X t− j + at (1)
j=1
Specific interest is devoted to the weights φ j for the p preceding steps used to
get the actual value in the time series and to the residual at , expected to belong to a
random Gaussian distribution with zero mean μa = 0 and variance σa2 .
A first step in the effort to compress information has consisted in the optimiza-
tion for the p value: the adopted tool has been the Bayesian Information Criterion
(BIC) [29], consisting in a likelihood method including a penalty as the number of
autoregressive coefficients increases, avoiding overfitting.
The founding steps of both AR and PCA are well known in damage detection: the
novelty here is their combined use. We chose to focus on implementation aspects and
on the main results. In the following, attention is devoted to a single sensor (1S33)
over a period between October 2017 and April 2018; this is considered representative
of the whole sensor set: the cloud resources had to be shared between ordinary
analysis, routinely performed on all sensors to provide real-time alarms, and the
present application.
Each record stored on the cloud consists of 200 s sampled at 100 Hz: the bridge is
always very busy, subject to a continuous random and high-level input excitation; the
AR approach has been applied to records of the same length as the file length. As the
accelerometers are g-sensitive, the average value has been high pass filtered. A data
normalization is considered a good practice, as the AR coefficients are representative
of the poles and not of the zeroes: the input data X̃ to AR models is:
X orig − μ X orig
X̃ = (2)
σ X orig
being μ X orig e σ X orig the mean and standard deviation of the original data X orig .
For the considered data, we have precious information about the ongoing damage,
occurring at subsequent steps; although the monitoring system has been installed on
a bridge already in service, we have considered the first dataset, up to the first known
damage event, as a healthy baseline.
The Bayesian Information Criterion (BIC) has been applied to a relevant quantity
of data: an example is given in Fig. 11, showing the estimated maximum order of
the AR model evaluated over 200 records belonging to the training set, in which the
bridge is assumed under reference undamaged conditions.
Although some scattering, the mean has been considered a reasonable compro-
mise.
To prove the quality of the adopted approach, Fig. 12 provides an original record
and its reconstruction given by the AR (34) model. A further confirmation about the
goodness of fit is proved by the Gaussian distribution of residuals. The next step has
then consisted in modeling every record of the training set by means of an AR (34)
model. Figure 13 provides the 34 φ coefficients evaluated over the period between
10/1/2017 and 4/7/2018 for the selected accelerometer, vertical axis.
480 I. Alovisi et al.
The 34 coefficients are not yet a synthetic damage feature; that is why a further step
consisted in compressing the available information in a lower number of coefficients
m p. Once evaluated the PCA over the whole AR coefficient set, the comparison of
each eigenvalue of the covariance matrix against their sum (the variance of the whole
set) provides the histogram in Fig. 14: the first three principal component values
provide around 97% of the total data variability. Instead of the 34 AR coefficients;
these three components are a much more compressed information set, allowing one
to preserve most of the desired information content (Fig. 15).
New Sensor Nodes, Cloud, and Data Analytics … 481
Fig. 15 Trend over time of the first three principal components of the autoregressive coefficients
Fig. 16 Score plot of the second and third principal components of the autoregressive coefficients
(November 2017, December 2017, February 2018, April 2018)
Fig. 17 Trend over time of the first three principal components of the autoregressive coeffi-
cients (from 10/01/2017 to 10/15/2017)
4 Conclusions
References
1. Farrar CR, Worden K (2012) Structural health monitoring: a machine learning perspective.
Wiley. ISBN: 978-1-119-99433-6
2. Cawley P (2018) Structural health monitoring: Closing the gap between research and indus-
trial deployment. Struct Health Monit 17(5):1225–1244. https://doi.org/10.1177/147592171
7750047
3. Augustin A, Yi J, Clausen T, Townsley WM (2016) A study of LoRa: long range & low power
networks for the internet of things. Sensors 16:1466. https://doi.org/10.3390/s16091466
4. Alonso L, Barbarán J, Chen J, Díaz M, Llopis L, Rubio B (2018) Middleware and communica-
tion technologies for structural health monitoring of critical infrastructures: a survey. Comput
Standards & Interf 56: 83–100, ISSN 0920-5489. https://doi.org/10.1016/j.csi.2017.09.007
5. Arcadius Tokognon C, Gao B, Tian GY, Yan Y (2017) Structural health monitoring framework
based on internet of things: a survey. IEEE Internet of Things J 4(3): 619–635. https://doi.org/
10.1109/JIOT.2017.2664072
6. Valeske B, Osman A, Römer F, Tschuncky R (2020) Next generation NDE sensor systems as
IIoT elements of Industry 4.0. Res Nondestruct Evaluat 31(5–6):340–369. https://doi.org/10.
1080/09349847.2020.1841862
7. Farrar CR, Park G, Allen DW, Todd MD (2006) Sensor network paradigms for structural health
monitoring. Struct Control Health Monit 13:210–225. https://doi.org/10.1002/stc.125
8. Sousa H, Wenzel H, Thöns S (2019) Quantifying the value of structural health information for
decision support guide for operators. COST Action TU 1402
9. Diamantidis D, Sykora M, Sousa H (2019) Quantifying the value of structural health
information for decision support guide for practicing engineers. COST Action TU 1402
10. Thöns S (2019) Quantifying the value of structural health information for decision support
guide for scientists. COST Action TU 1402
11. Ansari F (2007) Practical implementation of optical fiber sensors in civil structural health
monitoring. J Intell Mater Syst Struct 18(8):879–889. https://doi.org/10.1177/1045389X0607
5760
12. Wu T, Liu G, Fu S, Xing F (2020) Recent progress of fiber-optic sensors for the structural health
monitoring of civil infrastructure. Sensors 20(16):4517. https://doi.org/10.3390/s20164517
13. Cheung W-S (2020) Effects of the sample clock of digital-output MEMS accelerometers on
vibration amplitude and phase measurements. Metrologia 57(1)
14. Cigada A, Lurati M, Redaelli M, Vanali M (2007) Mechanical performance and metrolocigal
characterization of mems accelerometers and application in modal analysis. In: IMAC XXV
(International modal analysis conference), Orlando, 19–22 Feb 2007
15. JCGM 100 series—Guides to the expression of uncertainty in measurement (GUM series) ISO
16. ISO16063-16: Methods for the calibration of vibration and shock transducers—Part 16:
Calibration by Earth’s gravitation
17. ISO16063-21 Methods for the calibration of vibration and shock transducers—Part 21 Vibration
calibration by comparison to a reference transducer
18. Zhu J, Wang W, Huang S, Din W (2020) An improved calibration technique for MEMS
accelerometer based inclinometers. Sensors 20:452. https://doi.org/10.3390/s20020452
19. Levin A, Garion S, Kolodner EK, Lorenz DH, Barabash K, Kugler M, McShane N (2019)
AIOps for a cloud object storage service. In: 2019 IEEE international congress on big data
(BigData Congress). IEEE. https://doi.org/10.1109/BigDataCongress.2019.00036
20. Flynn EB, Todd MD (2010) A Bayesian approach to optimal sensor placement for structural
health monitoring with application to active sensing. Mechan Syst Signal Proces 24:891–903
21. Anania L, Badalà A, D’Agata G (2018) Damage and collapse mode of existing post tensioned
precast concrete bridge: the case of Petrulla viaduct. Eng Struct 162:226–244
22. Colajanni P, Recupero A, Ricciardi G, Spinella N (2016) Failure by corrosion in PC bridges:
a case history of a viaduct in Italy. Int J Struct Integr 7(2):181–193
484 I. Alovisi et al.
23. Minh H, Mutsuyoshi H, Niitani K (2007) Influence of grouting condition on crack and load-
carrying capacity of post-tensioned concrete beam due to chloride-induced corrosion. Constr
Build Mater 21(7):1568–1575
24. Levin A, Garion S, Kolodner EK, Lorenz DH, Barabash K (2019) IBM Research—Haifa AIOps
for a cloud object storage service. 2019 IEEE international congress on big data
25. Cigada A, Lucà F, Malavisi M, Mancini G (2021) A damage detection strategy on bridge
external tendons through long-time monitoring. In: Conference proceedings of the society for
experimental mechanics series, pp 159–168
26. Bertagnoli G, Lucà F, Malavisi M, Melpignano D, Cigada A (2020) A large scale SHM system:
a case study on pre-stressed bridge and cloud architecture. In: Conference proceedings of the
society for experimental mechanics series, pp 75–83
27. Pandit SM, Wu S-M (2000) Time series and system analysis with applications. Krieger
Publishing Company
28. Jolliffe IT (2002) Principal component analysis. Springer, Berlin
29. Ljung L (1999) System identification: theory for the user. Prentice-Hall, Englewood Cliffs