Artificial Intelligence in Construction Engineering and Management
Artificial Intelligence in Construction Engineering and Management
Artificial Intelligence in Construction Engineering and Management
Limao Zhang
Yue Pan
Xianguo Wu
Mirosław J. Skibniewski
Artificial
Intelligence in
Construction
Engineering and
Management
Lecture Notes in Civil Engineering
Volume 163
Series Editors
Marco di Prisco, Politecnico di Milano, Milano, Italy
Sheng-Hong Chen, School of Water Resources and Hydropower Engineering,
Wuhan University, Wuhan, China
Ioannis Vayas, Institute of Steel Structures, National Technical University of
Athens, Athens, Greece
Sanjay Kumar Shukla, School of Engineering, Edith Cowan University, Joondalup,
WA, Australia
Anuj Sharma, Iowa State University, Ames, IA, USA
Nagesh Kumar, Department of Civil Engineering, Indian Institute of Science
Bangalore, Bengaluru, Karnataka, India
Chien Ming Wang, School of Civil Engineering, The University of Queensland,
Brisbane, QLD, Australia
Lecture Notes in Civil Engineering (LNCE) publishes the latest developments in
Civil Engineering - quickly, informally and in top quality. Though original
research reported in proceedings and post-proceedings represents the core of
LNCE, edited volumes of exceptionally high quality and interest may also be
considered for publication. Volumes published in LNCE embrace all aspects and
subfields of, as well as new challenges in, Civil Engineering. Topics in the series
include:
• Construction and Structural Mechanics
• Building Materials
• Concrete, Steel and Timber Structures
• Geotechnical Engineering
• Earthquake Engineering
• Coastal Engineering
• Ocean and Offshore Engineering; Ships and Floating Structures
• Hydraulics, Hydrology and Water Resources Engineering
• Environmental Engineering and Sustainability
• Structural Health and Monitoring
• Surveying and Geographical Information Systems
• Indoor Environments
• Transportation and Traffic
• Risk Analysis
• Safety and Security
All books in the series now indexed by Scopus and EI Compendex database!
Artificial Intelligence
in Construction Engineering
and Management
Limao Zhang Yue Pan
School of Civil and Environmental School of Civil and Environmental
Engineering Engineering
Nanyang Technological University Nanyang Technological University
Singapore, Singapore Singapore, Singapore
© The Editor(s) (if applicable) and The Author(s), under exclusive license to Springer Nature Singapore
Pte Ltd. 2021
This work is subject to copyright. All rights are solely and exclusively licensed by the Publisher, whether
the whole or part of the material is concerned, specifically the rights of translation, reprinting, reuse
of illustrations, recitation, broadcasting, reproduction on microfilms or in any other physical way, and
transmission or information storage and retrieval, electronic adaptation, computer software, or by similar
or dissimilar methodology now known or hereafter developed.
The use of general descriptive names, registered names, trademarks, service marks, etc. in this publication
does not imply, even in the absence of a specific statement, that such names are exempt from the relevant
protective laws and regulations and therefore free for general use.
The publisher, the authors and the editors are safe to assume that the advice and information in this book
are believed to be true and accurate at the date of publication. Neither the publisher nor the authors or
the editors give a warranty, expressed or implied, with respect to the material contained herein or for any
errors or omissions that may have been made. The publisher remains neutral with regard to jurisdictional
claims in published maps and institutional affiliations.
This Springer imprint is published by the registered company Springer Nature Singapore Pte Ltd.
The registered company address is: 152 Beach Road, #21-01/04 Gateway East, Singapore 189721,
Singapore
Preface
The construction industry worldwide has been a late bloomer to adopting digital
technology, where construction projects are predominantly managed with a heavy
reliance on the knowledge and experience of construction professionals. Artifi-
cial intelligence (AI) works by combining large amounts of data with fast, itera-
tive processing and intelligent algorithms (e.g., neural networks, process mining,
and deep learning), allowing the computer to learn automatically from patterns or
features in the data. It provides a wide range of solutions to address many chal-
lenging construction problems, such as tunnel-induced damages, risk estimates, and
defect detection. A tremendous transformation has taken place in the past years with
the emerging applications of AI in the construction industry. This enables industrial
participants to operate projects more efficiently and safely, not only increasing the
automation and productivity in construction but also enhancing the competitiveness
globally.
This book highlights the latest technologies and applications of AI in the domain
of construction engineering and management (CEM). Various AI techniques have
been developed to make machines mimic human cognitive processes in terms of
learning, reasoning, and self-correcting. Relying on the important AI approaches,
we put emphasis on nine hot research topics of AI in CEM, including “Knowledge
Representation and Discovery” (second chapter), “Fuzzy Modeling and Reasoning”
(third chapter), “Time Series Prediction” (fourth chapter), “Information Fusion”
(fifth chapter), “Dynamic Bayesian Networks” (sixth chapter), “Process Mining”
(seventh chapter), “Agent-Based Simulation” (eighth chapter), “Expert Systems”
(ninth chapter), and “Computer Vision” (tenth chapter). Different AI approaches
have been proposed to achieve various kinds of functions which are beneficial to
CEM in terms of automation, safety, reliability, and digitalization. These applica-
tions serve as evidence to verify that AI techniques have created tremendous value
in revolutionizing the construction industry, leading to a more reliable, automated,
self-modifying, time-saving, and cost-effective process of CEM.
This book expects to (1) reveal complexities in conducting CEM activities and
unlock the potential of incorporating AI in supporting and improving CEM; (2)
display procedures and algorithms in a detailed manner on how to use AI to discover
knowledge and address CEM challenges; (3) present evidence-based practices to
v
vi Preface
Acknowledgements It is our sincere desire to thank the anonymous peer reviewers for this work.
We further want to thank Dr. Leilei Chang, Mr. Kai Guo, Ms. Ying Wang, and Ms. Weiyi Chen,
who are postdoctoral research fellows or Ph.D. candidates at the School of Civil and Environ-
mental Engineering, Nanyang Technological University, Singapore, for their cooperative research
and assistances with gathering literature in some of the methods and theories considered.
Contents
vii
viii Contents
Information Fusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
2 Structural Health Assessment (SHA) in Tunnels . . . . . . . . . . . . . . . . . . . . . 97
3 Information Fusion . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 99
4 A Hybrid Information Fusion Approach . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
4.1 Probabilistic SVM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 101
4.2 Improved D-S Evidence Rule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
4.3 Safety Risk Analysis and Assessment . . . . . . . . . . . . . . . . . . . . . . . . . 106
5 Information Fusion for SHA in Tunnels . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
5.1 Project Profile . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 107
5.2 SVM Model Training . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 109
5.3 Multi-classifier Information Fusion . . . . . . . . . . . . . . . . . . . . . . . . . . . 113
6 Merits Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
6.1 Uncertainty Modeling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 115
6.2 Comparative Study . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 119
7 Summaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 121
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 122
Dynamic Bayesian Networks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 125
2 BN and DBN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
2.1 BN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 126
2.2 DBN . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 127
3 Dynamics in Tunnel-Induced Damages . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 128
3.1 Tunnel-Induced Road Damage . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 129
3.2 Control Standard for Road Damage . . . . . . . . . . . . . . . . . . . . . . . . . . . 130
4 DBN-Enabled Dynamic Risk Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
4.1 Risk/Hazard Identification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 132
4.2 DBN Learning . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 133
4.3 DBN Analytics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 135
5 Dynamic Risk Analysis in Tunnel-Induced Damages . . . . . . . . . . . . . . . . . 137
5.1 Case Profile . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
5.2 DBN Model Development . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 137
5.3 Predictive Analysis and Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 141
5.4 Sensitivity Analysis and Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
5.5 Diagnostic Analysis and Control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 144
6 Summaries . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 145
References . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 146
Process Mining . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 147
2 Process Mining . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 149
3 Event Logs in BIM . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 151
4 Process Mining Framework . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 153
4.1 Process Discovery . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 154
4.2 Process Diagnosis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 156
x Contents
1 What Is AI
1.1 AI Development
AI is not a completely new concept. Its development has gone through three major
stages [3]. The early golden age of AI research is during 1956–1974. The name
“artificial intelligence” made its advent at the 1956 Dartmouth Conference. Since
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 1
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_1
2 Introduction to Artificial Intelligence
1.2 AI Techniques
Many AI techniques have been developed to make machines mimic human cogni-
tive processes in terms of learning, reasoning, and self-correcting. In other words,
with the help of AI techniques, machines can perform tasks under a certain degree
of intelligence. There are many ways to classify various developed AI techniques.
Herein, we category them into five major groups, as illustrated in Fig. 1.
2 What Is CEM
To ensure the resiliency and success of the construction industry, construction engi-
neering and management (CEM) highlights the significant contribution of theoretical
and technical knowledge in addressing key aspects of the infrastructure construction
project. CEM can be conceived as an executive process of the physical work on
the construction site [10]. Architectures, engineers, constructors, contractors, and
others work together to provide professional services in constructing buildings for
residential, commercial, or industrial applications and solving on-site problems.
2 What Is CEM 5
Since the assembly process is made up of several order events, there are complicated
patterns of individual time requirements along with restrictive sequential and coop-
erative relationships among the project’s many segments embedding in such process
[13]. The common interest of CEM chiefly arises from the following two sides.
For one thing, construction, as a large sector of the economy, plays a prominent
role in driving economic growth and long-term national development [4]. According
to a survey from McKinsey Global Institute in 2017, the global construction industry
makes up around 13% of the world’s Gross Domestic Product (GDP) and this number
is projected to rise to 15% in 2020. Meanwhile, construction projects create a broad
range of job opportunities for 7% of the world’s working population. Despite its
economic importance, an obvious issue is poor labor productivity during the construc-
tion procedure, negatively leading to the waste of manpower, material resources,
and financial resources. Since construction activities contribute a lot to our society
economically, it makes the most sense to take proper construction management to
improve product performance. If the construction productivity is enhanced by as
much as 50% to 60% or higher, it is estimated to bring an additional $1.6 trillion into
the industry’s value each year and further boost the global GDP.
For another, construction with inherent complexity is one of the most dangerous
industries worldwide, which is greatly susceptible to a variety of unpredictable
factors, such as participants in different roles, the changeable environment in large
uncertainty, and others [12, 15]. It is likely for workers to be exposed to a great threat
to health, and thus, they may suffer from mechanical injuries, degenerative disor-
ders, chemical hazards, psychosocial stress, and others [7]. Therefore, the construc-
tion industry tends to cause a small scale of fatal accidents with higher frequency
than other domains, which is even responsible for 30–40% of fatalities worldwide
[21]. In the Chinese construction market, the number of construction accidents and
fatalities remains stubbornly high [17, 18]. Numerous researches have revealed that
safety issues are tied up with management, emphasizing the necessity of construction
management for safety control and accident prevention [8]. Through identifying,
evaluating, and reacting to the potential risk in dynamic construction work at an
early stage, it is expected to eliminate safety hazards and then achieve a sustainable
reduction in fatalities in the construction industry.
and collaboration. Project managers are responsible for project management, who
coordinate closely with other participants to draw up plans, schema, timelines, costs,
personnel arrangements, and others. They monitor the entire progress of work and
concentrate on all aspects of the project (i.e., labor, capital, time, equipment, material,
risk), in order to give back corresponding instructions to lower the possibility of
delays, budget overruns, high risks, and great conflicts. In short, the main activities
in CEM can be categorized into three major phases of construction as follows.
(1) Planning
Prior to the start of physical construction, the conception of the project is proposed
by clients. Then, its feasibility is analyzed comprehensively to clear the project
objectives. For meeting such a specific goal, detailed plans for the project develop-
ment are created in terms of resources, schedules, budgets, and dependencies, which
can potentially help in reducing cost, duration, and irrational process in the prac-
tical project. For instance, the schematic design can be drawn to fully describe the
building systems. Scheduling can chronologically distribute multiple activities, and
then assign dates, workers, and resources to a certain activity. It aims to fit the plan
to a reasonable time scale and workflow. Cost estimation is the process of predicting
the needed funds and resources for conducting a project within a defined scope. That
is to say, the planning phase incorporates activities mainly for forming project plans,
drawing design, and allocating labors, sources, and time. These activities assist in
rationally streamlining the construction process. They can serve as a reference to
monitor the actual process and guide it to be finished punctually within the estimated
budget.
(2) Construction
This is a phase of executing the physical construction, and thus, the planning made
at the previous phase is expected to pay off. For construction workers, they are
the manual labor to perform the on-site construction, who conduct a range of
basic tasks in a general sequence, including layout marking, excavation, foundation
work, column casting, wall construction, lintel, roofing, plastering, fixing of doors,
windows, electrical and plumbing works, tiles laying, painting, and others. Some
skilled workers are also needed to operate sophisticated machine tools. Regarding
project managers, they take charge of the on-site monitoring and assessment. That is
to say, managers oversee the actual construction process in terms of scope, budget,
and schedule, and then compare the observations to the formed planning. If an incon-
sistency is detected, the corresponding action can be enacted to bring the process
back into conformance or adjust plans to cope with changes. Moreover, managers
are also expected to identify the exposure of risk and the associated impact on the
performance, schedule, or budget of projects. Furthermore, both the qualitative and
quantitative risk analysis can be carried out for quality control, which provides a better
understanding of the risk factors and their effects. As a result, timely responses can
be created to proactively address the potential issues, which can eventually minimize
the risk before it occurs.
2 What Is CEM 7
(2) Labor-intensive
Typically, a huge amount of manual labor should be required in a construction project,
who can offer great quantities of physical effort in charge of the project imple-
mentation, timely completion, and quality assurance. In general, the proportion of
labor costs can take up 30–60% of the total project budget, indicating that inten-
sive labor is a critical component in construction. These flexible workforces can be
accommodated to the changeable projects dynamically based on their initiative and
problem-solving ability. It has been found that 40 million workers were employed in
the construction sector in the year 2007, among whom 27% was skilled labor, 55%
represented the unskilled labor, and the rest denoted the support staff [7]. Currently,
there is fast growth in the demand for skilled and semi-skilled workers with more
skills and expertise, since they play an important role in improving the performance
and productivity of construction projects. In other words, this kind of labor force can
provide useful information and techniques to drive the project forward more easily,
efficiently, and safely.
(3) Dynamics
Although a project has been clearly defined from the beginning to the end, the actual
execution of the project is unable to always stick to the plan. That is to say, inevitable
changes or adjustments occur dynamically throughout the whole lifecycle of the
project, which are caused by various reasons. For instance, the project alternatives
need to be reformulated due to some human factors, including clients’ dissatisfaction,
designers’ and engineers’ mistakes, financial problems from contractors, and others.
Moreover, unforeseen conditions resulting from the undesired delays, bad weather,
complex geologic environment, additional demands of labor, equipment, materials,
and others are the major causes of changes. Besides, when additions or reductions are
applied to the project scope, the scheduling and budget should be updated accordingly
to adapt to the changing circumstances. It is known that managing changes properly
is the guarantee of success in a construction project, and thus, project managers need
to flexibly identify changes, cope with dynamics, and perform effective controls.
(4) Complexity
The complexity of construction projects can arise from two aspects, namely the task
and participants. For one thing, construction tasks are heavy and diverse, which can
possibly meet conflicting scheduling or performance problems. Therefore, it is not
a straightforward job to formulate proper scheduling to align interconnected tasks,
resources, and teams. To make the operation process run smoothly, a variety of
factors, such as safety, security, environment, weather, workers, and others, should
be taken into full consideration. The strict time limit can even negatively boost the
pressure in executing the planned tasks. When some unexpected things happen in the
construction process, there is no enough time to consider solutions [20]. Additionally,
although the advanced technologies and new materials are adopted for more sustain-
able development, they can possibly contribute to a higher degree of complexity. For
another, groups of workers in different roles and varying proficiencies are commonly
2 What Is CEM 9
involved, who can work together, communicate, and share information with each
other. This kind of multi-disciplinary collaboration will result in intricate interactions
in individuals and tasks. Besides, they tend to have conflicting business interests, and
thus, negotiations are needed to pursue the common goal.
(5) Uncertainty
Uncertainties are unknown before they actually occur, which can be regarded as
threats to unfortunately raise risk in project failure. Notably, a high level of uncer-
tainty is inherent in complicated construction projects, which is closely related to
various factors [19]. For instance, before the site construction, scheduling and cost
need to be estimated reasonably under great uncertainty. The improper estimation
will impede the progress of the project. As for the architecture design, some ques-
tions remain to be answered, such as whether the design can pass audits, whether
clients are satisfied with the design, and others. Moreover, a great deal of unavoid-
able uncertainty exists in the construction phase, such as the ground conditions,
soil-structure interaction, weather conditions, building material properties, design
changes, the reliability of suppliers, and others. If these uncertainties can be detected
and measured at an early stage, the project management can be optimized to act as
a reliable tool in mitigating potential risk in the construction project.
BIM has been considered as a digital backbone of future construction, which can
be reasonably integrated into AI to facilitate the digitalization of construction
projects. For BIM, it can provide a platform for not only collecting large data
about all aspects of the project, but also sharing, exchanging, and analyzing data
in real-time to realize in-time communication and collaboration among various
participants. Thus, all members have immediate access to the latest documents
and information, who can cooperatively examine the shared rich data sources. For
the AI techniques, they can deeply explore these massive amounts of data from
BIM to automate and improve the construction process. Based on the discovered
knowledge, strategic decisions can be informed for streamlining the complicated
workflow, shortening operation time, cutting costs, reducing risk, optimizing staff
arrangement, and others. In addition, the integration of BIM and AI can move
the paper-based work towards online management. For one thing, it can deliver
the most efficient and effective information to keep continuous updating of the
ongoing project. For another, it can take advantage of BIM data to make real-
time analysis, and thus, immediate reactions can be performed without delays to
prevent unexpected events.
• Computer Vision. Computer vision techniques have gradually taken the place
of the laborious and unreliable visual inspection in civil infrastructure condi-
tion assessment. Current advances in computer vision techniques lie in the deep
learning methods based on neural networks to automatically process, analyze,
and understand data in images or video through end-to-end learning, which
can conduct various tasks, like image classification, object detection, semantic
segmentation, and others. Towards the goal of intelligent management in the
construction project, computer vision plays an important role in two perspec-
tives, which are inspection and monitoring applications [14]. To be more specific,
inspection applications perform automated damage detection, structural compo-
nent recognition, unsafe behavior, and condition identification. Monitoring appli-
cations are non-contact methods to capture a quantitative understanding of the
infrastructure status, such as to estimate strain, displacement, and cracks’ length
and width. To sum up, the vision-based methods in CEM are comparatively cost-
effective, simple, efficient, and accurate, which are useful in robustly translating
image data into actional information for the purpose of structural health evaluation
and construction safety assurance.
This chapter briefly introduces the concept, development, and techniques of AI and
reviews activities, characteristics, and significance of CEM. The substantial benefits
of AI in CEM are summarized.
Chapter “Knowledge Representation and Discovery” describes the knowledge
representation and discovery and the relevant application in CEM. Many types
of knowledge representation and discovery methods are reviewed. One promising
12 Introduction to Artificial Intelligence
(BN) and DBN are reviewed. A DBN model consisting of various geological,
design, and mechanical variables is developed to reveal dynamics in tunnel-induced
damages over time. Results testify that the developed model is capable of performing
feed-forward, concurrent, and back-forward control respectively on a quantitative
basis.
Chapter “Process Mining” describes process mining and its application in CEM.
Process mining is regarded as an emerging data mining method, which leverages
various algorithms and mathematical models for the visualization and investiga-
tion of event logs and assists in discovering, evaluating, and improving the existing
process. Process mining is applied to deeply mining large volumes of event logs in
BIM project execution, aiming to improve the understanding and efficiency of the
operation process in the BIM-enabled project. A framework of process mining on
BIM event logs with step-by-step procedures is presented, where in-depth knowl-
edge discovery is conducted in terms of the process discovery, conformance checking,
social network analysis, and decision tree classification.
Chapter “Agent-Based Simulation” describes the agent-based simulation for
improved decision support in CEM. Agent-based simulation provides a near-real
solution to perform various what-if experimental scenario analysis, given some exper-
iments are impossible or hard to be conducted in reality, such as pedestrian evacu-
ation under an emergency. Several common methods in the agent-based simulation
are reviewed in this chapter. The promising social force model which can reflect the
anisotropic characteristics and dynamic features of pedestrian interaction is selected
to simulate and optimize pedestrian evacuation behaviors in a subway station. Several
route planning strategies are modeled in the simulation environment and validated
by actual observations. The developed approach can provide valuable theoretic and
practical insights into a deep understanding of route planning strategies during the
pedestrian evacuation, and thus, the improvement of safety and economic objectives
can further be achieved in the design or re-design of metro evacuation systems.
Chapter “Expert Systems” describes the expert system technique and its applica-
tion in CEM. The expert system provides a powerful tool for knowledge integration
and autonomous inference mechanism, which can solve complex problems to signif-
icantly reduce dependence on domain experts. Considered as a digital representation
of physical and functional characteristics of a facility, BIM has the capability of
generating, storing, exchanging, and sharing the building information in an interop-
erable and reusable way, providing an ideal solution for risk management. A novel
approach integrating BIM and expert systems is proposed to perceive the safety
risk at an early stage of the construction. Both case-based reasoning and rule-based
reasoning techniques are combined in the inference model to improve the flexibility
and comprehensiveness of the system reasoning capacity.
Chapter “Computer Vision” describes computer vision and its application in CEM.
Deep learning techniques are driving advances in computer vision, and different types
of deep learning techniques are reviewed in this chapter. A novel computer vision
approach named a spatial-channel hierarchical network (SCHNet) is proposed to
support the automated and reliable concrete crack segmentation at the pixel level.
14 Introduction to Artificial Intelligence
It not only considers the semantic interdependencies in spatial and channel dimen-
sions, but also adaptively integrates local features into their global dependencies.
Experimental results indicate that SCHNet can outperform state-of-the-art models.
It is robust to noises with a better generalization ability under various conditions,
including shadows, rough surfaces, and holes.
Chapter “Conclusions and Future Directions” summarizes the conclusions and
points out future directions. The AI’s functions of modeling and pattern detec-
tion, prediction, and optimization are highlighted, which create tremendous value
in revolutionizing the construction industry, leading to a more reliable, automated,
self-modifying, time-saving, and cost-effective process of CEM. Several emerging
applications of AI in CEM are put forward, which pave a more affordable and effec-
tive way to relieve the burden on manual labor and facilitate smart construction in
CEM.
References
1. Chang L, Zhang L, Fu C, Chen YW (2021) Transparent digital twin for output control using
belief rule base. IEEE Trans Cybern 1–15
2. Chui M (2017) Artificial intelligence the next digital frontier? McKinsey Comp Glob Inst
47(3):6
3. De Spiegeleire S, Maas M, Sweijs T (2017) Artificial intelligence and the future of defense:
strategic implications for small- and medium-sized force providers. The Hague Centre for
Strategic Studies
4. Giang DT, Pheng LS (2011) Role of construction in economic development: review of key
concepts in the past 40 years. Habitat Int 35:118–125
5. Hassanzad M, Orooji A, Valinejadi A, Velayati A (2017) A fuzzy rule-based expert system for
diagnosing cystic fibrosis. Electron Physician 9:5974
6. Hu Z-Z, Tian P-L, Li S-W, Zhang J-P (2018) BIM-based integrated delivery technologies
for intelligent MEP management in the operation and maintenance phase. Adv Eng Softw
115:1–16
7. Kulkarni G (2007) Construction industry: more needs to be done. Indian J Occup Environ Med
11:1–2
8. Li X, Yi W, Chi H-L, Wang X, Chan AP (2018) A critical review of virtual and augmented
reality (VR/AR) applications in construction safety. Autom Constr 86:150–162
9. Pan Y, Zhang L (2021) Roles of artificial intelligence in construction engineering and
management: a critical review and future trends. Autom Constr 122:103517
10. Pheng LS, Hou LS (2019) The economy and the construction industry. In: Construction quality
and the economy. Springer, pp 21–54
11. Purdy M, Daugherty P (2016) Why artificial intelligence is the future of growth. In: Remarks
at AI now: the social and economic implications of artificial intelligence technologies in the
near term, pp 1–72
12. Sacks R, Rozenfeld O, Rosenfeld Y (2009) Spatial and temporal exposure to safety hazards in
construction. J Constr Eng Manag 135:726–736
13. Sears SK, Clough RH, Sears GA (2008) Construction project management: a practical guide
to field construction management. Wiley
14. Spencer BF Jr, Hoskere V, Narazaki Y (2019) Advances in computer vision-based civil
infrastructure inspection and monitoring. Engineering 5:199–222
15. Waehrer GM, Dong XS, Miller T, Haile E, Men Y (2007) Costs of occupational injuries in
construction in the United States. Accid Anal Prev 39:1258–1266
References 15
16. West DM, Allen JR (2018) How artificial intelligence is transforming the world. Report, 24
Apr 2018
17. Wu C, Song X, Wang T, Fang D (2015) Core dimensions of the construction safety climate for
a standardized safety-climate measurement. J Constr Eng Manag 04015018
18. Zhang L, Wu X, Zhu H, AbouRizk SM (2017) Perceiving safety risk of buildings adjacent to
tunneling excavation: an information fusion approach. Autom Constr 73:88–101
19. Zhang L, Wu X, Zhu H, AbouRizk SM (2017) Performing global uncertainty and sensitivity
analysis from given data in tunnel construction. J Comput in Civ Eng 31(6):04017065
20. Zhang L, Wu X, Chen Q, Skibniewski MJ, Hsu SC (2014) Towards a safety management
approach for adjacent buildings in tunneling environments: case study in China. Build Environ
75:222–235
21. Zhou Z, Goh YM, Li Q (2015) Overview and analysis of safety management studies in the
construction industry. Saf Sci 72:337–350
Knowledge Representation
and Discovery
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 17
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_2
18 Knowledge Representation and Discovery
safety. In general, a common agreement on which party should play a leading role
of leadership in construction safety has not yet been reached.
The main objective of this chapter is to answer the following questions: (i) How
to perceive interactions among multiple stakeholders during construction safety
management practices? (ii) Who plays a leading role in safety leadership in a
changing project environment? In this chapter, an Structural Equation Model (SEM)-
based causal model is proposed for revealing relationships among multiple stake-
holders. The causal relationships among multiple stakeholders in safety leadership
are hypothesized and tested using data gathered via a questionnaire survey. Impacts
of path coefficients on construction safety performance are explored. This research
attempts to provide a dynamic perspective to perceive the updating interactions
among different stakeholders. A better understanding of the knowledge for managing
construction safety in a changing project environment is provided.
Safety leadership is the process of interaction between leaders and followers in order
to achieve organizational safety goals as well as increase motivation and perfor-
mance. There are mainly two aspects involved to perform the safety leadership
analysis, including the individual and organizational aspects. From the individual
aspect, there are mainly five factors, including personality, emotional intelligence,
self-efficacy, motivation, and experience. From the organizational aspect, there are
mainly six factors, including role demands, situational constraints, organizational
support, subcultures, migrant labor, and subcontractors. Not all factors need to be
to conduct a thorough safety leadership analysis. The selection of specific factors
should be determined according to practical conditions of construction projects. In
addition, the weights or contributions of different factors are also an important issue
that should be deliberately chosen.
Given “safety” and “leadership” are abstract concepts in nature, it is logical to
assume that a large quantity of the gathered data should be in the form of qualitative
form, which are mostly gathered through questionnaires. Therefore, it is important
to adopt strict mathematical measures and techniques to conduct surveys, and more
importantly, analyze the data.
In general, safety leadership has drawn tremendous attention from both construc-
tion practitioners and researchers. Safety leadership is considered as a significant
antecedent of safety culture, worker safety behavior, and project success. The causal-
ities between safety leadership and other internal and external variables and the inter-
actions among different stakeholders have not been fully investigated and validated
with panel data. The underlying mechanisms by which leadership may influence
safety are not yet well depicted and understood, where complex interactions among
stakeholders and their impacts on safety performance should be considered and
explored holistically.
3 Knowledge Representation, Learning, and Discovery 19
In recent years, the ramifications of construction accidents are growing with a trend
toward more large-scale and more complex construction projects, especially in devel-
oping countries, such as China. The rapid development of huge and complex infras-
tructure projects, such as urban transit systems and high-speed railways, increase
occupational fatalities in construction. Despite the existence of the development of
occupational health and safety provisions, the effects have been far from satisfactory
[28]. In this chapter, the Chinese construction industry is studied as an example to
investigate the relationship of safety leadership among different stakeholders, aiming
to explore effective strategies to enhance their safety leadership and management.
The owner, contractor, supervisor, and designer parties are the most important
stakeholders related to construction projects. Their safety management levels, profes-
sional cooperation, and coordination have great impacts on ensuring the satisfactory
performance of safe project delivery. Herein, it should be noted that the supervisor
party refers to the third-party supervisor who represents the owner to supervise the
contractor. Several causes in some typical accident cases in the Chinese construc-
tion market are revealed in terms of different stakeholders, as shown in Table 1. A
thorough understanding of accident causes and respective responsibilities of each
participant can provide practical guidance for exploring relevant influential factors
that affect construction safety performance on sites.
Table 1 Several causes in typical accident cases in the Chinese construction market
Participants Causes Construction accidents in different provinces, China
Hubeia Yunnanb Hunanc Henand Jiangxie
√ √
Owner (O) The bidding system is not
implemented strictly
√ √
Construction activities are
undertaken without a work
safety license
√ √ √
Contract terms on construction
performance are not stated
clearly
√ √ √
Contractor (C) Undertaking projects beyond
the certificate scope
√ √
Undertaking projects under the
name of other construction
enterprises
√ √ √
The illegal practice of
sub-contracting or
contract-transferring
√
Technical workers do not have
the required certificates
√
There is no valid construction
scheme on site
√ √
Protective staff provided to
workers are imperfect
√ √
Supervisor (S) The chief supervisor has a
long-term absence during
construction
√ √ √ √
Field supervisors are
unqualified, leading to
inadequate supervision
√ √ √
Examination of construction
organizational design is not
taken
√
Designer (D) Unreasonable design changes
cause misunderstanding and
arguments
√ √
Project design is not consistent
with safety standards
Note a Hubei [11], b SAWS [23], c News [19], d MOHURD [18], e Daily [6]
Owner
Project headquarters
Supervisor
Contractor
Designer
Contractual Supervisory
relationship relationship
Fig. 1 Contractual and supervisory relationships in the DBB mode in the construction industry
There are two types of models (i.e., measurement model and structural model) incor-
porated in an SEM model. Specifically, the measurement model addresses the reli-
ability and validity of the measurement of the latent variables, while the structural
model reveals the relationships between latent variables and describes the amount
of the explained and unexplained variances. An example of an SEM model is shown
in Fig. 2, where Fig. 2a represents a measurement model and Fig. 2b represents a
structural model. The measurement model could be defined by Eq. (1), while the
structural model can be defined by Eq. (2).
In the measurement model, there is a latent variable which cannot be directly
observed and can be revealed by certain observable variables. As shown in Eq. (1)
and Fig. 2a, the observable variables (i.e., Ai1 , Ai2 , Ai3 , and Ai4 ) are used to illustrate
the latent variable (i.e., Ai ) as a demonstration with some degrees of error. The
latent variables that are demonstrated in the measurement model would be used for
further computation in the structural model. The structural model aims to illustrate
relationships among latent variables by assessing the variance. As shown in Eq. (1)
and Fig. 2b, the target latent variable T is illustrated by the related latent variables
A1 , A2 , A3 , and A4 with their path coefficients λ1 , λ2 , λ3 , and λ4 , respectively, as
well as the residual error R1 [10].
Ai j = ϒ j Ai + E j (1)
n
T = λi Ai + R1 (2)
i=1
Fig. 2 A sample SEM model consisting of a a measurement model and b a structural model
24 Knowledge Representation and Discovery
In this research, almost 1000 questionnaires are distributed to workers in the construc-
tion industry by a variety of survey tools, including emails, mails, websites, and
hand-out types. All the surveyed respondents are front-line workers, including project
managers, on-site engineers, designers, supervisors, foremen, and others in construc-
tion projects, where the involved projects are mainly located in the south, central,
5 Knowledge Learning from Data and Validation 25
Table 2 (continued)
Stakeholders Items Descriptions
S6 Safety inspection (Safety inspection should be conducted on a routine
basis to reduce the hidden hazards)
Designer (D) D1 Compliance in design (Construction design should be in accordance
with the quality and safety standards in the construction industry)
D2 Design change management (Changing inquiries in the design should
be well handled, with the construction quality, progress and investment
taken into account)
D3 On-site service (Technical consulting services should be provided to
relevant departments on construction sites)
D4 Dispute resolution (Procedures for dispute resolution should be stated,
and designers should discuss with responsible departments to solve
disputes in case of incomplete, vague or contradictory requirement)
D5 Safety design (Safety issues should be considered in an efficient way in
the project design)
D6 Scientific innovation (Reasonable guidance should be developed when
new structures, materials or techniques are adopted in the project)
Construction CS1 Internal safety performance (Each stakeholder is well aware of
safety (CS) accident prevention to ensure construction safety)
CS2 Collaborative safety performance (Collaboration among stakeholders
for construction safety management is improved)
and northeast China. To have a global perspective of the safety performance across
multiple stakeholders, all the respondents are asked to indicate their role as an owner,
contractor, supervisor, or designer and answer all the stated questions in the survey.
For instance, a supervisor could answer the questions related to the owner on his or
her perception of the project involvement.
The collected data from the valid 464 respondents act as a foundational resource
for the model establishment and analysis in subsequent sections. According to the
results of statistical analysis among 464 valid respondents, it is clear that 37% of the
respondents are between the age of 30 and 39, 43% have 5–10 years of working expe-
rience in the construction industry, and 52% have bachelor degrees. As a result, the
quality of the questionnaire data is ensured, as the majority of the valid respondents
have good knowledge of actual construction practices.
The fact that initial variables have a strong correlation with each other is a neces-
sary pre-condition during the factor analysis. Accordingly, the Kaiser–Meyer–Olkin
(KMO) test and Bartlett test of Sphericity are conducted. Specifically, the KMO
test measures the adequacy of a sample in terms of the distribution of values for
the execution of factor analysis. Bartlett’s test of Sphericity determines whether the
correlation matrix is an identity matrix. Generally, the minimum desired value of
the KMO coefficient and the Bartlett test coefficient are is 0.8 and 0.01, respectively
[12].
5 Knowledge Learning from Data and Validation 27
Regarding the required sample size for the construction of an SEM, some of the
principles should be followed up (i) Thompson [25] indicated that the sample size
28 Knowledge Representation and Discovery
Table 4 Results of the total variance explained in the collected sample of 464 records
Component Total variance explained
Initial eigenvalues Extraction sums of squared loadings
Total % of variance Cumulative % Total % of variance Cumulative %
1 8.29 31.883 31.883 8.29 31.883 31.883
2 3.185 12.249 44.132
3 2.263 8.703 52.835
4 1.674 6.437 59.272
5 1.366 5.253 64.525
6 1.246 4.792 69.317
7 0.729 2.803 72.12
8 0.662 2.547 74.667
9 0.643 2.474 77.141
10 0.575 2.21 79.351
11 0.534 2.055 81.406
12 0.504 1.94 83.346
13 0.484 1.863 85.209
14 0.434 1.67 86.879
15 0.395 1.521 88.4
16 0.372 1.43 89.83
17 0.348 1.339 91.17
18 0.327 1.259 92.428
19 0.315 1.213 93.641
20 0.304 1.169 94.81
21 0.277 1.064 95.874
22 0.248 0.955 96.829
23 0.237 0.912 97.741
24 0.219 0.841 98.583
25 0.196 0.755 99.338
26 0.172 0.662 100
should be at least 10–15 times as much as that of the observed variables; (ii) Byrne
[4] noted that the maximum likelihood (that was the default method used in AMOS)
required the samples to be more than 400 as to better estimate the proposed model;
(iii) Tabachnick and Fidell [24] suggested that the sample number should be at least
200 to construct SEM to make empirical research on it, with the sample number
ranging from 200 to 500. For instance, Fernandez-Muniz et al. [7], Martinez-Corcoles
et al. [16], and Al-Refaie [1] conducted relevant investigations using SEM, and their
sample sizes were 455, 495, and 324, respectively. In this research, there are 26
5 Knowledge Learning from Data and Validation 29
observed variables (see Table 6), and thus, the sample with a size of 464 can meet all
the above requirements.
Based on the proposed hypotheses and data collected from 464 valid respon-
dents, the AMOS 21.0 software packages are used to test the hypotheses and gain
insights into a better understanding of the relationships between the above vari-
ables. To examine the impacts of different variables among different stakeholders
on construction safety leadership, many arrows representing causal relationships
between variables are established. Finally, the SEM for investigating the impact of
different stakeholders on construction safety performance can be established.
Two measures, Average Variance Extracted (AVE) and Composite Reliability (CR),
are used to evaluate the internal consistency in measurement models. Specifically,
AVE estimates the amount of variance captured by a construct’s measure that is
relative to random measurement, while CR is analogous to coefficient α. Bagozzi
and Yi [2] indicated that estimates of AVE great than 0.50 and CR greater than 0.60
could support the internal consistency. The results of the internal consistency test in
measurement models in the established SEM are presented in Table 5. The values
of AVEs range from 0.53 to 0.67, while those of CRs range from 0.87 to 0.92. As a
result, all constructs exhibit good internal consistency in the proposed SEM model.
Table 6 indicate that all the hypotheses as mentioned above are confirmed. In other
words, the established model as shown in Fig. 3 passes all the hypotheses proposed
earlier in this research.
(2) Goodness-of-fit verification
The assessment of the overall model fit is a critical issue in an SEM, and each
parameter in the hypothetical model should be successfully estimated. The SEM
model should pass through the following three types of assessment indices, namely
absolute fit indices, incremental fit indices, and parsimonious fit indices. There are
several specific indicators in each type of assessment indices. A critical review of
the description and evaluation criteria of those assessment indicators can be seen in
Xiong et al. [30]. Table 7 summarizes the results of the goodness-of-fit verification
in the developed construction safety model (see Fig. 3).
Taking the absolute fit indices as examples, the values of RMR (Root Mean Square
Residual), RMSEA (Root Mean Square Error of Approximation), GFI (Goodness-
of-Fit Index), and AGFI (Adjusted Goodness-of-Fit Index) are 0.048, 0.052, 0.941,
and 0.912, respectively, which are all in line with the recommended value. Among
all the indicators, only the value of RFI (that is 0.784) is below the recommended
value of 0.9. In conclusion, the model is considered to be acceptable, as most of the
goodness-of-fit indices are fairly good.
In the above sections, the SEM model for revealing the stakeholder’s relationships
within construction safety leadership is built and verified. Thus, it is rational and
reliable to apply the established model to many kinds of analysis. In this research,
to explore the cause-effect interaction within different stakeholders, the impacts of
path coefficient, stakeholder participation, and construction process on construction
safety performance are investigated.
Path coefficients are standardized versions of linear regression weights which can
be used in examining possible causal linkages between statistical variables in SEM.
Path coefficients analysis estimates the effects of variables in a causal system and
has been used to decompose the influences of one variable on another into direct,
indirect, and total effects [8]. In the established SEM (see Fig. 3), relevance and
mediation effects among latent variables (that represent different stakeholders) are
revealed. To verify effect significance, Table 8 illustrates direct, indirect, and total
effects among different stakeholders through the use of bootstrapping in AMOS.
Results are analyzed in detail as follows:
(1) Direct effects. Direct effects measure the strength of the association between
one variable and another variable that is directly collected. As shown in Table 8,
the contractor party displays a most significant direct effect (with a value of
0.3300), followed by the designer party with a direct effect of 0.2800, the owner
party (with a direct effect of 0.2100), and the supervisor party with a direct
effect of 0.1800. It is easy to understand that the contractor party is the direct
stakeholder that is responsible for project safety leadership in construction
projects. The designer and owner parties contribute a lot to project safety issues
Table 8 Results of direct, indirect, and total effects on causal relationships between different
stakeholders and construction safety performance
Relationships Direct effects Indirect effects Total effects Ranking
Owner → Construction 0.2100 0.1563 0.3663 1
safety
Contractor → Construction 0.3300 – 0.3300 2
safety
Supervisor → Construction 0.1800 0.0420 0.2220 4
safety
Designer → Construction 0.2800 0.0330 0.3130 3
safety
6 Knowledge Discovery in Safety Leadership 33
in terms of safety design and strict bidding, respectively. As for the supervisory
party, its direct contribution to the project safety is low.
(2) Indirect effects. Indirect effects measure the strength of the association of one
variable with another variable that is mediated through other variables in the
model. As shown clearly in Table 8, the owner party has the highest indirect
effect (with a value of 0.1563), followed by the supervisor party with an indi-
rect effect of 0.0420 and the designer party with an indirect effect of 0.0330.
No indirect path exists between the contractor party and the construction safety
performance. In total, there are 6 indirect paths leading to construction safety
performance. Taking the owner party as an example, Fig. 4 illustrates the distri-
bution of indirect paths between the owner party and construction safety perfor-
mance. The ranking list of indirect effects displays a great change, compared
to direct effects. This is because the causal influence mechanism varies greatly
within direct and indirect effects. In general, the contractor party leads a promi-
nent role in the direct influence on construction safety performance, while the
owner party leads a prominent role in the indirect influence.
(3) Total effects. Total effects sum up both direct and indirect effects on the target
variable in the model. As shown clearly in Table 8, the owner party has the
highest total effect (with a value of 0.3663), followed by the contractor party
(with a total effect of 0.3300), the designer party (with a total effect of 0.3130),
and the supervisor party (with a total effect of 0.2220). That is to say, the
owner party plays a significant role in construction safety leadership, with
both direct and indirect effects taken into account. This is mainly because
the current Chinese construction industry is a buyer’s market, and the owner
party with a powerful position has the capacity of exerting a big influence on
project performance in terms of both direct and indirect effects. As a result, the
owner party plays a leading role in ensuring the safety performance of a real
project, and more attention should be given to safety issues, such as contractor
bidding, design checking, and contract supervision and management for the
owner party.
0.20
Supervisor 0.18 Contractor
0.33
Construction
Owner 0.20 Supervisor 0.18
safety
0.23
0.33
Designer 0.10 Contractor
Fig. 4 Strength of indirect paths between the owner party and construction safety performance
34 Knowledge Representation and Discovery
Effective participation of all stakeholders is essential for the delivery of a safe project
throughout the life cycle. In SEM, the observed variables are seen as functions of
the latent variable, whereby changes in the latent variable are reflected or mani-
fested in the observed indicators. A high factor loading on an observed variable
implies that the latent variables are better manifested in the observed variable. To
identify the better reflective variables that impact the safety leadership performance
of each stakeholder, the four measurement models as shown in Fig. 3 are used to
reveal factor loadings of each variable in different stakeholders. The regression coef-
ficients between observed variables and a specific stakeholder are then extracted. The
observed variables that manifest well the latent variables (stakeholders) are picked
up and analyzed as follows.
(1) Owner party. The owner plays a leading role in the project-licensing, approving,
bidding, construction preparation, and fulfillment of the project target. The
ranking list regarding regression coefficients between observed variables (γOi ,
i = 1, 2, …, 6) and the owner’s safety leadership performance is presented
as followed, O4 (γO4 = 0.86) > O5 (γO5 = 0.79) > O6 (γO6 = 0.72) > O3
(γO3 = 0.69) > O2 (γO2 = 0.68) > O1 (γO1 = 0.61). Obviously, O4 (Permis-
sion achievement) and O5 (Bidding system) are the most remarkable observed
variables that reflect the safety leadership performance of the owner. In other
words, the owner’s involvement, including both participation and leadership
in construction safety management, is mainly manifested through the permis-
sion achievement mechanism (O4 ) and the bidding system (O5 ) based on the
perceptions of all the participants in construction. Thus, those issues should
be handled properly in order to maintain the role of the owner’s safety leader-
ship in a project. To be specific, construction activities should be carried out
when the owner has received all the required permissions from government
departments via legal means (O4 ), and the bidding system must be performed
strictly to ensure the openness, fairness, and justness of the contractor selection
process (O5 ). Results indicate that illegal operations in obtaining the required
licenses from the government and black-box operations in the bidding process
are both urgent problems that need to be solved in the current Chinese construc-
tion industry. Particularly, it is commonly accepted by all stakeholders that an
open, fair, and transparent market environment is a key matter to eliminate
subsequent safety hazards in construction projects.
(2) Contractor party. The contractor party is in charge of the entirety of a construc-
tion project in the DBB delivery mode. It needs to handle multi-objective
optimization problems to achieve ultimate goals for costs, schedules, quality,
and safety management standards. The ranking list regarding regression coef-
ficients between observed variables (γCi , i = 1, 2, …, 6) and latent variable (the
contractor’s safety leadership performance) is presented as follows: C2 (γC2 =
0. 86) > C4 (γC4 = 0.84) > C1 (γC1 = 0. 82) = C6 (γC6 = 0.82) > C3 (γC3 = 0.72)
= C5 (γC5 = 0.72). Such variables as safety incentive scheme (C2 ), qualified
6 Knowledge Discovery in Safety Leadership 35
employees (C4 ), legal certificates (C1 ), and safety check management (C6 )
exert remarkable impacts on the manifestation of contractor’s role in safety
leadership, as their regression coefficients are all greater than 0.80. Suprapto
et al. (2011) indicated that the effectiveness of incentive schemes played a
significant role in the owner-contractor working relationships. Ling and Kerh
[13] observed that contractors had higher incident rates with less qualified
employees. It is clear that there are a large number of management and harmo-
nization tasks in the working interface on construction sites, including orga-
nization, technology, economy, and contract issues, and thus, the contractor’s
safety management performance in leadership are excessive and complex.
(3) Supervisor party. The supervisor party is acting as a link or mediating role
between the owner and other parties, who is responsible for providing the owner
with professional management services, including project planning, resource
organizing, quality inspections, and supervisory. The ranking list regarding
regression coefficients between observed variables (γSi , i = 1, 2, …, 6) and
the supervisor’s safety leadership performance is presented as follows: S1 (γS1
= 0.83) > S3 (γS3 = 0.79) > S2 (γS2 = 0.75) > S4 (γS4 = 0.73) > S5 (γS5 =
0.66) > S6 (γS6 = 0.61). Obviously, the completion of construction supervision
tasks in an objective and fair manner (S1 ) is the most prominent observed vari-
able. In addition, the comprehensive quality and professional skills of the chief
supervisor (S3 ) and field supervisors (S4 ) should be also regarded as important
reflective factors. Blazsin and Guldenmund [3] concluded that supervisors’ fair
judgment was strongly and positively related to construction safety. Mahdavi
and Hastak [14] indicated that supervisors with professionally qualified skills
could largely help promote onsite workers’ understanding of task requirements,
operating procedures, safety compliance regulations, and reward and punish-
ment provisions. Thus, the supervisor party should pay close attention to the
development of satisfactory coordination strategies on sites, as well as excellent
supervisors.
(4) Designer party. The designer party is responsible for contributing ideas and
knowledge to create unique architectural solutions, especially for cost-effective
design and reasonable construction methods. The ranking list regarding regres-
sion coefficients between observed variables (γDi , i = 1, 2, …, 6) and the
designer’s safety leadership performance is presented as follows: D5 (γD5 =
0.87) > D4 (γD4 = 0.86) > D2 (γD2 = 0. 82) > D1 (γD1 = 0. 81) > D6 (γD6 = 0.78)
> D3 (γD3 = 0.77). Those factors, such as safety design (D5 ), dispute resolu-
tion (D4 ), design change management (D2 ), and compliance in design (D1 ),
are the most remarkable reflective variables indicating the designer’s role in
construction safety management. Globally, the concept of design for safety has
gained momentum as an intervention to improve worker safety. Olander [20]
indicated that the change order in design was a big threat to construction safety.
Thus, the design company should not only be responsible for the preliminary
and detailed design but also be necessary to offer timely and effective worksite
services and reject unfair requests from the owner in case of the violation of
the stipulations.
36 Knowledge Representation and Discovery
7 Merits of SEM
In this research, a total of 464 valid responses are collected and further used for the
model establishment, model verification, and model analysis. A systematic SEM-
based approach is proposed to explore causal-effect relationships on construction
safety performance from a stakeholder’s perspective. The advantages and capacities
of the developed approach are discussed as follows:
(1) Perceiving causal relationships. The causality is the relation between a set
of factors (causes) and a phenomenon. How to determine causal relation-
ships among relevant variables in an effective way, however, is a challenging
problem, as numerous complex processes and behaviors are involved in
construction safety leadership. Managing those complex processes/behaviors is
“tuned” successfully through experience, rather than through the application of
pure mathematics principles. In the developed approach, a causal SEM model
consisting of 26 variables is established to reveal relationships between behav-
iors of different stakeholders and construction safety performance according
to survey data. The established SEM model is considered acceptable, as most
of the verification indices are fairly good. In this way, the hypothetical model is
validated by using survey data, and it is rational and reliable to use the verified
SEM model to perform various sorts of analysis.
(2) Identifying leading roles. Many stakeholders are involved in the construc-
tion safety performance practice, and a large number of complex
processes/behaviors are involved in the safety management for different stake-
holders. The question on which factor plays a leading role in ensuring satisfac-
tory construction safety performance is unknown exactly. The developed SEM
approach enables to depict causal effects, including direct, indirect, and total
effects, between factors in terms of path coefficient analysis. From a global
perspective (stakeholders) or a local perspective (processes and behaviors of a
specific stakeholder), the factor with a high path coefficient in the fitted SEM
model can be identified as a leading factor in the construction safety perfor-
mance model. As shown in Table 8, the owner party has the highest total effect
(with a value of 0.3663), where the owner’s involvement in safety leadership is
mainly manifested through the permission achievement mechanism (O4 ) and
the bidding system (O5 ) based on the perceptions of all the participants in
construction. In this way, the leading roles and factors leading to construction
safety performance can be identified.
(3) Revealing construction dynamics. Causes and effects are typically related to
changes, events, or processes, which are the same in the dynamic construc-
tion process. The allocation of duty and responsibility for construction safety
management among different stakeholders may suffer a change as the construc-
tion process advances. The developed SEM-enabled approach can model the
construction safety performance at different phases, in order to reveal causal
relationships over time among different stakeholders. The direct, indirect, and
total effects on construction safety performance within different stakeholders
7 Merits of SEM 37
over time can be calculated and analyzed in a dynamic manner, based on the
fitted SEM model from the survey data. In this way, the dynamic nature of
construction activities can be taken into account to explore updates of the duty
and responsibility for a specific stakeholder over time.
8 Summaries
References
1. Al-Refaie A (2013) Factors affect companies’ safety performance in Jordan using structural
equation modeling. Saf Sci 57:169–178
2. Bagozzi RP, Yi Y (1988) On the evaluation of structural equation models. J Acad Mark Sci
16:74–94
3. Blazsin H, Guldenmund, F (2015) The social construction of safety: comparing three realities.
Safety Science 71:16–27
4. Byrne BM (2001) Structural equation modeling with Amos: basic concepts, applications and
programming. Lawrence Erlbaum Associates, New Jersey
5. Conchie SM, Moon S, Duncan M (2013) Supervisors’ engagement in safety leadership: factors
that help and hinder. Saf Sci 51:109–117
6. Daily P (2015) A collapse accident on construciton site in Jiangxi. http://news.cntv.cn/2015/
01/29/ARTI1422497971393589.shtml, 29 Jan 2015
38 Knowledge Representation and Discovery
32. Zuofa T, Ocheing EG (2017) Senior managers and safety leadership role in offshore oil and
gas construction projects. Procedia Eng 196:1011–1017
Fuzzy Modeling and Reasoning
1 Introduction
In recent years, the construction of subway systems and underground utilities has
increased dramatically due to population pressure and a lack of surface space. The
tunnel boring machine (TBM) has been observed with widespread applications in
tunnel construction, which can be used for excavating tunnels for nearly all types of
rock and geological conditions. However, occurrences of shield cutter-head failures,
cost overrun, and even catastrophic failures of drive motors take place frequently in
projects [26]. For instance, during the construction of the Lesotho Highlands Water
Project, vibration and impact loads on cutters damaged the cutter head, leading to a
serious delay [8]. TBMs experienced high cutter wear, low penetration and advance
rates, groundwater inflow, and tunnel face instability in the Deep Tunnel Sewage
System project, Singapore [27]. During the excavation of the Chengdu metro lines
in China, an extremely high cutter wear in TBMs was reported due to blocked disk
cutters [7]. In general, a decrease in the performance of TBMs can directly affect the
productivity of tunnel projects, leading to unacceptable damages.
TBMs have the advantages of limiting the disturbance to the surrounding ground,
which significantly reduce the cost and failures in lining the tunnel segments. A timely
and appropriate response to changing underground conditions plays a very important
role in ensuring satisfactory performance during the TBM excavation. Actually, one
big challenge is the mismatch between geological variables (e.g., soil density) and
operational variable settings (e.g., the gross thrust of the TBM), which exists in
most of the TBM failures [25]. To precisely predict geological conditions at every
location during a tunnel construction is nearly impossible. Complex TBM systems are
characterized by high dimensions and involve many kinds of variables or subsystems
that are strongly interconnected and mutually dependent. Given complex geological
conditions underground, the operation of TBM is generally “tuned” through expert
experience, where no pure mathematic principles are available [14]. In this situation,
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 41
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_3
42 Fuzzy Modeling and Reasoning
The core idea in fuzzy thinking is that “Borders are inexact”. Fuzzy modeling
and reasoning consider the degree of relationship and incorporates the updates for
dynamic process analysis. Fuzzy modeling is a fuzzy valued approach compared to
the two-valued data analysis and reasoning methods. The whole fuzzy set includes
more than two values representing the relationship ranging from weak to strong. In
fuzzy system modeling, the inference is achieved by approximate reasoning. Several
methods have been developed for conducting fuzzy modeling and reasoning, which
can be generally classified into the following four categories.
(1) Probability-Based Reasoning (PBR). PBR aims to combine the capacity of
the probability theory to handle uncertainty with the capacity of deductive
logic to exploit the structure of formal argument, which investigates a natural
extension of traditional logic truth tables, whereas the result is derived through
probabilistic expressions instead. The result is a more expressive formalism
with a broad range of possible application areas. Fault Tree Analysis (FTA) is
one of the most effective techniques for estimating the occurrence frequency of
hazardous events within PBR [5], which is highly recommended as a tool for
risk analysis and assessment in tunnel construction according to “Guidelines
for Tunnelling Risk Management” published by the International Tunneling
Association (ITA) in 2004.
(2) Rule-Based Reasoning (RBR). RBR is known as a reasoning technique with
powerful deductive inference capacity, which can be employed to deal with
complicated realistic problems, such as goal programming, scheduling, and
budgeting [10]. The basic form of a rule is in the form of “if <conditions>,
then <conclusion>”, where <conditions> represents the conditions of a rule and
<conclusion> represents its conclusion. The conditions of a rule are connected
between each other with logical connectives, such as AND, OR, NOT, and
others. When sufficient conditions of a rule are satisfied, the rule is acti-
vated and the conclusion is derived [2]. Fuzzy logic is one typical computing
approach within RBR, which provides a decision tool to solve engineering
problems under uncertainty because of its suitability for representing uncertain
information.
(3) Case-Based Reasoning (CBR). CBR is the process of solving new problems
based on the solutions of similar past problems, which attempts to seek an
approximatively similar case using an analogous reasoning technique, and then
make a corresponding adjustment for problem-solving. The reasoning process
is fast and efficient at the expense of in-depth analysis [10]. CBR has a self-
learning and self-improvement capability by adding new cases continually as
the system operates. When a new input case has to be dealt with, the case-based
system performs inference in four phases known as a CBR cycle: including (i)
retrieve, (ii) reuse, (iii) revise, and (iv) retain [1].
(4) Fuzzy Cognitive Map (FCM). FCM merges cognitive map theory used for
decision making in social science with fuzzy logic to improve causal reasoning.
44 Fuzzy Modeling and Reasoning
y = tanh(x)
Determine threshold
functions
Prior knowledge or Step 2. FCM
given observations Computation
Model dynamics
Predictive RCA
Hybrid RCA
FCMs are dynamic systems that have the topology of a directed fuzzy graph
consisting of nodes, arcs, permitting cycles, and feedbacks. Figure 2 illustrates a
simple FCM model. Nodes in the graph refer to the concepts that are used to describe
behaviors of the system, and they are connected by signed and weighted arcs repre-
senting causal relationships among concepts. It should be noted that all the values
in the graph are fuzzy, indicating that the concept nodes have a fuzzy nature. In the
FCM framework, each concept, defined by C i with an interval of [−1, +1], represents
an influential variable that has an impact on the system performance (e.g., events,
w13 w42
C3 w34 C4
46 Fuzzy Modeling and Reasoning
The causal interrelationships among concepts are declared using a linguistic vari-
able that takes a value in the universe U = [−1, +1]. It is known that the number
of linguistic intervals can largely affect the reliability of estimation. Specifically, a
narrow interval indicates that the statistic is precisely known, while a wide interval
3 A Holistic FCM Approach 47
0.5
0
-1.0 -0.75 -0.50 -0.25 0 +0.25 +0.50 +0.75 +1.0
Influence
Fig. 3 Membership functions for the 9-point linguistic scale regarding the degree of influence
between concepts
indicates great uncertainty. In order to reach a high precision of the influence estima-
tion among concepts, it is better to conduct a group of narrow intervals, rather than
wide intervals. According to Dawes’ experiment, the use of the 5–9 point scale is
likely to produce slightly higher mean scores relative to the highest possible attain-
able score, and besides, a larger number of the intervals is usually impractical [4]. For
instance, Stylios and Groumpos [21] used a 9-point linguistic scale to show the fuzzy
set of the linguistic variable influence terms. Accordingly, in this chapter, the 9-point
linguistic scale is used as a basis for an expert to describe in detail the influence of
one concept on another based on his/her knowledge. Figure 3 illustrates the corre-
sponding membership functions for these 9 linguistic terms, including negatively
very strong (μnvs ), negatively strong (μns ), negatively moderate (μnm ), negatively
weak (μnw ), zero (μz ), positively weak (μpk ), positively moderate (μpm ), positively
strong (μps ), and positively very strong (μpvs ), respectively.
In order to exploit the knowledge and experience of experts in the description
and modeling of a complex system, it is assumed that there are M experts who
evaluate every interconnection and describe it with a fuzzy rule inferring a linguistic
weight. The credibility weight for every expert is adopted to increase the objectivity
of the FCM development. In the traditional fuzzy decision making approaches, the
weights of experts are assumed to be equal to each other or fixed without any changes
[17]. However, the level of expert confidence can reveal the reliability of the data
obtained from interviews with various individuals. In the construction practice, it is
generally known that the judgment ability of individuals tends to become increasingly
sophisticated and stable with the accrual of educational background and working
experience.
In this chapter, all the experts are equally credible initially and have the same
credibility weight at the beginning. When determining the weight of the intercon-
nection in FCM, at least M/3 of the experts have to fully agree with suggestions, thus,
48 Fuzzy Modeling and Reasoning
an average weight of the interconnection is calculated using Eq. (1). Otherwise, the
credibility weight of an expert should be reduced by r% every time when there is a
wrong suggestion for an interconnection. Next, the suggested linguistic weights for
an interconnection may be aggregated. A defuzzification method, center of gravity
(COG), is used to produce an aggregated linguistic weight, and a numerical weight
for the interconnection is calculated.
M
k=1 bk × W k
ij
Wi j = (1)
M
where W ij refers to the aggregated fuzzy weight from C i to C j ; M refers to the
number of the participating experts in total; bk refers to the credibility weight of the
kth expert; W ij k refers to the fuzzy weight from C i to C j based on the judgment of
the kth expert.
1
f (x) = (3)
1 + e−λx
3 A Holistic FCM Approach 49
e x − e−x
f (x) = tanh(x) = (4)
e x + e−x
where f (x) is the threshold function, and x is the value of the node/concept in the
FCM framework.
The selection of the threshold function depends on the method that is used to
describe the concepts. In this chapter, we attempt to assess the impacts of root causes
on the reliability of the target event in terms of “positive influence” and “negative
influence”. As a result, the sigmoid threshold function as shown in Eq. (4) is selected
to be applied later in a case study in this chapter.
Once constructed, FCM can be used to simulate the behavior of a system and perform
what-if analysis, due to its capability in fuzzy modeling and reasoning [15]. Several
variables may be involved in the occurrence of a failure, however, which variable
has a greater contribution is unknown exactly. The FCM model should consist of
at least one concept representing the target event, while some concepts representing
the influential inputs. RCA is able to diagnose the root cause and predict the future
outcome of an event. In order to identify the root cause and effect of an event in a
complex system, we focus on the predictive, diagnostic, and hybrid RCA using the
FCM inference technique.
(1) Predictive RCA: Predictive RCA aims to forecast the future outcome or effect
of an event when some evidence or information is available. The propagation
of evidence of a possible cause concept allows an update of the outcome of the
target event in the FCM framework, leading to the change in the outcome of
the target event over time.
(2) Diagnostic RCA: Diagnostic RCA aims to identify the possible root cause(s)
given the occurrence of a failure is observed. The propagation of the observed
information of the target event allows an update of the outcome of other possible
cause concepts in the network, leading to the change in the outcome of the cause
concepts over time.
(3) Hybrid RCA: Hybrid RCA is an integration of both predictive and diagnostic
RCA. The given evidence information of both the target event and possible
causes allows an update of the outcome of both the target event and possible
causes in the FCM framework, leading to the change in their outcomes over
time.
It should be noted that the predictive, diagnostic, and hybrid RCA can be
conducted by using the same equation that is Eq. (2). The difference lies in that
only the evidence of certain root causes is observed in the predictive RCA, only the
evidence of the target event is observed in diagnostic RCA, while the evidence of
both the target event and certain root causes can be observed in hybrid RCA.
50 Fuzzy Modeling and Reasoning
In regard to underground excavation in soft soils, there are two main types of TBMs,
namely earth pressure balance (EPB) TBM and slurry TBM. Compared with the
slurry TBM, the EPB TBM is relatively inexpensive and does not need a large space
on working sites, and thus, it is commonly used with a broad range of applications,
especially in the construction of underground tunnels in urban areas. In this chapter,
we mainly focus on modeling, predicting, and diagnosing the EPB TBM performance
using FCM, where the updating dynamics of the TBM system over time are taken
into full consideration.
EPB TBM is a mechanized tunneling method where the spoil is admitted into TBM
via a screw conveyor arrangement. It allows the pressure at the face of the TBM to
remain balanced without the use of slurry. A screw conveyor transports the exca-
vated material from the base of the excavation chamber onto a belt conveyor. The
interaction between the screw conveyor and the TBM’s advance rate ensures that the
support pressure of the soil paste can be controlled precisely. Generally, TBM perfor-
mance is measured in terms of Advance Rate (C T ), which is the actual excavating
and supporting distance divided by the total time, including downtime for TBM
maintenance, machine breakdown, and tunnel failure. C T depends on various types
of factors ranging from machine design to geological features. Therefore, it may not
be possible to capture this complex relationship in an explicit mathematical formula.
In accordance with an extensive review of the literature, the most important param-
eters that affect the TBM performance can be divided into two major categories, as
follows:
• Geological variables: The construction of a tunnel project is very sensitive to
geological conditions due to complex tunnel-soil interactions. Some particular
geological features may increase the hazardous nature of the TBM advancing
process. Soil Cohesion (C 1 ), Friction Angle (C 2 ), Compression Modulus (C 3 ),
and Soil Density (C 4 ) are four variables frequently used to illustrate the geological
features.
• Operational variables: In the process of the TBM-driven tunneling excavation,
some pressure and speed sensors are installed on the top and middle of the cutter
head in advance. Engineers pay close attention to the measurement of those oper-
ational variables, in order to maintain the face stability of the excavation and
minimize the risk. Rational and appropriate Gross Thrust (C 5 ), Cuter Torque
(C 6 ), Earth Pressure (C 7 ), and Grouting Speed (C 8 ) are key to guarantee the
smooth excavation process. These variables can be further adjusted, depending
on geological conditions underground, in order to maintain the TBM’s satisfactory
performance under excavation.
4 TBM Performance in Tunnel Construction 51
0.5
0
-1.0 -0.5 0 +0.5 +1.0
Favor
Fig. 4 Membership functions for the 5-point linguistic scale regarding concept measurement
To demonstrate the effectiveness and practicality of FCM for analyzing the TBM
performance, a tunnel project, the construction of Metro Line 7 in the Wuhan metro
system in China, is used as a case study to demonstrate FCM’s capability in addressing
engineering problems with uncertainty. The Wuhan metro system is an elevated and
underground urban metro system in the city of Wuhan, Hubei, China. Several urban
metro lines are developed to serve the city of Wuhan, as shown in Fig. 5a. The Metro
Line 7 is a new line with 19 stations for Phase One in the Wuhan metro system,
and the first tunnel cross section of that line, namely Dongfangmacheng–Changfeng
Section (DCS) as shown in Fig. 5b, is used as a case to analyze and assess root causes
in case of poor TBM performance in this chapter. DCS is a double-track tunnel with a
length of 1714 m on the right track and 1717 m on the left track. Two EPB TBMs with
a diameter of almost 6.2 m are used to excavate soils from the Dongfangmacheng
Station to the Changfeng Station at an average depth of 20 m below the ground
surface. The excavation of the right track started on August 22, 2014, and that of the
left track started on September 15, 2014. Figure 5c is a photograph of EPB TBM at
the construction of DCS.
The increase in the complexity and sophistication of large-scale logistics systems
requires the implementation of new intelligent strategies. Capturing the heuristic
knowledge using FCMs may provide the foundation for a new direction in modeling
complex logistics systems. The main requirements for sophisticated systems are
the possession of human-like knowledge and expertise in specific domains [20]. To
explore the complex interactions/logics in the TBM operation system, the knowledge
and experience of a group of experts who operate, monitor, and supervise the system
were obtained through questionnaires. A questionnaire was prepared and distributed
among ten experts who participated in the construction of DCS, including seven
domain experts with at least five years of working experience and three professors
4 TBM Performance in Tunnel Construction 53
(a) (b)
Dongfangmacheng
Station
Changfeng Station
(c)
Fig. 5 Metro lines in the Wuhan metro system: a map of the Wuhan metro system for 2017;
b location of DCS at the Metro Line 7; c EPB TBM at the construction of DCS
in the domain. The experts were asked to evaluate the influence/weight between the
concept C i and the concept C j (i, j = 1, 2, …, 8, T; i = j) using the 9-point linguistic
terms as shown in Fig. 3. In accordance with the methodology, as mentioned in
Sect. 3, the linguistic inputs from the experts were transformed into precise values
using the defuzzification method of COG.
Taking the determination of the influence weight from C 1 to C T as an example,
among the above ten experts, there are seven experts who have a judgment of nega-
tively weak (μnw ), two experts have a judgment of negatively moderate (μnm ), and
only one expert has a judgment of zero (μz ). Figure 6 shows an illustration of how each
individual expert’s input is entered, and multiple inputs are automatically aggregated
and defuzzified. The weight of W 1T is then determined to be −0.275. With knowl-
edge from all the participated experts integrated in an effective manner, Fig. 7 depicts
the graphical representation of the obtained FCM model for the TBM performance
at the construction of DCS.
The operation of the TBM system plays a significant role in ensuring satisfactory
TBM performance in tunnel construction. It is, therefore, necessary to investigate
the cause-effect mechanism of the TBM operation and understand how the TBM
performance evolves quantitatively in case either geological or operational variables
54 Fuzzy Modeling and Reasoning
μnw
1.0
Membership
7×
Experts
0.5
7.0
0
-1.0 -0.75 -0.50 -0.25 0 +0.25
Influence
6.0
μnm
1.0
5.0
Membership
2×
Membership
0.5 4.0
Experts
3.0
0
-1.0 -0.75 -0.50 -0.25 0 +0.25
Influence
2.0
μz
1.0 1.0
Membership
COG = -0.275
1×
Expert
0.5
0
-1.0 -0.75 -0.50 -0.25
Influence
0 +0.25
0
-1.0 -0.75 -0.50 -0.25 0 +0.25
Note: Assume all the experts have the
Influence same credibility weight of 1.0.
undergo a change. After the FCM model is built, various sorts of what-if experiments
can be conducted to explore the cause-effect interaction of the TBM operation system.
In this chapter, we focus on the predictive, diagnostic, and hybrid RCA of the TBM
performance in tunnel construction.
Predictive RCA aims to forecast future outcomes when the evidence of a vari-
able/concept is observed. In the FCM framework, the propagation of evidence of
a possible cause concept allows an update of the outcome of the target concept in
the network in light of the newly found evidence. As mentioned above, the values of
concepts correspond to real measurements that have been transformed in the interval
[−1, +1] using the 5-point linguistic scale (see Fig. 4). The concept has a value
of 0 (−1 or +1), indicating it lies in a very favorable (negatively very unfavorable,
or negatively very unfavorable) range in response to the TBM performance. Hence,
when the evidence is set on a specific cause concept in the FCM model as shown in
5 FCM-Enabled TBM Performance Analysis 55
5 1.0
1.0 17
-0.
5
0.70
C8 -0.125 C1
68
0.5
0.8
22
0.
1
8
1.0 C7 0.4 C2 1.0
5
0.8
27
64
-0.
5
52
5
37
-0.
-0.
-0.35
5
0.42
0.2
5
CT
0.315
-0.63
6
-0.
1
8
0.
0.56
2
-0.
.575
81
0.35
0
8
1.0 C6 C3 1.0
5
0.37
0.
0.
42
25
42
0.
5
-0.636
5
0.12
C5 -0.22
C4
5
1.0 1.0
Fig. 7 Graphical representation of the obtained FCM model for the TBM performance at the
construction of DCS
Fig. 7, the evidence will propagate to the other concepts in the network, which will
cause the outcome of the target event to change.
In this experiment, we attempt to investigate the impacts of variations of a specific
cause concept in the predictive RCA on the TBM performance, in order to identify
the direct cause-effect relationship between the cause concept C i (i = 1, 2, …, 8)
and the target concept C T on a quantitative basis. To be specific, in the initial stage,
all the concepts are set to lie at a favorable range (the value is set to be 0), except
one specific concept C i (i = 1, 2, …, 8). It is assumed that the concept C i is set to
lie in sequence at a negatively very unfavorable level (−1.0), negatively unfavorable
level (−0.5), positively unfavorable level (+0.5), and positively very unfavorable level
(+1.0), respectively, and then we monitor the outcome of system behavior of the TBM
performance (that is the concept C T ) in different scenarios. Using Eqs. (2) and (4),
the value of the concept C T will evolve and then stabilize at a fixed point after several
interactions. Figure 8 illustrates the impacts of variations of geological variables on
the TBM performance in the predictive RCA. Figure 9 illustrates the impacts of
variations of operational variables on the TBM performance in the predictive RCA.
Table 2 shows the fixed values of C T after a set of iterations in different scenarios.
Taking the concept C 1 as an example (see Fig. 8a and Table 2), the concept C T tends
to have a fixed value of +0.7861 (or +0.6660) when the concept C 1 has a value of
−1.0 (or −0.5), and the concept C T tends to have a fixed value of −0.7861 (or −
56 Fuzzy Modeling and Reasoning
(a) (b)
(c) (d)
Fig. 8 Impacts of variations of geological variables on the TBM performance in the predictive
RCA: a the concept C 1 ; b the concept C 2 ; c the concept C 3 ; d the concept C 4
0.6660) when the concept C 1 has a value of +1.0 (or +0.5), respectively. It is clear
that a markedly negative correlation exists between C 1 and C T .
In general, the concepts C 1 , C 2 , C 3 , and C 4 all have markedly negative corre-
lations with the concept C T , while the concepts C 5 , C 6 , C 7 , and C 8 all perform
markedly positive correlations with the concept C T . With respect to the measure-
ment of this kind of correlation among different concepts, the concept C 4 displays
the strongest negative correlation with the concept C T , while the concept C 8 displays
the strongest positive correlation with the concept C T . That is to say, the concepts C 4
and C 8 are both very sensitive to exert impacts on the TBM performance, and thus,
they should be given significant attention in ensuring the satisfactory performance in
tunneling excavation. This is very consistent with what is observed in actual construc-
tion practice, indicating the established FCM model is validated to some extent. In
fact, to maintain the face stability of the excavation and minimize risk, engineers
need to adjust the operational parameters continuously during the TBM operation in
order to fit the surrounding geological conditions in a dynamic manner. The results
achieved from the predictive RCA can provide an increased understanding of how
the TBM performance evolves quantitatively in case either the geological or opera-
tional variables change, which can provide guidance for the reasonable adjustment
of operational variables in tunnel construction.
5 FCM-Enabled TBM Performance Analysis 57
(a) (b)
(c) (d)
Fig. 9 Impacts of variations of operational variables on the TBM performance in the predictive
RCA: a the concept C 5 ; b the concept C 6 ; c the concept C 7 ; d the concept C 8
Table 2 Fixed values of C T after a set of iterations in different scenarios in the predictive RCA
Items The fixed value of C T in different scenarios
P(C T | C i = −1.0) P(C T | C i = −0.5) P(C T | C i = +0.5) P(C T | C i = +1.0)
C1 0.7861 0.6660 −0.6660 −0.7861
C2 0.8374 0.7199 −0.7199 −0.8374
C3 0.9137 0.8105 −0.8105 −0.9137
C4 0.9426 0.8511 −0.8511 −0.9426
C5 −0.8263 −0.7078 0.7078 0.8263
C6 −0.8990 −0.7915 0.7915 0.8990
C7 −0.8570 −0.7417 0.7417 0.8570
C8 −0.9480 −0.8595 0.8595 0.9480
Diagnostic RCA aims to detect possible root cause(s) when a failure of the target
event occurs. In the FCM framework, the propagation of the observed information
of the target event allows an update of the outcome of other variables or causes in the
58 Fuzzy Modeling and Reasoning
network. As mentioned above, the values of the TBM performance correspond to real
measurements have been transformed into the interval [−1, +1] using the 5-point
linguistic scale (see Fig. 4). Hence, when a specific level of the TBM performance
(such as the negatively or positively unfavorable performance) is observed in the
target event (that is the concept C T ) in the FCM model, the observed evidence will
then propagate to the other cause concepts in the network, resulting in the change of
the causes affected accordingly.
In this experiment, we intend to identify the most likely root causes that affect
the reduction of the TBM performance in the diagnostic RCA. In the initial stage,
all the concepts are set to lie at a favorable range (the value is set to be 0), except
the target concept C T . It is assumed that the concept C T is set to lie at a negatively
very unfavorable level (C T = −1.0), a negatively unfavorable level (C T = −0.5),
a positively unfavorable level (C T = +0.5), and a positively very unfavorable level
(C T = +1.0) in turn, respectively, and then we monitor the outcome of behaviors
of other causes in different scenarios. Using Eqs. (2) and (4), the value of a specific
concept C i (i = 1, 2, …, 8) will evolve and then stabilize at a fixed point after several
interactions. The impacts of variations of the TBM performance on geological and
operational variables in the diagnostic RCA are shown in Fig. 10. Results indicate
that the values of geological variables C i (i = 1, 2, 3, 4) remain unchanged over
time since no arcs exist from the concept C T to the concept C i (i = 1, 2, 3, 4)
(a) (b)
(c) (d)
Fig. 10 Impacts of variations of the TBM performance on geological and operational variables in
the diagnostic RCA: a C T = −1.0; b C T = −0.5; c C T = +0.5; d C T = +1.0
5 FCM-Enabled TBM Performance Analysis 59
Table 3 Fixed values of operational variables after a set of iterations in different scenarios in the
diagnostic RCA
Scenarios Fixed value of C i in different scenarios
C5 C6 C7 C8
P(C i | C T = −1.0) −0.8374 −0.9006 −0.8476 −0.9426
P(C i | C T = −0.5) −0.7199 −0.7936 −0.7311 −0.8511
P(C i | C T = +0.5) 0.7199 0.7936 0.7311 0.8511
P(C i | C T = +1.0) 0.8374 0.9006 0.8476 0.9426
in the FCM model (see Fig. 7). This is due to the fact that the variations of the
TBM performance cannot lead to the change of geological conditions itself. The
fixed values of operational variables after a set of iterations in different scenarios
are presented in Table 3. Taking the scenario where the target concept C T suffers a
great negative reduction regarding the TBM performance (that is C T = −1.0) as an
example (see Fig. 10a), the concept C 8 tends to have the largest negative fixed value
(that is −0.9426). That is to say, this variable should be deemed as the suspected
cause which should be checked, given that a great negative reduction of the TBM
performance is observed.
In general, as seen in Table 3, the final fixed value of each concept varies greatly
in different scenarios, indicating that the degree of the cause-effect relation varies
a lot among different concepts. More specifically, the TBM performance is very
sensitive to the change of operational conditions. The concepts C 5 , C 6 , C 7 , and C 8
show positive correlation effects in response to the variation of the TBM performance.
Engineers should give much more attention to the adjustment of operational variables,
such as C 5 , C 6 , C 7 , and C 8 , since the values of geological variables cannot be changed
at the construction stage. The values of operational parameters, however, can be
adjusted to go up in case the TBM performance negatively reduces, and go down in
case the TBM performance positively reduces. In conclusion, the results achieved
from the diagnostic RCA can provide more insights on what kind of variables cause
the reduction of the TBM performance, as well as the direction of adjustment of
relevant variables to ensure the satisfactory performance of the TBM operation.
(a) (b)
Fig. 11 Evolution of TBM performance in the hybrid RCA in the case of different initial values of
the concept C 8 in different scenarios: a C T = −1.0; b C T = −0.5
5 FCM-Enabled TBM Performance Analysis 61
time, and the system components of TBM, such as the shield cutter head, are very
likely to be damaged otherwise.
As shown in Fig. 11a, in order to return the TBM performance from a negatively
very unfavorable level (C T = −1.0) to a favorable level (C T = 0) in a short ten
iterations, it would be a reasonable and sensible decision if we adjust the value of
C 8 to be 0.07. Similarly, as shown in Fig. 11b, to return the TBM performance from
a negatively unfavorable level (C T = −0.5) to a favorable level (C T = 0) in a short
ten iterations, it would be a reasonable and sensible decision if we adjust the value
of C 8 to be 0.05. This deduction is consistent with the actual tunneling construction
practice, since the real-time adjustment of operational variables is very necessary for
ensuring the satisfactory performance of the TBM operation in response to changing
geological conditions. The magnitude of the adjustment of operational variables
should be greater when the TBM performance is observed with a larger reduction.
In general, the observed evidence of operational variables plays a significant role
in the time required when the TBM performance returns to a normal state. The
TBM performance varies greatly, depending on both the initial value of the TBM
performance (C T ) and the adjusted value of operational variables. More specifically,
the value of one operational variable should be increased to be a certain number
when the TBM performance suffers a negative reduction, and the value should be
decreased to be a certain number when the TBM performance suffers a positive
reduction, respectively. At the same time, the capacity of the hybrid RCA can be used
to determine the exact value that the one operational variable should be adjusted to
be, in order to return the TBM performance from a reduction to a normal state in a
very short time.
6 Merits of FCM
This chapter presents an FCM-enabled RCA approach to assessing the TBM perfor-
mance in tunnel construction. The developed approach is capable of performing
predictive, diagnostic, and hybrid RCA. A tunnel case in the Wuhan metro system is
used to demonstrate the effectiveness of the developed approach. More discussions
are conducted in terms of the FCM interpretability, validation, and capacities, as
presented as follows.
[23], the greatest strength of the FCM approach is perhaps ease of use. FCM allows
multiple experts to collaboratively contribute to the construction of an FCM, which
enhances the confidence in its model development and further applications.
The FCM model considers expert feedback on the interrelationships between
concepts that impact the TBM performance. Through fuzzy weights in the FCM
framework, causal interconnections between concepts can provide an intuitive expla-
nation to end-users. In addition, one thing that needs to be pointed out is the measure-
ment of concepts in the developed model. Since the scope of favorable or unfavor-
able geological/operational variables varies greatly among different tunnel cases in
different areas, a challenge exists in determining the value of a specific concept
representing a corresponding geological or operational variable due to uncertainty.
To solve this problem, the values of all concepts have been initialized into a range
[−1, +1] using the 5-point linguistic scale in this chapter, where various kinds of
expert knowledge can be integrated in a unified platform.
For a particular project in a geological area, the magnitude of a specific concept
measurement (i.e., −1.0, −0.5, 0, +0.5, and +1.0) within the range [−1, +1] can
be determined according to the historical data in certain geological conditions, and
the values of concepts can then be calibrated by the Measurement-While-Drilling
(MWD) data in the real world. In this way, an immediate explanation can be obtained
regarding the measurement and adjustment of operational variables in real projects.
This is very convenient and simple for users who are not domain experts, but they
can accurately identify the root causes of an undesirable event or failure.
In most cases, to compare model outputs with observations using independent data
which is not used in the model development can provide a solution to validate the
developed model. It would be convincing that the simulation results can be compared
to the historical data of actual TBM failures for validation purposes. However, the
historical data regarding actual TBM failures are quite limited during the construc-
tion of tunnel projects in the city of Wuhan in China. This is partly because some
maintenance and repair works have been conducted ahead of the occurrence of actual
failures due to the consideration of excessive costs. In regard to this issue, the domain
experts have knowledge for telling when TBM failures are likely to occur according
to their working experience, which provides the potential to validate the simulated
results to some extent.
To validate the effectiveness of the established simulation model (see Fig. 7), the
above ten experts who participated in the questionnaire investigation were invited
again to evaluate the effectiveness of model outcomes in various what-if experiments.
They all considered that the results of model prediction and diagnosis (see Fig. 11 and
Tables 2 and 3) were fairly consistent with the expectation in the actual construction
practice. To some extent, the established model for the TBM performance can be
validated in this way. In addition, in accordance with the predictive and diagnostic
6 Merits of FCM 63
analysis results using the established simulation model, some corresponding control
measures and reasonable adjustment of operational variables were carried out in time
to mitigate the risk during the tunneling excavation. Finally, the TBMs successfully
passed through the right track of DCS on February 2, 2015, and the left track on
March 2, 2015, respectively, without an accident or duration delay.
FCM is considered as a soft computing method within the domain of fuzzy modeling
and reasoning. As mentioned above in Sect. 2, there are many kinds of methods for
fuzzy modeling and reasoning, including the probability-based reasoning methods
(e.g., FTA), rule-based reasoning methods (e.g., RBR), and case-based reasoning
(CBR) methods. Comparing FCM with these traditional methods, what are the advan-
tages of the developed FCM approach? As shown in Sect. 5, the successful imple-
mentation of the developed FCM approach indicates that this approach is capable of
performing predictive analysis, diagnostic analysis, and hybrid analysis. In addition,
the developed FCM approach is verified to be capable of modeling the dynamics of
system behaviors over time (see details to Figs. 8, 9, 10 and 11). For instance, the
evolution of TBM performance over time in the hybrid RCA in the case of different
initial values of the concept C 8 in different conditions is presented in Fig. 11.
From a perspective of the analysis capacity, a qualitative comparison among many
soft computing methods in fuzzy modeling and reasoning, including FCM, FTA,
RBR, and CBR, is presented in Table 4. It is shown clearly that all those four methods
can perform predictive analysis. However, the traditional soft computing methods,
including FTA, RBR, and CBR, have challenges in performing diagnostic analysis,
hybrid analysis, and modeling dynamics. Only FTA is able to conduct the diagnostic
analysis. In summary, the developed FCM approach provides a more competitive
solution in fuzzy modeling and reasoning than the other traditional methods.
Table 4 A qualitative comparison of the analysis capacity among several soft computing methods
Methods Analysis capacities
Predictive analysis Diagnostic analysis Hybrid analysis Modeling dynamics
√ √ √ √
FCM
√ √
FTA – –
√
RBR – – –
√
CBR – – –
√
Note “ ” means the method has the capacity, and “–” means that the method does not have the
capacity
64 Fuzzy Modeling and Reasoning
7 Summaries
The operation of the TBM system is very complex since various kinds of variables are
involved. To ensure the satisfactory performance of the TBM operation becomes a
challenging task. In this chapter, an FCM-based model consisting of nine concepts is
developed to analyze the TBM performance in tunnel construction, with construction
experience and knowledge from domain experts taken into account. The primary traits
of the holistic FCM method and the findings from the case study are summarized as
follows:
(1) From a predictive RCA perspective, a deep understanding of how the TBM
performance evolves quantitatively over time can be provided in case either
the geological or operational variables suffer a change. Through the case study,
we find that C 4 (Soil Density) displays the strongest negative correlation with
the concept C T (TBM Advance Rate); while C 8 (Grouting Speed) displays the
strongest positive correlation with C T . In other words, both C 4 and C 8 are
very sensitive to ensuring the satisfactory performance of the TBM operation,
which should be given significant attention in tunneling excavation.
(2) From a diagnostic perspective, more insights on what kind of variables cause
the reduction of the TBM performance can be detected. Through the case study,
we find that the TBM performance is very sensitive to the change of operational
conditions. The values of operational parameters, C 5 (Gross Thrust), C 6 (Cuter
Torque), C 7 (Earth Pressure), and C 8 (Grouting Speed) can be adjusted to go
up in case the TBM performance negatively reduces and go down in case the
TBM performance positively reduces.
(3) From a hybrid RCA perspective, we can identify the magnitude of the adjust-
ment scope of operational variables when the TBM operational performance
suffers a reduction. For instance, we can adjust the value of C 8 to be 0.07, in
order to return the TBM performance from a negatively very unfavorable level
(C T = −1.0) to a favorable level (C T = 0) in a short ten iterations. It would
be reasonable and sensible that we can adjust the value of C 8 to be 0.05, in
order to shortly return the TBM performance from a negatively unfavorable
level (C T = −0.5) to a favorable level (C T = 0).
The novelty of the proposed holistic FCM approach lies in that it is capable
of modeling dynamics of system behaviors over time and performing many kinds
of what-if scenario analysis, including predictive, diagnostic, and hybrid RCA. It
turns out to be a more competitive solution that deals with uncertainty, dynamics,
and interactions in the approximate reasoning process, compared to other traditional
methods (i.e., FTA, RBR, and CBR). The proposed approach can be used as a decision
support tool for ensuring the satisfactory performance of TBMs, and thus, it can help
increase the TBM productivity in a complex project environment.
References 65
References
25. Zhang L, Chettupuzha AJA, Chen H, Wu X, AbouRizk SM (2017) Fuzzy cognitive maps
enabled root cause analysis in complex projects. Appl Soft Comput 57:235–249
26. Zhang L, Wu X, Skibniewski MJ (2015) Simulation-based analysis of tunnel boring machine
performance in tunneling excavation. J Comput Civ Eng 30:04015073
27. Zhao J, Gong Q, Eisensten Z (2007) Tunnelling through a frequently changing and mixed
ground: a case history in Singapore. Tunn Undergr Space Technol 22:388–400
Time Series Prediction
1 Introduction
Over the past decades, the rapid development of big cities is raising the demands of
underground space utilization. One of the favorable options for urban development
is to build underground tunnels. Notably, a lot of tunnels are located at a low depth
in soil or soft rock zones under densely populated areas, and thus the excavation
works of shallow tunnels in the soft ground tend to result in both lateral and vertical
surfaces. There are two major characteristics of those shallow tunneling projects
including the lower depth and soft ground, which occasionally cause considerable
surface settlements. It should be highlighted the tunnel-induced ground settlement
poses a big threat to both surface and sub-surface facilities, which could act as a
critical factor in evaluating tunnel-induced damages. Therefore, a necessary step to
avoid excessive settlement is based upon the estimation, analysis, and controlling of
the tunnel-induced ground settlement accurately and timely, which has proven to be
a key procedure to ensure construction safety in tunnels [29].
Methods for estimating tunnel-induced ground settlement can be classified into
four main groups, including empirical [16], analytical [2], numerical methods [10],
and intelligent methods [20]. They have their respective merits in predicting tunnel-
induced ground settlement. However, some weaknesses still exist to cause the appli-
cation restrictions. For instance, Loganathan [13] claimed that the empirical methods
failed to give highly accurate results due to the limited adaptability in different ground
conditions. Chou and Bobet [4] indicated that analytical methods were prone to
underestimate the maximum soil deformations or overestimate the settlement trough,
since it took no account of time-dependent consolidation and creep ground loss
components. Łodygowski and Sumelka [12] mentioned that the numerical methods
suffered from low accuracy and validity, since simulation models could be sensitive
to boundary conditions (i.e., stress concentrations and local stresses) and difficult to
be validated. At the same time, most of the existing models are mainly concerned
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 67
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_4
68 Time Series Prediction
with the prediction of the maximum ground settlement without regard to the evolu-
tion of the ground settlement over time. The static results are incapable of estimating
tunnel-induced settlement in a dynamic and real-time manner [5].
At present, an increase in the number of urban tunneling projects can bring about
a large amount of project-related data, which is helpful in developing a more accurate
approach with improved prediction performance in the tunnel-induced daily ground
settlement. In this regard, a novel hybrid approach is proposed under the integration
of Packet Transform (WPT) and Least Square Support Vector Machine (LSSVM),
which combines the complementary advantages of both WPT and LSSVM. That
is to say, WPT with the process of time-frequency signal analysis is carried out to
handle the monitored tunnel-induced settlement sequences, while LSSVM can make
a reliable prediction by learning the high- and low-frequency information. A case
study is provided to estimate, test, and analyze the ground-induced settlement of a
real tunnel project in China in detail, aiming to demonstrate the feasibility of the
proposed approach and its potential in future implementation.
Excavation work of shallow tunnels in the soft ground is bound to generate soil move-
ments, and thus unwanted settlement will occur through developing above the tunnel
on the ground surface. A key problem is to protect existing structures and under-
ground utilities against damages by excessive ground movement. The understanding
of the influence degree and scope of tunneling excavation is particularly an impor-
tant task to address the problem. Typically, the following four methods, including
empirical, analytical, numerical, and intelligent methods, have been developed to
accurately estimate the tunnel-induced ground settlement.
The well-established empirical methods available to date are used primarily to esti-
mate ground settlements in soft soils. Martos [14] firstly proposed that a Gaussian
or normal distribution curve could well describe the shape of the daily settlement
trough. Later, Peck [16] analyzed the settlement data collected from several tunnels
and mines by fitting Gaussian curves, which could provide suggestions about the
equation of settlement, as shown in Eq. (1).
2 Estimation of Tunnel-Induced Ground Settlement 69
⎧
⎪
⎪ x2
⎪
⎪ Sv (x) = Sv,max exp − 2
⎪
⎪ 2i x
⎪
⎪
⎪
⎪
⎨ Sv,max = √ VS
2π i x (1)
⎪
⎪
⎪
⎪ VS = Vl π R 2
⎪
⎪
⎪
⎪
⎪
⎪ Z
⎩ ix = √
2π tg(45◦ − ϕ/2)
where S v (x) is the ground settlement at a transverse distance x from the tunnel center-
line, S v,max is the maximum transverse settlement at x = 0, x is the horizontal distance
between the location and the tunnel centerline in the transverse direction, and ix is
the horizontal distance from the tunnel centerline to the inflection point of the curve.
V S is the stratum loss caused by tunneling excavation, V l is the stratum loss ratio,
and ϕ is the friction angle. Based on the hypothesis of Peck, many scholars offered
further suggestions about the function of ix [15].
However, these empirical methods fail to generate highly accurate results due to
two important limitations as follows. One is that their applicability may be limited
due to different ground conditions and construction techniques, while the other is
that horizontal movements and subsurface settlements cannot be predicted using the
established empirical relationships [13]. Besides, Ding et al. [5] indicated that the
field has a limited number of settlement markers available, which could be another
reason for the unsatisfying estimation accuracy. Celestino et al. [3] claimed that the
real ground settlement was sometimes significantly greater than that obtained from
the approximately fitted curves.
Multiple attempts have been devoted to developing analytical solutions for estimating
tunneling-induced ground settlements. For example, Sagaseta [17] presented closed-
form solutions to obtain the strain field in an initially isotropic and homogeneous
incompressible soil due to near-surface ground loss from tunneling. Verruijt and
Booker [25] proposed an analytical solution for tunnels in homogeneous elastic
half-spaces using an approximate method suggested by Sagaseta [17] for the case
of ground loss. This solution is applicable for the incompressible case and arbitrary
values of Poisson’s ratio. The effect of tunnel lining deformation in the long term is
taken into account [13]. Equation (2) expresses the Verruijt and Booker’s closed-form
solution for estimating settlements.
z1 z2 z 1 (kx 2 − z 22 ) z 2 (kx 2 − z 22 )
Uz = −ε R 2 2
+ 2 + δ R2 +
r1 r2 r14 r24
70 Time Series Prediction
2ε R 2 (m + 1)z 2 mz(x 2 − z 22 )
+ +
m r22 r24
2
x − z 22 m 2zz 2 (3x 2 − z 22 )
− 2δ R h
2
+ (2)
r24 m+1 r26
where U z is the estimation of the vertical settlement, ε is the uniform radial ground
loss, δ is the long term ground deformation due to the ovalization of the tunnel lining,
R is the tunnel radius, H is the tunnel depth, and ν is the Poisson’s ratio of soil. z1 =
z − H, z2 = x + H, r 21 = x 2 + z21 , r 22 = x 2 + z22 , m = 1/(1 − 2 ν), and k = ν/(1 − ν).
In most cases, analytical solutions cannot consider the excavation process with
sufficient approximation. Many simplifications, such as a plain strain assumption,
elastic behavior, and/or soil isotropy, are adopted in those solutions. Nevertheless,
those solutions take no account of time-dependent consolidation and creep ground
loss components, it is likely to underestimate the maximum soil deformations or
overestimate the settlement trough [4]. Therefore, those methods were only suitable
for short-term undrained conditions, which could be restricted to cases with small
ground deformations, such as stiff clays and rocks.
particular, Karakus [8] presented that the settlement trough produced by numerical
analysis was always much shallower and wider than the measured data.
Along with the rapid development of artificial intelligence (AI) technologies, intel-
ligent methods have been gaining increasing attention around the world due to their
powerful capacities and comprehensive adaptabilities. Those intelligent methods,
like artificial neural networks (ANNs) and support vector machines (SVMs), can be
helpful in solving complex real-world problems. Especially for ANN, it acts as a
typical class of intelligent methods to model nonlinear relationships between input
and output variables. Therefore, it is reasonable to apply this method to the analysis
of the ground settlement, which has become a popular topic.
In an existing work [20], the neural network has been constructed based on a
series of inputs, including tunnel depth, tunnel diameter, soil cohesion, volume loss,
and groundwater level, to effectively predict the output of the maximum surface
settlement. However, a great challenge of using ANN is to collect multiple variables
relevant to the tunnel-induced ground settlement [23]. Besides, how to properly
determine an optimal combination of controlling parameters in ANN models, such
as the number of hidden layers/nodes and learning rates, remains another question.
2.5 Comparison
Moving the existing research forward, the importance of accurately assessing tunnel-
induced deformations has been emphasized in recent years to tackle an increasing
number of urban tunnel projects. However, most of the existing methods for deter-
mining ground movements could be somewhat inadequate, over-simplistic, or overly
complicated [13]. It is increasingly necessary to develop a more accurate approach
to better predict tunnel-induced ground settlement. Although many existing models
have focused on the maximum ground settlement, they pay less attention to dynamic
and real-time predictions [26]. This kind of static results is insufficient to meet the
dynamic and real-time requirements regarding updates of the ground settlement over
time [5]. For a clear comparison, Table 1 lists the advantages and disadvantages of
different methods in estimating tunnel-induced ground settlements.
In pursuit of a reasonable solution, some assumptions have been made in tradi-
tional empirical, analytical, and numerical methods. Nevertheless, large amounts
of randomness and uncertainties are bound to exist in the assumptions and actual
construction practice due to the great level of dynamic and complexity in underground
conditions, which possibly lead to remarkable deviations between the predicted and
observed effects. To resolve the above-mentioned inefficiencies, the field monitoring
technique can be a useful solution, which implements the scientific method to check
72 Time Series Prediction
During the process of tunnel excavation, the ground settlement for a specific cross-
section is measured continuously, and thus, the data points collected at regular inter-
vals constitute a time series sequence to track the excavation process over time.
Notably, there are some potential internal structures, such as autocorrelation, trend,
cyclicity, stationarity, and seasonality, embedded in a time series. The implementa-
tion of a time series analysis could act as an essential task to discover meaningful
statistics and characteristics from such serially correlated data with timestamps, By
making full use of changes that occurred in the past, it is possible to predict future
conditions reliably and inform tactical decisions at an early stage.
As computer and machine learning technologies develop, a lot of prediction
algorithms have been proposed over the past decade. As reviewed, the forecasting
approaches can be grouped into two categories, namely the linear methods and non-
linear methods. Early research work pays more attention to the smoothing technique,
error component model, and auto-regressive integrated moving average (ARIMA)
based on the assumption of linearity and stationarity to infer future trends. Due
to the simplicity of these models, they are more suitable for exploring short-term
predictions. In particular, ARIMA is one of the most widely-used models in time
series forecasting, which adopts a minimal number of parameters to ensure relatively
high prediction accuracy. However, some previous research claims that ARIMA is
improper to interpret and model nonlinear relations of many real-world problems,
leading to the increased errors in prediction problem about the nonlinear ground
settlement.
Compared with linear methods, the non-linear prediction methods specialize in
describing non-linear characteristics, aiming to achieve more accurate forecasting
performance. Algorithms belonging to the artificial intelligence field, like artificial
neural networks (ANN) and support vector machine (SVM), have received increasing
attention due to their powerful capacities and comprehensive adaptabilities. It has
been seen that ANN is adopted as an alternative approach for analyzing the ground
settlement, which has become increasingly popular [20, 23]. However, a difficult task
74 Time Series Prediction
Original
settlement data
j
fj1 (t) fji (t) fj2 (t)
Feedback
Predicted
settlement data
where i, j, and k are the modulation, scale, and translation parameters of the wavelet
function ψ ij,k (t), respectively. The first wavelet is the so-called the mother wavelet
function, which is defined as ψ 1 (t) = ψ (t). h(k) and g(k) are the discrete filters
that are quadrature mirror filters associated with the scaling function and the mother
wavelet function.
The great challenge in wavelet analysis lies in the selection of the proper wavelet
function as well as the decomposition level of the signal. Several factors should be
taken into account, including (i) orthogonal or nonorthogonal, (ii) complex or real,
(iii) width, and (iv) shape. Ding et al. [5] presented a selection criterion to choose
the optimal wavelet function in the analysis of tunnel-induced ground settlement.
They developed a few wavelet functions under different decomposition levels and
conducted a unit root and normality tests to compare the performance of each alterna-
tive. Finally, it is concluded that decomposing with DB16 performed best. Therefore,
it is reasonable to employ DB16 in the case of this chapter.
In the wavelet framework, the decomposition process is executed by a recursive
filter-decimation operation. Figure 2 illustrates the full WT and WPT of a time-
domain signal f (t) up to the 3rd level of the decomposition. It can be observed that
the WT contains one high-frequency term from each level and one low-frequency
residual from the last level of the decomposition, while the WPT is composed of
complete decomposition at every level to make it achieve a higher resolution in the
high-frequency region. As a more detailed description of the WPT decomposition,
the recursive relations between the jth and the j + 1th level components are expressed
in Eqs. (4) and (5), respectively.
4 A Hybrid Time Series Prediction Approach 77
f31 (t)
1
f2 (t)
f11 (t) f32 (t)
2
f (t) f2 (t)
2
f1 (t)
(a)
H
H f31 (t)
1
f2 (t)
H f32 (t)
1
f1 (t) G
H
f33 (t)
f22 (t)
G f34 (t)
G
f (t) H
H f35 (t)
3
f2 (t)
f36 (t)
2 G
f1 (t) H
G f37 (t)
4
f2 (t)
G f38 (t)
G
(b)
2i−1
f j+1 (t) = H f ji (t)
(4)
2i
f j+1 (t) = G f ji (t)
⎧ ∞
⎪
⎪
⎪
⎪ H {.} = h(k − 2t)
⎪
⎨
k=−∞
(5)
⎪
⎪
∞
⎪
⎪ = g(k − 2t)
⎪
⎩
G{.}
k=−∞
where H and G are filtering-decimation operators and are related to the discrete
filters h(k) and g(k), respectively. h(k) and g(k) are the quadrature mirror filter (QMF)
associated with the predefined scaling function and the mother wavelet function. See
more details to Gao and Yan [6].
From Fig. 2, it is clear that the calculation in WPT contains more steps than WT.
Although WPT could probably suffer from a higher computational complexity, it is
not an issue for the current computing system. Especially in the construction industry,
the amount of the monitored data is relatively limited in real construction practices,
and engineers have more concerns about the accuracy of prediction results, instead
of the calculation complexity. Each component in the WPT tree can be viewed as
78 Time Series Prediction
the output of a filter that is tuned to a particular basis function. For this reason, the
whole tree can be taken as a filter bank. It can be seen from Fig. 2 that a three-level
resolution of the original signal will produce eight time series. After j levels of the
WPT decomposition, the original signal f (t) can be presented by a set of time series
f ij (t), (i = 1, 2, …, 2j ). The wavelet packet component signal f ij (t) can further be
expressed by a linear combination of wavelet packet functions ψ ij,k (t) in Eq. (6).
⎧ ∞
⎪
⎪
⎪ f i
(t) = cij,k ψ ij,k (t)
⎨ j
k=−∞ (6)
⎪
∞
⎪
⎪
⎩ cij,k = f (t)ψ ij,k (t)dt
−∞
where cij,k is the wavelet packet coefficient of the ith wavelet packet component
signal f ij (t), (i = 1, 2, …, 2j ). The wavelet packet functions are orthogonal, which
satisfies ψ mj,k (t)ψ nj,k (t) = 0 if m = n. See more details to Wu and Liu [27].
In the last stage, the original data is decomposed into several time series, and then
prediction models can be built to estimate the target time series within high- and low-
frequency regions. For this purpose, the support vector machine (SVM), a popular
machine learning tool relying on statistical learning theory, is adopted to solve multi-
dimensional functions. The least-squares version of SVMs is called LSSVM, which
is a set of supervised learning methods to analyze data and detect patterns for classi-
fication and regression analysis. Let a given training set of LSSVM be formulated as
N data points {xk , yk }k=1
N
with input data x k ∈ RN and output yk ∈ r, where RN is the
N-dimensional vector space and r is the one-dimensional vector space. As given in
Eq. (7), the LSSVM approach formulates the regression as an optimization problem
in the primal weight space.
⎧
n
⎨ Minimi ze : 1 ωT ω + γ
⎪
e2
2 2 i=1 i (7)
⎪
⎩
Subject to : y(x) = ωT ϕ(xk ) + b + ek , k = 1, 2, . . . , N
where ek is the error variable, ω is the weight vector, and γ is the regularization
parameter.
In general, it is not a straightforward task to directly handle the weight vector ω
due to its high dimensionality. In order to simplify the computation, a dual formation
is constructed and a kernel function is adopted. Then, the solution of the regression
problem can be obtained according to Eq. (8).
4 A Hybrid Time Series Prediction Approach 79
N
y(x) = ak K (x, xk ) + b (8)
k=1
where K is the kernel function satisfying Mercer’s condition, x k is the training data,
and x represents the new input data. See more details to Xu [28].
Different from Lagrange multipliers, the kernel and its specific parameters
together with regularization parameter γ in LSSVM cannot be set from the opti-
mization problem. To determine and fine-tune these model parameters, some candi-
date methods, like Vapnik–Chervonenkis bounds, cross-validation, independent opti-
mization set, or Bayesian learning, can be carried out. In this chapter, the widely-used
radial basis function (RBF) is chosen as a kernel function, which has been formulated
in Eq. (9). To search optimized values of the two parameters including the regular-
ization parameter γ and the RBF kernel function parameter σ, genetic algorithm
(GA) is utilized as a search heuristic that mimics the process of natural selection in
the field of AI.
(x − xk )(x − xk )T
K (x, xk ) = exp − , k = 1, 2, . . . , N (9)
2σ 2
where σ is the width of RBF as a kernel function, and K(x, x k ) is the kernel function.
As shown in Fig. 1, the prediction task is performed separately in regard to each
specific decomposed sequence. Given the length of the decomposed sequence in the
last stage is T, the ith decomposed sequence can be defined as a vector {f ij (1), f ij
(2), …, f ij (T )} (i = 1, 2, …, 2j ). For prediction and evaluation purposes, the whole
sequence can be divided into two clusters, where one is for training and the other is
for testing. To deal with the limited observed data in the construction industry, the
rolling forecast technique is applied for the time series analysis in this chapter.
In each sample, the length of the input and output is defined to be N and 1, respec-
tively, which means that the output is decided based on the past N observations in a
time-series perspective. In this case of estimating tunnel-induced ground settlement,
the value of N is defined as 5. In other words, it is assumed that the current ground
settlement could be closely associated with the observed settlement information from
the past five days (Note: the tunnel-induced ground settlement data is monitored and
recorded on a daily basis). In other words, when the size of the training set is L, the
size of the testing set will be T-L-N. The samples in the training set and the testing
set have been given in Tables 2 and 3, respectively. The optimized values of γ and σ
are set by the GA-based search method in the training set. Subsequently, predictions
defined by f ij * (N + l), (l = 1,2, …, T − N), can be obtained, as shown in Tables 1
and 2.
80 Time Series Prediction
… … … …
L f ji (L), f ji (L + 1), . . . , f ji (L + N − 1) f ji (L + N ) f ji∗ (L + N )
… … … …
T-L-N f ji (T − N ), f ji (T − N + 1), . . . , f ji (T − 1) f ji (T ) f ji∗ (T )
This stage intends to obtain the predicted tunnel-induced ground settlement over time
through reconstructing the predicted sequences. In both the training set and testing
set, the predicted tunnel-induced ground settlement at the ith decomposed sequence
can be gained. Based on Eq. (10), all the decomposed sequences can be reconstructed
by following a level-by-level way, which is similar to the decomposition process in
the first stage [27]. Figure 3 shows an example of a three-level WPT reconstruction
for a predicted time-domain signal.
⎧ i∗
⎪
⎪ f (t) = f j+1
2i−1∗
(t) + f j+1
2i∗
(t)
⎨ j
, t = 1, 2, . . . , T
j
2 (10)
⎪ ∗
⎩ f (t) =
⎪ f ji∗ (t)
i=1
where f ji∗ (t) is the predicted ith decomposed sequence at the jth level, and f ∗ (t) is
the predicted sequence, with all the decomposed sequence reconstructed.
4 A Hybrid Time Series Prediction Approach 81
f31* (t)
f21* (t)
2*
f3 (t)
f11* (t)
3*
f3 (t)
f22* (t)
4*
f3 (t)
f * (t)
5*
f3 (t)
f23* (t)
6*
f3 (t)
f12* (t)
f37* (t)
f24* (t)
8*
f3 (t)
When the predicted results are obtained, prediction performance needs to be numer-
ically evaluated to reveal the accuracy of the developed prediction model, which can
rely on several standard evaluation measures presented in the literature [21]. This
chapter refers to two indicators, namely Mean Absolute Error (MAE) and Root Mean
Square Error (RMSE), to illustrate the correspondence between individual pairs of
model predictions and observations. To be more specific, MAE describes the average
magnitude of the error between the predicted and observed distributions. When the
model is performing well, the value of MAE will arrive at zero. Its value larger than
zero indicates that the predicted value is more or less than the observed value. As
for RMSE, it measures the standard deviation between the predicted and observed
distributions. RMSE serves to aggregate the magnitudes of the errors in predictions
for various times into a single measure of predictive power. Similarly, RMSE is zero
when the model precisely predicts the observed value. For its value greater than zero,
it means that the model predictions have certain deviations from observations.
In general, MAE reflects the mean prediction accuracy, while RMSE quantifies
the dispersion degree of the prediction performance. According to MAE and RMSE
provided in Eqs. (11) and (12), it is expected to effectively assess the prediction about
the time series of the tunnel-induced settlement.
1
T
M AE = f (t) − f ∗ (t)∗
× 100%, t = 1, 2, . . . , T (11)
T t=1
1 T
RMSE = ( f (t) − f ∗ (t)∗ )2 , t = 1, 2, . . . , T (12)
T t=1
82 Time Series Prediction
where f (t) is the observed sequence of the tunnel-induced settlement, and f ∗ (t) is
the predicted sequence, with all the decomposed sequences reconstructed.
Based on predicted results, decision analysis can be performed to make better
decisions in terms of various aspects, such as techniques for structuring a decision,
uncertainty measurement, alternative assessment, option evaluation for arriving at
the course of actions. That is to say, effective analysis and management in decisions
contribute to improving project safety and performance. In this case, the tunnel-
induced ground settlement is accurately predicted on a daily basis, and then relevant
control measures can be proposed and conducted in time to deal with the risk situation
that the settlement is going to exceed the warning standard. Besides, the safety
warning standard herein is set as 2 mm/d regarding the daily ground settlement
according to the industry standard, “Technical code for monitoring measurement of
subway engineering (DB11/490-2007)”. Meanwhile, the previous stages can receive
feedback constantly to guide the model modification and updating.
Fig. 5 Evolution of the measured daily ground settlements from 13th April to 11th June in 2015
at DK26460 and DK26580 during the H-H tunnel construction
84 Time Series Prediction
In this case, the measured daily ground settlements at DK26460 and DK26580 from
13th April to 11th June (see Fig. 5) are taken as two original time-domain signals to
implement the detailed computation process in the developed approach. According
to the WPT process, the DB16 wavelet is used to decompose each original signal f
(t) (t = 1, 2, …, 60) into two levels with four separate sequences f 12 (t), f 22 (t), f 32
(t) and f 42 (t) (t = 1, 2, …, 60). As a result, the decomposition sequences of original
daily ground settlements at DK26460 and DK26580 have been displayed in Fig. 6.
Regarding each sequence f i2 (t) (i = 1, 2, 3, 4; t = 1, 2, …, 60), the length of the input
(that is N as mentioned in Table 2) and output of each sample is set to be 5 and 1,
respectively, indicating that each sample is made up of 6 data points. Besides, when
the size of the training set (that is L as mentioned in Table 2) is defined to be 35, the
size of the testing set can be calculated to be T-L-N = 60-35-5 = 20. From the rolling
forecast technique, the obtained data at the first 40 days (from 13th April to 22th
May) in each time sequence constitute 35 samples which serve as the training set,
while the obtained data at the last 25 days (from 18th May to 11th June) constitute
20 samples which are used as the testing set. Then, the sequence in both the training
set and the testing set can be predicted based on LSSVM. Figure 7 illustrates the four
separate sequences predicted at DK26460 and DK26580, respectively, which can be
reconstructed into a comprehensive sequence according to Eq. (10).
Subsequently, it is necessary to evaluate the prediction performance of the devel-
oped WPT-LSSVM approach by analyzing the correspondence between individual
pairs of model predictions and observations. The scatterplot in Fig. 8 shows the obser-
vations and WPT-LSSVM enabled predictions for the tunnel-induced daily ground
settlement at DK26460 and DK26580, respectively. It can be seen that the observed
daily ground settlement is in general low when the predicted value is low. Similarly,
the observed daily ground settlement is high when the predicted value is high at
both DK26460 and DK26580. To better visualize the average agreement between
actual observations and model predictions, regression lines are fitted in both training
data and testing data. Figure 8 provides the fitted regression lines along with rele-
vant parameters. Special focus needs to be paid on the testing data, where the fitted
regression lines at DK26460 and DK26580 have a gradient of 0.9507 and 0.9630,
respectively. Since the gradients are both very close to 1, it validates the effectiveness
of the WPT-LSSVM model in accurately predicting the average magnitude of the
daily ground settlement.
For a more comprehensive assessment, the hybrid WPT-LSSVM model is
compared to the traditional LSSVM. In the same way, the originally measured ground
settlement sequence is divided into a training set with a size of 35 and a testing set
with a size of 20. LSSVM is applied to directly predict the ground settlement on a
daily basis with no procedure of decomposition and reconstruction. Also, the scatter-
plot is drawn in Fig. 9 to display observations and LSSVM enabled predictions for the
tunnel-induced daily ground settlement at DK26460 and DK26580, respectively. In
terms of the fitting results in the testing data from Fig. 9, the fitted regression lines at
5 A Realistic Tunnel Case 85
f21 (t)
f22 (t)
f23 (t)
f24 (t)
12/04 17/04 22/04 27/04 02/05 07/05 12/05 17/05 22/05 27/05 01/06 06/06 11/06
Date
(a)
f21 (t)
f22 (t)
f23 (t)
f24 (t)
12/04 17/04 22/04 27/04 02/05 07/05 12/05 17/05 22/05 27/05 01/06 06/06 11/06
Date
(b)
Fig. 6 Decomposition sequences of original daily ground settlements from 13th April to 11th June:
a DK26460; b DK26580
DK26460 and DK26580 have a gradient of 0.5248 and 0.5077, respectively, which are
significantly lower than that with the developed WPT-LSSVM approach (see Fig. 8).
In other words, the developed WPT-LSSVM approach in this chapter can return
predictions in significantly higher accuracy than the traditional LSSVM approach.
To sum up, the prediction results strongly advocate the possibility of hybridizing the
86 Time Series Prediction
f21 (t)
f22 (t)
f23 (t)
f24 (t)
12/04 17/04 22/04 27/04 02/05 07/05 12/05 17/05 22/05 27/05 01/06 06/06 11/06
Date
(a)
f21 (t)
f22 (t)
f23 (t)
f24 (t)
12/04 17/04 22/04 27/04 02/05 07/05 12/05 17/05 22/05 27/05 01/06 06/06 11/06
Date
(b)
Fig. 7 Predicted sequences of daily ground settlements using LSSVM: a DK26460; b DK26580
WPT and LSSVM algorithm to attain more preciseness and reliability in estimating
the tunnel-induced ground settlement.
5 A Realistic Tunnel Case 87
3 3
1 1
Predictions
Predictions
0 0
Training data Training data
Slope=0.9761 Slope=0.9824
-1 Intercept=-0.0020 -1 Intercept=0.0087
R2=0.9867 R2=0.9881
Fig. 8 Scatterplot of observations and WPT-LSSVM enabled predictions for tunnel-induced daily
ground settlement in both training data and testing data: a DK26460; b DK26580
3 3
1 1
Predictions
Predictions
0 0
-2
Testing data -2
Testing data
Slope=0.5248 Slope=0.5077
Intercept=-0.0738 Intercept=0.1742
R2=0.5158 R2=0.4841
-3 -3
-3 -2 -1 0 1 2 3 -3 -2 -1 0 1 2 3
Observations Observations
(a) (b)
Fig. 9 Scatterplot of observations and LSSVM enabled predictions for tunnel-induced daily ground
settlement in both training data and testing data: a DK26460; b DK26580
To further test the accuracy of the developed prediction approach, some traditional
prediction approaches are selected to perform the prediction for the same data sets.
Comparisons among different methods are made in terms of their prediction perfor-
mance, where the measured ground settlement provides a benchmark. Since one
typical class of intelligent methods named ANN has been widely applied to analyze
88 Time Series Prediction
the tunnel-induced ground settlement in the last decade, two representatives of the
ANN-based approach, including Back Propagation (BP) ANN and Radial Basis
Function (RBF) ANN, are chosen for comparison purposes. As a result, Fig. 10
visualizes the Comparisons between the actual measured values and the predicted
12/04
(a)
12/04
(b)
Fig. 10 Comparisons between the actual measured values and the predicted values using different
methods concerning the tunnel-induced daily ground settlement: a DK26460; b DK26580
5 A Realistic Tunnel Case 89
values using different methods at DK26460 and DK26580, and Table 4 measures the
prediction performance. Those results are further analyzed and discussed as follows.
(1) From a perspective of the evolution of the ground settlement, all the prediction
methods, including BP, RBF, LSSVM, and WPT-LSSVM, are competent to
capture the rough trend of the tunnel-induced ground settlement evolution in the
real world. For DK26460 as shown in Fig. 10, in the testing set, the maximum
positive and negative daily ground settlement is predicted to appear on 3rd June
and 25th May, respectively. For DK26580 in the testing set, the positive and
negative maximum daily ground settlement is predicted to appear on almost
11th June and 3rd June, respectively. These predictions are quite consistent
with the actual observations. Either the maximum positive or negative ground
settlement at DK26460 appears earlier than that at DK26580. That is because
DK26460 is at the front of DK26580 along the H-H tunnel section, and the
TBM first passes through the cross-section at DK26460, and then DK26580
in a sequential order. At DK26460, it is found that the maximum positive and
negative settlement values almost exceed the safety warning line, and thus some
corresponding control measures are necessarily carried out in time to mitigate
the risk. Accordingly, the magnitude of tunnel-induced ground settlement at
DK26460 is expected to quickly go back to the normal level.
(2) From a comparison between ANN-based and SVM-based approaches, the
prediction performance of the SVM-based approach is proven to be greater
than the ANN-based approach in this case. For DK26460 and DK26580,
Table 4 reveals that the average values of MAE by using BP and RBF are
0.3593 and 0.3655, respectively, while the average values of MAE by using
LSSVM and WPT-LSSVM are 0.3000 and 0.1006, respectively. The average
values of RMSE by using BP and RBF are 0.4881 and 0.4687, respectively,
while the average values of RMSE by using LSSVM and WPT-LSSVM are
0.4179 and 0.1538, respectively. Clearly, the SVM-based approach results in
a lower value of MAE and RMSE, indicating that the SVM-based approach is
prone to generate higher accuracy in the prediction task than the ANN-based
approach. Besides, these two popular AI approaches (ANN and SVM) are
significantly different in theory and applications. More specifically, the devel-
opment of ANNs follows a heuristic path with applications and extensive exper-
imentation preceding theory, while the development of SVMs involves sound
theory, implementation, and experiments. ANNs can suffer from multiple local
minima, while SVMs training always finds a global minimum. A possible
90 Time Series Prediction
reason for the outstanding performance of SVM over ANNs in practice is that
SVMs can tackle the biggest problem with ANNs, since ANNs use empirical
risk minimization, while SVMs use structural risk minimization. SVMs utilize
kernels to represent nonlinear relationships by a linear model, which can be
less likely to be overfitting [26]. Besides, two more advantages of SVMs can be
highlighted as the simple geometric interpretation and the flexibility to provide
sparse solutions [18]. In short, SVMs are suitable to reach relatively high clas-
sification accuracy even in the case with high-dimensional data and a small
set of training patterns. Therefore, this method can assist in the time series
analysis of tunnel-induced ground settlement, where the size of the training set
is always small, particularly for one specific tunnel section.
(3) From a comparison between LSSVM and WPT-LSSVM methods, the predic-
tion accuracy of the WPT-LSSVM method exceeds the LSSVM method. For
DK26460 and DK26580 listed in Table 4, the average values of MAE and RMSE
by using LSSVM are 0.3000 and 0.4179, respectively, while the average values
of MAE and RMSE by using WPT-LSSVM are 0.1006 and 0.1538, respectively.
Due to the smaller value of MAE and RMSE from the WPT-LSSVM method,
it can be concluded that the WPT-LSSVM method has a better estimation
performance with higher accuracy than the LSSVM method. Therefore, the
improvement of prediction accuracy is likely to root in the WPT process. As
for the original time sequence of the observed daily ground settlement in Fig. 5,
it is impossible to capture any patterns during this time-dependent settlement
process, since there are some unavoidable measurement errors. The measure-
ment error always presents a high frequency in a similar way to noise, while the
actual settlement presents a low frequency. Accordingly, we can then separate
the monitored data into the high- and low-frequency parts with the support of
WPT, resulting in four separate sequences f 12 (t), f 22 (t), f 32 (t), and f 42 (t) (t =
1, 2, …, 60) that come from the original signal f (t) (t = 1, 2, …, 60). Based on
the sequence within the lowest frequency (see f 12 (t) (t = 1, 2, …, 60) in Fig. 6,
we can capture a key pattern that the ground surface experiences an upheaval
first and then is followed by continuous settlements. This is due to the fact of
the improvement effect on the surrounding soil because of plastic flows as the
EPB TBM progresses [23]. In conclusion, the proposed WPT-LSSVM model
possesses the following two advantages in the task of predicting tunnel-induced
ground settlement based on the past ground settlement. For one thing, WPT-
LSSVM has a high capability in multi-resolution decomposition, which can
offer better localization properties in both spatial and frequency domains and
provide flexibility for the signal representation to achieve better classification
accuracy. For another, it can effectively avoid disturbance between strong and
weak signals during the training and learning of prediction models, and thus
the accuracy and reliability of prediction results can be further enhanced.
6 Summaries 91
6 Summaries
In the end, the novelty of this chapter can be summed up from two aspects: (a) the
state of knowledge by proposing a hybrid method that integrates signal processing
method named WPT and a popular machine learning algorithm called LSSVM,
aiming to significantly improve the prediction accuracy and reliability; (b) the state of
practice by precisely estimating the tunnel-induced ground settlement as the construc-
tion project progresses toward completion, which can help to dynamically detect the
possible risk and sound a warning ahead of time.
References
19. Samui P, Sitharam T (2008) Least-square support vector machine applied to settlement of
shallow foundations on cohesionless soils. Int J Numer Anal Meth Geomech 32:2033–2043
20. Santos OJ, Celestino TB (2008) Artificial neural networks analysis of Sao Paulo subway tunnel
settlement data. Tunn Undergr Space Technol 23:481–491
21. Sun Y, Leng B, Guan W (2015) A novel wavelet-SVM short-time passenger flow prediction in
Beijing subway system. Neurocomputing 166:109–121
22. Sun Z, Chang C (2004) Statistical wavelet-based method for structural health monitoring. J
Struct Eng 130:1055–1062
23. Suwansawat S, Einstein HH (2006) Artificial neural networks for predicting the maximum
surface settlement caused by EPB shield tunneling. Tunn Undergr Space Technol 21:133–150
24. Suykens JA, Vandewalle J (1999) Least squares support vector machine classifiers. Neural
Process Lett 9:293–300
25. Verruijt A, Booker J (1996) Surface settlements due to deformation of a tunnel in an elastic
half plane. Géotechnique, pp 753–756
26. Wang F, Gou B, Qin Y (2013) Modeling tunneling-induced ground surface settlement
development using a wavelet smooth relevance vector machine. Comput Geotech 54:125–132
27. Wu J-D, Liu C-H (2009) An expert system for fault diagnosis in internal combustion engines
using wavelet packet transform and neural network. Expert Syst Appl 36:4278–4286
28. Xu Q (2013) Impact detection and location for a plate structure using least squares support
vector machines. Structural Health Monitoring, 1475921713495083
29. Zhang L, Wu X, Ji W, AbouRizk SM (2017) Intelligent approach to estimation of tunnel-
induced ground settlement using wavelet packet and support vector machines. J Comput Civil
Eng 31:04016053
30. Zhang L, Wu X, Skibniewski MJ, Fang W, Deng Q (2015) Conservation of historical buildings
in tunneling environments: Case study of Wuhan metro construction in China. Constr Build
Mater 82:310–322
Information Fusion
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 95
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_5
96 Information Fusion
outlined as follows [1]: (1) SVM has excellent prediction and generalization capabil-
ities; (2) SVM is global and unique, and thus, it can reach a global minimum to tackle
the overfitting problem; (3) SVM does not need too much concern about the dimen-
sionality of the input space; (4) Although SVM is a black-box model, the post-hoc
explainability techniques can be adopted to describe its inner structure mathemat-
ically. This research, therefore, intends to implement an SVM-based classification
for safety risk assessment using multi-sensory data.
When several SVM-based classifiers are deployed to review a set of data gathered
by sensors and predict the structural damage at different locations, it remains an
open question to perform a decision-level information fusion to realize an overall
risk assessment in the whole structure system. Specifically, data recorded by sensors
could be incorrect and incomplete with varying degrees of uncertainties, and thus,
evaluation based on a single SVM is hard to meet the demand of high prediction accu-
racy. To handle this issue, a typical data fusion tool named the Dempster-Shafer (D-S)
evidence theory is taken into account. It relies on certain combination rules to aggre-
gate all available evidence and model uncertainty, imprecision, and incompletion,
resulting in a degree of belief.
The D-S evidence analysis has been successfully applied to fault diagnosis in
complex systems. For instance, Bao et al. [2] carried out the D-S evidence anal-
ysis on a 20-bay rigid truss structure for structural damage detection, confirming
that the combination of damage detection results from different individual data sets
was better than individual results from each test data separately. Zhang et al. [28]
combined the D-S evidence theory with the cloud model and Monte Carlo simulation
to perceive the safety risk of buildings adjacent to the tunneling excavation, indicating
that the developed information fusion approach was able to synthesize multi-source
information for higher prediction accuracy and a more optimal decision making. Pan
et al. [16] proposed a hybrid risk analysis method based on interval-valued fuzzy
sets, the D-S evidence theory, and fuzzy Bayesian networks to fuse multiple expert
judgments in linguistic expressions, in order to efficiently evaluate the risk level in a
complex project environment under uncertainty. Additionally, it is indeed useful for
classification problems by treating individual classifiers differently according to the
evidence theory. It is noticeable that the Basic Probability Assignment (BPA) has
a great effect on the accuracy of the fusion results. However, there is no universal
standard for BPA calculation in fuzzy and complex conditions, which needs further
attention [28].
In this chapter, a hybrid approach that integrates SVM and the D-S evidence theory
is developed to facilitate risk analysis using multi-sensory data with uncertainties,
conflicts, and even bias. That is to say, the SVM classifier is built to detect the risk level
of a certain structure part based on a great deal of sensor data involving uncertainty
and incorrectness, and then the D-S evidence theory is utilized to combine predic-
tive results from different SVM models for achieving an overall risk assessment.
The main questions are: (i) How to train multi-class SVM with the most suitable
kernel functions and parameters? (ii) How to construct BPAs from the observed
evidence to fuse information from multi-classifiers at the decision level, leading to
1 Introduction 97
more accurate prediction results? (iii) How to verify the effectiveness of the devel-
oped approach under uncertainty? (iv) How to measure the global sensitivity of risk
factors contributing to system safety?
The proposed risk perception approach based on the hybrid information fusion
method is validated in a realistic operational tunnel case in China, which can reliably
evaluate the magnitude of the structural health risk. Results indicate the developed
SVM-DS approach is capable of fusing multi-classifier information effectively from
different SVM models and exhibiting a more outstanding classification performance
than the single SVM model under a high bias. Since the proposed reliable risk analysis
method can efficiently fuse multi-sensory information with ubiquitous uncertainties,
conflicts, and bias, it provides a comprehensive assessment for the structural health
status together with the most critical risk factors, and thus, proper remedial actions
can be taken at an early stage to mitigate the structural failure.
With the acceleration of urbanization, more and more underground facilities have
been built for the greater use of underground space, which can relieve the stress of
land scarcity and cope with the increased demand for residence and transportations
in cities. As a major part of underground facilities to support human activities, the
tunnel system is experiencing a quick and mass extension, which brings a lot of
benefits, such as to alleviate both traffic congestions and environmental pollution
problems, to avoid the acquisition of costly land or property, and others [13]. For
example, metro networks in Singapore are planned to be doubled from the current
180–360 km by 2030. China is going to invest more than 1 trillion dollars to construct
90 metro lines in 44 cities with a total mileage of 4448 km from 2018 to 2030 [11].
It should be noted that structure diseases, like cracks, water leakage, settlement, and
others, are common to tunnels in the complex underground environment.
Unlike the surface infrastructures which have received enough regulatory scrutiny,
the prior knowledge about the underground geological conditions and environmental
parameters of underground structures is extremely limited, raising difficulties in the
early risk detection and perception. Since any kind of structural disease will incur
cascading structural failures in the whole tunnel system and eventually threaten the
normal operation of the tunnel, they should be evaluated carefully and timely. In this
regard, a network of smart sensors has gained traction in SHA. They are installed
in various structures of civil infrastructure to continuously collect data about the
structure and its environments, such as temperature, stress, delamination, strain,
vibration, and humidity [18], which can increase the diversity of the data resources.
Subsequently, these large amounts of information from several sources need to be
fused properly for forming a unified picture regarding the health status of the targeted
infrastructure, which can be defined as the multi-sensor data fusion. Hence, it raises
a critical issue that how to assess the structural health accurately and reliably using
98 Information Fusion
B1: Material deterioration B2: Seepage damage B3: Segment deformation B4: Exterior load risk
data from multiple sensors, in order to offer a prospective insight into the structural
health status of the underground facilities.
In general, the structural health status of operational tunnels can be divided into
three levels, which are I (Safe), II (Low risk), and III (High risk). It means that when a
structure is evaluated as a level III, it tends to take a very high risk of structural failure.
Working under the complicated underground environment, the structural health of
operational tunnels is impacted by a number of factors. Commonly, risk factors,
which help in perceiving the overall safety risk of the monitoring sections in the
operational tunnel (T), are determined by actual probe data, domain expert judg-
ment, accumulated engineering, and Chinese technical specifications and manuals.
For a clearer framework, these factors can be grouped into four major structural
risk categories, namely material deterioration (B1), seepage damage (B2), segment
deformation (B3), and exterior load risk (B4). Figure 1 presents an SHA framework
consisting of 16 risk factors extracted from sensor data. A detailed description of
those factors is presented as follows.
(1) Material deterioration (B1): Cracks and peeling-off areas in the structure are
the two most visible forms of material deterioration on the structural surface.
They are prone to reduce structure durability and even increase the likelihood
of structural failure. Thus, the information about cracks and peeling-off areas
plays an important role in evaluating concrete structural performance. Relevant
factors, including the measured crack length and width (C1), crack density
(C2), peeling-off area (C3), and peeling-off depth (C4), are used to characterize
the degree of material deterioration precisely.
(2) Seepage damage (B2): The great amount of underground water can unavoid-
ably inflow and leak through the tunnel weak points, like cracks, joints, and
others, leading to structural erosion. In other words, seepage water has a long-
term negative influence on the structure. However, seepage hazards are easily
overlooked by engineers, since they are too tiny to be detected at the initial
stage. Basically, the monitored records, including PH deviation value of the
neutral (C5), leakage point number per 100 ml (C6), leakage area per point
(C7), and leakage water amount (C8) are considered to estimate the severity
of seepage damage, More specifically, C5 shows the chemical property of the
underground water. A poor PH value will inevitably accelerate rock damage,
2 Structural Health Assessment (SHA) in Tunnels 99
and then the saturated rock will give rise to structural diseases. C6 describes
the leakage distribution, and C7, C8 reflect the leakage volume.
(3) Segment deformation (B3): The structural health also depends on the quality
of joints. It is known that water leakage is more likely to occur in too large
joints, and thus, both the width of the transverse joint (C9) and longitudinal
joint (C10) should be concerned. Besides, the transverse (C11) and longitudinal
slab staggering (C12) are important, since the proper slab staggering can serve
to avoid the concentration of load in the concrete section.
(4) Exterior load risk (B4): The stability of the tunnel structure is very sensitive to
the exterior load. Notably, unbalanced pressure applied from the exterior load
tends to trigger considerable deformation and displacement in tunnel founda-
tion, valut, and reinforcement. The occurrence of structural failures can root
in the exterior load, which exceeds the allowable limit value. Generally, four
factors are refined out from the sensor data to measure the magnitude of the
exterior load risk, which are differential settlement (C13), cumulative settle-
ment (C14), water-earth pressure over valut (C15), and a pressure difference
of reinforcement inside and outside (C16).
3 Information Fusion
In general, there are three levels of information fusion, namely the data level, feature
level, and decision level. At the data level, the information fusion process directly
utilizes raw data in form of signal refinement as inputs, such as structure vibration
velocity and acceleration. It can not only offer more accurate data than the individual
sources, but also provide a preliminary understanding of the observed situation. At
the feature level, characteristics of features are retrieved from the raw data, which are
then fused into a single concatenated feature vector. These extracted features can be
fed into machine learning models for further analysis. At the decision level, decisions
from different features are integrated under particular combination rules, resulting
in a more robust and accurate evaluation along with more specific inferences about
the structural health status.
To incorporate different kinds of risk information from multi-sensors into the
risk evaluation process, numerous soft computing methods using mathematical
models have been developed, such as the probabilistic approach [27], fuzzy logic
[7], interval theory [15] and analytic hierarchy process (AHP) methods [12]. For
example, Chileshe et al. [3] established the degree of risk by evaluating the likeli-
hood of risk occurrence in construction projects. Wang et al. [23] adopted the fuzzy
synthetic evaluation method to assess the structural health status of the tunnel, using
data from different types of sensors, where the predicted results were consistent
with the actual operational situation. Hyun et al. [10] performed a risk analysis by
adopting FTA and AHP, reaching a 90% predictability. Notably, the prerequisite of
these above-mentioned risk analysis methods is the high-quality data, but it is impos-
sible for actual projects to always implement precise data acquisition without errors
100 Information Fusion
and missing values. In other words, those traditional soft computing methods are in
difficulties to address important aspects of randomness, noisy, incompleteness, and
uncertainties stemming from the complex dynamic circumstance, human factors, and
information conflicts [28, 29]. Unfortunately, these adverse factors tend to generate
remarkable deviations in decision-making.
Various machine learning algorithms have been developed to learn large data
from different sources for classification, regression, and clustering. It is worth noting
that artificial neural network (ANN) and support vector machine (SVM) are two
promising supervised learning algorithms in the machine learning domain, which
are proven useful in classification tasks [24]. To be specific, ANN can mimic human
learning processes by establishing relationships between process input and output,
which is inspired by the biological neural network with interconnected neurons [14].
However, it still presents some limitations, such as a black box, long computing time,
overfitting, dimensional disaster, and local minimum [9]. As a method in parallel with
ANN, SVM performs classification by constructing optimal hyperplanes in higher-
dimensional space to search for a global solution [22]. According to comparisons
between SVM and ANN, SVM can exhibit greater performance than ANN in some
cases. For example, Farooq et al. [6] made a comparative study to perform a structural
evaluation of damage detection and identification using a set of static strain data with
Gaussian noise by SVM and ANN, respectively. It was found that the prediction
accuracy of SVM was better than ANN under noise contamination. Dushyanth et al.
[5] concluded that SVM stood out due to a relatively small number of training data
and a dramatic reduction in computational time. Some variations of SVM have been
developed to fuse multi-sensor data in multi-classification problems [30].
Due to the effectiveness of SVM and the D-S evidence theory in information
fusion, we aim to propose a hybrid information fusion method based on these two
advanced soft computing methods. Besides, we intend to implement the proposed
approach in SHA where collected data is usually in a high degree of uncertainty and
impreciseness. Currently, the single or simple structures, like a beam, steel frame, and
others, are studied, and very few studies have concentrated on SHA for underground
facilities in a complicated and changeable environment. Moreover, the information
fusion with high conflicts has not gained much consideration in applications, which
tends to make the data fusion meaningless. Thus, a novel risk analysis approach inte-
grating the probabilistic SVM and an improved D-S evidence theory is developed in
this research, in order to provide insights into a better understanding of the structural
risk analysis and assessment subjected to uncertainty and conflicts.
Based on the SVM algorithm and the D-S evidence theory, a hybrid information
fusion approach is developed for achieving an overall structural safety risk assess-
ment. Figure 2 depicts the flowchart of the proposed hybrid approach for risk analysis.
To be specific, several SVM-based classifiers are built to learn data from different
4 A Hybrid Information Fusion Approach 101
Probabilistic
Feature extraction SVM
Dempter Weighted
fusion rule mean rule
End
sensors and output posterior probabilities, and then an improved D-S evidence theory
is utilized to refine and synthesize different classification results generated from
probabilistic SVM models. There are mainly three steps incorporated, as elaborated
below.
The core idea of SVM is to create the optimal line or hyperplane to separate data into
classes as far as possible, which can tackle both the linear and non-linear problems.
For the linear problem, it is observed that data is linearly separable. Linear SVM
can make linear classification by maximizing the distance (margin) between the
102 Information Fusion
Table 1 Equations of
Kernel functions Equation
commonly used kernel
functions Linear kernel K(xi , x) = xi · x
Polynomial kernel K(xi , x) = (xi · x + 1)d
Gaussian kernel K(xi , x) = exp(−x − xi 2 /2σ2 )
Radial basis function kernel K(xi , x) = exp(−γx − xi 2 )
Sigmoid kernel K(xi , x) = tanh(αxi T x + c)
Note d is the degree of the polynomial kernel function, and γ is
the parameter of the RBF kernel function
hyperplane and the closest data point (support vector). Moreover, the kernel trick
makes SVM suitable for the non-linear problem, which relies on the appropriate
kernel function to map the original non-linear data in a low-dimensional space into a
higher dimension. Table 1 lists some commonly used kernel functions of SVM, which
define the inner product in the transformed space to avoid the expensive calculations
in the new dimensions [19]. Equation (1) gives the nonlinear decision function of the
hyperplane. It is known that the type of kernel function and its related parameters
greatly affect the classification accuracy of the SVM model. For example, large kernel
parameters are prone to cause overfitting. The penalty parameter C of the error term
controls the tolerance of the systematic outliers. A small value of C tends to ignore
some data near the decision boundary, while a large value of C can minimize the
training error as much as possible. However, if the value of C is too large, it is likely
to lose some generalization properties of the classifier and take a long calculation
period.
N
f (x) = sign αi yi K(xi , x) + b (1)
i=1
where αi is Lagrange multiplier, N is the size of the training data set, b is a threshold
parameter based on the training set, and K(xi , x) is the kernel function.
Regarding the linear SVM, it has shown many strengths, like the great generaliza-
tion capability, clear margins of separation between classes, effectiveness in handling
high-dimensional datasets. However, the model can be very sensitive to uncertain-
ties, such as noise, error-prone sensor data, and others, and thus, it is inefficient in
datasets with a high level of noise. Additionally, it does not take the confidence of the
prediction into account. To address the limitations of the linear SVM, a probabilistic
SVM is developed to give the classification result in a form of the posterior proba-
bility. Platt [21] put forward a method to map the output of SVM into the interval
[0, 1]. According to Eq. (2), a posterior probability can be estimated using a sigmoid
function.
1
P(y = 1|x) ≈ PAS BS (f(x)) = (2)
1+ e(AS f(x)+BS )
4 A Hybrid Information Fusion Approach 103
where x is the input data, and f(x) is the output of the standard SVM from Eq. (1). AS
and BS are parameters calculated by minimizing the negative log-likelihood function
as follows:
min − t(xi )log(P(xi )) + (1 − t(xi ))log(1 − P(xi )) (3)
i
N
+ +1
y
N+ +2 i
= +1
ti = 1
y
N− +2 i
= −1
i = 1, 2, . . . , l (4)
where P(xi ) = P(y = 1|x), N+ is the number of positive data (yi = +1), and N− is
the number of negative data (yi = −1).
A data fusion method is required to deal with multiple classification results that
are generated from different SVM models. It should be noted that the D-S evidence
theory specializes in tackling uncertainty and fusing multi-source information, which
quantifies the degree of belief in each hypothesis [25]. A frame of discernment is
defined as = {A1 , A2 , . . . , AN } to be a finite nonempty set of mutually exhaustive
and exclusive hypotheses, where N refers to the number of hypotheses. A power set
2 is the set of all subsets of , which includes itself and the empty set ∅. A basic
probability assignment (BPA) is a function m: 2 → [0,1] (where 2 is the power
set of ), which satisfies the following conditions:
m(∅) = 0
(5)
A∈2 m(A) = 1
where m(A) is the basic probability assignment of A for A ⊆ . More specifically, the
belief measurement m(A) between [0, 1] denotes the ratio of all related and available
evidence to support the claim that a certain element in belongs to the set A but not
to the subset of A. The degree of the belief in the interval [0, 1] is assigned to every
possible hypothesis under given evidence.
Based on the BPA, the belief function and plausibility function can be expressed
by Eqs. (6) and (7) to measure the confidence level. Figure 3 visualizes Bel and Pl
in a more comprehensible way. That is to say, Bel refers to the least probability to
hold the proposition A, while Pl is interpreted as the maximum uncertainty value of
A. The greater the uncertainty and impreciseness of the fused information, the larger
the distance between Bel(A) and Pl(A).
104 Information Fusion
Pl ( A)
Bel(A) = m(B) (6)
B⊆A
Pl(A) = m(B) (7)
B∩A=∅
where n is the number of pieces of evidence, j is the jth piece of evidence, N is the
number of hypotheses, k is the kth hypothesis, K is termed as the conflict coefficient
in the range of [0, 1] to represent the conflict level among multiple pieces of evidence,
and m1 ⊕ m2 ⊕ · · · ⊕ mn means the orthogonal sum of evidence.
To deal with high-conflict evidence, a hybrid combination rule is proposed based
on the Dempster’s rule and the weighted mean rule, which can aggregate pieces
of evidence directly at a time rather than a stepwise method [26]. The developed
weighted mean rule is given in Eqs. (10)–(12). More specifically, Eq. (10) measures
the similarity degree in different pieces of evidence by the Euclidean distance. Equa-
tion (11) calculates the weight of the ith piece of evidence. Equation (12) provides the
modified BPA m(A) to represent the belief measure of the target A in all hypotheses.
It is assumed that K > 0.95 (1–0.05 = 0.95) indicates an unacceptable high conflict,
4 A Hybrid Information Fusion Approach 105
di−1
wi = i=n −1
(11)
i=1 di
mi∗ (Ak ) = wi mi (Ak )
∗ (12)
mi () = 1 − N k=1 mi (Ak )
where n is the number of evidence, i (or j) is the ith (or the jth) piece of evidence, N is
the number of hypotheses, and k is the kth hypothesis. Specifically, mi () quantifies
the degree of uncertainty after the evidence fusion. The smaller mi () is, the more
reliable the fusion result is.
Herein, a toy problem is taken as an example to intuitively explain the
improved D-S evidence rule. There are two pieces of evidence, namely,
m1 (A) = {m1 (A1 ), m1 (A2 ), m1 (A3 ), m1 ()} = {0.98, 0.01, 0, 0.01}, and m2 (A) =
{m2 (A1 ), m2 (A2 ), m2 (A3 ), m2 ()} = {0, 0.02, 0.96, 0.02}. Their conflict coeffi-
cient K can be derived from Eq. (13). If we only refer to the classical Dempster’s
rule in Eqs. (8) and (9), we can calculate the BPA for A1 according to Eq. (14) and
obtain the mass function for A2 , A3, and in the same way. The combination result
turns out to be m1,2 (A) = {0.653, 0.02, 0.32, 0.007}. However, K = 0.97 is larger
than 0.95, which means there is a very high conflict between the given evidence. As
a solution, the weighted mean rule in Eqs. (10)–(12) should be triggered. Since only
two pieces of evidence exist, the distances d 1 and d 2 are the same, as expressed in
Eq. (15). Also. the weights w1 and w2 for the first and second pieces of evidence
are equal, as shown in Eq. (16). Equation (17) returns the new fusion result as
m1,2 (A) = {0.49, 0.015, 0.48, 0.015}, which gives less belief to A1 and A2 .
1
m1 ⊕ m2 (A1 ) = [m1 (A1 ) · m2 (A1 ) + m1 (A1 ) · m2 () + m1 () · m2 (A1 )]
1−K
1
= (0.98 × 0 + 0.98 × 0.02 + 0.01 × 0) = 0.653
0.03
(14)
106 Information Fusion
d1 = d2 = (0.98 − 0)2 + (0.01 − 0.02)2 + (0 − 0.96)2 + (0.01 − 0.02)2 = 1.372
(15)
1.372
w1 = w2 = = 0.5 (16)
1.372 + 1.372
In this step, when all information from different SVM models is fused at the decision
level, we can reach the final safety risk degree. Equation (18) presents a compound
decision-making rule to evaluate the belief of the fusion results, where the maximum
BPA determines the risk label for a certain sample. The confidence indicator m() is
utilized to measure the credibility of fusion results. That is, the smaller the value of
m() indicates the more credible the fusion result. Based on the perceived magnitude
of the safety risk, corresponding measurements can be taken in advance to prevent
potential structural failures.
m(Aw ) = max{m(Ai )}
(18)
m() < θ
magnitude of fuzzy sets by a defuzzied value. Then, the Monte Carlo (MC) simula-
tion approach is employed to simulate a series of input data {xi1 , xi2 , . . . , xiQ } by a
certain probability distribution. After Q times of simulation iterations, the relevant
outputs {t1 , t 2 , . . . , t Q } can be obtained. Finally, the Spearman’s rank correlation
coefficient expressed in Eq. (20) is calculated, which is not dependable on distribu-
tions with a linear relationship or similar shape. Since it quantifies the contribution
of each risk feature C i to the overall risk level, the most sensitive risk factors can be
determined. Accordingly, decision-makers can give more attention to these critical
factors for lowering the risk of structural failure.
⎧
⎨ t = 0.5 ϕ (Aj )−1 + j
ϕ (Aj )+1
i=n
⎩ ϕ Aj = i=j m(Ai ) (19)
i=j
i=1 m(Ai )
Q q q
q=1 (R xi − R(xi ))(R(t ) − R(t ))
q q
GSA(Ci ) = q
q 2 Q (20)
Q q 2
q=1 (R xi − R xi ) q=1 (R(t ) − R(t ))
q
A realistic operational tunnel case in China is used to verify the effectiveness and
feasibility of the developed information approach. That is to say, multi-sensory infor-
mation with ubiquitous uncertainties, conflicts, and bias can be reasonably fused to
perceive structural risk level, leading to a better understanding of the structural health
status together with the most critical risk factors. Based on the in-depth analysis,
proper remedial actions can be taken at an early stage to reduce the potential risk in
the operational tunnel as much as possible.
The research object is an operational tunnel located at the Wuhan Metro System,
China. To be specific, the tunnel case (short for the C-H tunnel) is a double-spool
tunnel connecting the C-Station and H-Station in the Wuhan Metro Line 2. The whole
metro line begins operation in Dec. 2016. The C-H tunnel section has a length of
1174 m. The tunnel is mainly buried in the layer of muddy-silty clay with silt and
108 Information Fusion
silty sand, and some sections pass through the artesian aquifer at 2.2–3.5 m below
the ground. The layout of the C-H tunnel case is provided in Fig. 4. Due to the under-
ground environment in the nature of complexity and uncertainty, various risks from
erosion, cracks, aging, and others are inevitably involved in the operational tunnel.
Therefore, the evaluation of its structural health status is not a straightforward task,
which could be considered as a complex multi-attribute decision problem. Figure 5
shows some examples of diseases in the operational tunnel.
With the help of sensor technologies and data acquisition platforms, various
sensors can be installed in the operational tunnel to continuously monitor the physical
behavior of the structure. In this case, we divide the whole tunnel into 35 monitoring
Fig. 5 Examples of structural diseases: a water seepage; b diseases in the cross section of segments;
c cracks; d differential settlement
5 Information Fusion for SHA in Tunnels 109
sections (i.e., No. 1, No. 2, …, No. 35). Each section is 5 m in length along with
the tunnel central axis, which has several sensors with different functions to gather
a wide variety of data, including the displacement sensor, pressure sensor, and other
sensors. Subsequently, these large amounts of information from multiple sources
need to be fused properly for forming a unified picture regarding the health status
of the targeted tunnel, which can be defined as the multi-sensor data fusion. To be
more specific, four SVM models are performed directly on the features extracted
from the data collected by sensors. Classification results from SVM models need
to be further fused using the improved D-S evidence theory, aiming to achieve an
overall structural health status with high reliability and quality.
There are two major steps in SVM model training, which are described in detail
below:
(1) Choice of kernel function and parameters
The kernel function, kernel function parameters, and penalty parameter C exert a great
impact on the performance of SVM, which all contribute to finding out an optimal
hyperplane for separation in a high-dimensional feature space. Figure 6 visualizes
the scatter points about two factors C3 and C4 in the structure risk category B1 and
compares the classification results based on different kernel functions. From Fig. 6a,
b, it is observed that the linear separation is poor in classifying the two groups.
This is because there is an obvious overlapping in C3 and C4. In contrast, the RBF
kernel in Fig. 6c works well in separating points into three classes with the fewest
errors, which outperforms the linear kernel in Fig. 6a, b and the polynomial kernel
in Fig. 6d. In other words, the RBF kernel better fits the training data than others,
which is selected as the proper kernel function in this case.
Regarding the RBF-SVM model, two key parameters, namely the radius of RBF
gamma γ and the penalty parameter C, need to be fine-tuned by cross-validation.
Due to the limited input data, the sevenfold cross-validation is conducted here to
determine the best value of γ and C, aiming to establish an optimal SVM model in
pursuit of more reliable probability outputs. To be specific, the validation data set
(i.e., 35 sets of sample data (C1, C2, …, C16) from multiple sensors in 35 monitoring
sections) is randomly split into 7 mutually exclusive groups in equal size. For each
training, one of the 7 groups is taken out as a test set, while the remaining 6 groups
are put together as a training set. That is to say, all the data can be utilized for training
and testing, which can expectedly reduce the errors from bias. After 7 iterations of
training and validation, every group can be used as a test set once and as a training set
six times. In the end, the model performance is determined by the average accuracy
from the 7 iterations.
Pairs of (C, γ) in different values are tested in the RBF-SVM model and their
corresponding classification accuracy for four structural risk categories B1, B2, B3,
110 Information Fusion
C3 C3
(a) (b)
C3 C3
(c) (d)
Fig. 6 SVM classification results for the classification of the feature B1 using factors C3 and C4
based on different kernel functions: a linear kernel; b linear SVM; c RBF kernel; d polynomial
kernel
and B4 are shown in Fig. 7. In Fig. 7b–d, it is obvious that the green point has the
highest accuracy than others, which denotes the value of C = 10, γ = 5. As the
value of γ in the green line increases, the accuracy value goes up first, reaches the
peak value at γ = 5, and then decreases gradually. It means that a too large curvature
of the decision boundary will decrease the classification accuracy. Although the best
classification performance in terms of the structural risk category B1 in Fig. 7a does
not originate from the pair of (C, γ) in (10, 5), C with a value of 10 can still reach
the highest accuracy under the condition that γ is equal to 0.5 or 1. Consequently,
the value of parameter C and γ is set as 10 and 5, respectively, in the RBF-SVM to
exhibit the robust classification performance.
(2) Calculation of posterior probability
Based on the constructed SVM model, the safety levels (i.e., I, II, and III) can be
assigned to the risk categories (i.e., B1, B2, B3, and B4). The confusion matrix in
Fig. 8 summarizes the number of true and false predictions, where a deeper color
refers to a larger number. It offers a clear insight into the performance of the SVM
classification model. From a qualitative point of view, the classification accuracy can
be calculated by the proportion of the sum of the value along the diagonal to the total
value, which works out to be 82.86% for B1, 94.29% for B2, 91.43% for B3, and
94.29% for B4, respectively. In other words, the overall accuracy for B1 is at least
8.57% lower than other categories, that is because the parameters C = 10, γ = 5 are
not the optimal ones in the SVM model of B1.
5 Information Fusion for SHA in Tunnels 111
0.90 0.96
C=0.01 C=0.01
C=0.1 0.94 C=0.1
0.88 C=1
C=1
0.92
C=10 C=10
0.86 C=100 0.90 C=100
0.88
Accuracy
Accuracy
0.84
0.86
0.82
0.84
0.80 0.82
0.80
0.78
0.78
0.76 0.76
0 2 4 6 8 10 0 2 4 6 8 10
γ γ
(a) (b)
0.95
0.90
0.85 0.90
0.80
0.85
Accuracy
Accuracy
0.75
0.70
0.80
0.65 C=0.01 C=0.01
C=0.1 C=0.1
0.60 C=1 0.75 C=1
C=10 C=10
0.55 C=100 C=100
0.70
0 2 4 6 8 10 0 2 4 6 8 10
γ γ
(c) (d)
Fig. 7 SVM classification accuracy based on pairs of (C, γ) for four risk features: a B1; b B2;
c B3; d B4
To take full advantage of the SVM outputs, they can be mapped into the poste-
rior probabilities by a sigmoid function using Eq. (2). Table 2 takes the monitoring
sections No. 31, No. 32, No. 33, No. 34, and No. 35 as an example, which lists
their probability results corresponding to safety levels I, II, and III, respectively. As
a result, the predicted label is determined by the highest probability in bold font in
Table 2. Although the established four SVM models for different risk categories can
reach comparatively promising classification accuracy greater than 82.86%, they are
likely to return different predictions for the same monitoring section. For instance,
the predictions for risk categories B1, B2, and B4 in the monitoring section No. 31
all turn out to be a level II, while the risk category B3 is labeled as level III. What’s
worse, the monitoring section No. 35 suffers from a more significant inconsistency
in the classification predictions. Only predictions for B3 and B4 are the same (level
II), while the SVM models for B1 and B2 output contradictory results as level I and
level III, respectively.
It is revealed that SVM models tend to generate conflicting and imprecise predic-
tions for the same monitoring sector, which will inevitably cause confusion in the
112 Information Fusion
18
15
12
0
Level I Level II Level III Level I Level II Level III
True label True label
(a) (b)
16
12
0
Level I Level II Level III Level I Level II Level III
True label True label
(c) (d)
Fig. 8 Confusion matrix for four risk factors: a B1; b B2; c B3; d B4
evaluation of the overall structure health status. Notably, the highest and second-
highest probability in an SVM model could be very close to unfortunately reduce
the credibility of the predicted result. The monitoring section No. 35 is taken as an
example. There is a very small discrepancy in the probability of a level I (0.581) and
level III (0.406) for risk category B1, implying that the conclusion drawn from such
results does not have a very high confidence level. The same problem also appears
in other monitoring sections. In this regard, when two or three risk levels have the
same proportion, the safety level can be defined by the combinations as (I, II), (II,
III), (I, III), and (I, II, III) and then transferred into a corresponding framework
= {I, II, III} with seven subsets except for the empty set for further investigation.
5 Information Fusion for SHA in Tunnels 113
According to the improved combination rule introduced in Sect. 3.2, the predicted
results from four SVM classifiers can be fused continually at the decision level
until all of them have been combined. It can generate more robust results than the
fusion at the raw data level and feature level, which is expected to enhance the
reliability and quality of the safety risk perception. Noticeably, BPA based on the
belief degree is critical in the information fusion process. Following the works from
[31], the outputs of SVMs in the form of posterior probabilities can be directly
converted into BPA. The frame of the discernment = {I, II, III} is defined to
refer to three risk levels. In consequence, the largest value among m (I), m (II),
and m (III) determines the label assigned to the targeted monitoring section, and
m() measures the uncertainty in the assessment result. The results of BPAs for all
monitoring sections after multi-classifier information fusion are outlined in Table 3,
which are analyzed as follows.
(1) The proposed data fusion approach can effectively reduce the uncertainty and
increase the precision in the overall structural health evaluation. Clearly, only
114 Information Fusion
Table 3 Results of BPAs for all monitoring sections after multi-classifier information fusion
Monitoring section m(I) m(II) m(III) m(Θ) Predicted label True label
No. 1 0.0222 0.9398 0.0379 1.05 × 10−4 II II
No. 2 0.0207 0.9366 0.0426 1.21 × 10−4 II II
No. 3 0.0207 0.9366 0.0426 1.21 × 10−4 II II
No. 4 0.2735 0.6766 0.0500 4.85 × 10−6 II II
No. 5 0.0236 0.9407 0.0356 9.48 × 10−5 II II
No. 6 0.0200 0.9384 0.0415 1.31 × 10−4 II II
No. 7 0.0287 0.7895 0.1811 6.35 × 10−4 II II
No. 8 0.0596 0.9051 0.0353 4.47 × 10−5 II II
No. 9 0.0596 0.9051 0.0353 4.47 × 10−5 II II
No. 10 0.0192 0.9503 0.0305 9.27 × 10−5 II II
No. 11 0.2385 0.7242 0.0372 5.87 × 10−6 II II
No. 12 0.0192 0.9503 0.0305 9.27 × 10−5 II II
No. 13 0.0383 0.7640 0.1970 7.00 × 10−4 II II
No. 14 0.0099 0.9778 0.0122 3.91 × 10−5 II II
No. 15 0.0305 0.9240 0.0452 2.22 × 10−4 II II
No. 16 0.0305 0.9240 0.0452 2.22 × 10−4 II II
No. 17 0.0628 0.8099 0.1269 3.84 × 10−4 II II
No. 18 0.0104 0.9865 0.0030 1.45 × 10−5 II II
No. 19 0.0160 0.9628 0.0211 7.11 × 10−5 II II
No. 20 0.2368 0.7358 0.0273 1.74 × 10−6 II II
No. 21 0.0192 0.9617 0.0190 6.17 × 10−5 II II
No. 22 0.0178 0.9479 0.0342 1.05 × 10−4 II II
No. 23 0.0140 0.9202 0.0655 2.85 × 10−4 II II
No. 24 0.0126 0.9508 0.0364 1.25 × 10−4 II II
No. 25 0.0375 0.8418 0.1204 3.58 × 10−4 II II
No. 26 0.0971 0.7031 0.1994 3.46 × 10−4 II II
No. 27 0.1473 0.8015 0.0511 2.94 × 10−5 II II
No. 28 0.1109 0.8327 0.0564 3.64 × 10−5 II II
No. 29 0.0416 0.5083 0.4483 1.79 × 10−4 II II
No. 30 0.0100 0.9685 0.0214 7.96 × 10−5 II II
No. 31 0.0386 0.7636 0.1972 6.70 × 10−4 II II
No. 32 0.8416 0.0375 0.1205 3.60 × 10−4 I I
No. 33 0.0290 0.1884 0.7791 3.43 × 10−4 III III
No. 34 0.0207 0.9366 0.0426 1.21 × 10−4 II II
No. 35 0.3276 0.3144 0.3578 1.75 × 10−4 III II
5 Information Fusion for SHA in Tunnels 115
the monitoring section No. 35 cannot be evaluated correctly, which means that
the classification accuracy based on SVM and the D-S evidence theory can
rise up to 97.14% (34/35). As for the confidence indicators m(), its value of
the 35 monitoring sections all gets close to zero (less than 1.0 × 10−4 ). That
is because the case has no combination (I, II, III) in Table 2.
(2) Data fusion has a powerful capability in dealing with the conflicting results
from four SVM models, and thus, it serves as a promising tool to fully synthe-
size multi-source information with high quality. The monitoring section No.
32 is taken as an example. In Table 2. there is no agreed result about the overall
safety level from the four SVM models (B1, B2, and B4 belong to the level I,
while B3 belongs to the level III). After the information fusion, Table 3 indi-
cates that the BPA value for safety level I (that is m (I)) is equal to 0.8416,
resulting in an ultimate prediction of the level I under a very high confidence
level.
(3) For all correct prediction results, No. 33 is rated at the high-risk level (level
III), which can be considered as the most dangerous section in this opera-
tional tunnel. This evaluation result consists of the facts, since the leakage
water amount recorded by sensors is 7.695 L/m2 /day above the limit value of
1 L/m2 /day. Therefore, more attention should be paid to the monitoring section
No. 33 for improving its safety performance. To mitigate risk at No. 33, it is
essential to take proper actions especially for controlling the seepage damage.
To be specific, some common methods, such as using the waterproof materials,
grouting to block the water, discharging water, and others, can be implemented
to restrict the amount of water leakage below the limit value.
6 Merits Analysis
The effectiveness of the proposed hybrid information approach based on the SVM
model and the improved D-S evidence theory is demonstrated from two aspects.
On the one hand, it can return a robust and comprehensive evaluation result, which
is useful in the structural safety analysis. On the other hand, it is superior to the
single SVM model in handling high-bias data, which is possible to achieve higher
classification accuracy. Discussions are presented as follows.
Fig. 9 Results of BPAs after 1000 iterations at the monitoring section No. 33 under different bias
levels: (1) 5%; (2) 10%; (3) 15%; (4) 20%
acceptable. There are still 79.2% of predictions at No. 32 falling in the correct
label (that is a risk level I (Safe)) and 70.1% of predictions at No. 33 belonging
to a risk level III (High risk). As for the condition with data in a 5% bias,
the model can realize a nearly 100% classification accuracy, verifying that
bias limited less than 5% has little impact on the reliability of the prediction.
Therefore, the strong robustness of the hybrid information fusion method is
validated to some extent. Besides, it is worth noting that No. 34 is free of bias
since the prediction accuracy under various bias levels all reach up to 100%.
That is because the four SVM models of No. 34 give back consistent results,
namely the safety level II (Low risk).
(3) Global sensitivity analysis is conducted through the 1000 times iterations at
the monitoring section No. 33, since No. 33 is evaluated as the most dangerous
section (level III) and deserves special attention. Global sensitivity analysis
aims to identify the critical risk factors contributing more to the structural health
status. Accordingly, proper actions can be taken ahead of time for structural
failure prevention. The Spearman’s rank correlation coefficient from Eq. (20)
118 Information Fusion
Risk level I Risk level II Risk level III Risk level I Risk level II Risk level III
1000 1000
800 800
Frequency
Frequency
600 600
400 400
200 200
0 0
5% 10% 15% 20% 5% 10% 15% 20%
Bias Bias
(a) (b)
Risk level I Risk level II Risk level III Risk level I Risk level II Risk level III
1000 1000
800 800
Frequency
Frequency
600 600
400 400
200 200
0 0
5% 10% 15% 20% 5% 10% 15% 20%
Bias Bias
(c) (d)
Fig. 10 Results of safety risk assessment after 1000 iterations under different bias levels at four
monitoring sections: (1) No. 31; (2) No. 32; (3) No. 33; (4) No. 34
measures the contribution of risk factors to the overall safety level. Figure 11
presents the global sensitivity analysis of 16 risk factors at the monitoring
section No. 33. It is clear that C7, C10, C13, and C15 turn out to be the top
four sensitive risk factors with the greatest positive global sensitivities, which
should be given more concern for reducing the risk magnitude at No. 33.
For C10, C13, and C15, their sensitivities gradually decline as the bias level
grows. When the bias increases to 20%, the most sensitive risk factors remain
unchanged, which verifies the robustness of the proposed method once again.
In contrast, C3, C4, and C11 perform the lowest sensitivity, which has little
influence on the structural system safety.
6 Merits Analysis 119
0.5
Bias 5%
Bias 10%
0.4 Bias 15%
Bias 20%
Global Sensitivity Measure
0.3
0.2
0.1
C3 C12 C16
0.0
C1 C2 C4 C5 C6 C7 C8 C9 C10 C11 C13 C14 C15
-0.1 Risk Factors
-0.2
It is known that the standard SVM model also owns the ability to estimate the overall
structural safety level. A comparison between the SVM model and the proposed
SVM-DS approach is presented below. To address unavoidable uncertainty, the
Monte Carol simulation is adopted, and thus, the calculation is repeated throughout
1000 times for the single SVM model.
(1) The single SVM model can effectively deal with input data at a very low bias
level, but it exhibits very poor classification performance at a high bias. Using
the standard SVM model, Fig. 12 describes the frequency distribution of the
total correct classification number among 35 monitoring sections at various bias
levels (i.e., 5%, 10%, 15%, and 20%). It is clear that nearly the 1000 iterations
can return correct prediction for 34 or even all the 35 monitoring sections
at 5% and 10% bias with a low standard deviation, representing a precise
and stable prediction performance. However, when the bias level grows up to
15% and 20%, respectively, the frequency distribution significantly changes.
The range of the correct number falls in the range [20, 35], which no longer
concentrates on the value of 34 or 35. In addition, the standard deviations at
15% and 20% bias increases to 4.09 and 5.83, respectively, which indicates
that the total correct number is spread out over a wide range, leading to great
fluctuations and unstable results. In Fig. 12c at a bias level of 15%, although the
largest frequency is in total correct number 35, the value of frequency is only
344. That is to say, only 34.4% of iterations can make the right classification
for all 35 monitoring sections. In Fig. 12d under the 20% bias, more than
120 Information Fusion
1000 1000
Mean=35.00 Mean=34.90
Standard DeviaƟon=0.00 Standard DeviaƟon=0.36
800 800
Frequency
Frequency
600 600
400 400
200 200
0 0
20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35
Total number of correct monitoring points Total number of correct monitoring points
(a) (b)
350
Mean=28.92 Mean=22.26
500
300 Standard DeviaƟon=5.83 Standard DeviaƟon=4.09
250 400
Frequency
Frequency
200
300
150
200
100
100
50
0 0
20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35
Total number of correct monitoring points Total number of correct monitoring points
(c ) (d)
Fig. 12 Frequency distribution of the total correct classification number after 1000 iterations using
the standard SVM method under different bias levels: (1) 5% bias; (2) 10% bias; (3) 15% bias; (4)
20% bias
Classification Accuracy
0.8
0.6
0.4
0.2
0.0
5% 10% 15% 20%
Bias
the single SVM model owns 2.93% and 3.37% higher classification accuracy
than the hybrid method within a bias below 10%, its accuracy goes an apparent
decline when the bias is greater than 15%. More specifically, there is a sudden
drop in a 20% bias, reaching a classification accuracy of 63.59%. It implies a
poor classification performance of the single SVM model, since approximately
40% of data is classified incorrectly. Therefore, it can be concluded that the
single SVM model is very sensitive to bias. On the contrary, the proposed
SVM-DS model can still classify 87.99% of data correctly at a 20% bias,
whose classification accuracy only decreases 10% from a 5% bias to a 20%
bias. Since the underground environment of the operational tunnel is generally
filled with much complexity and uncertainty, it is more reasonable to imple-
ment the proposed SVM-DS model, which is proven to effective in tackling
high-bias information and then giving more accurate results of the safety risk
perception.
7 Summaries
The main objective of this research is to present a hybrid information fusion approach
with the combination of the probabilistic SVM and the improved D-S evidence theory.
Special emphasis should be put on the improved D-S evidence theory integrating the
Dempster’s rule and the weighted mean rule. The proposed risk analysis approach
serves as a decision tool to provide a reliable and comprehensive assessment of
structural safety level in the operational tunnel under uncertainty, which takes the
following three steps: (1) A total of 16 risk factors ranging from four risk categories
are identified and extracted from multiple sensors, which greatly affect the reliability
of the operational tunnel; (2) Four probabilistic SVMs are well trained by selecting the
appropriate kernel function and parameters; (3) The improved D-S evidence theory
122 Information Fusion
makes the best of predictions from the four SVM models and fuse them at the decision
level. As a result, it is expected to achieve a more accurate safety assessment for all
the monitoring sections. Moreover, the Monte Carlo simulation is adopted to address
uncertainty for robustness validation and global sensitivity analysis. For verifying
its feasibility and effectiveness, it has been successfully applied in an operational
tunnel in the Wuhan metro system in China.
In view of the case, the hybrid SVM-DS method is helpful in fusing multi-source
information, which can considerably improve the accuracy and stability of the risk
level evaluation. The meaningful findings can be summarized as follow: (1) The
prediction results from the SVM-DS model are robustly reliable for the observed data
from the underground environment with a great deal of uncertainty and variability. It
can better synthesize highly conflicting classification results from four probabilistic
SVM models, reaching an overall accuracy of 97.14% for the targeted infrastructure.
(2) The SVM-DS model has a strong robustness to bias. Even when the bias level
grows to 20%, the proposed method can maintain the classification accuracy above
70% for each monitoring section. That is to say, SVM-DS has a superior bias tolerant
ability to minimize the adverse influences from bias, in order to ensure the reliability
and accuracy of the prediction results. (3) The SVM-DS model outperforms the single
SVM model in handling information with high conflicts and bias. Especially in the
condition of a 20% bias, the single SVM model only correctly classify 63.84% of
samples, which is 24.15% less than the accuracy of the SVM-DS model. In addition,
global sensitivity analysis is carried out to recognize the most critical risk factors on
structure risk throughout a series of repeated iterations. The identified most sensitive
factors can be the major concern for safety control measurements, and thus, managers
can tackle the potential risk in advance to prevent the occurrence of structural failures.
The novelty of the developed approach lies in two aspects. On one hand, it
contributes to the state of the knowledge by training data from various sensors by
SVM and synthesizing multi-sensory information using the improved D-S evidence
theory at the decision level to obtain a more precise and reliable assessment. On the
other hand, it contributes to the state of the practice by providing a promising tool for
evaluating the structural health and perceiving the potential risk of the operational
tunnel in the subsurface environment with great complexity and uncertainty, in order
to inform data-driven decisions for mitigating the risk of structural failure at an early
stage.
References
1. Arrieta AB, Díaz-Rodríguez N, Del Ser J, Bennetot A, Tabik S, Barbado A, García S, Gil-
López S, Molina D, Benjamins R (2019) Explainable artificial intelligence (XAI): concepts,
taxonomies, opportunities and challenges toward responsible AI. arXiv preprint arXiv:.10045
2. Bao Y, Li H, An Y, Ou J (2012) Dempster-Shafer evidence theory approach to structural damage
detection. Struct Health Monit 11:13–26
3. Chileshe N, Boadua Yirenkyi-Fianko A (2012) An evaluation of risk factors impacting
construction projects in Ghana. J Eng Des Technol 10:306–329
References 123
4. Chong JW, Kim Y, Chon KH (2014) Nonlinear multiclass support vector machine–based health
monitoring system for buildings employing magnetorheological dampers. J Intell Mater Syst
Struct 25:1456–1468
5. Dushyanth N, Suma M, Latte MV (2016) Detection and localization of damage using empirical
mode decomposition and multilevel support vector machine. Appl Phys A 122:250
6. Farooq M, Zheng H, Nagabhushana A, Roy S, Burkett S, Barkey M, Kotru S, Sazonov E (2012)
Damage detection and identification in smart structures using SVM and ANN, Smart Sensor
Phenomena, Technology, Networks, and Systems Integration 2012. International Society for
Optics and Photonics, p. 83461O
7. Feng SJ, Yao AM, Dong YF (2014) Fuzzy comprehensive evaluation of open-pit slope stability.
Adv Mater Res 962–965:1029–1033
8. Ghiasi R, Torkzadeh P, Noori M (2016) A machine-learning approach for structural damage
detection using least square support vector machine based on a new combinational kernel
function. Struct Health Monit 15:302–316
9. Huang Y (2009) Advances in artificial neural networks–methodological development and
application. Algorithms 2:973–1007
10. Hyun KC, Min S, Choi H, Park J, Lee IM (2015) Risk analysis using fault-tree analysis (FTA)
and analytic hierarchy process (AHP) applicable to shield TBM tunnels. Tunnel Undergr Sp
Technol Incorp Trenchless Technol Res 49:121–129
11. Li M, Yu H, Jin H, Liu P (2018) Methodologies of safety risk control for China’s metro
construction based on BIM. Saf Sci 110:418–426
12. Manca D, Brambilla S (2011) A methodology based on the analytic hierarchy process for the
quantitative assessment of emergency preparedness and response in road tunnels. Transp Policy
18:657–664
13. Pan Y, Ou S, Zhang L, Zhang W, Wu X, Li HJRE (2019) Modeling risks in dependent systems:
a Copula-Bayesian approach. Reliab Eng Syst Saf 188:416–431
14. Pan Y, Zhang L (2020) Data-driven estimation of building energy consumption with multi-
source heterogeneous data. Appl Energy 268:114965
15. Pan, Y., Zhang, L., Li, Z., Ding, L., 2019b. Improved fuzzy Bayesian network-based risk
analysis with interval-valued fuzzy sets and DS evidence theory. IEEE Transactions on Fuzzy
Systems.
16. Pan Y, Zhang L, Li Z, Ding L (2020) Improved fuzzy Bayesian network-based risk analysis
with interval-valued fuzzy sets and DS evidence theory. IEEE Trans Fuzzy Syst 28:2063–2077
17. Pan Y, Zhang L, Wu X, Skibniewski MJ (2020) Multi-classifier information fusion in risk
analysis. Inform Fus 60:121–136
18. Pan Y, Zhang L, Wu X, Zhang K, Skibniewski MJ (2019) Structural health monitoring and
assessment using wavelet packet energy spectrum. Saf Sci 120:652–665
19. Patle A, Chouhan DS (2013) SVM kernel functions for classification. In: Advances in
Technology and Engineering (ICATE), 2013 International Conference on. IEEE, pp 1–9
20. Peng Y, Lin J-R, Zhang J-P, Hu Z-Z (2017) A hybrid data mining approach on BIM-based
building operation and maintenance. Build Environ 126:483–495
21. Platt J (1999) Probabilistic outputs for support vector machines and comparisons to regularized
likelihood methods. Adv Large Margin Classif 10:61–74
22. Vapnik V (2013) The nature of statistical learning theory. Springer Science & Business Media
23. Wang B, Mo C, He C, Yan Q (2017) Fuzzy synthetic evaluation of the long-term health of
tunnel structures. Appl Sci 7:203
24. Worden K, Manson G (2007) The application of machine learning to structural health
monitoring. Philos Trans 365:515
25. Yuan K, Deng Y (2019) Conflict evidence management in fault diagnosis. Int J Mach Learn
Cybern 10:121–130
26. Zhang L, Ding L, Wu X, Skibniewski MJ (2017) An improved Dempster-Shafer approach to
construction safety risk perception. Knowl-Based Syst 132:30–46
27. Zhang L, Skibniewski MJ, Wu X, Chen Y, Deng Q (2014) A probabilistic approach for safety
risk analysis in metro construction. Saf Sci 63:8–17
124 Information Fusion
28. Zhang L, Wu X, Zhu H, Abourizk SM (2016) Perceiving safety risk of buildings adjacent to
tunneling excavation: an information fusion approach. Autom Constr 73:88–101
29. Zhang W, Sun K, Lei CZ, Jr, Bill FS (2014b) Fuzzy analytic hierarchy process synthetic
evaluation models for the health monitoring of shield tunnels. Comput Aided Civil Infrast Eng
29:676–688
30. Zhao J, Xie X, Xu X, Sun S (2017) Multi-view learning overview: Recent progress and new
challenges. Inform Fusion 38:43–54
31. Zhou Q, Zhou H, Zhou Q, Yang F, Luo L, Li T (2015) Structural damage detection based on
posteriori probability support vector machine and Dempster-Shafer evidence theory. Appl Soft
Comput 36:368–374
Dynamic Bayesian Networks
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 125
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_6
126 Dynamic Bayesian Networks
of the ordinary BN, which allows the explicit modeling of changes subjected to time
series or sequences. DBN is capable of revealing complex multivariate time series,
where the relationships between multiple time series can be represented in the same
model. DBN is developed to unify and extend traditional linear state-space models
(e.g., Kalman filters), linear and normal forecasting models (e.g., ARMA), and simple
dependency models (e.g., hidden Markov models) into a general probabilistic repre-
sentation and inference mechanism for arbitrary nonlinear and non-normal time-
dependent domains [8]. That is to say, DBN can model the evolution of the proba-
bilistic dependencies within a random system. This chapter attempts to use DBN to
deal with the potential uncertainty and randomness underlying safety management
in tunnel construction.
A systemic decision approach based on DBN is developed, aiming to provide
guidelines for dynamic risk analysis of the tunnel-induced damage over time. The
proposed approach can be used to conduct various analysis tasks, including predic-
tive, sensitivity, and diagnostic analysis. Finally, a case study concerning the dynamic
risk analysis in a tunnel project in China is used to verify the applicability of the
proposed DBN-based approach.
2 BN and DBN
2.1 BN
Y
2 BN and DBN 127
computed. The relations represented by DAG allow the JPD to be specified locally by
the conditional probability distribution for each node. Assuming Pa(X i ) is the parent
node of X i in the DAG, the conditional probability distribution of X i is denoted by
P(X i |Pa (X i )). The JPD of P(X 1 , …, X N ) can then be written as Eq. (1).
P(X 1 , . . . , X N ) = P(X i |Pa(X i ) ) (1)
X i ∈{X 1 ,...,X N }
where X i is the ith node in the BN model, Pa(X i ) is the parent node of the ith node,
P(X 1 , …, X N ) is the joint probability of the node set {X 1 , …, X N }, and P(X i |Pa(X i ))
is the conditional probability of X i given Pa(X i ) has been observed.
2.2 DBN
N
P( X t |X t−1 ) = P( X ti Pa(X ti )) (2)
i=1
where X ti refers to the ith node at time t (i = 1, 2, …, N), and Pa(X ti ) refers to the
parents of X ti in the directed acyclic graph.
The nodes in the first slice of the 2TBN do not have any parameters associated
with them, however, each node in the second slice of the 2TBN has associated a
N
conditional probability distribution for variables which define P( X ti Pa(X ti ))
i=1
for all t > 1 [6]. The distribution given by a 2TBN can be divided into two aspects,
namely, (i) the inter-slice distribution that models the probability of variables in X t
with parents at time t-1; (ii) the intra-slice distribution that models the probability of
variable in X t with parents in the same time slice.
Generally, we assume that the parameters of the conditional probability distribu-
tion are time-invariant, and the model is time-homogeneous. The parents of a node
Pa(X ti ) can either be in the same time slice or the previous time slice [6]. The arcs
between slices are from up to down, which represent causal flows of time. If there
i
is an arc from X t−1 to X ti , this node is called persistent. The arcs within a slice are
arbitrary, and one can put arcs in any way one wants to. That is to say, their relation-
ships are unconstrained, as long as the intra-slice-model is acyclic. The parameters
of the conditional probability distribution used by the proposed model are assumed
time-invariant. When we unroll the 2TBN until we have T time slices, the resulting
joint distribution probability can be defined by Eq. (3).
T
N
P(X 1:T ) = P( X ti Pa(X ti )) (3)
t=1 i=1
where X 1: T refers to the joint events of the observation for X 1 from time t = 1 to T.
of piles for adjacent bridges [23]. Several cases of collapses and sinkholes in road-
ways resulted from tunneling construction were reported [18]. Moreover, the ground
movement due to tunneling together with the aging corrosion issues may result in
deformation and fracture of pipelines, leading to the interruption of transportation of
water, oil, gas, and other supplies [22]. Most of the times, tunnels are built under the
urban trunk roads. Tunnel-induced road damage attracts more and more attention,
since the reliability and safety of the road system are directly affected. Typically, the
road damage displays obvious time-dependent characteristics. Thus, it is necessary
and meaningful to investigate the safety risk mechanism of the tunnel-induced road
damage, which can provide a reference for risk modeling and further investigation.
The soil plays a mediating role in the tunnel-induced damage. The tunnel excavation
in soft soils inevitably leads to ground movement, greatly influencing the road surface
damage and further ground transportation safety. In accordance with the actual tunnel
construction practice, the tunnel-induced ground settlement displays a typical non-
stationary pattern along with time series [19]. Generally, the deformation tendency
following the excavation progress can be divided into three phases, including the
initial deformation, fast development, and deterioration phases. Figure 3 illustrates
the tunnel-induced ground settlement within three phases.
(1) Initial deformation phase. In this phase, both the accumulative ground settle-
ment and the settling rate begin to gradually grow from zero. Mostly, the ground
settlement takes place before the shield machine arrives at the excavation face.
The stress release of the surrounding soil is generally considered as the primary
cause of the ground settlement during this phase.
ground settlement
Accumulative
Excavation direction
Fig. 3 Patterns of the tunnel-induced ground settlement following the excavation progress
130 Dynamic Bayesian Networks
(2) Fast development phase. In this phase, the accumulative settlement is contin-
uously increasing at a high speed, while the subsidence rate starts to decrease
gradually. The ground settlement takes place when the shield machine passes
exactly through the excavation face from the shield head until the shield nail.
The generated settlement is mainly due to the increased shearing stress, which
is caused by the friction resistance between the shield machine body and the
surrounding soil.
(3) Deformation deterioration phase. In this phase, the settlement continues to
grow at a very low speed. The subsidence rate is gradually reduced to zero.
Normally, it is the secondary consolidation settlement of the soft soil that
contributes to the settlement. Generally, the maximum ground settlement is
observed at this phase.
The soil can be disturbed by tunneling activities of a shield machine, and the road
surface damage can be inflicted when the settlement deformation exceeds a certain
threshold. The impact of tunnel excavation on the road surface deformation and
damage can be presented in three aspects, namely cut-in ruts, dents, and carvings.
These aspects are very likely to appear on the road surface due to uneven surface
settlement, which could further affect the performance and durability of the road
pavement. According to statistics, more than 75% of existing roads are made up of
asphalt pavements [21]. Thus, the road surface damage assessment is mainly focused
on asphalt roads, and a case concerning the safety analysis of an asphalt pavement
induced by tunneling excavation is presented later.
To assess the road surface damage, the indicator, Pavement Condition Index (PCI),
is commonly used in the construction engineering practice. An industry standard
“Technical Specifications for maintenance of Highway Asphalt Pavement (JTJ073.2-
2001)” was issued in October 2001 in China, where PCI was calculated by Eq. (4).
Moreover, based upon the calculated results, the health condition of an existing road
surface can be rated into four different levels, namely Excellence (85–100), Good
(70–85), Fair (55–70), and Poor (0–55). The higher the value of PCI, the better the
condition of the road surface.
⎧
⎨ PC I = 100 − 15D R 0.412
⎩ D R = A1 × K 1 + A2 × K 2 + A3 × K 3 × 100%
(4)
A
where DR refers to the damage rate of the pavement condition; A1 , A2 , and A3 refer
to the damaged area of cut-in ruts, dents, and caving, respectively; A refers to the
total area of the investigated road surface; K 1 , K 2 , and K 3 refer to the conversion
coefficient representing the degree of damage due to cut-in ruts, dents, and caving,
respectively. The values (i.e., K 1 , K 2 , and K 3 ) can be obtained from the standard.
3 Dynamics in Tunnel-Induced Damages 131
where S max refers to the maximum ground settlement induced by tunnel excavation,
S 0 refers to the normal ground settlement in the free-field where no ground transport
is presented, and i refers to the width coefficient of the tunnel.
Following Eq. (5), a large number of monitoring records of prior cases are used
to conduct an intuitional representation of relations between PCI and S max , and the
fitted curve between the two is displayed in Fig. 4. The fitted curve is approximately
linear when S max falls in a certain range, and tends to be flattened when S max reaches
up to 38 mm. As for S max with a general distribution range of 0–90 mm, we can follow
up the suggested PCI division to separate S max into four different levels accordingly,
including I (Very Safe, 0–15 mm), II (Safe, 15–38 mm), III (Dangerous, 38–71 mm),
and IV (Very Dangerous, 71–90 mm). Specifically, the level I of S max is in response
Fitting Curve
Pavement Condition Index (PCI)
I
Smax vs PCI
SSE: 27.6
R-square: 0.9869
II Adjusted R-square: 0.9868
RMSE: 1.285
III
IV
to the excellent road condition (85–100), the level II of S max is in response to the
good road status (70–85), the level III of S max is in response to the fair road condition
(55–70), and the level IV of S max is in response to the poor road condition (0–55).
Risk/hazard identification involves determining which risks (or risk factors) may
affect the project safety and documenting their characteristics. It is considered a diffi-
cult task for a complex system, particularly in tunnel construction. With the devel-
opment of the tunnel construction practice worldwide, large amounts of scattered
knowledge resources are accumulated, including explicit and tacit knowledge. For
the risk/hazard identification in tunnel construction, three procedures are explicitly
illustrated in this stage.
Start
Predictive analysis
Fig. 5 Flowchart of the proposed DBN-based dynamic risk analysis in complex projects
4 DBN-Enabled Dynamic Risk Analysis 133
aspects: namely, i) the accepted knowledge from existing design and construc-
tion standards, and ii) empirical knowledge from statistical models in the
literature, and iii) the tacit knowledge from domain experts.
(2) Risk mechanism analysis. The risk mechanism analysis for the target risk
event can be conducted by consulting metro design and construction standards
and technical manuals and retrieving literature. Relevant variables that repre-
sent or indicate system safety conditions, flow process, and performance states
can be determined.
(3) Relationship discovery. Based on the potential failure-consequence scenarios
in the risk mechanism analysis, explicit relationships and interdependencies
among the above variables can be revealed. Here, the range of each variable’s
observation data is defined. The state of each variable with a different attribution
is usually defined based on both historical data and expert estimation. The
failure modes, internal variables, exogenous factors, the explicit cause, and
effect relationships are determined.
R R
Bayesian
Qualified
network
fault trees X1 X2 X1 X2 structure
(b) OR gate
R R
X1 X1
(c) NOT gate
from the given data using some kinds of learning algorithms. Learning BN’s
parameters from complete data have received considerable attention in the
last decade. When modelers know the structure of BN and the data set has
no missing values, the maximum a posterior (MAP) estimation method is
usually used to learn BN’s parameters [20]. Taking the Dirichlet distribu-
tion as a prior distribution, the MAP method first counts the occurrence of
each value of a given variable to obtain the relative frequency. Then modelers
search for the best BN parameters using a practical search-and-score proce-
dure. Many search-and-score algorithms are available to evaluate the fitness of
the network, such as Greedy search algorithms (i.e., Cooper and Herskovits’
K2 algorithm [4]) and Markov Chain Monte Carlo (MCMC) method (i.e.,
Simulated Annealing and Minimum Description Length Principle). For more
detailed information, see Onisko et al. [14].
One of the practical learning problems is to deal with incomplete data. Compared
to the complete data, the occurrence of missing values leads to the analytical
intractability. It is tempting to simply delete incomplete records or fill in missing
values, because this leads to biased results or an over-confidence in the follow-
up analysis. To avoid such an ad hoc dilemma, some approximation methods can
be adapted to improve the increased uncertainty due to missing values, such as
Expectation-Maximization (EM), Data Augmentation (DA), and so on.
(3) Markov chain process. A discrete-time Markov chain is a mathematical
system that simulates transitions from one state to another in the real-world. It
is a random process usually characterized as memoryless, indicating that the
next state depends only on the current state and not the sequence of events that
precede it. Based on the Markov chain process as illustrated in Eq. (6). The
DBN model is a probability distribution function on the sequence of T hidden-
state variables X = {x 0 ,…, x T-1 } and the sequence of T observables Y = {y0 ,
…, yT-1 }that has the following factorization, which satisfies the requirements
for DBN that the state x t depends only on the state x t-1 .
4 DBN-Enabled Dynamic Risk Analysis 135
−1
T −1
T
Pr ( X 1:T | Y1:T ) = Pr ( xt |xt−1 ) × Pr ( yt |xt ) × Pr (x0 ) (6)
t=1 t=0
where Pr(x t |x t-1 ) is the state transition probability density function, Pr(yt |x t ) is the
observation probability density function, and Pr(x 0 ) is the initial state distribution.
The DBN model can be used to conduct various types of analysis after it is estab-
lished. The most important use of DBN lies in updating probabilities in lights of
actual observations. It is, therefore, possible to predict the probability distribution
of potential safety risks and identify the most likely potential causes, given undesir-
able events are observed. In this stage, three sorts of analysis, including predictive,
sensitivity, and diagnostic analysis using the Bayesian inference capabilities, are
presented.
(1) Predictive analysis. The predictive analysis aims to predict the probability
distribution of a risk event (defined by R) under a combination of risk factors
(X 1 , X 2 ,…, X n ). The state of each risk factor, certain or uncertain, is treated as
evidence input into the DBN model. Compared to the traditional FTA, predic-
tive analysis in DBN models does not need to get minimal cut sets, leading
to the improvement of the computational efficiency. The probability distribu-
tion of R, represented by P(R = r), is given by Eq. (7). P(R = r) can serve
as an indicator to evaluate the risk of R, and decision-makers can take proper
preventions in advance if the risk magnitude of P(R = r) is very high.
P(R = r ) = P( R = r |X 1 = x1 , X 2 = x2 , . . . , X N = x N )
× P(X 1 = x1 , X 2 = x2 , . . . , X N = x N )
r = {r1 , r2 , . . . , r P }, xi = xi1 , xi2 , . . . , xiQ i , i = 1, 2, . . . , N (7)
Qi
1
j
P(R = r |X i = xi ) − P(R = r )
S P M(X i ) = , i = 1, 2, . . . , N (8)
Q i j=1 P(R = r )
1,...,qi −1,qi +1,...,Q i
P(R = r |X = x j ) − P(R = r |X = x qi )
1 i i i
S P M(X i ) = i
Qi − 1 q
P(R = r |X i = xi i )
j=1
i = 1, 2, . . . , N (9)
where r refers to the state of a risk-prone event R with P states; x i refers to the state
j
of risk factor X i with Qi states, and xi refers to the jth state of the risk factor X i .
Generally, SPM(X i ) can be used as an indicator to measure the degree of sensitivity
of the root node X i in case of the occurrence of a risk event. X i is more likely to
become the sensitive factor of the risk event R when SPM (X i ) is close to one, and
then should be given more attention.
(3) Diagnostic analysis. The diagnostic analysis aims to obtain the posterior prob-
ability distribution of each risk factor when an accident occurs. The suspected
causes can be detected in a very short period using a posterior probability
distribution, assisting in the real-time fault diagnosis. Compared to traditional
risk-based methods, such as FTA, CFEM, and NN, the feature of the diagnostic
analysis technique is matchless in the DBN reference [9]. The posterior prob-
ability distribution of risk factor X i , represented by P(X i = x i |R = r), can be
calculated by Eq. (10).
P(X i = xi ) × (R = r |X i = xi )
P( X i = xi |R = r ) = , i = 1, 2, . . . , N (10)
P(R = r )
where r refers to the state of a risk-prone event R with P states, and x i refers to the
state of risk factor X i with Qi states. Normally, X i is more likely to become the direct
cause of the occurrence of the risk event R when P(X i = x i |R = r) is close to one.
Based on the DBN-based risk analysis results, the critical risks and risk factors can
be determined, respectively. Then relevant safety control measures can be proposed
4 DBN-Enabled Dynamic Risk Analysis 137
in advance for risk mitigation. Furthermore, safety monitoring and review can be
conducted to secure the rational status of the whole system or its system components.
In addition, the implementation of the control measures can provide feedback and
suggestions on the improvement and optimization of the established DBN model.
During the tunneling excavation process, the road surface damage is regarded as a
prominent example of tunnel-induced damages, which can significantly affect the
safe operation of the ground transport system. A case concerning dynamic risk anal-
ysis in the construction of the Wuhan Yangtze Metro Tunnel (WYMT) in China
is used to verify the applicability of the proposed approach. A DBN model is
developed to reveal the causal relationships among different influential variables
in tunnel-induced damages. Various types of what-if scenario analysis, including
predictive, sensitivity, and diagnostic analysis and control, are performed to analyze
the probabilistic risk in tunnel-induced damages in a dynamic manner.
Wuhan Yangtze Metro Tunnel (WYMT) is known as the first metro tunnel across
the Yangtze River in China, which acts as an important route connecting Wuchang
and Hankou districts in the city of Wuhan. It is a double-spool tunnel with a total
length of almost 6200 m. The tunnel construction started in April 2008, and a slurry
shield machine with a diameter of 6.52 m in cutter diameter was adopted to push
the tunnel from the Jiyu Bridge Station in Hankou to the Jianghan Road Station in
Wuchang. A specific monitoring section where the tunnel passed exactly under the
Wuchang Riverside Avenue (WRA) was taken as an example to present the detailed
DBN-based risk analysis.
Figure 7 presents the spatial relationship between WYMT and WRA. WRA was
built in 1927 and further upgraded several times. Due to the close distance to the
bank of the Yangtze River, WRA played a significant role in flood defenses and port
transport development in Wuhan. Therefore, no fault was allowed while the tunnel
was crossing under the existing road surface, leading to a great challenge for safety
management in the construction of WYMT.
Legend
Tunnel axis
Tunnel outline
Monitoring point
Fig. 7 Spatial relationship between WYMT and WRA in the studied case
Table 2 Transition probability table for the leaf node R over time
R(t − 1) R(t)
I II III IV
I 0.60 0.20 0.10 0.10
II 0.05 0.60 0.25 0.10
III 0.05 0.10 0.60 0.25
IV 0.05 0.05 0.10 0.80
X1 X5
Legend
B1
Root node
R 1
Intermediate node
B2
Leaf node
X6 X10
X7 X9
X8
5 Dynamic Risk Analysis in Tunnel-Induced Damages 141
Predictive control aims to use the predictive Bayesian inference technique to predict
the probability distribution of the occurrence of tunnel-induced road damage before
an accident occurs. Some reasonable control measures can be implemented ahead
of time for risk mitigation. In construction practice, many accidents occur when the
risk grows to be extremely high, and no measures can prevent the occurrence of the
accident.
According to the established RSDBN (see Fig. 9), we are able to calculate the
probability of a specific risk, as well as the tendency of the risk evolution over time.
We first enter the evidence information of risk factors (X i ) (i = 1, 2, …, 10) into
RSDBN within 18 time slices. Using the predictive analysis technique in the DBN
inference, the outcome of the risk levels of R is presented in Fig. 10. Typically, colors
are used to represent the degree of the risk. The color of Green, Blue, Orange, and
Red is in correspondence with the risk level of I, II, III, and IV, respectively. As seen
in Fig. 10, the Level I (Very Safe) tends to decrease rapidly at the beginning, while
the Level IV (Very Dangerous) tends to rise quickly. The tendency indicates that the
potential safety risk is growing fast in the initial stage. At the 9th slice, the potential
safety risk jumps up to the Level III (Dangerous), since the predicted probability of
the Level III is higher than that of other levels then. In this situation, early-warning
signals should be released to decision-makers, and the risk magnitude tends to be
smooth and steady afterward.
Furthermore, the interface displaying the risk evolution process as shown in Fig. 10
can be used as a dashboard to check and track the status of the potential safety risk.
0.4
I (Very safe)
II (Safe)
III (Dangerous)
IV (Very dangerous)
0.3
Predicted probabilities
0.2
0.1
0.0
0 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18
Time slices
Figure 11 illustrates a dashboard for checking and tracking the status of the potential
safety risk in different scenarios. To be specific, the Scenario A represents the safety
status of R in the 1st time slice, where the predicted probability of the level of I is
calculated to be 0.296 and higher than that of the other three levels; the Scenario B
represents the safety status of R in the 4th time slice, where the potential risk suddenly
increases to the level of II; the Scenario C (or D) represents the safety status of R in
the 9th (or 11th) time slice, where the potential risk suddenly reaches up to the level
of III (or IV). As a result, the dynamics in the safety risk evolution over time can
be revealed, which can provide a visualized platform for decision makers to update
corresponding control measures in real-time.
1 0 1 0
0 1
0 1
0 1
0 1
90 90
0 1 0 1
(a) (b)
1 0 1 0
0 1
0 1
0 1
0 1
90 90
0 1 0 1
(c) (d)
Fig. 11 Dashboard for checking and tracking the status of the potential safety risk in different
scenarios: a Scenario A; b Scenario B; c Scenario C; d Scenario D
5 Dynamic Risk Analysis in Tunnel-Induced Damages 143
Sensitivity analysis and control aims to use the sensitivity analysis technique of the
Bayesian inference to identify critical and sensitive factors to the occurrence of risk-
prone events. As a matter of fact, the values of some geological and design variables
are fixed and can not change at the construction stage. Thus, engineers turn to give
much more attention to mechanical variables which have the flexibility to be adjusted
for ensuring construction safety. The most sensitive mechanical variables can be
determined as key checking points during the shield-driven tunneling excavation.
Under different excavation sections, the values of the input variables are varied
under given evidence. This results in the variation of the sensitivity of the involved
variables. The actual observations of influential variables X i (i = 1, 2, …, 10) are
entered into the model as the observed evidence. Given the risk magnitude of the
target event under different risk levels, Eq. (9) is used to calculate the performance
sensitivity of mechanical variables X i (i = 6, 7, …, 10). Figure 12 illustrates the
results of sensitivity analysis for mechanical parameters in different risk levels, with
the normalization process being adopted.
As shown in Fig. 12, under the risk level of I and II (indicating the low safety risk
level), X 6 (Torque Ratio) and X 7 (Cut Slurry Pressure) are identified as top sensitive
factors, since PSM (X 6 ) and PSM (X 7 ) are higher than that of other variables in such
conditions. Under risk level of III and IV (indicating the high safety risk level), X 8
(Drilling Speed), X 9 (Thrust Ratio), and X 10 (Grouting Speed) are identified as top
sensitive factors, since they are at the top of the ranking list regarding the performance
sensitivity. In general, the key checking points should be updated given the target
0.45
R=I
0.40 R = II
R = III
Sensitivity Performance Measure (SPM)
R = IV
0.35
0.30
0.25
0.20
0.15
0.10
0.05
0.00
X6 X7 X8 X9 X10
Mechanical variables
Fig. 12 Results of sensitivity analysis for mechanical parameters under different risk levels
144 Dynamic Bayesian Networks
event is observed at different risk levels. Accordingly, the major focus for safety
management strategies should be updated over time during the tunneling excavation
process.
Diagnostic control aims to use the diagnostic analysis technique in the DBN inference
to determine the most possible direct causes which have greater impacts on the
occurrence of risk-prone events. In case of an accident, the posterior probability of
each risk factor can be obtained, and the most suspected causes contributing to the
accident can be detected. More attention can then be paid to those suspected causes
in actual fault diagnosis.
Indeed, the values of some mechanical variables should be updated in accor-
dance with the changing geological conditions during the tunneling excavation
process. However, how to update these mechanical variables is not known exactly.
For instance, during the construction of WYMT, the tunnel-induced ground settle-
ment was observed to be 40 mm, indicating that the potential safety risk of R was
rated at a level of III (Dangerous). It is, therefore, necessary to reduce the risk and
perform the fault diagnosis. In this situation, the determined geological and design
variables were entered into the model as given evidence, and Eq. (10) was used to
obtain posterior probabilities of mechanical variables in the case of P(T = III) = 1.
Figure 13 illustrates comparisons of posterior probabilities at different states among
0.5 P(Xi=N|R=III)-P(Xi=L|R=III)
P(Xi=N|R=III)-P(Xi=H|R=III)
0.4
0.3
0.2
Posterior probabilities
0.1
0.0
X6 X7 X8 X9 X10
-0.1
Mechanical variables
-0.2
-0.3
-0.4
Suspected
Suspected causes
causes
-0.5
mechanical variables, where P(X i = N|R = III), P(X i = L|R = III), and P(X i = H|R
= III) referred to the posterior probability of the variable X i (i = 5, 6,.., 10) at the
state of Normal, Low and High, respectively.
When the result of P(X i = N|R = III) − P(X i = L|R = III) (or P(X i = N|R = III) −
P(X i = H|R = III)) is below zero, the variable X i can be identified as the suspected
variable, which is required to lower its value (or increase its value). Results as shown
in Fig. 13 indicate that X 8 (Drilling Speed) and X 9 (Thrust Ratio) were deemed as
the suspected variables which should be checked and updated in the case of P(T =
III) = 1. For this reason, the practical fault diagnosis and checking should then focus
on the reliability of these two factors and their reasonable parameter values until the
high risk was under control. That is to say, the posterior probabilities can provide
new evidential information for fault diagnosis in real time, and the values of relevant
variables can be updated in a dynamic manner. In this way, the suggestions on which
factor should be checked and how to adjust the suspected factor can be provided on
a quantitative basis.
6 Summaries
Underground tunnel transport has become a popular solution to relieve the pressure of
ground transportation systems worldwide, especially in developing countries such as
China. However, the tunnel-induced road damage has grown to be a public concern
due to its close relation to public safety in recent years. A DBN-based dynamic
risk analysis of the tunnel-induced road damage is developed in this paper. A case
study concerning the dynamic safety analysis of tunnel-induced road damage in the
construction of the Wuhan Yangtze Metro Tunnel in China is presented. Results
demonstrate the feasibility of the proposed approach.
In lights of the case study, some conclusions are drawn up. DBN provides a
powerful tool for knowledge representation and reasoning in a dynamic and complex
project environment. The proposed DBN-based approach can accurately illustrate
the dynamics of geological, design, and mechanical variables as the construction
progress evolves. This overcomes the deficiencies of traditional static fault anal-
ysis methods, such as FTA, CFEM, SCL, and influence diagrams. The proposed
DBN-based approach can perform predictive, sensitivity, and diagnostic analysis
and control on a quantitative basis. The critical and sensitive risks and risk factors
can be determined, and some corresponding safety control measures can then be
proposed in advance for risk reduction before an accident occurs. Moreover, the
suspected variables can be detected in real time in case an accident occurs, assisting
engineers in carrying out the fault diagnosis in a dynamic manner.
To summarize, the proposed DBN-based decision analysis approach can present
the causal relationships with time-dependent variables, as well as perform predic-
tive, sensitivity, and diagnosis analysis. The changes over nodes of DBN can reveal
the evolution in the construction progress, resulting in higher prediction accuracy.
The capability of DBN to model dynamics and uncertainty in the system enables
146 Dynamic Bayesian Networks
References
1. Alaeddini A, Dogan I (2011) Using Bayesian networks for root cause analysis in statistical
process control. Expert Syst Appl 38:11230–11243
2. Cawley GC, Talbot NLC (2007) Preventing over-fitting during model selection via bayesian
regularisation of the hyper-parameters. J Mach Learn Res 8:841–861
3. Cheng CY, Dasari GR, Chow YK, Leung CF (2007) Finite element analysis of tunnel-soil-pile
interaction using displacement controlled model. Tunn Undergr Space Technol 22:450–466
4. Cooper GF, Herskovits E (1992) A Bayesian method for the induction of probabilistic networks
from data. Mach Learn 9:309–347
5. Franco VH, Gitirana GdFN, de Assis AP (2019) Probabilistic assessment of tunneling-induced
building damage. Comput Geotech 113:
6. Hu J, Zhang L, Ma L, Liang W (2011) An integrated safety prognosis model for complex system
based on dynamic Bayesian network and ant colony algorithm. Expert Syst Appl 38:1431–1446
7. Hu Q-F, Huang H, Cheng L (2005) The risk evaluation on effect of shield tunnel construction
on urban asphalt pavement. Underground Space 2:263–267
8. Khakzad N (2019) Modeling wildfire spread in wildland-industrial interfaces using dynamic
Bayesian network. Reliab Eng Syst Saf 189:165–176
9. Khakzad N, Khan F, Amyotte P (2011) Safety analysis in process facilities: comparison of
fault tree and Bayesian network approaches. Reliab Eng Syst Saf 96:925–932
10. Langseth H, Portinale L (2007) Bayesian networks in reliability. Reliab Eng Syst Saf 92:92–108
11. Liao SM, Liu JH, Wang RL, Li ZM (2009) Shield tunneling and environment protection in
Shanghai soft ground. Tunn Undergr Space Technol 24:454–465
12. Liu W, Wu X, Zhang L, Zheng J, Teng J (2017) Global sensitivity analysis of tunnel-induced
building movements by a precise metamodel. J Comput Civil Eng 31:04017037
13. Montani S, Portinale L, Bobbio A, Codetta-Raiteri D (2008) RADYBAN: a tool for reliability
analysis of dynamic fault trees through conversion into dynamic Bayesian networks. Reliab
Eng Syst Saf 93:922–932
14. Onisko A, Druzdzel MJ, Wasyluk H (2001) Learning Bayesian network parameters from small
data sets: application of Noisy-OR gates. Int J Approximate Reasoning 27:165–182
15. Peck RB (1969) Deep excavations and tunnelling in soft ground. In: Proceedings of 7th
International Conference on SMFE, pp 225–290
16. Smith NJ, Merna T, Jobling P (2009) Managing risk: in construction projects. Wiley-Blackwell
17. Špačková O, Straub D (2013) Dynamic Bayesian network for probabilistic modeling of tunnel
excavation processes. Comput-Aided Civil Infrastruct Eng 28:1–21
18. Thomas T (2018) Perth TBMs on hold after sinkhole appears. Tunnel J
19. Wu X, Liu H, Zhang L, Skibniewski MJ, Deng Q, Teng J (2015) A dynamic Bayesian network
based approach to safety decision support in tunnel construction. Reliab Eng Syst Saf 134:157–
168
20. Yang J, Zhang X, Suo H, Lu L, Zhang J, Yan Y (2011) Maximum a posteriori linear regression
for language recognition. Exp Syst Appl 4287–4291
21. Yao G, Sun H, Wei Q (2008) Research on the damaging control criterion of blacktop due to
underground tunnel construction. J Beijing Jiaotong Univ 32:40–43
22. Zhang L, Wu X, Qin Y, Skibniewski MJ, Liu W (2016) Towards a Fuzzy Bayesian network based
approach for safety risk analysis of tunnel-induced pipeline damage. Risk Anal 36:278–301
23. Zhang Y, Zhang L, Wu X (2019) Hybrid BN approach to analyzing risk in tunnel-induced
bridge damage. J Perform Construct Facil 33:04019048
Process Mining
1 Introduction
BIM, as a new digital revolution for civil engineering, provides a collaborative plat-
form to facilitate information exchange and sharing among participants in different
roles for better decision-making [31]. The Architecture, Engineering, and Construc-
tion (AEC) industry has seen a rapid growth of BIM implementation. That is to say,
around 28% of projects use BIM in 2007 and the proportion increases to 71% in
2012 [3]. Notably, it is unreasonable to simply consider BIM as a three-dimensional
(3D) parametric model. Instead, BIM acts as a digital representation of building
characteristics, which allows users to gain quick and easy access to model informa-
tion. BIM can bring benefits more than the improved visualization due to the easy
retrieval of information, such as to enhance productivity, facilitate collaboration, and
reduce mistakes, conflicts, costs, and others. With the growing applications in BIM,
a huge amount of data can be generated and stored in the BIM platform to passively
track detailed activities in the modeling and construction process, leading to some
features of “big data” [16]. Therefore, more focus should be put on deeply mining
such large volumes of BIM data, in order to open a new way for better understanding
the project.
However, knowledge discovery from BIM data is not an easy task. In this regard,
the emerging data mining methodology called process mining can be implemented
as an optional solution to realize the data-driven decision support in terms of
work scheduling, task assignment, and productivity improvement for smart project
management. This can expectedly reduce the project’s risk of failure and improve
project performance.
It should be noted that the BIM event log associated with cases and events is
an important type of BIM data, which contains rich process-specific information
about project evolution [15]. They are shared and updated across the BIM platform
among users for information integration and interoperability. To logically organize
the event log data, a process is defined to consist of a collection of cases, where a case
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 147
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_7
148 Process Mining
In this chapter, several process mining techniques are conducted to uncover knowl-
edge from available BIM event logs in large volumes, aiming to offer a better under-
standing of the dynamics in the design and construction process. It is supposed to
create long-lasting positive impacts on the construction industry that has traditionally
suffered from chronic productivity problems. The main objective is to answer the
following questions: (1) How to automatically capture the behavior from event logs
and construct an understandable process model to describe this observed behavior;
(2) How to analyze the process model for checking compliance, detecting deviations,
predicting delays; (3) How to explore collaborative characteristics to reveal the flow
of work, information, and communication in the organization. Based on the results
obtained from process mining, project managers can easily identify the bottleneck,
collaboration characteristic, and performance preference relying on such valuable
results from process analysis. Insights can are provided to help managers arrange
and adjust workflows more efficiently and rationally, in order to strategically smooth
and speed up the operation process along with the increasing collaboration cohesion
in BIM projects.
2 Process Mining
process is executed. That is to say, the process discovery aims to provide a fast, reli-
able, and cost-effective way for managers to identify large-scale processes with less
manual efforts, which serves as a basis for further analysis. Besides, animations of
the discovered process model can be produced to simulate the execution of events.
(2) Process conformance
Process conformance compares the modeled behavior in the process model and the
observed behavior in the event logs to check whether they are compliant with each
other. It mainly focuses on measuring the quality of the discovered process model
under four dominant dimensions, namely the fitness, precision, generalization, and
simplicity [26]. Regarding the application benefits, it is used for business alignment
and auditing, token replay, and footprints comparison, which assists in identifying
differences between a process model and the actual process in event logs. The detected
deviation can be explained as the undesired exception on the process model, which
does not conform to reality. It can serve as evidence to refine or improve the model.
(3) Process enhancement
Process enhancement is primarily responsible to extend or improve the existing data-
driven process model based on event logs for optimization potential. For instance,
when deviations are detected from the conformance checking, corresponding modi-
fications can be executed on the process model to make it replay most of the event
log. To enrich the process model described by the control flow, a more new perspec-
tive, such as the organizational, case, and time perspective can be added in by cross-
correlating the model with event logs. More specifically, these three new perspectives
concern the organizational structure, properties of cases, and timing and frequency
of events, respectively. Moreover, predictive approaches can also work together with
the process mining for various purposes, such as to cluster traces, identify the root
cause, and predict the activities. They can formulate suggestions to improve process
performance and mitigate risks of bottlenecks in advance.
Many software products for process mining are available to efficiently convert
event logs into process-related views and deliver insightful analytics. The following
software is commonly used for research and practice, including ProM, Disco (Flux-
icon), Celonis, ARIS Process Mining, Myivenio, and others. They aim to clearly
describe the step-by-step process by creating a visual map and then realize the func-
tions of diagnosis, checking, exploration, prediction, recommendation, and others.
At the list of tools, ProM and Disco, which are representative of the non-commercial
and commercial tools, respectively, attract more attention than other tools. ProM [27]
is the leading open-source framework to support process mining, which is easy to
extend. A variety of plug-ins (more than 1500) can be added to the ProM framework,
enabling users to implement the newest methods on their own data. But it could
be unfriendly for beginners. Disco [7], as the commercial process mining software,
is easy-to-use and fast for automated process discovery, process map animation,
deviation controlling, and performance optimization.
It should be noted that process mining is not merely a theoretical subject, which
has played important roles in a wide range of industries. Process mining has been
2 Process Mining 151
The starting point of process mining is the event logs, which map the complicated
process holistically from beginning to end. Typically, an event log is a list of cases,
where a case is a related set of ordered events identified by the unique ID, and thus,
it can be adopted to differentiate different patterns of activities. Some additional
information is added to the event, such as the timestamp to indicate the sequence of
events, the activities, and the resources to represent the participants or other assets
involved in the execution of an activity. Since the BIM event log also belongs to
the log data, it can be defined as a rich source of process-specific information to
capture flows of activities in chronological order, which contain several attributes,
including cases, activities, persons, and time [28]. Herein, the event logs from several
users over the course of working on multiple projects are taken as an example. They
can be generated passively by the Autodesk Revit to record all modeling activities
of designers in executing commands, which are then stored in the Program Files
directory under the Revit Product version folder called Journals [20]. To make good
use of the Revit logs in plain texts, it is necessary to extract several information items
about design activities from the event logs. Figure 1 displays a workflow of the data
extraction process from BIM design event logs, which contains three main steps,
namely data harvesting, data parsing, and data cleaning.
(1) Data Harvesting: In the BIM-based design project, a group of designers work
together on the BIM Revit platform. Each designer generates thousands of
access logs within Revit throughout the whole project. Several projects run
in parallel within a design firm, leading to the generation of a large number
of journal files. To streamline design logs across different projects from
numerous designers, a cloud-based database is developed to regularly harvest
the distributed log files onto the Server Centre via the firm’s Intranet.
(2) Data Parsing: BIM Revit logs are stored in a text format, which makes it chal-
lenging for data analysis. A journal file parser is developed to automatically
152 Process Mining
140901_atlcn0102_John_journal.0001.txt
1. Data
Journal 140901_atlcn0102_John_journal.0002.txt
Harvesting
files 140901_atlcn0102_Mike_journal.0003.txt
created 140901_atlcn0102_Mike_journal.0004.txt Server Logs
by Revit 140901_atlcn0102_Mary_journal.0005.txt 2. Data
140901_atlcn0102_Pete_journal.0006.txt Parsing
User ID
Open
Jrn.Directive "WindowSize" _
, "..\800654_UNT_Student_Union.rvt]", CSV Data File
"Floor Plan: Level 1" _
, 1580, 733 3. Data
Timestamp 'E 03-Oct-2014 14:51:18.036; 0:< Cleaning
Jrn.MouseMove 0 , 28 , 4
Design 'E 03-Oct-2014 14:51:18.036; 0:<
Command Jrn.Command "Ribbon" , "Create a wall ,
ID_OBJECTS_WALL" Cleaned Logs
parse through a folder of journal files and read Revit logs [19]. Several infor-
mation items, such as user ID, project ID, timestamp, and design command,
are pulled out and stored in a Comma Separated Values (CSV) format.
(3) Data Cleaning: In the parsed CSV file, each line represents a design command
that is executed by a designer at a time point. A number of lines are involved
with noise, such as empty or meaningless commands, which creates a need
for removing noisy commands. For instance, in Lines 5 and 12 in Table 1,
Table 1 Example of records in the parsed CSV from BIM design event logs
Line User ID Timestamp Project ID Design command
1 John 4/1/2014 10:16:55 800,645 “Create a wall”
2 John 4/1/2014 16:26:51 801,320 “Pick lines”
3 John 4/2/2014 10:36:45 801,344 “Finish Sketch”
4 John 4/2/2014 11:12:25 801,344 “Create a wall”
5 John 4/2/2014 11:16:25 801,344 “Cancel the current operation”
6 Mike 4/8/2014 09:10:15 801,344 “Pick lines”
7 Mike 4/8/2014 10:01:05 800,645 “Split walls and lines”
8 Mike 4/9/2014 13:21:25 800,645 “Manage Links”
9 Mike 4/9/2014 14:21:11 801,320 “Move selected objects or their
copies”
10 Mary 4/9/2014 08:25:31 800,645 “Finish Sketch”
11 Mary 4/9/2014 12:28:13 801,320 “Create a wall”
12 Mary 4/9/2014 14:22:31 801,320 “Cancel the current operation”
3 Event Logs in BIM 153
the command “Cancel the current operation” shows that the designer presses
the “ESC” key while moving the mouse or in the middle of the execution of a
command, which has no specific meaning for the design and should be removed
from further analysis. Command lines are continuously evaluated, and those
containing valid commands without any null values or errors will be kept as
new data and are placed into the clean logs file.
Worker1 217
217 217
Designer 1 Log File 1 211 2
165 212
5 212
Project 1 8 1
13
1
1
2 115
114
13
Worker4
12
13
2
8
44 Process model
Worker7 115
Designer 2 Log File 2 Intranet 115
13
Project .. 182
180
182
17
2
• Process discovery
Server Log Data
Designer .. Log File ..
Worker3
182
• Conformance checking
(b) Process discovery and analysis • Bottleneck analysis
Project M 51 3 6
50 11
5
13
48 17
Social network
47 19
46 20
15
140901_atlcn0102_John_journal.0001.txt 45
4
21
22 8
Journal 140901_atlcn0102_John_journal.0002.txt 0
44
• Macro-level analysis
43 27 10 24
files 140901_atlcn0102_Mike_journal.0003.txt 42 25
created 140901_atlcn0102_Mike_journal.0004.txt 41 26
by Revit 140901_atlcn0102_Mary_journal.0005.txt
40
39
16 14
29
28
140901_atlcn0102_Pete_journal.0006.txt 37
36
35
12
34 33
32
31
• Correlation analysis
User ID
(c) Social Network Analysis
Jrn.Directive "WindowSize" _
, "..\800654_UNT_Student_Union.rvt]", Decision tree
"Floor Plan: Level 1" _
, 1580, 733
Timestamp 'E 03-Oct-2014 14:51:18.036; 0:<
• Performance
Jrn.MouseMove 0 , 28 , 4 prediction
Design 'E 03-Oct-2014 14:51:18.036; 0:<
Command Jrn.Command "Ribbon" , "Create a wall ,
ID_OBJECTS_WALL"
The first task in process mining is process discovery, which extracts knowledge from
event logs to build rational process models. The discovered model demonstrates
the way the process actually works, which is taken as the basis for further process
analysis and optimization. One of the simple and fundamental algorithms is called
the α-algorithm, which can adopt relations, such as direct succession, causality,
parallel, and others to output the Petri net along with animations. Although the α-
algorithm is sufficient to catch the concurrency in the process, it is unable to deal
with problems arising from the unstructured process, including noise, incomplete
information, non-local dependencies, short loops, and others. In the meantime, the
α-algorithm tends to generate useless spaghetti-like models containing complete
processes with all details, which is hard to distinguish important and non-important
process information in noisy and less-structured logs. To overcome these limitations,
some advanced algorithms for process discovery have been developed. Herein, two
popular algorithms, namely fuzzy mining and inductive mining are introduced.
(1) Fuzzy mining. For model simplification and interpretability, fuzzy mining [6,
8] is proposed to display suitable abstractions or aggregations of the observed
process graphically using a map metaphor. It pays attention to the subsets of
the most significant behavior within the process, which is especially effective
in visualizing correlations of activities in a less-structured process. The funda-
mental idea of fuzzy mining lies in configuring two metrics named significance
and correlation. Significance representing the relative importance of behavior
can be easily quantified by frequency of events and routings, while correlation
estimates the closeness degree between two events. An efficient method for
simplifying the process model is to only preserve the high-level information.
Therefore, it is necessary to delete undesirable events and relations with both
low significance and correlation. As for the less significant but highly corre-
lated behavior, they need to be aggregated into clusters. From the map-like view
of abstract process models, primitive and cluster nodes are linked by edges in
different widths and colors representing relative significance and correlation
after conflict resolution and edge filtering. Besides, fuzzy mining has taken
effect in some practical applications, especially for interactively simplifying
models and investigating frequency and time duration [9, 10, 18]. However,
fuzzy mining is prone to suffer from unfitness and unsoundness due to its
deliberately imprecise model.
(2) Inductive mining: As an improvement over α-algorithm and fuzzy mining, a
promising technique termed inductive mining [11] is developed to tackle infre-
quent behavior and huge models. It belongs to a variation of the α-algorithm,
which can create the block-structured process in great fidelity, soundness, and
fitness. The implementation of inductive mining starts by recursively splitting
original event logs into sub logs. Then, each sub log is used to construct a
directly-follow graph which is defined by a set of activities by nodes and their
execution sequences by directed edges. As a result, the directly-follow graph
4 Process Mining Framework 155
is cut into smaller ones to partition the set of nodes into distinct subsets under
four kinds of cuts: exclusive choice, sequential composition, parallel compo-
sition, and the re-do loop. Overall, the splitting procedure will repeat until
every subset is only comprised of one node (activity). It has been proved that
inductive mining is flexible in creating a reasonable process model with no
duplicated activities, which can be fit and sound to the observed behavior in
the event log.
A process model serves as a concise description of the complicated process based
on the facts from event logs. It is typically in the graphical format to record the
workflow along with the related process knowledge, which helps people to better
understand the execution sequences and dependencies in a series of activities. To
represent the discovered process model, there are three common types, namely Petri
net, business process modeling notation (BPMN), and process tree. They can provide
different views of the discovered process in reality.
(1) Petri net: Petri net originally developed in the late 1960s is one of the oldest and
the most basic process modeling languages under the combination of mathe-
matical formalism and a graphical representation [17]. It is a bipartite graph
consisting of place, transition, arc, and token. To be specific, places and tran-
sitions are connected by a collection of directed arcs. Tokens flow through
the network, which are hold by places. The superiority of the Petri net lies
in its ability in demonstrating both the concurrency and asynchrony nature of
processes. Figure 3a shows a simple example of the Petri net model, where
circles represent the places, squares denote the transitions, and arcs indicate
various associations.
(2) BPMN. BPMN is a flow chart method widely used in business process manage-
ment. Two kinds of notations are incorporated to visually describe a detailed
sequence of business activities and information flows. One is a kind of activity
nodes on behalf of business events, which is comparable to a place in the
Petri net. The other is a kind of control nodes, where sequential and message
flows are used to indicate the flows and logic between activities. BPMN is
analogous to the Petri net to some extent, but it provides a more comprehen-
sive set of elements to express the flow behavior. That is to say, BPMN is a
high-level notation to offer an intuitive visual representation of the compli-
cated processes, which can be easier to understand even for people with no
professional knowledge. The BPMN in Fig. 3b has a similar meaning as the
Petri net in Fig. 3a.
(3) Process tree. Since the Petri net is likely to experience deadlocks and some
anomalies, the process tree can be another optional graph notation to ensure
the soundness of representations, which is especially beneficial to the inductive
process discovery. To be specific, a process tree is a block-structured workflow
net in a hierarchical structure, where the inner nodes refer to operators and the
leaves are labeled with activities. The Petri net in Fig. 3a is convertible to the
process tree in Fig. 3c.
156 Process Mining
p3 t4 t5 p6
(a)
XOR-split Gateway
t2 XOR-join Gateway
AND-split Gateway AND-join gateway
t1 t3 t6
t4 t5
(b)
AND
XOR SEQ
t2 t3 t4 t5
(c)
Fig. 3 Examples of: a Petri nets; b BPMN; c process tree. Note AND means parallel composition,
XOR means exclusive choice, and SEQ means sequential composition
Process diagnosis aims to check the conformance and determine the discrepancies
between the observed process in event logs and the discovered process model. It is
closely relevant to the quality evaluation of the discovered model. Herein, two major
quality dimensions called fitness and precision are used to check the model quality,
as presented below [2].
(1) Fitness. Fitness is the most dominant metric to measure the model’s competence
in replaying the event log. It can be defined by Eq. (1) to clarify how much of
the observed behavior can be demonstrated in the process model. Based on an
alignment-based calculation, the fitness can be obtained by counting deviations
in replaying events on the top of the model [24]. During the process of aligning
events to the process model, the cost should be given when events are skipped
or activities are inserted with no expectation. The better fitness is, the value of
Qf is closer to 1 where the cases in the model can well replay the log with fewer
problems. On the contrary, the fitness of 0 signifies that the process model is
4 Process Mining Framework 157
unable to replay traces in the log. For example, a process model in the fitness
of 0.95 means that it fails to explain 5% of the observed behavior in the event
log. It should be also noted that a process model with a comparatively high
fitness may be overfitting.
f cost (L , M)
Qf = 1− (1)
move L (L) + |L| × move M (M)
where fcost(L,M) represents the total alignment cost for event log L and model
M. For example, if fcost(L, M) = 0, it means that the model M can perfectly
replay the log L. For the denominator, it refers to the maximal possible cost,
where moveL (L) is the cost of moving through logs rather than the model,
and moveM (M) is the cost only in the model. It is used to normalize the total
alignment cost.
(2) Precision: Precision in Eq. (2) is the proportion of actual behavior in logs to
the behavior allowed by the process model, indicating how well the simple
process model with minimal behavior can describe the event log [24]. It can
be observed that a poor precision approaching 0 can be caused by |enL (e)|
|enM (e)|, possibly leading to underfitting. This would imply that behaviors in
the process model are quite different from the event log. When almost all of
the behavior in the process model can be actually seen in the log, it returns
a high precision approaching the value of 1. That is to say, the lack of great
fitness or precision is likely to cause an oversimplified process model.
1 |en L (e)|
Qp = (2)
|E| e∈E |en M (e)|
where |enM (e)| represents the number of activities enabled in the model M,
|enL (e)| refers to the number of actual activities executed in event logs L under
the similar context, e ∈ E is events, and |E| is the number of events in logs L.
Fig. 4 Example of a
decision tree Rainy ?
classes. The strategy of the node splitting heavily influences the predictive accuracy.
Herein, the widely used decision criterion named the Gini index is introduced, which
calculates the probability of a feature classified incorrectly when selected randomly.
As defined in Eq. (3), the Gini index can be derived from subtracting the sum of the
squared probability of each class from one, which is relatively easy to implement
even for large partitions. In particular, it is deployed in the classical CART algorithm
to create binary splits in a reasonable manner, which also performs well for nonlinear
datasets. The value of the Gini index is between 0 and 1, where 0 indicates the purity
of classification. Inversely, 1 means the data on the node is not pure at all, and thus,
several new nodes need to be generated from this node.
n
Gini = 1 − ( pi )2 (3)
i=1
where pi represents the probability of an element being classified for a distinct class.
Since a group of people will involve in the construction project to finish the common
task together, their complicated collaboration can be reasonably mapped into a social
network for deeper analysis. A social network is a graphical model to intuitively
describe the interaction structure and characteristic, where nodes standing for people
who are connected by directed and weighted links. Besides, the network can be
mathematically expressed by a weighted sociogram (P, R, W) representing relations
among multiple performers, where P, R, W refer to performers, relations, and weights
of relations, respectively. In subsequence, social network analysis (SNA) can be
implemented on the built network to explore hidden knowledge about information
spreading and performers’ behavior from the following two levels:
(1) SNA at macro-level: SNA at the macro-level aims to analyze and measure
the characteristics of the entire social network established. Measurements at
4 Process Mining Framework 159
the macro-level include the network size (S), density (D), average degree
(AD), mean distance (MD), diameter (ND), and clustering coefficient (CC),
as defined in Eqs. (4)–(8) [22]. Information can flow more easily in dense
networks compared to sparse networks, as there are multiple paths between
the nodes in the dense network.
E
D= (4)
N (N − 1)
E
AD = (5)
N
i= j d(n i , n j )
MD = (6)
N (N − 1)
N 2ei
CC = (8)
i=1 ki (ki − 1)
where E and N are the total number of ties and nodes in the network, respec-
tively, d(ni , nj ) is the shortest distance between the ith node and jth node, and
k i is the number of neighbors for the ith node.
(2) SNA at micro-level: SNA at the micro-level analysis aims to analyze and
measure characteristics of individuals at the established social network. It can
be conducted in two aspects. For one thing, the degree centrality C D can be used
to evaluate the position of a single node in the whole network. Equation (9)
defines the degree centrality for a certain node ni , which counts the number
of immediate links of an individual with others. A higher degree of centrality
indicates a higher importance of the node. For another thing, the Spearman
rank correlation can uncover the correlation between the degree centrality and
the production performance of individuals. It attempts to discover whether
individuals at the center of the interaction map are those most reachable for
others and assess the correlation between their centrality performance and
productivity. We will investigate whether the designers with high productivity
records in historical projects will gradually grow to be placed in core positions
in the working process.
j
1 ei, j
C D (n i ) = (9)
N −1
where j is the number of nodes linked to the node ni , and ei,j is the edge between
the ith node and the jth node.
160 Process Mining
Some typical applications of process mining techniques are introduced in this section,
which can fully explore the BIM design event log from different perspectives. The
main objective is to facilitate a better understanding of the design process and to
enhance productivity in BIM-enabled collaborative practices.
A large dataset of event logs recording BIM Revit design applications throughout
several projects is collected from a major international design firm in the construction
industry. It is a Structured Query Language (SQL) database built in the design firm
for data sharing and storage. The design event logs are harvested and stored at the
Server Center on a weekly basis. The timeline of the design event logs collected is
from March 2013 to June 2014. The Journal File Parser is used to automatically read
the unstructured text files in order to capture the records of design commands. After
the data parsing process, 687,143 lines of design commands are generated and stored
in a CSV file, of which each line corresponds to the information of user ID, project
ID, timestamp, and the context of the command executed. By removing empty and
meaningless commands, 620,492 lines of valid commands are kept. The cleaned
dataset indicates that 51 designers are working together on 82 different projects.
Within the studied period, the production of designers varies significantly.
Figure 5a illustrates the scatter of the number of valid commands executed by different
90000 12
Y (Number of commands executed) 11.38 (max)
80000 11
Group I:
Y (Number of commands executed)
70000 High-producƟon
10 designers
60000 9.19
9
50000 Group II:
In (Y)
8 Median-producƟon
40000 In (Y) designers
Linear Fit of In (Y) 7.00
7
Equation y = a + b*x
30000 In (Y)
Plot
Weight No Weighting Group III:
6 Intercept 11.239 [0.128]
20000 Slope -0.119 [0.004] Low-producƟon
Residual Sum of Square 9.94769 designers
Pearson's r -0.96975
10000 5 R-Square(COD) 0.94042
Adj. R-Square 0.9392
4.80 (min)
0 4
0 4 8 12 16 20 24 28 32 36 40 44 48 52 0 4 8 12 16 20 24 28 32 36 40 44 48 52
Designer ID Designer ID
(a) (b)
Fig. 5 Classification of designers by accounting the number of valid commands executed: a scatter
of the number of commands executed by different designers; b three groups of designers (high-
production, median-production, and low- production designers). Note In Fig. 5b, the red dot line
is a linear fit of the scatter diagram, where the ordinate values are natural logarithms of ordinate
values (number of commands executed) as shown in Fig. 5a for purpose of a better classification
resolution
5 Typical Applications in BIM Process Discovery 161
designers, where all the 51 designers are represented by “#1”, “#2”, …, “#51”, respec-
tively. It is clear to see some designers, like “#1” ~ “#10”, contribute more than
20,000 valid commands, while some designers, like “#40” ~ “#51”, contribute less
than 2000 valid commands. To further investigate the correlation between network
characteristics and the production performance of designers, all the 51 designers are
classified into three separate groups for comparison study presented in later sections.
Figure 5b illustrates the results of the classified three groups of designers, namely
high-production, median-production, and low-production designers, where there are
17 members in each group.
Herein, we take a part of the prepared BIM design event log concerning seven workers
as an example. The targeted event log records how the tasks are transmitted among
the seven workers in detail, which is then fed into the inductive mining algorithm
for process discovery. To emphasize the critical relationships in the collaboration,
workers who are responsible for less than 2% of tasks are automatically removed
from the process. As a result, the process model can be visualized in three forms,
namely the Petri net, BPMN, and process tree. Workers 1, 3, 4, 7 are taken into
account in the discovered model, which work together to complete around 96.61%
of tasks. As shown in Fig. 6, all three process models can intuitively describe the
activities executed by workers from a “worker-centered” view, which are explained
below.
The Petri net in Fig. 6a is the basic process model, which has 10 places, 17 transi-
tions, and 34 arcs. More specifically, transitions in boxes represent workers to model
the possible process status. When tokens are input into the place, the corresponding
transition is activated. In this case, the important causal relationship between workers
is the XOR operation. Based on the XOR node in Fig. 6a, c and the XOR-split
gateway in Fig. 6b at the beginning of the process model, we can simply divide the
seven workers into three clusters. That is to say, the first cluster is made up of Workers
1 and 4, the second cluster contains Worker 7, and the remaining workers are in the
third cluster. It should be noted that the worker in a certain cluster is unable to take
over tasks executed by workers from other clusters.
Since the quality of the process model largely influences the reliability of analytical
results, it is an essential step to check how well the process model can reflect the
behavior captured in the event log. In this regard, the plug-in in the ProM tool for
a conformance checker is used to generate the Petri net annotated with diagnostic
information about the deviations between the log and the model, which is provided
162 Process Mining
in Fig. 7. To be more specific, the number in the box indicates how many tasks the
worker performs, and the number above the arc states how many times the process
traverses between participants. The red dash lines highlight the part in the process
model, which cannot conform with the event log. It is known that two types of
deviations exist in the process model [12]. The event, which appears on the event log
but does not reflect in the process, causes the log-move deviation. The event, which
is required by the model but does not occur in reality, is the model-move deviation.
It is clear to see that Fig. 7 only owns the log-move deviation, which is repre-
sented by the self-arc. The metrics of fitness and precision (expressed by Eqs. (1)
and (2), respectively) can be calculated to measure the discovered process model in
a numerical way. Table 2 outlines the evaluation results. Since no model-move devi-
ation appears in the model, the corresponding fitness reaches a high value of 1. The
log-move fitness is only 0.791, which is still acceptable for the model to demonstrate
a great re-discoverability property. In addition, all kinds of precision are greater than
0.85, indicating that there is no obvious underfitting in the model. Therefore, the
discovered process model from inductive mining is proven reasonable.
An effective method to ensure a smooth process is to shorten the unnecessary
waiting times as much as possible. Therefore, an important aspect of process analysis
can investigate the frequency and duration, in order to identify the key parts and
potential bottlenecks within the process. The Disco Fluxicon tool based on fuzzy
mining is deployed for this goal. It can output two simplified process maps as shown
in Fig. 8, which are analyzed as follows.
Fig. 8 Simplified process map fuzzy mining focusing on: a absolute frequency; b median duration
5 Typical Applications in BIM Process Discovery 165
(1) Absolute frequency view. The frequency in Fig. 8a states how often an activity
occurs. It is assumed that the higher frequency can indicate the greater impor-
tance of the worker or the process. Clearly, Workers 7, 1, and 3 are the three
most critical persons playing the central role in the construction project, who
are more active to finish about 36.34%, 33.98%, and 23. 70% of the total
tasks, respectively. As for the core process paths, the top two with the highest
frequency are from Worker 1 to Worker 1 (80), and from Worker 3 to Worker 1
(30). That is to say, the core paths are more likely to be executed by the critical
workers.
(2) Median duration view. Figure 8b concentrates on the performance metric
named median duration, which reflects the median time between the start and
the end of an activity. Since the deeper red color denotes a longer execution
time, it is observed that Worker 7 takes more service time than others. This is
reasonable since Worker 7 is the most key person in the collaboration. Besides,
the thicker and redder arrows highlight the potential bottlenecks, where the
longer waiting time is spent on task transmission between two workers. As
shown in Fig. 7b, the path from Worker 1 to Worker 6, from Worker 1 to
Worker 11, and from Worker 11 to Worker 1 are the three most problem-
atic sequences, which spend comparatively a longer time than others to cause
severe performance bottlenecks. Therefore, these identified bottlenecks can
guide the measurements for improving process efficiency. It is worth noting
that the execution process in paths going in or out Worker 1 tends to take a
longer time, which means the path related to Worker 1 is possible to exert great
impacts on the delay. Since Worker 1 has been regarded as a critical person,
it can be reasonably assumed that bottlenecks are likely to arise from the key
participants who are expected to conduct more tasks.
In this section, the decision tree model is used to predict the worker’s active time,
which can provide evidence to guide the manager in planning reasonable work
arrangements in a data-driven manner. All records about a certain worker named
“Tom” are extracted from the BIM design event log as a dataset for the case study.
The prepared dataset has in total 3240 records, from which two features called Time
(x1) and Weekday (x2) are extracted and one label termed is Work are prepared as
the foundation of prediction application. Figure 9 visualizes the distribution of two
extracted features. To be more specific, Time (x1) denotes one hour in a day, which
is given the range 0–23 on behalf of a one-hour time slot. For example, a value
of 9 indicates the time interval of 9:00–10:00. Weekday (x2) refers to Monday to
Sunday, which is represented by the index 1–7. Apart from features, each record also
owns one label termed isWork for the classification task. The label has two values:
1 means ‘worked’ and 0 means ‘didn’t work’. That is to say, the decision tree model
166 Process Mining
300
400
Count
Count
200
200
100
0 0
0 10 20 2 4 6 8
Time Weekday
will intelligently learn data about the two features, in order to make predictions about
whether the worker will work at a certain time.
To train and test the decision tree model, the prepared dataset needs to be split
into an 80%–20% ratio, which is a common practice for building and evaluating a
decision tree automatically. Herein, we take out a part of the decision tree diagram
as an example, which is provided in Fig. 10. Observably, the root node and the lead
nods all include criteria to be answered, which are typically extended to two branches
demonstrating the flow from the question to the certain answer. Regarding the root
node, it raises the question that whether the designer works before 7:30 am, which
needs to answer “True” or “False”. The “gini” indicates the Gini index at this node.
The term “value = [2490, 750]” states that when no further judgment is made, the
possibility of working before 7:30 am can be roughly calculated as 2490/3240. In
subsequence, the leaf node offers the “True” response, which continues to make
decisions on the new criteria “x1 ≤ 6.5”. As a result, the outcomes from the decision
tree are determined by the hierarchical relationships in a series of decisions. The
decision tree is easy to follow and comprehend, which provides a unique opportunity
in predicting the working time of a targeted worker.
For the model validation, Table 3 presents the promising classification perfor-
mance of the developed decision tree. There are in total 2683 records to be classified
correctly, resulting in an overall accuracy of 0.828. For all records that are labeled
as “work”, 97.7% of them indicate the worker is active at that time. For all records
that truly represent “work”, 79.5% of them can be given the correct labels. F1-score
balances the precision and recall by considering both the false positives and false
negatives, which has an acceptable value of 0.877. That is to say, the decision tree is
proven to reliably predict the worker’s active time. It serves as convincing evidence
for managers to not only understand worker’s working performance, but also define
an appropriate staffing strategy for a project with the expectation of raising project
efficiency.
For the social network analysis, the whole event log is taken into account, which
actually contains information about the executed commands by 51 designers from
the design firm. They cooperate to work on 82 projects and contribute 620,492
commands to the projects. Figure 11 displays an example of the collaboration among
four designers across multiple projects during the design phase. A social network
extracted from the whole event logs is displayed in Fig. 12. It is an undirected
network, where points represent all the designers involved, and the width of edges
#1 1 Project 1
2
2 (800645)
2
#2 1
Project 2
(801320)
1
Designers #3 2 Projects
1 2 Project 3
(801344)
#4 2
Collaboration relations
Fig. 11 An example of collaboration during the modeling process. Note The number in the arrows
indicate total valid commands a designer contributes to a project
168 Process Mining
High-production designers
Median-production designers
Low-production designers
Fig. 12 Map of the discovered social network among 51 designers in BIM-based collaborative
design
represents the weight of the relation. Table 4 summarizes the characteristics of the
network structure at the macro-level. The average degree (10.078) and the density
(0.202) indicate that the whole network of collaboration is finely intensive. The value
of the mean distance between nodes (1.984) points out that a worker can easily be
linked with another one in a few short steps. As for the network density and diameter,
Table 4 Characteristics of
Characteristics Value
the social network from the
event logs Number of nodes 51
Number of links 514
Density (D) 0.202
Average degree (AD) 10.078
Mean distance (MD) 1.984
Diameter (ND) 4.00
Clustering coefficient (CC) 0.771
5 Typical Applications in BIM Process Discovery 169
11 51 3 6
50 11
3 9
49 13
51 19
21 5
48 17
42 47 4 25 5 1
47 19
1
46 20
50 8 7 15
46 27
48 14 45 21
49 20
18 4
22 8
44 23
34 2 6
0.00 0.07 0.14 0.21 0.28 0.35 0.42 0.49 0.56 0.63 24 0 219 438 657 876 1095 1315 1534 1753 1972 2
12 32 43 10 24
44 36 27
33 17
42 25
7
45 23
41 26
41 37
30 10
22 9 13 40 28
31 29
16 14
39 29
40 15 38
38 30
39 18
43 28
26 16 37 31
12
High-producƟon designers 36 32
35 34 33
35 Median-producƟon designers
Low-producƟon designers
(a) (b)
Fig. 13 Concentric map of centrality analysis among 51 designers by a degree centrality, and b link
centrality
20.2% potential links actually appear in the network, and the longest length of all
the shortest paths between node pairs is 4. The above-mentioned metrics all verify
that there is a cohesive collaboration hidden in the established network. Besides, the
clustering coefficient of the network (0.771) implies that the network can possibly be
divided into several clusters to facilitate better collaboration, and thus, connections
among designers in the same cluster are more intense.
The micro-level SNA concerns the degree centrality to investigate the embedded-
ness of each node and the position of the node in the whole network. Two measures
in this type are computed, including degree centrality by link and by weight. To
be more specific, degree centrality by link measures the width of the collabora-
tion between a designer and others, while degree centrality by weight measures the
depth of collaboration. Figure 13 illustrates the concentric map of centrality analysis
among 51 designers by degree centrality and link centrality. It can be seen that the
top three designers ranked in the degree centrality by the link are designers “#24”,
“#2”, and “#18”, while the top three designers ranked in the degree centrality by
weight are designers “#2”, “#7”, and “#4”. This indicates that these designers are
highly active and interact frequently with other designers in the network. These
designers are located in the center of the concentric map (as seen in Fig. 13) and
other designers would be likely to build up interactions with them. For instance,
designer “#2” is considerably active not only in the width but also the depth of
the collaborative network. In general, all these designers, including designers “#2”,
“#24”, “#7”, “#24”, and “#4”, can be regarded as “stars” in the network of BIM-based
collaborative design.
Moreover, the relationship between network characteristics and the production
performance of designers deserves further exploration. Correlation analysis can be
conducted between the measured node centralities and the recorded commands of the
170 Process Mining
Number of commands (Production) Linear Fit Number of commands (Production) Linear Fit without outliers
100000 100000
90000 90000
Y: Number of commands (Production)
70000 70000
Linear F it: Linear F it without outliers :
60000 60000
Y = 26.73 * X + 4405 Y = 24.82 * X + 2196
50000 Adj.R-Square = 0.3790 50000 Adj.R-Square = 0.7113
40000 40000
30000 30000
20000 20000
10000 10000
0 0
-200 0 200 400 600 800 1000 1200 1400 1600 1800 2000 2200 2400 -200 0 200 400 600 800 1000 1200 1400 1600 1800 2000 2200 2400
Fig. 14 Linear fits between network characteristics (node degree centrality by weight) and produc-
tion performance of designers (number of commands executed): a with all data points; b with
outliers masked
involved designers within the organization. Considering the nature of the underlying
data, nonparametric Spearman correlation analysis is used to assess the relationships
between degree centrality and production performance of designers (represented
by the number of commands executed during the studied period). It turns out that
the correlation coefficients for the measured node degree centrality by link and by
weight are 0.253 (at a 10% significance level), and 0.805 (at a 1% significance
level), respectively. Thus, the calculated degree centrality by weight has the greatest
association with the production performance of designers, which can be described
by a linear regression test. Figure 14 illustrates the results of the linear fit. When all
data points are considered, an increase of one unit in the degree centrality by weight
generates an increase of 26.73 units in the production performance, associated with
an adjusted R-Square of 0.3790 (see Fig. 14a). When two outliers (marked in red
in Fig. 14b) are masked, the value of the adjusted R-Square is sharply increased
to 0.7113, indicating a strong interpretation for the regression fit. In other words,
the measured node degree centrality by weight is capable of explaining the greatest
percentage of variations (71.13%) in the production performance of designers. An
increase of one unit in the degree centrality by weight probably leads to an increase
of 24.82 units in the production performance of designers.
6 Summarizes
With the massive growth of BIM applications in the AEC/FM industry, large volumes
of computer-generated event logs are available to provide rich data resources. It is
worth noting that process mining provides a good opportunity to make good use
of these event logs. Various techniques from the process mining, including process
discovery, conformance checking, social network analysis, and decision tree, are
6 Summarizes 171
carried out to extract valuable information from event logs, in order to capture and
study the nature of the complicated workflow and collaboration. The analytical results
provide an objective approach to assess worker’s performance, which can assist
managers to make critical staffing decisions and optimize the workflow for enhancing
the likelihood of success of the BIM-enabled project.
In light of the application cases, the process mining has been successfully imple-
mented in a real BIM design event log to improve the understanding of the building
design development process. The benefits of the proposed approach can be summa-
rized as (1) Key processes hidden in the event logs can be efficiently mapped into the
process model based on the advanced process discovery algorithm, such as induc-
tive mining and fuzzy mining; (2) Conformance checking can not only verify the
correctness the discovered process model by metrics (i.e., fitness, precision), but also
identify the potential bottlenecks causing undesired delays for process optimization;
(c) Social network can be extracted from the event logs, which can be deeply analyzed
at different levels to reveal characteristics of collaboration within an organization; (d)
Decision tree as a supervised learning algorithm can intelligently predict a worker’s
work behavior during the design process, which provides managers some reference
to evaluate the worker’s performances and make appropriate work schedule planning.
From a broad perspective, process mining from event logs mainly provides two
kinds of merits. For one thing, hidden knowledge behind event logs can be revealed
and visualized, which can be used to monitor and analyze the project process in a
data-driven manner. For another, new insights can be provided into the execution
phase of the project to detect the potential deviations, delays, collaboration patterns,
and active time, and thus, project managers can reasonably adjust the process and
allocate workers to make the project perform better.
References
1. Ailenei I, Rozinat A, Eckert A, van der Aalst WM (2011) Definition and validation of process
mining use cases. In: International conference on business process management. Springer, pp
75–86
2. Buijs JC, Van Dongen BF, van Der Aalst WM (2012) On the role of fitness, precision, gener-
alization and simplicity in process discovery. In: OTM confederated international conferences
on the move to meaningful internet systems. Springer, pp 305–322
3. Construction M-H (2012) The business value of BIM in North America: multi-year trend
analysis and user ratings (2007–2012). Smart market report, pp 1–72. https://bimforum.
org/wp-content/uploads/2012/2012/MHC-Business-Value-of-BIM-in-North-America-2007-
2012-SMR.pdf
4. De Weerdt J, De Backer M, Vanthienen J, Baesens B (2012) A multi-dimensional quality
assessment of state-of-the-art process discovery algorithms using real-life event logs. Inf Syst
37:654–676
5. dos Santos Garcia C, Meincheim A, Junior ERF, Dallagassa MR, Sato DMV, Carvalho DR,
Santos EAP, Scalabrin EE (2019) Process mining techniques and applications–a systematic
mapping study. Expert Syst Appl 133:260–295
6. Günther CW (2009) Process mining in flexible environments
7. Günther CW, Rozinat A (2012) Disco: discover your processes. BPM (Demos) 940:40–44
172 Process Mining
8. Günther CW, Van Der Aalst WM (2007) Fuzzy mining–adaptive process simplification based
on multi-perspective metrics. In: International conference on business process management.
Springer, pp 328–343
9. Gurgen Erdogan T, Tarhan A (2018) A goal-driven evaluation method based on process mining
for healthcare processes. Appl Sci 8:894
10. Jaisook P, Premchaiswadi W (2015) Time performance analysis of medical treatment processes
by using disco. In: 2015 13th international conference on ICT and knowledge engineering (ICT
& Knowledge Engineering 2015). IEEE, pp 110–115
11. Leemans SJ, Fahland D, van der Aalst WM (2013) Discovering block-structured process models
from event logs-a constructive approach. In: International conference on applications and theory
of Petri nets and concurrency. Springer, pp 311–329
12. Leemans SJ, Fahland D, Van Der Aalst WM (2014) Process and deviation exploration with
inductive visual miner. BPM (Demos) 1295:8
13. Pan Y, Zhang L (2020a) BIM log mining: exploring design productivity characteristics. Autom
Constr 109:102997
14. Pan Y, Zhang L (2020b) BIM log mining: learning and predicting design commands. Autom
Constr 112:103107
15. Pan Y, Zhang L, Skibniewski MJ (2020) Clustering of designers based on building information
modeling event logs. Comput-Aided Civ Infrastruct Eng 35:701–718
16. Peng Y, Lin J-R, Zhang J-P, Hu Z-Z (2017) A hybrid data mining approach on BIM-based
building operation and maintenance. Build Environ 126:483–495
17. Petri C (1962) Kommunikation mit Automaten//Ph. D. thesis. Universitat Bonn, Schriften des
Instituts fur Instrumentelle Mathematik, Germany (in German)
18. Premchaiswadi W, Porouhan P (2015) Process modeling and bottleneck mining in online
peer-review systems. Springerplus 4:1–18
19. Revit A (2011) J File Parser. http://revitclinic.typepad.com/my_weblog/2011/11/journal-file-
parser.html. 18 Nov 2011
20. Revit A (2017) About J Files. https://knowledge.autodesk.com/support/revit-products/getting-
started/caas/CloudHelp/cloudhelp/2017/ENU/Revit-GetStarted/files/GUID-477C6854-2724-
4B5D-8B95-9657B636C48D-htm.html
21. Rovani M, Maggi FM, De Leoni M, Van Der Aalst WM (2015) Declarative process mining in
healthcare. Expert Syst Appl 42:9236–9251
22. Scott J (2017) Social network analysis. Sage
23. Van der Aalst W (2016) Data science in action. Process Mining. Springer, Heidelberg.
24. Van der Aalst W, Adriansyah A, van Dongen B (2012) Replaying history on process models
for conformance checking and performance analysis. Wiley Interdisc Rev Data Min Knowl
Disc 2:182–192
25. Van der Aalst WM (2011) Using process mining to bridge the gap between BI and BPM. IEEE
Comput 44:77–80
26. van der Aalst WM (2012) What makes a good process model? Softw Syst Model 11:557–569
27. Van Dongen BF, de Medeiros AKA, Verbeek H, Weijters A, van Der Aalst WM (2005) The
ProM framework: a new era in process mining tool support. In: International conference on
application and theory of petri nets. Springer, pp 444–454
28. Yarmohammadi S, Pourabolghasem R, Castro-Lacouture D (2017) Mining implicit 3D
modeling patterns from unstructured temporal BIM log text data. Autom Constr 81:17–24
29. Zhang L, Ashuri B (2018) BIM log mining: discovering social networks. Autom Constr 91:31–
43
30. Zhang L, Wen M, Ashuri B (2018) BIM log mining: measuring design productivity. J Comput
Civ Eng 32:04017071
31. Zhang L, Wu X, Ding L, Skibniewski MJ, Lu Y (2016) Bim-based risk identification system
in tunnel construction. J Civ Eng Manage 22:529–539
Agent-Based Simulation
1 Introduction
In recent years, with the continuous and rapid development of the urban economy,
a large number of people have migrated into cities, especially in developing and
thriving countries such as China and India, imposing a rigorous challenge on urban
transportation systems [12]. Due to the advantages of large transportation capacity,
low pollution, high speed, easy traffic, and low energy consumption, subways are
regarded as the modernized urban rail transportation system, playing a significant
role in transiting large passenger flows in large cities [28]. However, the predominant
disadvantages of the subway systems are densely distributed occupants and highly
restrained evacuation since a number of people usually reach durative summits in
density during the operating time, especially traffic peaks and holidays [21]. Due to
the heavily crowded population and the situation of underground space, the magni-
tude of injuries and casualties will be overwhelmingly disastrous in case of an emer-
gency [4]. In general, how to guide pedestrians to find out a more efficient route
for evacuation under emergency conditions attracts the public’s attention, and it is,
therefore, necessary to investigate a solution for optimizing evacuation route planning
strategies in metro stations.
The evacuation of pedestrians in metro stations is identified as a key issue in the
safety management of subway systems [29]. The challenges lie in the pedestrians
who are generally considered as the largest uncertainties since their motives and
actions are dynamic and independent [5]. Evacuation route planning is a strategy
that seeks to minimize the loss of life or harm to public security. In recent years,
the pedestrian evacuation has attracted more and more attention so far, where the
behavioral and psychological features of pedestrians and evacuation models have
been studied.
Existing pedestrian movement and behavior models mainly focus on pedestrian
evacuation simulation and route choice selection. Those models are most often
stochastic and oriented towards pedestrian dynamics and seldomly incorporate the
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 173
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_8
174 Agent-Based Simulation
interactions between pedestrians and the traffic [27]. It is easy to understand that one
pedestrian will move fast when the traffic is smooth while the pedestrian will move
slowly when the traffic is crowded. How to consider the interactions between pedes-
trians and the traffic to plan a more efficient evacuation route in heavily over-crowded
metro stations is a challenging task. Metro stations are generally situated under-
ground, where a large number of pedestrians get out from trains underground and
stream into escalators to upstairs in a constant manner, particularly in metro junction
hubs. If the pedestrians are not evacuated properly, especially in rush hours, newly
arrived passengers from the trains will increase the total volume of the pedestrians
stranded underground, which is very likely to result in stampede accidents.
The main research questions include: (i) What kind of factors should be considered
to come up with the best evacuation planning strategy during the pedestrian evacu-
ation in metro stations; (ii) How to validate the pedestrian movement and behavior
models that are stochastic and oriented towards pedestrian dynamics; (iii) How to
assess and improve the efficiency of the evacuation performance. To address those
research questions, a systematic multi-attribute decision approach with detailed step-
by-step procedures is developed for route choice planning. One metro station in
China is selected as a case to simulate the pedestrian evacuation process, and obser-
vations are compared with simulation results for model validation. The performance
efficiency of the proposed evacuation route planning strategy is assessed using the
proposed performance metrics.
2 Agent-Based Simulation
For systems with a complex structure and over-numbered entities, e.g., people, vehi-
cles (cars/trains/planes), equipment, etc., it is difficult to directly observe and model
from the system level whereas the behaviors of the entities are relatively accessible
from the entity level [11]. For such complex systems, the agent-based simulation
(ABS) approaches have been proposed by considering the entities as agents [14] by
adopting the “bottom-to-top” perspective that focuses more on the dynamics and
procedures of the targeted system [15].
Normally, ABS is implemented in an iterative fashion [19]. First, the experts
determine the parameters based on their knowledge and understanding of the current
conditions. Next, the simulation process would go on for a while until it yields a
procedural result or stops at a pre-set pause point. Then, the experts would determine
whether to adjust their parameters to continue with the simulation or produce the
final output. ABS is especially fit for problems with dynamic procedural features
and interactive requirements, e.g., engineering construction with multiple phases,
combatant operations with varied scenarios, etc. Moreover, the results of ABS are
likely to be different in each round of simulation because the parameter configurations
are different. Therefore, the results of ABS are often given in a probability distribution
rather than a deterministic value [13].
2 Agent-Based Simulation 175
There are multiple advantages of ABS, including the ability to predict emergent
phenomena, natural description, flexibility, cost-effective, and time-saving. (1) Emer-
gent phenomena prediction [15]. ABS by nature can predict emergent phenomena
that cannot be modeled and predicted using analytical approaches because such
emergent phenomena cannot be reduced from the comprehensive system level to the
elementary and component level whereas ABS can make full advantage of designing
behaviors of single agents. (2) Natural description [13]. It is hard to directly design
a complex system from the system level because it is rather difficult to gather all of
the relevant information due to its scarcity. Comparatively, it is relatively easier to
obtain information on the entities or agents at the component level. (3) Flexibility
[18]. By designing agents rather than the whole system, more detailed information
can be used as inputs of ABS which provides much more flexibility to the modeling
process. (4) Cost-effective and time-saving [9]. Even if we can obtain information on
a complex system, it can be expected that the cost would be huge, and the resources
required would be enormous. ABS provides an effective and time-saving means by
exploring from the most elementary level that can be understood by stakeholders
from varied backgrounds. In this sense, adopting ABS would eventually save costs
in comparison with direct experiments.
It is admitted that there is a certain level of uncertainty in adopting ABS as the
final results are mostly given in terms of probability distributions [15]. This type of
uncertainty is almost unavoidable for systems in such a complex structure and with
over-numbered entities [22]. In fact, a deterministic result would almost be counter-
intuitive. By providing results in a probability distribution, along with the ability to
depict agent dynamics and other advantageous features, ABS has been successfully
applied in simulating and solving many problems with dynamic procedural charac-
teristics and interactive requirements from a wide range of theoretical and practical
backgrounds.
Pedestrians during evacuation exhibit complex and variable patterns of behavior, and
understanding these patterns is extremely important for improving evacuation proce-
dures and relevant regulations. The development of pedestrian dynamics for evacu-
ation is an interesting and ongoing research topic. To simulate the uncertainties and
dynamics of pedestrians’ behaviors, several famous models have been put forward
in the domain of evacuation, such as the cellular automaton model [1], discrete
event simulation [17], social force model [21], and others. The cellular automaton
is a discrete dynamic system with a grid-based motion decision, which consists of
a regular grid of cells, each in one of a finite number of states. Both the cellular
automation and discrete event simulation models are in a discrete state space, while
the social force model is in a continuous space and introduces the desired force to
describe the inner drive of pedestrians to escape, especially under stressful situations
[27].
176 Agent-Based Simulation
Among the aforementioned models, the social force model is considered as one
of the widely accepted models for simulating individual movement behaviors [21],
especially in over-crowded situations. This is due to the fact that: (i) The social force
model is suitable to describe crowd dynamics since it can qualitatively reproduce
some self-organizing phenomena, such as lane formation and arching [6], (ii) The
social force model as a continuous space model can allow pedestrians to move contin-
uously within a pre-defined place by defining some forces while the discrete space
model, such as a cellular automaton, cannot allow pedestrians to move around in
an unrestricted manner [26], and (iii) the social force model is capable of importing
psychological and sociological factors in the real world by considering physical and
motivation forces [24]. Thus, the social force model is adopted in this chapter to
explore the crowd dynamics during the evacuation process in subway stations.
The conception of “social force” has been introduced to keep reasonable distances
among individuals. The social force model, proposed by Helbing et al. [8], is based
on the fluid dynamics equation, where the pedestrians were driven by different forces.
The original social force model is improved by [10], where an evolutionary optimiza-
tion algorithm is used to determine optimal parameter specifications for the social
force model based on suitable video recordings of interactive pedestrian motions. A
simple social force model is shown in Fig. 1, where the corresponding mathematical
expression of each pedestrian i can be represented by Eq. (1).
d vi (t)
mi = fi0 + fi j + fiw (1)
dt j (=i) w
vi Wall
f ij
vi0 ei0
Pedestrian i
Target
f iw fi 0
Pedestrian j
Wall
where m i refers to the mass of pedestrian i; vi (t) refers to the actual walking velocity;
fi0 represents the diagram of the desired force; fi j represents the interaction force
between pedestrians i and j; fiw represents the interaction force between pedestrian
i and walls.
Generally, three kinds of forces are considered in the social force model, including
the desired force, the interaction force between pedestrians, and the interaction force
between a pedestrian and walls. The desired force reflects the pedestrian’s willingness
to achieve the desired velocity, and this force can be expressed by Eq. (2).
where vi0 is the desired speed with a certain direction ei0 , and τi is a constant of
adaptation time to adjust the pedestrian’s actual walking velocity.
The interaction force between pedestrians i and j can be expressed by Eq. (3),
which mainly consists of the psychological force, fisj , and the physical force, fi j .
p
fi j = kg(ri j − di j )
p
n i j + κg(ri j − di j )v tji ti j (3)
⎪
⎪
⎪
⎪ 0, i f x < 0
⎪
⎩ g(x) =
x, i f x ≥ 0
The social force model can reflect the anisotropic character and dynamic features of
pedestrian interaction, and thus, it provides a basis for simulating the pedestrian evac-
uation process. Aiming to improve the effectiveness of the route planning strategies
during the pedestrian evacuation, a systematic simulation-based decision approach
is developed to investigate how to simulate, assess, and improve the efficiency of the
evacuation route planning. The overall workflow is shown in Fig. 2. In the proposed
approach, the following four steps are adopted.
The evacuation network is an abstract and geometric graph that reveals the spatial
distribution of all the possible evacuation paths. Basically, an evacuation network
can be denoted by a directed graph, G = (N, A, W ). Herein, N = {1, 2, …, n}
represents the set of evacuation nodes, such as escalators, turnstiles, and exits in a
metro station. Each evacuation node can be used to store the information, such as
evacuation speed, and pedestrian density at that specific area. A = {a1 , a2 , …, ak }
refers to the set of evacuation paths representing links between evacuation nodes.
Each evacuation path can be used to store information, such as the length and the
evacuation capacity of that specific path. Also, the time when a pedestrian enters and
leaves the specific evacuation path can be recorded. W = {w1 , w2 , …, wk } refers to
Fig. 2 A systematic simulation-based decision approach for optimizing the route planning strategy
during the pedestrian evacuation
4 Simulation-Based Route Planning and Optimization 179
a1 a5
a4
S a2 E
a3 a6
O
the set of impedances (or weights) of paths in the graph. The value of path impedance
can be used to represent the uncertainty and complexity of a specific evacuation path.
Usually, the more uncertain (or complex) the evacuation path is, the higher the path
impedance is.
Figure 3 illustrates a simple evacuation network, where there are four evacuation
nodes and six evacuation paths in total. The nodes S and E refer to the start node
and end node, respectively. Besides, there are four combinations of paths during the
evacuation from the start node to the end node, namely a1 − a5 , a1 − a4 − a6 , a2 ,
and a3 − a6 .
Several factors, such as the evacuation length, evacuation time, risk perception of
the individual, and individual psychology, can affect individual behavioral decisions
in the evacuation process. The impacts of these factors on individual behaviors can
change over time. According to the experience-driven mindset, in emergency egress,
pedestrians are more likely to choose a route with the least evacuation time, since
this often coincides with the shortest path, as the shortest path is much easier to be
computed. Pedestrians often walk the shortest path if they do not want to think a lot
about their choice. However, pursuing either the shortest evacuation length or the
least evacuation time solely can easily result in the “faster-is-slower” phenomenon
[7] because of incomplete information.
The essence of the “faster is slower” phenomenon is a lack of complete informa-
tion. One pedestrian lacks a complete understanding of the crowded condition and the
required time for evacuation in real-time. As a result, s/he is unable to make a system
optimum decision. Accordingly, too many pedestrians choose the same evacuation
path based upon their incomplete information, and the congestion occurs due to high
pedestrian density in certain areas. That is to say, apart from the length of evacuation
route and the time of evacuation process, the density of pedestrian flow is another
significant factor that affects the efficiency of the evacuation process. The definition
and calculation of those three major factors are described in detail as follows.
180 Agent-Based Simulation
(1) Length of evacuation route(L). This factor aims to record the total length of an
evacuation route from the start node to the end node. Basically, an evacuation
route may include several links between nodes in the evacuation network (see
Fig. 2), which can be a combination of some evacuation paths. Given the
importance and complexity of evacuation paths taken into account, the value
L at a specific evacuation route is represented by Eq. (5).
k
Lr = wi × ai × qi , r = 1, 2, ..., p (5)
i=1
k
Tr = (wki + wti ) × qi , r = 1, 2, ..., p (6)
i=1
n
Dr = ρ j × ξ j , r = 1, 2, ..., p (7)
j=1
4 Simulation-Based Route Planning and Optimization 181
Most models have been developed in the literature to simulate the evacuation process
according to the principles of the least evacuation time or the shortest distance, which
is easy to cause the “faster-is-slower” effect. In order to avoid that, pedestrians are
acknowledged that they should consider more than one factor in the choice of an
evacuation route, rather than rely on a sole factor. For instance, when too many
pedestrians choose the shortest or fastest evacuation route without considering the
crowded condition, the congestion is very likely to occur due to high pedestrian
density in certain areas. As a result, pedestrians cannot move forward and will waste
much more time. With many influential factors taken into account, a multi-attribute
decision approach should be used to discover the optimal route planning strategy for
guiding pedestrians.
In regard to the issue of the multi-attribute decision, there are generally two kinds
of methods, namely the comprehensive additive and multiplicative methods [25].
To be specific, the comprehensive additive method, also known as the weighted
arithmetic mean method, is suitable in situations where factors are basically inde-
pendent of each other. The comprehensive multiplicative method, by contrast, is
relatively more suitable in situations where the interaction among factors is high. In
this research, as mentioned above, each of the factors, such as L, T, and D, plays a
significant role in the decision of a reasonable evacuation route, and these factors are
highly related to each other. Thus, the comprehensive multiplicative method is used
to develop a multi-attribute decision strategy for evacuation route planning.
In the real world, to get rid of the “faster-is-slower” phenomenon, people might
give up a routine with the least evacuation time or the shortest distance, but choose a
routine with a lower density. In this research, two different route planning strategies
for pedestrian evacuation are proposed in the simulation environment. One is to
simulate the real-world pedestrian evacuation process for validation purposes, while
the other one is to improve the efficiency of the pedestrian evacuation. In the former
route planning strategy, defined by Scenario I, we assume that all passengers choose
an evacuation route which has both the shortest evacuation length and the least
evacuation time. The expression of the route planning strategy in Scenario I is given
by Eq. (8). Since this strategy is very consistent with the motivation of pedestrians
in the real world, the real situation of the pedestrian evacuation process can then be
simulated and validated via Scenario I.
182 Agent-Based Simulation
⎛ ⎞
n
min Z r (t) = min ⎝ (L r (t))α × (Tr (t))β ⎠ (8)
j=1
where Z r (t) refers to the target value that colligates the condition of the influential
factors, such as the evacuation length and evacuation time, at the rth evacuation route
at time t; L r (t) and Tr (t) refer to the value of the evacuation length and the evacuation
time at the rth evacuation route at time t, respectively,
In the latter route planning strategy, defined by Scenario II, we assume that all
passengers choose an evacuation route which has the shortest evacuation length, the
least evacuation time, and the lowest pedestrian density. The expression of the route
planning strategy in Scenario II is given by Eq. (9). The factor of pedestrian density is
considered in the route planning strategy in Scenario II, to improve the performance
efficiency of the current pedestrian evacuation strategy specified in Scenario I.
⎛ ⎞
n
min Z r (t) = min ⎝ (L r (t))α × (Tr (t))β × (Dr (t))γ ⎠ (9)
j=1
where Z r (t) refers to the target value that colligates the condition of the influential
factors, such as the evacuation length and evacuation time, at the rth evacuation
route at time t; L r (t), Tr (t) and Dr (t) refer to the value of the evacuation length,
the evacuation time, and the pedestrian density at the rth evacuation route at time
t, respectively, and α, β, and γ refer to weights of the above factors, respectively.
For simplicity, we assume the above three factors, namely L r (t), Tr (t), and Dr (t)
share the same level of contributing to the emergency efficiency in this research, and
their values are defined to be 1.0 in Eqs. (8) and (9), respectively.
Route planning aims to identify an optimal path from a specified origin node to
other destination nodes for each pedestrian at time t. To search for the minimum
target value Z r (t) at time t, the most commonly used algorithms are the Dijkstra
algorithm [23], Floyd algorithm, and genetic algorithm. Chen et al. [2] indicated that
the Dijkstra algorithm was recognized as the best algorithm among them. Thus, the
Dijkstra algorithm is used to find the optimal evacuation route in this research. See
details on the Dijkstra algorithm to [20].
A number of route planning strategies have been proposed, including the above two
strategies. It is, therefore, necessary to assess the performance efficiency of a route
planning strategy and determine which route planning strategy is more effective. In
this research, four key performance indicators (KPIs), namely average pedestrian
density, average evacuation length, average evacuation time, and average evacuation
4 Simulation-Based Route Planning and Optimization 183
n
AveD = Aveρ j /n (10)
j=1
Ttotal
where Aveρ j (t) refers to the average pedestrian density at the jth evacuation
node (j = 1, 2, …, n), which can be defined by Eq. (11). ρ j (t) refers to the
pedestrian density at the jth evacuation node (j = 1, 2, …, n) at time t, and
T total refers to the duration of the whole evacuation process in total.
(2) Average evacuation length(AveL). This indicator aims to assess the distance
that a pedestrian walks during the evacuation on average. The value of AveL
can be calculated by Eq. (12). Generally, the lower the value of AveL is, the
better the evacuation efficiency is.
p
r =1(L r × Sr )
AveL = p (12)
r =1 Sr
Herein, T r refers to the time a pedestrian spends at the rth evacuation route (r
= 1, 2, …, p) [see Eq. (6)].
(4) Average evacuation capacity(AveC). This indicator aims to assess how many
pedestrians can be evacuated within a certain period on average. The value of
AveC can be calculated by Eq. (14). Generally, the higher the value of AveC
is, the better the evacuation efficiency is.
184 Agent-Based Simulation
p
AveC = (Sr /Ttotal ) (14)
r =1
where S r refers to the number of pedestrians that evacuate along with the rth
evacuation route (r = 1, 2, …, p), and T total refers to the duration of the whole
evacuation process in total.
In the past ten years, tunnel construction has presented a powerful momentum for
rapid economic development worldwide, especially in developing countries, like
China. To simulate and analyze pedestrian evacuation behaviors during rush hours,
one metro station located at the Wuhan metro system in China is used as a case for
modeling the pedestrian evacuation process.
The Wuhan metro system is an elevated and underground urban metro system in the
city of Wuhan, Hubei, China. It is a rapid transit system that serves to accommodate
increasing cross-borough traffic. By the end of the year 2020, there are nine metro
lines in operation, consisting of 240 metro stations with a total operation mileage
of 360 km. The Hongshan Square Station, located in the southeast of Wuhan, is
an interchange station of Line 2 and Line 4 in the Wuhan metro system. Figure 4
presents an internal 3D view of the Hongshan Square Station. There are three floors
below the ground in the Hongshan Square Station. Pedestrians get out of trains on the
second and/or third floor below the ground, go to the first floor (hall) using escalators,
and finally go to the ground through exits. Figure 5 illustrates the layout of the hall
of the Hongshan Square Station. In total, there are four escalators (i.e., Escalators 1,
2, 3, and 4), two turnstiles (i.e., Turnstiles A and E), and five exits (i.e., Exits B, C,
D, G, and F) at the hall of that station.
As shown in Fig. 5, Exit F is not visible. To make good use of exits for improving
pedestrian evacuation, the location of Exit F is made visible with the help of direc-
tional signs and optical clues. Pedestrians are persuaded that the Exit F locates in the
place of the relocated Exit F, and thus, the location of the physical Exit F is moved
to that of the relocated Exit F in the simulation model (see Fig. 5).
5 Pedestrian Evacuation Simulation 185
Fig. 4 3D view of the Hongshan Square Station in the Wuhan metro system
Escalator
Turnstile
Exit
0
W
00
C
Evacuation inflow
42
Unit mm
12
00
0
00
Ex 00
r3
Re
84
it F
to
loc
la
at
ca
ed
Es
r1
Ex
to
it F
la
ca
Es
B
it
Es
Ex
ca
PAID AREA
la
to
r2
A
ile
st
Escalator 4
rn
30000
Tu
Ex
it
C
Exit G
Turnstile E
0
00
55
160000
Exit D
One of the biggest challenges for the construction of the evacuation model lies in
the modeling of complex pedestrian dynamics. To address this problem, AnyLogic
provides a dynamic simulation tool with a pedestrian library for simulating pedestrian
186 Agent-Based Simulation
F 1
a1
2
4 a9
G a2
a10 B
a4 3 a3 a5
Escalator A a6
Turnstile E a7
a8
Exit
Key node
D C
Fig. 6 Excavation network for the hall of the Hongshan Square Station
Table 2 Description of
No. Excavation route Total length (m) Key nodes
twelve excavation routes at
the hall of the metro station 1 1-A-B 42.0 A, B
2 2-A-B 38.5 A, B
3 3-A-B 33.5 A, B
4 1-A-C 55.5 A, C
5 2-A-C 52.0 A, C
6 3-A-C 47.0 A, C
7 1-A-D 67.5 A, D
8 2-A-D 64.0 A, D
9 3-A-D 59.0 A, D
10 4-E-D 50.5 E, D
11 4-E-F 60.0 E, F
12 4-E-G 64.5 E, G
lengths of all the evacuation routes. For instance, the 1st evacuation route “1-A-B”
is a combination of the path link “a1 ” (from Escalator 1 to Turnstile A) and the path
link “a4 ” (from Turnstile A to the Exit B). As shown in Table 1, these two path links
have a path impedance of 0.5 and 1.0, respectively. The total length of the evacuation
route “1-A-B” is then calculated to be 42.0 m using Eq. (5).
This model simulates the pedestrian evacuation process in a continuous manner.
Pedestrians successively get out of trains on the second or third floor below the
ground and go to the first floor (hall) using escalators. In this model, we assume
that pedestrians appear continuously at a flow rate of 30 p/min at the Escalator 1, 25
p/min at Escalators 2 and 3, and 50 p/min at Escalator 4. There are three passages
(or evacuation options) at each turnstile, and the width of each exit is set to be 1.8 m
in the simulation model. Figure 7 illustrates a flowchart for path selection in the
evacuation simulation for pedestrians located at Escalator 1. The bottom left box
188 Agent-Based Simulation
pedGoTo2 pedSink2
Selection strategies:
Condition1: Z1<=min (Z4, Z7)
Condition2: Z4<=min (Z1, Z7) pedSink3
pedGoTo3
Condition3: Z7<=min (Z1, Z4)
Fig. 7 Flowchart for path selection in the evacuation simulation for pedestrians located at the
Escalator 1
in Fig. 7 shows the route selection strategies that are incorporated in the flowchart
in detail. For instance, a pedestrian at time t would select the 1st evacuation route,
rather than the 4th or the 7th evacuation route, once the calculated target value of Z 1
is less than that of Z 4 and Z 7 using Eq. (9) [or Eq. (8)].
example, Fig. 8 illustrates the statistics of the pedestrian density located at the node
of A in Scenario I, and the average pedestrian density is calculated to be 2.66 p/m2 .
At the same time, the pedestrian evacuation process at the key evacuation node of
A in the real world is monitored by using video cameras. Pedestrian densities at a
specific evacuation node are then recorded by drawing frames of the video at the same
position at different time within six minutes. Figure 9 illustrates some photographs
for monitoring the pedestrian density at the key evacuation node A in the real world
at different time. The observed average pedestrian density at the key evacuation node
A is calculated to be 2.69 p/m2 . With all the seven key evacuation nodes taken into
account, Fig. 10 illustrates the scatterplot of simulation results and observations in
70
60
50
40
30
20
10
0
1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 3.2 3.4 3.6 3.8
Pedestrian density
Fig. 9 Observed pedestrian density at the key evacuation node A in the real world at different time:
a T = 9: 00 am; b T = 9:01 am; c T = 9:02 am
190 Agent-Based Simulation
2.6 A
2.4
D
Simulation
2.2 B
E
2.0
F
1.8
G Pedestrian density
1.6 C Linear of fit
1.4
1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0
Observation
Fig. 10 Scatterplot of simulation and observation for the average pedestrian density at seven key
evacuation nodes (unit: p/m2 )
terms of the average pedestrian density. Results indicate that, in general, the observed
average pedestrian density is low when the simulated mean is low, and similarly, the
observed average pedestrian density is high when the simulated mean is high. The
fitted regression line has a gradient of 0.98 approaching 1, which indicates that the
established evacuation model can simulate the average magnitude of the pedestrian
density in a metro station well. In other words, the central tendency of the model
simulation results is not significantly different from that of the observations.
When the established evacuation model has been validated in terms of pedestrian
densities located at key evacuation nodes, it is believed that it is rational and reliable
to apply the established model to simulate the pedestrian evacuation process in the real
world. Aiming to improve the efficiency of the current route planning strategy, a novel
route planning strategy that is specified in Scenario II is developed. Thus, the route
planning strategy specified in Scenario I is replaced by the route planning strategy
specified in Scenario II. All passengers would choose an evacuation route that has
the shortest evacuation length, the least evacuation time, and the lowest pedestrian
density during the evacuation [see Eq. (9)]. The main purpose is to compare the
strategies specified in Scenarios I and II and identify which strategy is better during
pedestrian evacuation.
6 Route Planning Optimization 191
Fig. 11 Distribution of pedestrian densities in Scenario II (unit: p/m2 ). Note Exit F is moved to the
center to simplify the layout of the hall of the station
The pedestrian density is a very important KPI for the performance assessment of
the evacuation efficiency since the pedestrians cannot move forward if the pedes-
trian density is too high. More time would be wasted, and the “faster-is-slower”
phenomenon is likely to occur. Following the simulation of the pedestrian evacua-
tion process within six minutes in both Scenarios I and II, pedestrian densities at seven
key evacuation nodes are recorded every 0.01 min. Equation (11) is used to obtain
the average pedestrian density at each key evacuation node. Figure 12 illustrates the
statistic of pedestrian densities at those nodes in Scenarios I and II.
Taking Fig. 12a as an example, the average pedestrian density at the key evacuation
node of A is calculated to be 2.66 p/m2 in Scenario I and 1.94 p/m2 in Scenario II,
respectively. Generally, the pedestrian density in Scenario II is lower than that in
Scenario I at almost all the key evacuation nodes, except for the nodes of C and G.
At the nodes of C and G, the pedestrian density in Scenario II is slightly higher than
that in Scenario I. In terms of deviations displayed on those curves in Fig. 12, the
central tendency of pedestrian densities in Scenario II is more apparent than that in
192 Agent-Based Simulation
Frequency
Frequency
100 80
Frequency
140
90 130
70
120
80
110
60
70 100
90 50
60
80
50 70 40
40 60
50 30
30 40
20
20 30
20
10 10
10
0 0 0
1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 3.2 3.4 3.6 3.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 3.2 3.4 3.6 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0
Frequency
Frequency
140 110
130 80
100
120 70
110 90
100 80 60
90 70
80 50
60
70 40
60 50
50 40 30
40 30
30 20
20
20 10
10 10
0 0 0
1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 3.2 3.4 3.6 3.8 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8 3.0 3.2 3.4 3.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8
60
50
40
30
20
10
0
0.8 1.0 1.2 1.4 1.6 1.8 2.0 2.2 2.4 2.6 2.8
Pedestrian density
(g)
Fig. 12 Statistic results of pedestrian densities in Scenarios I and II at seven key evacuation nodes
(unit: p/m2 ): a Turnstile A; b Exit B; c Exit C; d Exit D; e Turnstile E; f Exit F; g Exit G
Scenario I, indicating that a more balanced fashion regarding the crowded condition
is achieved in Scenario II, with that in Scenario I compared.
Given the above seven critical evacuation nodes taken into account, Eq. (10) is used
to obtain the average pedestrian density as a whole, and the results are shown in Table
3. Results indicate that the value of the average pedestrian density is reduced from
2.11 p/m2 in Scenario I to 1.83 p/m2 in Scenario II. In other words, the performance
Table 3 Comparisons of the KPIs during the pedestrian evacuation in Scenarios I and II
KPIs Scenario I Scenario II Improved Efficiency (%)
Average pedestrian density (p/m2 ) 2.11 1.83 0.28 13.27
Average evacuation length (m) 56.47 55.58 0.89 1.58
Average evacuation time (min) 1.55 1.26 0.29 18.71
Average evacuation capacity (p/min) 92 122 30 32.61
6 Route Planning Optimization 193
of the crowded condition during the evacuation is increased by 13.27% (i.e., (2.11–
1.83)/2.11) using the improved route planning strategy specified in Scenario II.
The evacuation length measures the distance that a pedestrian walks during the
evacuation. Along with the pedestrian evacuation simulation within six minutes as
mentioned above, the number of pedestrians evacuating by the rth evacuation route
(r = 1, 2, …, 12) is recorded in both Scenarios I and II. Equation (12) is used to
obtain the average evacuation length in Scenarios I and II, respectively. Figure 13
illustrates the distribution results of the evacuation length for all pedestrians during
the evacuation in Scenarios I and II.
It is clear to see that in Scenario I, a total of 552 pedestrians excavate from the
metro station within six minutes, and the length of all the evacuation routes in total
reaches up to 31,172.5 m. As a result, the average evacuation length in Scenario I
is 31,172.5/552 = 56.47 m. In Scenario II, a total of 728 pedestrians excavate from
the metro station within six minutes, and the length of all the evacuation routes in
total reaches up to 40,460.5 m. As a result, the average evacuation length in Scenario
II is 40,460.5/728 = 55.58 m. That is to say, compared with Scenario I, a distance
70
Scenario I
65 Scenario II
60
55
50
Evacuation length (m)
45
40
35 Scenario I
30 Evacuated pedestrians=552
Total evacuation length=31172.5 m
25
Average evacuation length=56.47 m
20 Scenario II
15 Evacuated pedestrians=728
Total evacuation length=40460.5 m
10
Average evacuation length=55.58 m
5
552 728
0
0 50 100 150 200 250 300 350 400 450 500 550 600 650 700 750
Pedestrians during the evacuation
Fig. 13 Distribution results of evacuation length for all pedestrians during the evacuation in
Scenarios I and II. Note The records of evacuation length of each pedestrian are sorted from smallest
to largest
194 Agent-Based Simulation
The evacuation time measures the time that a pedestrian spends during the evacuation
from the start node to the end node. Along with the pedestrian evacuation simulation
within six minutes as mentioned above, the time that a pedestrian evacuates on
average by the rth evacuation route (r = 1, 2, …, 12) is recorded in both Scenarios
I and II. Figure 14 illustrates the statistic of average evacuation time among 12
evacuation routes during the pedestrian evacuation in both scenarios.
Results indicate that the values of the average evacuation time among the above
12 routes in Scenario II do not vary greatly and achieve a balance to some extent,
while the average evacuation time varies greatly in Scenario I. Given the above 12
routes taken into account, Eq. (13) is used to obtain the average evacuation time as a
whole. Results, as shown in Table 3, indicate that the value of the average evacuation
time is reduced from 1.55 min in Scenario I to 1.26 min in Scenario II. In other words,
0.29 min that is about 18 s can be saved for every pedestrian during the evacuation
using the improved route planning strategy specified in Scenario II.
2.6
Scenario I
2.4 Scenario II
2.2
Average evacuation duration (min)
2.0
1.8
1.6
1.4
1.2
1.0
0.8
0.6
0.4
0.2
0.0
1 2 3 4 5 6 7 8 9 10 11 12
Route paths
Fig. 14 Statistic results of average evacuation time among 12 evacuation routes during the
pedestrian evacuation in Scenarios I and II
6 Route Planning Optimization 195
The evacuation capacity measures the number of pedestrians who can be evacu-
ated within a certain period. During the pedestrian evacuation simulation within six
minutes, the number of pedestrians that evacuate by the rth evacuation route (r = 1,
2, …, 12) is recorded in both Scenarios I and II. Equation (14) is used to obtain the
average evacuation capacity in both scenarios I and II, respectively. Table 3 illustrates
the statistic of average evacuation capacity in both scenarios. Results indicate that
the value of the average evacuation capacity is increased from 92 p/min in Scenario I
to 122 p/min in Scenario II. That is to say, the performance of the evacuation capacity
during the evacuation is increased by 32.61% (i.e., (122–92)/92) using the improved
route planning strategy specified in Scenario II.
As mentioned above, pedestrians appear continuously at a flow rate of 30 p/min
at the Escalator 1, 25 p/min at Escalators 2 and 3, and 50 p/min at Escalator 4.
Generally, the input flow rate is 130 p/min in total. It is clear to see that 38 (i.e.,
130–92) and 8 (i.e., 130–122) pedestrians would be stranded at the station hall every
minute in Scenarios I and II, respectively. In general, the route planning strategy
as specified in Scenario II provides a better strategy for pedestrian evacuation than
that in Scenario I. In other words, the evacuation performance can be significantly
improved, when the route planning strategy specified in Scenario II is implemented
during the pedestrian evacuation.
8 Summaries
To improve the effectiveness of route planning strategies during the pedestrian evac-
uation in rush hours, a systematic simulation-based decision approach is developed
to investigate how to simulate, assess, and improve the performance efficiency of
the route planning strategy during the evacuation. Scenario I is used to simulate the
pedestrian evacuation process in the real world. Scenario II is used to improve the
performance efficiency of the pedestrian evacuation process. Four KPIs are proposed
to evaluate the evacuation performance within different route planning strategies.
One metro station located at the Wuhan metro system in China is used as a case
to demonstrate the applicability and effectiveness of the proposed route planning
strategy for improving the evacuation performance efficiency. Results indicate that
the developed approach is capable of (i) simulating and validating pedestrian evac-
uation dynamics given the complex interaction between pedestrians and the traffic
is taken into account, (ii) optimizing the route choice planning strategy during the
evacuation, and (iii) assessing the evacuation performance. All the three factors,
including the length of evacuation route (L), time of evacuation process (T ), and
density of pedestrian flow (D), are regarded as significant factors that affect the
efficiency of the evacuation process. Furthermore, valuable theoretical and prac-
tical insights can be provided into a deep understanding of route planning strategies
during the pedestrian evacuation, enabling to achieve the improvement of safety and
economic objectives in the design or re-design of the metro evacuation systems.
The major novelty of the proposed ABS approach in this chapter is that the uncer-
tainties and dynamics underlying the pedestrian behaviors during the evacuation
process are modeled, when the complex interactions between pedestrians and the
traffic can be taken into account. A better route planning strategy has been devel-
oped, which has been demonstrated to be effective in optimizing the pedestrian
evacuation performance. Some useful KPIs are proposed to compare and benchmark
the evacuation efficiency among different route planning strategies.
References
1. Alizadeh R (2011) A dynamic cellular automaton model for evacuation process with obstacles.
Saf Sci 49:315–323
2. Chen Y-Z, Shen S-F, Chen T, Yang R (2014) Path Optimization Study for Vehicles Evacuation
based on Dijkstra Algorithm. Proc Eng 71:159–165
198 Agent-Based Simulation
3. Chung TH, Mohamed Y, AbouRizk S (2006) Bayesian updating application into simulation in
the North Edmonton Sanitary Trunk tunnel project. J Constr Eng Manag 132:882–894
4. Fang Z, Li Q, Li Q, Han LD, Wang D (2011) A proposed pedestrian waiting-time model for
improving space–time use efficiency in stadium evacuation scenarios. Build Environ 46:1774–
1784
5. Guo R-Y, Huang H-J, Wong S (2012) Route choice in pedestrian evacuation under conditions
of good and zero visibility: Experimental and simulation results. Transp Res Part B Methodol
46:669–686
6. Helbing D, Farkas I, Vicsek T (2000) Simulating dynamical features of escape panic. Nature
407:487–490
7. Helbing D, Farkas IJ, Molnar P, Vicsek T (2002) Simulation of pedestrian crowds in normal
and evacuation situations. Pedestrian Evacuation Dyn 21:21–58
8. Helbing D, Molnar P (1995) Social force model for pedestrian dynamics. Phys Rev E 51:4282–
4286
9. Jahangirian M, Eldabi T, Naseer A, Stergioulas LK, Young T (2010) Simulation in manufac-
turing and business: a review. Eur J Oper Res 203:1–13
10. Johansson A, Helbing D, Shukla PK (2007) Specification of the social force pedestrian model
by evolutionary adjustment to video tracking data. Adv Complex Syst 10:271–288
11. Kiesling E, Günther M, Stummer C, Wakolbinger LM (2012) Agent-based simulation of
innovation diffusion: a review. CEJOR 20:183–230
12. Lei W, Li A, Gao R, Hao X, Deng B (2012) Simulation of pedestrian crowds’ evacuation in a
huge transit terminal subway station. Phys A 391:5355–5365
13. Leng N, Corman F (2020) The role of information availability to passengers in public transport
disruptions: an agent-based simulation approach. Transp Res Part A Policy Pract 133:214–236
14. Lohmer J, Bugert N, Lasch R (2020) Analysis of resilience strategies and ripple effect in
blockchain-coordinated supply chains: an agent-based simulation study. Int J Prod Econ
228:107882
15. Marzouk M, Al Daour I (2018) Planning labor evacuation for construction sites using BIM and
agent-based simulation. Saf Sci 109:174–185
16. Musolino G, Vitetta A (2014) Calibration and validation of a dynamic assignment model in
emergency conditions from real-world experimentation. Proc Soc Behav Sci 111:498–507
17. Na HS, Banerjee A (2019) Agent-based discrete-event simulation model for no-notice natural
disaster evacuation planning. Comput Ind Eng 129:44–55
18. Railsback SF, Lytinen SL, Jackson SK (2006) Agent-based simulation platforms: review and
development recommendations. SIMULATION 82:609–623
19. Rozo KR, Arellana J, Santander-Mercado A, Jubiz-Diaz M (2019) Modelling building emer-
gency evacuation plans considering the dynamic behaviour of pedestrians using agent-based
simulation. Saf Sci 113:276–284
20. Skiena S (1990) Dijkstra’s algorithm. Implementing discrete mathematics: combinatorics and
graph theory with mathematica. Addison-Wesley, Reading, MA, pp 225–227
21. Wan J, Sui J, Yu H (2014) Research on evacuation in the subway station in China based on the
combined social force model. Phys A 394:33–46
22. Xiong C, Zhu Z, Chen X, Zhang L (2018) Optimal travel information provision strategies: an
agent-based approach under uncertainty. Transportmetrica B Transp Dyn 6:129–150
23. Xu M, Liu Y, Huang Q, Zhang Y, Luan G (2007) An improved Dijkstra’s shortest path algorithm
for sparse network. Appl Math Comput 185:247–254
24. Xu S, Duh H-L (2010) A simulation of bonding effects and their impacts on pedestrian
dynamics. IEEE Trans Intell Transp Syst 11:153–161
25. Xu Z (2015) Uncertain multi-attribute decision making: methods and applications. Springer
26. Yuen J, Lee E, Lam W (2014) An intelligence-based route choice model for pedestrian flow in
a transportation station. Appl Soft Comput 24:31–39
27. Zhang L, Liu M, Wu X, AbouRizk SM (2016) Simulation-based route planning for pedestrian
evacuation in metro stations: a case study. Autom Constr 71:430–442
References 199
28. Zhang L, Wu X, Chen Q, Skibniewski MJ, Hsu S-C (2014) Towards a safety management
approach for adjacent buildings in tunneling environments: case study in China. Build Environ
75:222–235
29. Zhong M, Shi C, Tu X, Fu T, He L (2008) Study of the human evacuation simulation of metro
fire safety analysis in China. J Loss Prev Process Ind 21:287–298
Expert Systems
1 Introduction
Tunnel construction entails a highly complicated project with large potential risks,
which can bring enormous dangers to public safety. Numerous accidents have led
to growing public concerns about prior risk identification and assessment in rela-
tion to tunnel construction safety. Risk identification plays an important role in the
safety assurance process, aiming to reveal the potential safety risk and determine
risk factors’ contribution to the occurrence of an accident. Critical potential risks
and risk factors can be discovered to assist project engineers in determining critical
safety checking points for the improved safety management [31].
Generally, the frequently used methods for risk identification can be divided into
qualitative and quantitative risk analysis tools, such as Safety Check List (SCL),
Delphi, Fault Tree Analysis (FTA), Event Tree Analysis (ETA), Hazard and Oper-
ability study (HAZOP), and others. Šejnoha et al. [25] conducted a quantitative risk
assessment of the tunneling excavation process in a Czech Republic tunnel project
using FTA. Hong et al. [14] employed ETA to provide a quantitative risk evaluation
in a tunnel project passed under the Han River in South Korea. Using these risk
analysis tools, domain experts are commonly invited to identify the potential safety
risks according to their expertise accumulated from years of experience. Generally,
the values of risk factors are captured by manually reading engineering drawing
documents.
There are three kinds of deficiencies existing in the current safety risk identifica-
tion process: (1) The domain experts’ expertise is excessively relied on due to a lack of
autonomous inference capacity in the current approach. Domain experts are generally
considered as a scarce resource that is unable to provide universal consultation and
real-time guidance due to time constraints. (2) The process of the engineering param-
eter extraction involves many time-consuming and error-prone activities, especially
in reading engineering drawing documents, and therefore, the accuracy and effec-
tiveness of the risk identification results can be significantly affected to some extent.
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 201
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_9
202 Expert Systems
2 Expert Systems
as the concept name, the possible values of each attribute, and the procedural
information of the target problems. Thus, the frame-based ES owns the ability
to store and process descriptive and behavioral information. Due to this advan-
tage, the frame-based ES can deal with more complex problems, compared to
the rule-based system.
(3) Fuzzy logic-based ES. Integrating the fuzzy theory into systems, the fuzzy
logic-based ES is established. The fuzzy logic is the basis of the reasoning,
and it is used in the system to produce a rough fuzzy judgment for solving
problems. Because of the integration of the fuzzy logic, the fuzzy logic-based
ES has high robustness, and it can perform tentative and heuristic reasoning.
(4) Neural network-based ES. The neural network-based ES is based on the
rapid development of machine learning algorithms, and this feature makes
it fundamentally different from other ESs. With the integration of machine
learning algorithms, the neural-network-based ES can automatically acquire
the needed information, instead of manual processing. Along with the applica-
tion of machine learning in ESs, the reasoning performance of ESs is greatly
promoted.
The rapid development and wide applications of ESs provide effective tools to
solve complex problems, and accordingly, the demand and applications with ESs are
increasing in various domains, such as finance, education, public health, military,
transportation, and manufacturing industries. For instance, ESs are used in organiza-
tions for solving problems, such as financial analysis, consultant, strategic planning.
Bohlouli et al. [4] developed an ES for data information management in the orga-
nization and it was found that the developed ES could effectively solve complex
cases, where the management of them normally needs more time for human experts
to solve. Goksu [10] introduced ESs into the teaching and learning systems to create
a dynamic environment and help facilitate the learning experience. Abu-Naser and
Aldahdooh [1] integrated ESs into a diagnosis platform, which could identify disease
symptoms and provide an immediate, accurate response for the patients.
ESs have also been widely used in the construction industry. Chou [5] proposed a
linear model-based ES to estimate the cost of transportation projects. The proposed
system could effectively provide detailed information for the cost changing track for
the transportation projects. Similarly, Chou and Tseng [6] developed an ES for data
analysis and cost prediction in transportation projects, which could help process the
management information and produce preliminary budgets for the agencies. To deal
with the problem of construction disputes, Elziny et al. [8] presented an ES that can
evaluate the overall dispute settlement procedure for the construction projects, and
four cases studies were performed to test the effectiveness of the proposed system,
and it was found that the developed ES can efficiently facilitate the decision makings
in the dispute resolution process. Latif et al. [17] developed a knowledge-based ES
for the material cost control of building construction projects. The proposed system
could identify the impacts, the causes, and construct the relationships between the
influential factors and aims. A case study was conducted to present the advantage
of the proposed system for the improvement of the project monitoring and process
2 Expert Systems 205
Safety is a major issue across global construction industries [28] noted that construc-
tion safety should be a prime consideration in the conceptual and preliminary design
phases, since the ability to influence construction site safety is progressively lost as
the project moves into the construction phase. Behm [3] investigated the relationship
between construction fatalities and design for construction safety. A total of 224
fatality investigation reports were examined, where the results suggested that 42%
of the fatalities reviewed were linked to design issues. Gambatese et al. [9] noted
that design for safety is one significant element in minimizing construction project
risk and enhancing worker safety through multi-level risk assessment and hazard
prevention mechanisms. Recent research indicates that designers need to indicate
knowledge and understanding of how risks and hazards affect health and safety can
arise in construction and how they can be avoided or reduced through design.
In the actual construction practices, contracts and regulatory requirements from
the Occupational Safety and Health Administration (OSHA) place the burden for
206 Expert Systems
worker safety solely on the constructor [3]. This approach is still widespread across
many countries, especially in developing countries. Zhou et al. [33] claimed that
designers began to appreciate each other’s concerns by working with their colleagues
who were responsible for the construction of the project in design-build firms.
Good ideas would be shared and applied in subsequent projects. Nevertheless, many
designers, who were not part of design-build firms, noted that they lack the skills and
training to address construction safety. The best construction practice highlights the
importance of multi-party collaboration for improved design for safety. However,
there is a lack of responsive tools and resources to assist designers in addressing
construction safety risks.
Managing construction safety is challenged due to the lack of efficient tools to
facilitate multi-party collaboration in construction management practices. To guar-
antee construction safety, a range of new tools have been developed. BIM is one of
the fastest-growing tools that has gained wide acceptance in the AEC/FM industry,
offering great solutions to address the inefficiencies of the current paper-based
processes. BIM is widely applied as a central technology for site hazard preven-
tion and safe project delivery. Hardin [13] noted that the implementation of BIM
in construction practice promised to improve the communication and collaboration
between participants through higher interoperability of data. Opportunities are iden-
tified to promote safety with the application of BIM by involving multi-partners. BIM
can be used as a support tool for constructability review, site management, safety
planning, and the identification and resolution of time–space conflicts.
Considered as the core of the construction safety management, safety risk iden-
tification is quite difficult given the complexity and size of the building systems
designed, such as the tunnel construction. Tunnel construction is characterized as
a highly complicated project with high potential risks, which integrates multiple
design disciplines, sub-projects, and interfaces. Information technology has proven
to be crucial to the success of a project by effectively controlling the safety risks.
Information flow from design to construction is critical and allows for design-build
and other integrated project delivery methods to be favored. BIM is regarded as
the information carrier of engineering characteristics, structure design, and structure
regulations. The information concerning structural component attributes, constraint
relations, and the interaction can be retrieved from a BIM model.
Most safety risks in tunnel construction are related to the engineering informa-
tion, and thus, the domain experts and engineers can identify risks and risk factors
through reading the conventional 2D engineering drawings. However, BIM itself
cannot identify the potential safety risks due to the lack of inference capacity, but
can provide the related information needed in the risk identification process.
The Industry Foundation Classes (IFC) standard has been developed as an open
standard for common data structures on information capturing and exchange. The IFC
3 Relationships Between BIM and Construction Safety 207
IfcGridPlacement
ObjectPlacement 1
ObjectPlacement
IfcLocalPlacement
IfcProduct
IfcPhysicalRepresentation
Representation
1
Represenataion IfcAbstractRepresentation
IfcMatrialRepresentation
Fig. 1 Relationships between the IfcProduct class and its related geometric properties
Tunnel construction risk results from various kinds of uncertainty, which can lead to
economic loss, project delay, human injury, and damage. The key to the risk identi-
fication process is to identify risk events and their mutual relationships. Safety risks
are related to many influential factors in tunnel projects, such as structural forms,
208 Expert Systems
The prior expert expertise and historical cases can provide a wealth of knowledge
resources for safety risk analysis and identification. With the development of tunnel
construction practice worldwide, large amounts of scattered knowledge are accumu-
lated, including explicit and tacit knowledge. Sherehiy and Karwowski [26] claimed
that both explicit and tacit knowledge was equally important in knowledge base
4 Knowledge Base Development for Construction Safety 209
Table 1 Relationships between the extracted information and the related risks
Risk category BIM elements Engineering Risk factors Safety risks
parameters
Technical risks Station contour Outline length Large foundation Instability of steel
Standard width pit width strut
Number of tip
wells
Bored pile Pile diameter Poor effect of Seepage from side
Distance waterproofing wall of foundation
between piles Short seepage pit
paths
Jet grouting Pile diameter
pile Distance
between piles
Bored pile Pile length Bored pile not Flowing sand from
State of embedded into the bottom of
embedded into rocks foundation pit
rocks Embedded length Instability of
ratio < 0.7 support system
Embedded Soft ground not
length ratio reinforced
Supporting Steel strut stories Over-limited Instability of steel
system slenderness ratio strut
Structure Construction Special restriction Construction
design method of construction method risk
Structure safety method
grade
Foundation pit Reinforcing Soft ground not Flowing sand/soil
bottom stratum scope reinforcement from bottom of
Reinforcing foundation pit
depth
Geological risks Foundation pit Soil layer Soft ground of Sand/soil gush or
bottom stratum foundation pit upheaval from the
Artesian water Artesian water bottom bottom of
level Harmful geological foundation pit
conditions within Seepage of harmful
Harmful Silt stratum the scope of gas
geological Harmful gas excavation
conditions
Cave
Environmental Building Structural type Close to foundation Building tilt or
risks Foundation type pit cracking
Shallow foundation
Building height embedment
Foundation Bad structure
depth health
(continued)
210 Expert Systems
Table 1 (continued)
Risk category BIM elements Engineering Risk factors Safety risks
parameters
Distance to the
foundation pit
Underground Pipe type Close to foundation Underground pipe
pipe Pipe material pit cracking and
Pressure-bearing seepage
Distance to the pipes
foundation pit Rigid joint
Pipe direction
Pipe burial depth
Pipe connection
style
Construction complexities Safety-based knowledge Flowing sand of foundation pit (f1) Retaining pile embedded
into rocks (e1)
Large deformation of steel strut (f2)
Water head difference of
Complexity in Large deformation of steel strut (f3) foundation pit (e2)
Technical Risks
construction techniques
Explicit Knowledge Instability of support system (f4) Geological conditions of
foundation pit (e3)
Complexity in Instability of foundation pit (f5)
Geological Risks Reinforcement state of
geological conditions foundation pit (e4)
Unpredicted underground caves (f6)
Tacit Knowledge Embedded length ratio of
Complexity in Unpredicted harmful gases (f7)
Environmental Risks bored pile (e5)
surrounding environments
Large ground settlement (f8) Bed thickness of
foundation pit (e6)
Building tilt or cracking (f9)
Underground pipe cracking (f10)
and when hydrodynamic pressure is equal to or greater than the submerged unit
weight of soil, given the existence of water head difference. Consequently, the
terms or keywords concerning the specific risk “water gushing at foundation
pit” can be preliminarily achieved.
2. Tacit knowledge. Tacit knowledge is generally developed by individuals
through years of working experience. Dialogues and communications among
individuals are basic means of knowledge sharing. Therefore, questionnaires,
depth interviews, and group decision-making methods can be used to reveal and
transfer the tacit knowledge in a reusable manner. Also, numerous researchers
built simulation models for the safety analysis, serving as a valid reference for
the relation discovery within various influential factors.
With the support of safety-based knowledge, including explicit and tacit knowl-
edge resources, the projects in tunnel construction are analyzed through the risk
mechanism process. Figure 3 lists the results of risk mechanism analysis in tunnel
construction. There are mainly 10 kinds of safety risks in total, among which
the hazard occurring in “Flowing sand of foundation pit (f 1 )” is considered as a
commonly occurring risk event. A total of six risk factors are identified to contribute
to the hazard occurrence. A case study concerning this safety risk is presented later
in this chapter.
risk related parameter lead to the different credibility degree of the factor
being considered as the evidence in accidence occurrence. CF(e) takes a value
of zero when the evidence is unreliable, while CF(e) takes a value of one
when the evidence is completely reliable. The more difficult to become the
evidence, the smaller its CF(e) is. In accordance with the safety-related facts
in tunnel construction, the facts are divided into two types, namely certain
and uncertain facts. Accordingly, the method to determine the corresponding
CF(e) of different types would vary as follows.
• Certain fact. A certain factor refers to the risk factor which has discrete
values. For a risk factor with a discrete value, its fact CF(e) is a discrete
value. Taking the state of “Retaining pile not embedded into rocks (e1 )” as
an example, CF(e1 ) is represented by “1” in the “Not embedded into rocks”
state, otherwise as “0”, as seen in Eq. (1).
1 N ot mbedded into r ocks
C F(e1 ) (1)
0 Embedded into r ocks
• Uncertain fact. An uncertain factor refers to the risk factor which has an
interval in measurement due to randomness, fuzziness, and greyness. For
a risk factor with a continuous value within an interval, its fact CF(e) can
be represented by a sectional function. The sectional threshold value can
be obtained from theoretical calculations or empirical formulas. Taking the
state of “Water head difference between the inside and outside of a foun-
dation pit (e2 )” as an example, the water head difference x is a continuous
value, which has an approximately linear relationship with the evidence
credibility of e2 . Therefore, the sectional function expression for CF(e2 ) is
given by Eq. (2).
⎧
⎨ 0.2 x < 2m
C F(e2 ) = 0.8 x−2 + 0.2 2m ≤ x < 4m (2)
⎩ x−2
0.8 x ≥ 4m
During the safety risk identification in tunnel construction, both certain and uncer-
tain facts exist in the establishment of the fact base. Taking the risk “Flowing sand
of foundation pit (f 1 )” as an example, the safety-related fact base is constructed
according to explicit and tacit knowledge, as seen in Table 2.
(2) Rule base
In tunnel construction projects, safety risk identification knowledge involves
various kinds of theories and practical knowledge, such as construction
methods, engineering geology, structural design, accident analysis, and reli-
ability theory. Most of the knowledge is empirical, displaying patterns of
diversity, hierarchy, inheritance, and uncertainty. Existing knowledge repre-
sentation methods consist of predicate logic, semantic network, production
4 Knowledge Base Development for Construction Safety 213
Table 2 Safety-related fact base for the risk “Flowing sand of foundation pit (f 1 )”
Risk factors Fact code Certain? Fact attributes CF(ei )
Retaining pile FID.025 Yes Embedded 0
embedded into rocks Not embedded 1.0
(e1 )
Water head difference FID.104 No <2 m 0.2
of foundation pit (e2 ) 2–4 m 0.8(x − 2)/(4 − 2) + 0.2
≥4m 0.8
Geological conditions FID.100 Yes Medium-coarse sand 0.6
of foundation pit (e3 ) Fine sand 0.7
Silt sand 0.8
Powder sand 0.5
Others 0.2
Reinforcement state of FID.101 Yes Reinforcement 0
foundation pit (e4 ) Non-reinforcement 1.0
Embedded length ratio FID.026 No <0.4 0.6
of the bored pile (e5 ) 0.4–0.7 0.6(0.7 − x)/(0.7 − 0.4)
+ 0.2
≥0.7 0.4
Bed thickness of FID.102 No <400 mm 0.8
foundation pit (e6 ) ≥400 mm 0.8e−(x − 400)/400
rule, and frame method. Owing to that the production rule “IF (premise)
THEN (conclusion)” provides a powerful tool for knowledge representation
and reasoning under uncertainty, the rule can be employed to describe the
empirical knowledge for safety risk identification.
However, due to the lack of classification and relevance, this type of rule is
likely to cause combination explosion and low reasoning efficiency. Therefore,
an extended rule representation as seen in Eq. (3) is adopted, with the hierarchy
and uncertainty of rules being fully considered. The causal relationship between
the premise (e) and the conclusion (h) is represented by Eq. (4).
where λ refers to the rule threshold. The rule can be activated if and only if
the rule evidence CF(e) ≥ λ. Generally, λ is defined within a range [0.5, 1.0],
depending on the importance of the project.
The rule credibility CF(h, e) with a range [0, 1] is related to the credibility
degree of the rule. CF(h, e) can be calculated by Eq. (5).
214 Expert Systems
P(h|e )−P(h)
P(h|e ) ≥ P(h)
C F(h, e) = 1−P(h)
P(h|e )−P(h) (5)
P(h)
P(h|e ) < P(h)
where P(h) refers to the prior probability of the conclusion h, and P(h, e) refers
to the conditional probability of the conclusion h when the premise e is true.
(3) Case base
A large number of accident records are accumulated, which provide prior
knowledge for risk identification in tunnel construction. By matching the char-
acteristics of a specific project with cases in the case base, the very similar
accident event can be detected. Compared to rules, we do not need to construct
explicit rules as seen in Eq. (2) for case representation. Meanwhile, the case
base is an open system, which is easy to maintain. There is no need to perform
dependency and consistency checks when adding new cases into the case base.
To facilitate the efficiency during the case matching process, cases are required
to be structured in accordance with the characteristic of domain cases. There
are numerous methods for the case representation, including feature vectors,
object, frame, and category representation methods. Accident cases in tunnel
construction generally are characterized by multi-level and multi-attribute
patterns. Therefore, a combination of frame and category representation
methods can satisfy the demand, which is adopted to represent the acci-
dent cases in tunnel construction. A complete case record consists of several
attributes, including Construction Project Data (C 1 ), Accident Data (C 2 ), etc.
Each attribute can then be further refined. A simplified structure for a tunnel
case is presented in Fig. 4.
Initially, either rules or cases are stored in text forms, which are likely to result in
inconvenience in retrieval and traversal operations for knowledge-based reasoning.
Due to the simple structures, high independence, and strong description ability, the
relational database can be widely employed to store various kinds of knowledge.
4 Knowledge Base Development for Construction Safety 215
Rule
Rule premise Rule code A10 Rule conclusion
Rule premise desc A100
R_Premise code A10 R_Conclusion code A10
Worksite type A10
R_Premise name A100 R_conclusion desc A100
Construction method A10
Sign A10 CF conclusion A100
Related Rule type A10 Reasoning
Value A100 Memo A100
Rule group number A10
Fact R_Premise CF N4.2 Rule CF N4.2
Reference Rule threshold N3.1 R_Conclusion code <pi>
Fact code A10 R_Premise code <pi>
Conclusion code A10
Fact attribute A100
Y/N A100 Rule code <pi>
Explained
Fact CF N4.2
Fact SIM N5.2
Case
Fact code <pi> Case conclusion
Reference Case premise Case code A10
Case premise desc A100 C_Conclusion code A10
C_Premise code A10
Worksite type A10 C_conclusion desc A100
Construction data A100
Construction method A10 SIM conclusion A100
Accident data A100 Related Case type A10 Reasoning Risk name A200
Problem reasons A100
Case group number A10 Consequence A200
Case SIM N5.2
Case SIM N5.2 Response solution N5.2
C_Premise code <pi> Case threshold N3.1 Case application A200
Conclusion code A10
C_Conclusion code <pi>
Case code <pi>
Therefore, to improve data matching and reasoning efficiency, rules and cases in
tunnel construction can be stored in a relational database, which can be retrieved
using a structured query language (SQL). The database structure of the knowledge
base is shown in Fig. 5.
Both rule and case premises are stored in the fact table, aiming to ensure the
consistency of terms in the knowledge base. The rule premise table describes the
various states of risk factors. The rule table describes a basic structure of a production
rule, representing the relation between the case premise and its conclusion. The case
premise table describes the various attributes of case records. The case table describes
the structure of a case, representing the relation between case characteristics and their
consequences. The case consequence table describes the consequence of the accident
case and provides explanation information for the risk identification results and the
corresponding control measures.
Application layer
Data layer
MS SQL
JAVA Tomcat
Server
(1) Data layer. B-RIES is developed in the JAVA programming language and
running in the Tomcat container. Mainstream data warehouses consisting of
the Microsoft SQL Server and Oracle database system are compatible with
the system. In addition, four built-in databases, namely Fact Base, Rule Base,
Case Base, and Project Base, are used to provide data support for the intelligent
inference for the safety risk identification process.
(2) Standard and protocol layer. This layer provides a unified platform solution
for data communication, processing, and the interface between the system and
external environments using four main types of protocols. To be specific, the
IFC standard is employed to extract the information of model components
from the input BIM of a specific project. The role-based access protocols aim
to set up authorities among dispersed clients, experts, and maintainers. The
regulatory and security protocols are used to ensure the security of the network
environment, protecting the system against external attacks.
(3) Application layer. This layer provides all functions and services for the system
application, mainly consisting of three subsystems, including BIM Extraction
Subsystem (BES), Knowledge Base Management Subsystem (KBMS), and
Risk Identification Subsystem (RIS). Specifically, BES aims to extract the
engineering parameters of BIM models, KBMS aims to acquire, represent,
5 BIM-Based Risk Identification System (B-RIES) 217
store, and update the risk-based knowledge in tunneling fields, and RIS aims
to calculate the risk occurrence probability and propose control measures in
real-time. Each subsystem is further composed of several functional modules.
B-RIES can be used to automatically determine construction safety risks and the
corresponding control measures. B-RIES is a Browser/Server (B/S) system. Figure 7
illustrates the overall workflow for the safety risk identification process in B-RIES.
In general, B-RIES will go through three steps during the safety risk identification
process, as elaborated below.
BIM
Convert
IFC XML
Credibility
Factor CF(e)
Reasoning
Strategies
CBR RBR
Control
Risk Category
Measures
Technical risks Reinforcement measures
Geological risks Management measures
Step 3
Environmental risks Prevention measures Results Output
Forewarning Extremely Very
Risk Ranking Dangerous Dangerous
Releasing
Dangerous Safe
Fig. 7 The overall workflow for the safety risk identification process in B-RIES
218 Expert Systems
where CF(ei ) (or CF(ci )) refers to the credibility factor of the ith parameter
for the specific project (or the optional case) while being matched with the fact
base.
(2) RBR: RBR is incorporated into the knowledge reasoning mechanism, aiming
to find out valid rules which can be executed in the reasoning process. At first,
some alternative rules are selected from the rule base using rule matching.
Next, the valid rule is determined if and only if CF(e) is greater than the
given threshold λ. In general, the premise evidence e is a combination of risk
factors ei (i = 1,2, …, n), including disjunction, conjunction, and weighted
combinations. Accordingly, CF(e) is calculated by Eqs. (7)–(9), respectively.
Finally, the valid rule is executed to calculate the credibility degree of the
conclusion of the risk event using the former Eq. (4).
n
C F(e1 (w1 ) ∧ e2 (w2 ) . . . ∧ en (wn )) = wi × C F(ei ) (9)
i=1
Based on the calculated results through the knowledge reasoning mechanism, the
safety risk analysis and corresponding control measures can be worked out. To eval-
uate the degree of the identified safety risk, during the CBR approach, the degree
of the identified risk is equal to the risk level in the target case. During the RBR
approach, the conclusion credibility CF(h), which can be calculated by Eq. (2), is
used as an indicator for the risk evaluation.
With regard to CF(h) with a range of [0, 1], we divide its calculated value into
the following four levels, namely “I (Safe, [0, 0.25)), II (Dangerous, [0.25, 0.5)),
III (Very Dangerous, [0.5, 0.75)) and IV (Extremely Dangerous, [0.75, 1.0))”. These
four levels are defined by four different colors, namely, “Green, Yellow, Orange and
Red”, respectively. According to the risk identification results, some corresponding
safety control measures can be determined, and early-warning signals can be released
in a visualization scenery.
6 System Application 221
6 System Application
To relieve the pressure of urban traffic jams across the Yangtze River, the construc-
tion of Wuhan Metro Line Two (WMLT) in the city of Wuhan, China, started on
August 28, 2006. The 27.7 km route, with 21 stations and a total investment of nearly
US $3.2 billion, passed underground on a northwest-southeast alignment between
the Hankou and Wuchang districts. Due to complexities in construction environ-
ments, high potential safety risks existed during the excavation progress of WMLT.
In this section, B-RIES is implemented to identify safety risks at the pre-construction
stage of tunnel construction, aiming to provide guidelines for safety assurance at the
construction stage. A realistic case in the Mingdu station located at WMLT in China
is used as an example for demonstrating the system application.
The Mingdu station is an underground 2-story station. This station was started on
November 26, 2008, with an outlined length of 241.3 m, an average width of 18.88 m,
and a total floor area of 11,932.1 m2 . A foundation pit with a depth of around 15 m was
constructed using the cut and cover excavation method. The aerial view of the BIM
model of Mingdu station is shown in Fig. 8. The retaining structure was composed
of bored piles and jet grouting piles. Two high-rise buildings were adjacent to the
deep foundation excavation. Specifically, the Baoli Huadu Building was located in
the north of the foundation pit, while the Tibetan Middle School was located in the
south. The excavation space was limited due to the narrow working site. The working
site of the foundation pit excavation of the Mingdu station is shown in Fig. 9.
Fig. 8 Aerial view of BIM for the Mingdu station in Wuhan, China
222 Expert Systems
(a) (b)
Fig. 9 Working site of the excavation of the Mingdu Station: a steel support; b foundation pit
0.8
0.6
0.4
0.2
0.0
FID.025 FID.036 FID.100 FID.101 FID.126 FID.128
Fact Code
ground subsidence and construction delay. The hazard level of this accident
was evaluated to be a level of III during the accident investigation. Based on
the causal relationship and control process of the accident in the target case,
the project engineers can have a deep understanding of the potential safety risk
at the Mingdu station. Also, the risk level is identified and some relevant safety
control measures can be proposed.
(3) RBR calculation
Based on the aforementioned procedures in RBR, the rule base was first
matched to select suitable rules. It is known that the worksite type was
“station"(represented by No. 10) while the construction method was “cut and
cover” (represented by No. 1110). Table 5 listed some alternative rules for the
risk of water gushing. Using Eqs. (7)–(9), CF (e) of R1-11 and R1-14 was
less than the threshold λ, and therefore, both were discarded. Next, R1-26 was
activated and CF(e) was calculated to be 0.76 as follows:
6 System Application
Thus, CF(e) > λ = 0.5, indicating that the rule R1-26 was valid. Accordingly,
Eq. (4) was used to calculate the conclusion credibility as follows:
As a result, the risk of “water gushing at foundation pit bottom” was rated at
a risk level of III (Very dangerous).
(4) Risk identification and report
The results can be testified when RBR and CBR worked in a parallel way. In this
case, the results were consistent between these two approaches. The safety risk
analysis and control measures can then be determined, including the risk type,
description, location, possible consequences, risk level, and control measures.
The report of the risk of “water gushing at foundation pit bottom” at the Mingdu
station is shown in Fig. 11. It is clear to see that the possible consequences
of this risk event include the instability of the retaining structure and the steel
structure, as well as the casualties of workers. As a response, retaining piles
should be embedded into rocks, the water head difference should be reduced,
and the bottom of the foundation pit should be reinforced.
Fig. 12 Risk event of water gushing at the Mingdu Station: a location of water gushing, and b a
precipitation well
According to the risk identification results from B-RIES, the designers gained insights
into a better understanding of the safety risk in the construction of the Mingdu station.
The contractor was suggested to strictly implemented the work of dewatering on
the working site during the construction stage, controlled the water head difference
between the inside and outside of the foundation pit, and reinforced the foundation pit
bottom to reduce the risk limit. Also, emergency supplies, equipment, and personnel
were prepared in advance to cope with the risk events which were very likely to
occur.
On September 14, 2009, a risk event of water gushing was witnessed at the
construction site of the Mingdu station, as shown in Fig. 12a. The volume of the
gushing water was small, and its impacts on the safety of nearby retaining piles
and adjacent buildings were limited, since some prevention measures were adopted
ahead of time. Subsequently, a precipitation well was built up at a location of the
water gushing and continued to be effective until the main structural work of the
Mingdu Station was completed, as shown in Fig. 12b. As a matter of fact, with effec-
tive safety control measures implemented, the construction of the Mingdu station
went smoothly, and no accidents were reported during the excavation construction
process. Finally, the main underground structural work was completed without any
project delay on August 9, 2010.
7 Summaries
This chapter describes an innovative approach that integrates BIM and ESs to address
deficiencies in the traditional safety risk identification in tunnel construction. B-RIES
is consisting of three main built-in subsystems, namely BIM extraction, knowledge
base management, and risk identification subsystems, providing real-time support
228 Expert Systems
for decision making in the tunnel risk management. Firstly, engineering parameter
information related to risk factors should be acquired from BIM models of a specific
project, where the IFC standard acts as an intermediary between the BIM data and
risk-related information. Then, an integrated knowledge base, composed of fact, rule
and case bases, is built up to systematize the fragmented explicit and tacit knowledge
for knowledge sharing and communication among dispersed clients and domain
experts. A hybrid reasoning approach with CBR and RBR combined can be used to
improve the flexibility and compatibility of the system reasoning capacity. Finally, the
results of safety-related risks and their occurrence probability and control measures
can be obtained in real-time. B-RIES can be used by practitioners in the industry as
a decision support tool by providing guidelines on risk assessment and management
for multi-parties in tunnel construction, enabling to increase the likelihood of a
successful project in a complex environment.
Traditionally, only experienced engineers and experts can identify potential safety
risks and be aware of unexpected dangerous situations and accidents. This system
provides a solution to automate safety risk identification in tunnel construction. One
benefit of the developed system is that the dependence on domain experts can be
reduced, and the knowledge sharing and communication in construction management
among dispersed clients and domain experts can be facilitated. Furthermore, the low-
efficiency in traditional information extraction from 2D drawing documents can be
significantly improved. Due to the implementation of the developed B-RIES and the
efforts of all participants, no accidents are reported during the construction of the
WMLT.
A large number of rules serving safety risk identification were obtained from
domain experts for system development. Numerous domain experts participated in
the collection, editing, reorganizing work of the safety-related knowledge resources,
making an essential contribution to securing a qualified rule base for the development
of B-RIES. Automatic knowledge acquisition and decision rule generation regarding
different knowledge resources should be of significance in developing a real-time
expert system for improved decision making.
References
1. Abu-Naser SS, Aldahdooh R (2016) Lower back pain expert system diagnosis and treatment.
J Multidisc Eng Sci Stud 4:441–446
2. Beckman TJ (1999) The current state of knowledge management. CRC Press
3. Behm M (2005) Linking construction fatalities to the design for construction safety concept.
Saf Sci 43:589–611
4. Bohlouli M, Mittas N, Kakarontzas G, Theodosiou T, Angelis L, Fathi M (2017) Competence
assessment as an expert system for human resource management: a mathematical approach.
Expert Syst Appl 70:83–102
5. Chou J-S (2009) Generalized linear model-based expert system for estimating the cost of
transportation projects. Expert Syst Appl 36:4253–4267
6. Chou J-S, Tseng H-C (2011) Establishing expert system for prediction based on the project-
oriented data warehouse. Expert Syst Appl 38:640–651
References 229
31. Zavadskas EK, Turskis Z, Tamošaitiene J (2010) Risk assessment of construction projects. J
Civ Eng Manag 16:33–46
32. Zhang L, Wu X, Ding L, Skibniewski MJ, Lu Y (2016) Bim-Based Risk Identification System
in tunnel construction. J Civ Eng Manag 22:529–539
33. Zhou W, Whyte J, Sacks R (2012) Construction safety and digital design: a review. Elsevier,
Amsterdam, pp 102–111
Computer Vision
1 Introduction
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 231
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_10
232 Computer Vision
and others, aiming to boost efficiency and reliability in the domain of structural
health monitoring.
In general, classical approaches in computer vision can be considered as feature
engineering containing two major aspects. The first one is to extract features from
images, which can be represented by a compact feature vector. The other is feature
matching to establish feature correspondence between two images about the same
object. However, the quality of hand-crafted extracted features will greatly influ-
ence prediction performance, and thus such a method is still limited by subjectivity
and domain expertise. Also, it requires a great number of preprocessed images to
make the detection process unadaptable, tedious, and inefficient. Even worse, these
hand-crafted features have no adequate ability to differentiate cracks and complex
backgrounds in low-level image cues, which are less applicable in images with large
variations.
Up to date, deep learning techniques are driving advances in computer vision
for resolving weaknesses of classical crack detection models. It aims to realize an
automated process in capturing intricate structures of large-scale data by models
with multiple processing layers [22]. Remarkably, one of the most important models
for automated feature learning and supervised classification is called the convolu-
tional networks (CNNs/ConvNets), which belong to a kind of neural network to take
images as inputs. Its unique properties in the partial connections, sharing weights,
and pooling layers make CNNs to automatically capture the grid-like topology of
images under fewer computations, and then generate promising detection results [4].
Implementation of CNN-related methods has been proven effective in classifying
multiple damage types of structural deterioration [24]. Moreover, some more compli-
cated CNN-based architectures have been developed and applied in the task of crack
detection, including U-net, SegNet, CrackNet, and others. Among these advanced
models, fully convolutional networks (FCNs) [16] have attracted great attention,
which can be useful in transforming image pixels to pixel categories. The core idea
behind FCNs is to convert the last fully connected layer into CNNs to a convolution
layer for training an end-to-end pipeline. FCNs have shown their promising perfor-
mance of semantic segmentation in some works [12]. However, it is likely for FCNs
to ignore the relationships between pixels, and then to poorly influence the reliability
of scene segmentation.
To address the above-mentioned issue from FCNs, the attention mechanism that
has brought great breakthroughs from natural language processing (NLP) can be
deployed in the computer vision task. More specifically, the role of the attention
mechanism is to assign different weights to features according to their importance,
allowing for making the network quickly focus on the more relevant parts of the
input. Therefore, the emerging self-attention is good at catching long-range depen-
dencies with no increase of parameter number, contributing to boosting the perfor-
mance of image classification, object detection, semantic segmentation, and others
[3]. Inspired by the successful application of FCN and attention mechanism, this
chapter attempts to design a hybrid network structure for a more credible process of
1 Introduction 233
Inspired by the structure and function of biological neural networks in human brains,
deep learning encompassing multiple processing layers enables the computer to
perform a process of learning and problem solving. The promise of deep learning has
been highlighted in capturing intricate structures of large-scale data and addressing
complicated problems with no human intervention in an end-to-end learning frame-
work. Currently, there is an increasing interest in applying deep learning to drive a
great revolution in traditional computer vision. A variety of deep learning algorithms
have developed to support the machine to process and interpret the unstructured visual
data from digital images or videos automatically, and thus, data representations with
different levels of abstraction can be extracted and learned effectively resulting in
improved accuracy and robustness [13]. In general, the deep learning-based computer
vision task (i.e., classification, detection, and segmentation) follows four key steps,
including feature extraction, forward propagation, loss calculation, and backpropa-
gation. Some recent deep learning-based studies with great performance have been
listed in Table 1, which are divided into three categories, namely the ConvNet-based
model, Region-based model, and FCN-based models.
(1) ConvNet-based models
The fundamental and mainstream deep learning model named ConvNet has
three kinds of neural layers (convolutional layers, pooling layers, and fully
connected layers). which has reached remarkable achievements in computer
vision. For one thing, ConvNet shows a great capability of automatically
extracting meaningful features from a great number of raw images. For another,
since ConvNet needs less image pre-processing, it can be easily trained and it
owns fewer parameters and connections compared to the standard feedforward
neural networks. Generally speaking, the ConvNet-based models regard crack
detection as a binary classification problem, where each pixel of the image is
labeled by one class. However, an issue to note is that the process of training the
234 Computer Vision
Table 1 (continued)
Method Limitation Example
Authors Remarks
Cheng and Wang [7] Faster RNN is
designed to detect
sewer pipe defects
from closed-circuit
television (CCTV)
images, which is
trained in 1260 images
with the resolution
between 1440 × 720
and 320 × 256 pixels
Deng et al. [10] Concrete cracks from
real-world images
with complex
backgrounds are
identified
automatically based
on Faster R-CNN. 160
images in 3264 ×
2448 pixels are used
for training and testing
FCN-based models Lack of consideration Yang et al. [23] The FCN model with
of relationships VGG 19 is
implemented to
identify and measure
diverse cracks
concurrently at the
pixel level, whose
performance is
examined in 800
images with the
number of pixels
varying from 1 to 100
Dung and Anh [12] An FCN-based crack
detection method
under VGG16 encoder
is developed for
semantic
segmentation on 500
concrete crack images
in 227 × 227 pixels
(continued)
236 Computer Vision
Table 1 (continued)
Method Limitation Example
Authors Remarks
Bang et al. [2] The encoder-decoder
network based on
FCN is built as a
pixel-level detection
method to identify
road cracks from 527
images with a
resolution of 2560 ×
1440, which are taken
from black-box
cameras
This chapter proposes a new structure termed SCHNet to distinguish crack and non-
crack images at the pixel level. The novelty of SCHNet lies in the combination of
the VGG base net and a self-attention mechanism, which helps in better capturing
long-range contextual information in both the spatial and channel dimensions. On
the one hand, VGG is a deep convolutional neural network model supporting up to
19 layers. Its performance in object recognition has been proven outstanding on the
ImageNet dataset. On the other hand, a self-attention mechanism displayed in Fig. 1
consists of three parallel modules, including a feature pyramid attention module, a
spatial attention module, and a channel attention module. That is to say, after the
process of inputting raw images into the feature pyramid attention module, mean-
ingful features will be extracted automatically and then fed into both the channel and
spatial attention module. Subsequently, the two output attention maps are integrated
using an element-wise summation operation, resulting in segmentation images in the
size of input images. In short, these three attention modules are the key parts of the
proposed SCHNet, contributing to the success of the high-level feature learning of
crack images. The detailed description of these modules along with the evaluation
metric and training process are provided below.
238 Computer Vision
Feature
Feature pyramid attention pyramid
module
Input
attention attention
matrix matrix
reshape reshape
Output
CONV
Spatial Channel
Attention Attention
Module Module
Fusion
Fig. 1 The framework of the proposed SCHNet approach for automated crack detection. Note
CONV represents the convolutional layer
It is known that many recent methods adopt skip connections or pyramidal shapes
to associate low-level feature maps for improving segmentation performance. But
these methods are prone to suffer from low efficiency and require high memory, since
independent predictions are made at all levels. To deal with the problem, a feature
pyramid attention module is deployed to quickly extract and combine features by
a top-down pathway and lateral connections, resulting in strong semantics at all
scales. As shown in Fig. 2, the output features from the base net VGG are input
into a convolutional layer in a kernel size of 1 × 1. Then, it will be multiplied
with the pyramid attention features in three different pyramid scales of 3 × 3, 5
× 5, 7 × 7, which are created by a U-shape structure. In other words, different
scale context features can be fused in the feature pyramid attention module, which
are then multiplied with original feature maps pixel-wisely. Eventually, the pixel-
level attention information for high-level feature maps can be built, contributing to
improving the feature extraction.
3 Computer Vision Framework 239
Basenet
1X1
CONV 7X7 7X7
CONV CONV
Global downsample
pooling upsample
1X1
X + 5X5 5X5
16X16 CONV CONV
CONV
downsample
8X8 +
upsample 4X4 3X3 3X3
upsample CONV CONV
+ output
Fig. 2 Feature pyramid attention module structure. Note All the convolution layers are followed
by batch normalization. 4 × 4, 8 × 8, where 16 × 16 means the resolution of the feature map. Red
and blue color represent the process of upsampling and downsampling, respectively
Notably, the context relationship in FCNs plays a key role in feature representation,
which helps to capture global spatial dependencies regardless of locations. Never-
theless, it is more likely for local features of traditional FCNs to misclassify objects
and stuff [9]. To improve the effectiveness in modeling rich contextual dependen-
cies over local feature representation, a spatial attention module in Fig. 3 is utilized,
which can encode a wider range of contextual information into local features. To
be more specific, after the feature map A ∈ R c ×h ×w is fed into three convolution
layers with the same kernel size of 1 × 1 and batch normalization, three new feature
maps B, C, and D ({B, C, D ∈ R c ×h ×w }) are generated. In particular, B and D
n×c
are reshaped to R , where n = h × w is the number of pixels in each channel.
Concurrently, the transposition of the new feature map labeled by C is multiplied by
B to get a self-attention map S ∈ R n×n . Then, the result of S × D goes through a
softmax layer to calculate the self-spatial attention maps E ∈ R c ×h ×w (Fu et al.). As
for the softmax function, it aims to map the non-normalized output to a normalized
probability distribution over predicted classes, which has been widely adopted in the
last layer of a neural network-based classifier.
Herein, an example is provided based on an image that is in a size of 64 × 64 × 3
pixels and contains three channels called RGB (R is red, G is green, and B is blue).
After the convolutional operations with a kernel size of 1 × 1, the output feature
map is reshaped to a two-dimensional matrix of 3 × 4096. Thus, the new shape
derived from the reshape operation is compatible with the original shape to reserve
all information.
240 Computer Vision
B C D
+
X
output feature maps self-attention map
F
X
self-spatial
attention maps E
Fig. 3 Structure of the proposed spatial attention module. Note that the capital letter in bold italics
is the label of feature maps
B C D
X +
channel
attention map output feature maps
X F
Fig. 4 Structure of the proposed channel attention module. Note that the capital letter in bold italics
is the label of feature maps
3 Computer Vision Framework 241
X ∈ R c ×c can be directly calculated from the original feature maps A ∈ R c ×h ×w .
That is to say, A is reshaped to C ∈ R c ×n , and then it is multiplied with the trans-
position D. Finally, a softmax layer expressed in Eq. (1) is utilized to output the
channel attention map X ∈ R c ×c [14]. After the multiplication of X and B, the result
c ×h ×w
E is reshaped to R . As formulated Eq. (2), the final output F ∈ R c ×h ×w
is generated based on the element-wise sum operation of A and E scaled by the
parameter β. In other words, the final representation feature of each pixel can be
obtained via a weighted sum of the features of all pixels and original features, which
is helpful in modeling the rich contextual dependencies over image regions. Besides,
it can be employed to strengthen class-dependent feature maps and boost feature
discriminability.
exp(Am · An )
xnm = C (1)
m=1 exp(Am · An )
C
Fj = β x q p A p + Aq (2)
p=1
where x nm refers to the mth pixel’s impact on the nth pixel, and β gradually learns a
weight from 0.
Regarding the training process, features in the spatial and channel dimensions can
be captured from the two parallel attention modules. Then, these obtained features
are aggregated to model long-range dependencies, resulting in the improved feature
representation for pixel-level segmentation. More specifically, outputs of two atten-
tion modules are fused through an element-wise summation, which will then be put
into a convolutional layer in the kernel size of 3 × 3 to learn the final prediction
map. Cross-entropy loss (also known as log loss) based on the natural logarithm is
employed as the loss function to measure the difference in two probability distribu-
tions from the predictions and ground truth. Since there are only two classes in the
image (the crack and the background), we rely on the binary cross-entropy in Eq. (3)
to deal with the binary classification task in this chapter. The value of cross-entropy
decreases when the predicted class probability is much closer to the actual label. Our
goal is to minimize the cross-entropy during the model optimization process and
continuously update the model parameters through gradient descent.
1
N
Fig. 5 An example of computing IoU for various bounding boxes: a Excellent segmentation (with
an IoU of 0.91); b good segmentation (with an IoU of 0.82); and c poor segmentation (with an IoU
of 0.67)
where N is the total number of pixels, yi is the predicted value of the ith pixel, yi is
the corresponding actual value of the ith pixel.
For quantitatively assessing the segmentation performance of the developed
model, the metric called Intersection over Union (IoU) in the form of a Jaccard
index is adopted rather than the commonly used metrics like precision and recall in
the traditional machine learning-based classification tasks. That is because the (x,
y)-coordinates of the predicted bounding box cannot exactly match the ground-truth
bounding box, meaning that the simple consideration of object detection as a binary
problem is irrational. For this concern, IoU is defined in Eq. (4) as a ratio of the over-
lapping region in two boxes to their total combined regions, aiming to effectively
measure the similarity in the predicted bounding box and the ground truth bounding
box. A larger IoU can, therefore, represent the greater segmentation performance.
As an example, three pairs of ground truth and predictions under different values of
IoU are displayed in Fig. 5, which numerically indicate that how much the predicted
bounding box is close to the actual cracks.
Ar ea o f Overlap
I oU = (4)
Ar ea o f U nion
Besides, when cracks are detected in a large number of images, another metric
termed Mean IoU is formulated in Eq. (5) to evaluate the segmentation accuracy
among various images averagely.
N
Mean I oU = (1/N) n i_crack / n i_crack + n i_noncrack (5)
i=1
where N refers to the number of input images, ni_crack is the number of pixels that are
correctly segmented into cracks in each image, and ni_noncrack represents the number
of crack pixels that are segmented into the background.
4 Computer Vision for Automated Crack Detection 243
For the validation of the proposed SCHNet, a series of comprehensive experiments are
conducted on an open concrete crack image dataset named SDNET2018 [17]. There
are over 56,000 cracked and non-cracked images about concrete bridge decks in this
adopted dataset. The collection of original full images is based on a 16 MP Nikon
camera at a working distance of 500 mm without zoom, where the International
Organization for Standardization (ISO) sensitivity is 125, the image resolution is
4068 × 3456 px, and the surface illumination is between 1500 and 3000 lx. Each
original full image actually presents a physical area of about 1000 mm × 850 mm,
which is then segmented into several 256 × 256 px images on behalf of a physical
area in approximately 60 mm × 60 mm.
Since the early-age cracking in bridge deck concrete is a severe problem causing
the reduction of bridge life, this case study mainly concentrates on images of bridge
decks for enhancing bridge structure monitoring and assessment. To this end, a total
of 11,000 cracked and non-cracked images about the bridge deck at 256 × 256 px are
selected from SDNET2018. The length of cracks in images is in the range from 0.06
to 25 mm. Notably, various obstructions, like shadows, surface roughness, scaling,
edges, holes, and background debris, are inevitable in images from the dataset to
raise difficulties of crack detection.
Another step in data preparation is manual annotation, Herein, we rely on an online
annotation tool LabelMe to annotate a new object by clicking along the boundary of
the desired object, aiming to make images own semantic information. Some examples
of images used in the concrete crack dataset are given in Fig. 6, where the first,
second, and third column provides original images, images with manual labels, and
the corresponding ground truth, respectively. The main goal of this chapter is to
segment cracks on images under the condition that all cracks on images are treated
as one class.
For improving the robustness of the network, two key steps are conducted in the
model training process. One is to fine-tune all layers by backpropagation through
the whole network, and the other is data augmentation. For the first steps, the initial
learning rate is updated by a poly learning rate policy, which is multiplied by a factor
0.9
1 − total−iter
iter
after each iteration. This given factor expressed by a polynomial
decay function helps to drive the initial learning rate to the end learning rate within
the given decay steps. In this case, the initial value of three important parameters
named learning rate, momentum, and weight decay coefficients are set as 0.1, 0.9,
0.0001, respectively.
244 Computer Vision
(a)
(b)
(c)
(d)
Fig. 6 Examples of images used in the concrete crack dataset with a holes; b shadows; c rough
surfaces; d smooth surfaces
20% is responsible for verifying the model performance. The parameters are zero-
initialized since the random initialization will bring no significant impact on the
performance and convergence of the model. Besides, the dropout is included in the
original classifier nets.
The developed model is implemented by Tensorflow and trained on three NVIDIA
GFORCE 1080TI GPUs with an 11 GB memory. In avoidance of the out-of-memory
problem, the batch size is set to be less than 16 during the training and tuning process.
In experiments, the resolution of images and ground truths is downsampled to be 256
× 256 px. The approximately estimated training time is about 1–2 days.
The proposed SCHNet model has demonstrated a fast convergence speed and great
stability. To choose the best initial learning rate for configuring the neural network,
we set the initial value as 10–2 , 10–3 , 10–4 , and 10–5 in sequence. Figure 7 displays the
loss of the developed approach under the four candidate learning rates. Observably,
the loss all converges after 1000 training epochs, indicating a fast training speed and
high effectiveness of SCHNet. In particular, the lowest loss can achieve based on the
learning rate of 1 × 10–2 , which is smaller than half of the learning rate 1 × 10–5 . It
reveals that the adoption of a too small learning rate is improper, since a relatively
1.8 1.8
1.6 1.6
1.4 1.4
Loss
1.2
Loss
1.2
1.0
1.0
0.8
0.8
0.6
0.6
0.4
0.4
0.2
0 1000 2000 3000 4000 5000 6000 0 1000 2000 3000 4000 5000 6000
Epoch Epoch
(a) (b)
2.4
2.2 learning rate 1e-4 learnin rate 1e-5
2.0 2.2
1.8
2.0
Loss
1.6
Loss
1.8
1.4
1.2
1.6
1.0
1.4
0.8
0.6 1.2
0 1000 2000 3000 4000 5000 6000 0 1000 2000 3000 4000 5000 6000
Epoch Epoch
(c) (d)
Fig. 7 Cross entropy loss under different initial learning rates: a 10–2 ; b 10–3 ; c 10–4 ; d 10–5
246 Computer Vision
long training time will be taken to reach the minimum of the loss function. Therefore,
we set the initial learning rate of 1 × 10–2 as the optimal value, enabling the model
to learn the problem stably and efficiently.
50
100
50
100
Fig. 8 Visualization of the differences in segmentation results with feature pyramid attention
module (FPA), channel attention module (CAM), and spatial attention module (SPM): a orig-
inal image; b without any attention module; c with FPA and CAM; d with FPA, CAM, and SPM;
e Ground truth. Note X- and Y-axis represents the width and length of an image (unit: pixel),
respectively
4 Computer Vision for Automated Crack Detection 247
0 0 0 0
First
Layer
100 100 100 100
0 0 0 0
Third
Layer
100 100 100 100
0 0 0 0
Fifth
Layer
100 100 100 100
0 0 0 0
Eighth
Layer
100 100 100 100
Fig. 9 Learned features at different convolutional layers from four testing images: a Image 1;
b Image 2; c Image 3; d Image 4. Note X- and Y-axis represents the width and length of an image
(unit: pixel), respectively
is made among SCHNet, DeepLab-v3 [6], PSPNet [26], and FCN [16]. The
runtime speed is measured on a single GTX1080TI and summarized in Table
2. All the models are evaluated without any pre- or post-processing to ensure
fairness. As listed in Table 2, DeepLab-v3 shows the slowest computational
speed (8.5 frames per second (FPS) due to the ultra-deep VGG model as the
backbone network. On the contrary, FCN runs the fastest (24.1 FPS) because of
its small model size. But Table 5 finds out that the segmentation performance
of FCN is relatively poorer than SCHNet. Therefore, it can be concluded that
SCHNet can run at an acceptable speed (16.3 FPS) to reach the highest Mean
IoU (85.31%).
(3) SCHNet can learn features automatically and reliably through updating the
weights of receptive fields, which does not require additional feature extraction
techniques that are the necessity in other image-based studies for segmentation.
Figure 9 takes four crack images as examples to visualize the receptive field,
which is acknowledged as the learned features from networks. The number of
receptive fields of the designed architecture is 24 with 64 × 64 × 3 dimensions.
The visualized features indicate whether the network needs more training and
what kinds of features are recognized by the trained network. It can be seen
248 Computer Vision
Fig. 10 Some qualitative crack segmentation results by using SCHNet: a original images; b ground
truth; c segmentation
Table 2 Comparison of
Models FPS
running time with different
models DeepLab-v3 8.5
PSPNet 10.2
Dilated FCN 24.1
SCHNet 16.3
Note that FPS is the abbreviation of frames per second
4 Computer Vision for Automated Crack Detection 249
from Fig. 9 that spots and lines become clearer as the convolutional layer
goes deeper. Herein, eight convolutional layers in the developed network are
suitable to train the network well in this experiment. Thus, features in the last
row of Fig. 9 can be rationally taken as crack features, while features in the
second and third rows are speculated as concrete surface cavities or aggregates
in the training data set. Besides, although receptive fields of a well-trained
network generally have smooth patterns, noisy features with various colors are
unavoidable due to the complex and arbitrary patterns in concrete surfaces. But
the segmentation results in the last column of Fig. 10 are very consistent with
the ground truth in the middle column, which implies that the proposed model
is able to well deal with noise to reduce the negative effects from noise.
Table 3 Results of the ablation study on the concrete crack validation set
Models Base net FPA SAM PAM Mean IoU (%)
Dilated FCN VGG16 70.03
√
SCHNet VGG16 72.63
√ √
SCHNet VGG16 76.62
√ √
SCHNet VGG16 77.37
√ √ √
SCHNet VGG16 80.13
Dilated FCN VGG19 72.54
√
SCHNet VGG19 74.16
√ √
SCHNet VGG19 79.31
√ √
SCHNet VGG19 79.92
√ √ √
SCHNet VGG19 83.42
Note FPA represents feature pyramid module; SAM represents spatial attention module; PAM
represents channel attention module
250 Computer Vision
Table 4 Performance comparison between different strategies on the concrete crack validation set
Models Base net DA Multi-grid MS Mean IoU (%)
SCHNet VGG19 83.42
√
SCHNet VGG19 84.31
√ √
SCHNet VGG19 84.86
√ √ √
SCHNet VGG19 85.31
Note DA represents data augmentation with random scaling. Multi-grid represents employing a
multi-grid method, and MS represents multi-scale inputs during inference
Table 5 Segmentation
Models Basenet Mean IoU (%)
results using different models
on the concrete crack U-net [20] – 71.4
validation set DeepLab-v3 [6] – 75.7
PSPNet [26] VGG19 77.8
Ding [11] VGG19 71.60
Dilated FCN [16] VGG16 70.30
Dilated FCN [16] VGG19 72.54
SCHNet VGG16 80.13
SCHNet VGG19 85.31
Note “–” indicates the base net is not applied in the model
combination of the feature pyramid and spatial attention (or channel atten-
tion) modules can increase the value of Mean IoU to 76.62% (or 77.37%).
When the three attention modules work together, the Mean IoU can reach
up to the highest (80.13%). Moreover, when the base net VGG19 is used
to initialize the parameters, the network with three attention modules can
further enhance Mean IoU to peak at 83.42%. In contrast to the new base-
line model (FCN with VGG19), the feature pyramid attention, the spatial
attention, and the channel attention module under VGG 19 can individually
increase Mean IoU by 1.62% (74.16–72.54%), 5.15% (79.31–74.16%), and
5.76% (79.92–74.16%), respectively.
(2) Three new strategies, including data augmentation, multi-grid method, and
maps, help to increase the Mean IoU of SCHNet to 85.31%. From experimental
results in Table 4, it suggests that the developed approach experiences a 0.89%
increase by using the data augmentation with random scaling. In other words,
enriching the scale diversity of training data can act as an effective way to
improve network performance. Besides, an additional 0.55% improvement in
Mean IoU can be brought by the multi-grid strategy. The idea behind such a
method is to employ a hierarchy of grids with different sizes in the last VGG
net, and thus the greater feature representation can be gained from a pre-trained
network.
4 Computer Vision for Automated Crack Detection 251
Apart from the accuracy of the proposed model in crack detection, it is necessary
to focus on two other crucial aspects including flexibility and robustness. To eval-
uate SCHNet more comprehensively, the concrete crack dataset is divided into four
scenarios, including Scenario I with holes on surfaces, Scenario II with shadows on
surfaces, Scenario III with rough surfaces, and Scenario IV with smooth surfaces.
Figure 11 presents some typical images in each scenario. Figure 12 provides the
segmentation results for images in the first column of Fig. 11. Discussions are
presented below.
(1) The developed SCHNet model is robust to different levels of noise, which
means that it is flexible and valid in different scenarios with holes, shadows,
and rough surfaces. Based on results from Table 6, the segmentation perfor-
mance in the four scenarios can be ranked as follows: Scenario IV > Scenario I
> Scenario III > Scenario II. More specifically, Scenario IV yields the highest
Mean IoU (88.67% with the base net VGG19 and 87.43% with the base net
VGG16, respectively). This duction is consistent with the reality, since images
in Scenario IV are more smooth with less noise. Inversely, the value of Mean
IoU is prone to be smaller in Scenario I, Scenario II, and Scenario III, which
is around 10% less than Scenario IV. It means that holes, shadows, and surface
roughness do exert a certain impact on crack segmentation. However, the Mean
IoU for the four scenarios all remains above 76%, meaning that influence from
noise is quite limited. Figure 12 gives examples about the segmentation perfor-
mance under four different scenarios in the test dataset, where the first, second,
and third column is the original image, segmentation results, and ground truth,
respectively. Even in Scenario I, II, and III under the relatively lower Mean
IoU, the developed model can reliably provide promising segmentation results
(see Fig. 12a–c). Therefore, the robustness of the model in crack detection is
validated.
(2) The developed SCHNet model with the base net VGG19 owns a greater gener-
alization ability in crack detection than the base net VGG16, as outlined in
Table 6. Taking Scenario IV as an example, its value of Mean IoU with the
base net VGG19 turns out to be 88.67%, while that with the base net VGG16
is 1.24% lower (87.43%). In a comparison of the segmentation performance
252 Computer Vision
Fig. 11 Typical images in the dataset under four different crack scenarios: a Scenario I (with
holes on surfaces); b Scenario II (with shadows on surfaces); c Scenario III (with rough surfaces);
d Scenario IV (with smooth surfaces)
between VGG16 and VGG19 under the same scenario, the value of Mean
IoU with the base net VGG19 is increased by 0.31% in Scenario I, 0.86% in
Scenario II, 0.74% in Scenario III, and 1.24% in Scenario IV, respectively. That
is due to the multi-nonlinear convolution layers in VGG19 that help to extract
more complex and expressive features. Therefore, SCHNet with the base net
VGG19 is capable of obtaining more representative features of images, which
eventually generate a more satisfying segmentation.
5 Merits of the Proposed Computer Vision Approach 253
50
100
50
100
50
100
50
100
Fig. 12 Segmentation performance in the dataset under four different crack scenarios in the dataset:
a Scenario I (with holes on surfaces); b Scenario II (with shadows on surfaces); c Scenario III (with
rough surfaces); d Scenario IV (with smooth surfaces)
254 Computer Vision
Table 6 Segmentation
Scenario Basenet Mean IoU (%)
results of Mean IoU on the
testing set under four Scenario I VGG16 77.93
different crack scenarios Scenario I VGG19 78.24
Scenario II VGG16 76.31
Scenario II VGG19 77.17
Scenario III VGG16 77.12
Scenario III VGG19 77.86
Scenario IV VGG16 87.43
Scenario IV VGG19 88.67
6 Summaries
of SCHNet lies in its ability to not only consider the semantic interdependencies in the
spatial and channel dimensions but also adaptively integrate local features into their
global dependencies. Except for the high value of Mean IoU, it should be noted that
SCHNet is robust to noises with a better generalization ability. Due to the promising
performance in crack detection, the application of SCHNet can be further extended
to the practical engineering project rather than the public dataset to make the model
more practical and effective. In addition, unmanned aerial vehicles (UAVs) can be
deployed to prepare the customized dataset. Plenty of photos or videos about the
cracks in various structural systems can be taken by UAVs, which will then be fed
into the developed deep learning approach for pixel-level automated crack detection.
References
15. Li S, Zhao X (2019) Image-based concrete crack detection using convolutional neural network
and exhaustive search technique. Adv Civil Eng 2019:1–12
16. Long J, Shelhamer E, Darrell T (2015) Fully convolutional networks for semantic segmentation.
In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 3431–
3440
17. Maguire M, Dorafshan S, Thomas RJ (2018) SDNET2018: a concrete crack image dataset for
machine learning applications
18. Pan Y, Zhang G, Zhang L (2020) A spatial-channel hierarchical deep learning network for
pixel-level automated crack detection. Autom Constr 119:103357
19. Prasanna P, Dana KJ, Gucunski N, Basily BB, La HM, Lim RS, Parvardeh H (2016) Automated
crack detection on concrete bridges. IEEE Trans Autom Sci Eng 13:591–599
20. Ronneberger O, Fischer P, Brox T (2015) U-net: convolutional networks for biomedical image
segmentation. In: International conference on medical image computing and computer-assisted
intervention. Springer, Berlin, pp 234–241
21. Shurun F (2018) Historic achievements of highway and bridge construction in China. http://
www.chinahighway.com/news/2018/1191607.php. Access Date: 19/11/2018
22. Voulodimos A, Doulamis N, Doulamis A, Protopapadakis E (2018) Deep learning for computer
vision: a brief review. Comput Intell Neurosci 2018:1–13
23. Yang X, Li H, Yu Y, Luo X, Huang T, Yang X (2018) Automatic pixel-level crack detection and
measurement using fully convolutional network. Comput Aided Civil Infrastruct Eng 33:1090–
1109
24. Zhang A, Wang KC, Li B, Yang E, Dai X, Peng Y, Fei Y, Liu Y, Li JQ, Chen C (2017) Automated
pixel-level pavement crack detection on 3D asphalt surfaces using a deep-learning network.
Comput Aided Civil Infrastruct Eng 32:805–819
25. Zhang L, Yang F, Zhang YD, Zhu YJ (2016) Road crack detection using deep convolutional
neural network. In: 2016 IEEE international conference on image processing (ICIP). IEEE, pp
3708–3712
26. Zhao H, Shi J, Qi X, Wang X, Jia J (2017) Pyramid scene parsing network. In: Proceedings of
the IEEE conference on computer vision and pattern recognition, pp 2881–2890
Conclusions and Future Directions
1 Main Conclusions
Recent decades have witnessed the rapid development of digital technology and
the growth of big data in the construction industry. In particular, AI implemen-
tation, which attempts to equip machines with human-like intelligent behavior and
reasoning, has gained a lot of attention. Currently, various AI techniques have created
tremendous value in revolutionizing the construction industry, leading to a more
reliable, automated, self-modifying, time-saving, and cost-effective process of CEM
[12]. In contrast to traditional computational methods and expert judgments, the
distinct advantage of AI lies in tackling complex and dynamic problems under great
uncertainty and intensive data, which is more likely to return accurate and convincing
results for improved decision-making. In general, AI-based CEM follows the key
steps including data acquisition and preprocessing, data mining based on appro-
priate models, and knowledge discovery and analysis. As a result, AI provides deeper
insights into big data, which commonly achieves value for itself from the following
three points.
(1) Modeling and pattern recognition
Modeling is a process of creating conceptual models in a standard, consistent, and
predictable manner, which can be known as a key prerequisite for further knowl-
edge interpretation and reasoning to resolve complex construction problems. It is
essential to ensure the high quality and reliability of the established model, since
the analytical results heavily rely on the model. There have been a wide variety of
knowledge representation methods based on rules, logic, and schemas that help to
construct models in an understandable form for computers. Also, rich information
of the real project, including declarative, procedural, structural, meta, and heuristic
knowledge, should be necessarily incorporated into the model to make it meaningful
and authentic. Apart from modeling, another way to extract information from data is
© The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd. 2021 257
L. Zhang et al., Artificial Intelligence in Construction Engineering and Management,
Lecture Notes in Civil Engineering 163,
https://doi.org/10.1007/978-981-16-2842-9_11
258 Conclusions and Future Directions
applied across the entire lifecycle of the project. For example, when the project
objectives are optimized at the planning stage, proper plans concerning strategy,
operation, and schedule can be developed to act as an important premise for the
project’s success. As for the phase of executing construction tasks in a compli-
cated site, optimization demonstrates its great value in better allocating resources,
arranging staff, determining layouts of facilities, and making corresponding adjust-
ments in a reasonable and timely manner. Lastly, O&M optimization intends to both
run the day-to-day operational tasks responsibly and perform maintenance measures
suitably at the optimal cost, and thus the infrastructure is likely to be kept in a satis-
factory state. Also, it contributes to reducing waste and building energy consumption
to support sustainability.
2 Future Directions
In the future, it is believed that CEM will continue to undergo rapid digital transfor-
mation. Therefore, more and more advanced technologies inspired by AI will serve
as the technological innovator in the domain of CEM. They will be implemented
and spread to the full lifecycle of the project ranging from the phase of design to
construction to O&M, which can potentially pave a more affordable and effective way
to relieve the burden on manual labor and facilitate smart construction management.
For future studies, we hold the opinion that the following six important directions that
have already been or tend to be hotspots will become the research focus, contributing
to further addressing various issues existing in laborious, complex, or even dangerous
tasks within the CEM domain.
(1) Smart robotics
Smart robotics have been progressing rapidly to drive a wide range of semi- or fully-
autonomous construction applications. Notably, ground robots and aerial robots are
two main types of robotics [1]. For the ground robots in different functions, they
have been developed to automate some manual processes and take over repeatable
tasks, such as brick-laying, masonry, prefabrication, model creation, rebar tying,
demolition, and others. They can also protect workers from work-related injuries
and accidents. To sum up, these robots can provide opportunities to solve issues
of labor shortage, lower operation costs, and ensure overall quality, productivity,
and safety. For the aerial robots, the representative is the unmanned aerial vehicles
(UAV) equipped with image acquisition systems (i.e., camera, laser scanner, go-
pros). Demands of aerial robots are going up to make the procedure of land survey,
site monitoring, and structure health monitoring easier, safer, more efficient and
affordable. More specifically, UAVs fly over the construction site or even fly into
the building structure in place of manual inspecting, which assist in taking high-
resolution images, capture real-time videos, and conduct laser scanning remotely to
maintain the safety of employees and detect structural defects, like cracks, erosion,
blister, spall, and others. Besides, robots can be trained by various machine learning
260 Conclusions and Future Directions
algorithms, and thus they can be equipped with the talent to learn from data and
conditions.
(2) Cloud virtual and augmented reality (Cloud VR/AR)
The evolutionary path of VR/AR is towards the cloud. Due to the development of
the fifth-generation (5G) networks and edge cloud technologies, cloud VR/AR solu-
tions have been generated to accelerate VR/AR applications, leading to better users’
experience. On the one hand, VR/AR as the information visualization technology
can realize easier interactions between the physical and cyber worlds, where VR
simulates the whole situation and AR integrates the information about the real enti-
ties with computer-generated images. Particularly, a distinct advantage of VR/AR
is to offer an engaging and immersive environment. Therefore, VR/AR has been
tentatively adopted to simulate hazardous construction scenarios, aiming to help
managers quickly identify dangers and issues in the working environment and then
formulate reasonable plans and measures ahead of accidents in a visual and interac-
tive way [8]. In addition, construction engineering education and training is another
common application area of VR/AR [17], which well trains workers based on both
visualization and experience in real time instead of courses taught by professionals.
It is expected to strengthen workers’ cognitive learning and safety consciousness
and even raise the overall productivity. On the other hand, the 5G evolution is fast
enough to stream VR and AR data from the cloud, bringing about great benefits in
cloud computing and interactive quality networking. In other words, the 5G tech-
nology can effectively enhance the data processing capability from the local computer
to the cloud and then make real-time perception along with responsive interactive
feedback. Real-time information exchange will also promote collaboration among
different participants.
(3) Artificial Intelligence of Things (AIoT)
The new generation of IoT is called AIoT, incorporating AI techniques into IoT
infrastructure to facilitate more efficient IoT operation and data analysis. IoT can
be defined as a network of interconnected physical devices, like sensors, drones,
3D laser scanners, wearable and mobile devices, and radio frequency identification
devices (RFID). As reviewed, various IoT devices have attached to construction
resources to collect real-time data about the operational status of the project. Smart
IoT-based sensing systems have become a research focus in many existing studies to
feasibly track the progress, monitor the worksite, which eventually support contin-
uous project improvement and accident prevention [6]. In the meantime, a huge
amount of recorded data can be shared over a network. To deeply explore such
rich data sources, various AI methods can be implemented to offer actional insights
for better supervision and decision making. That is to say, AIoT solutions for the
construction industry largely rely on real-time data transformation and instantaneous
data analysis. Empowered by AI, AIoT is, therefore, superior over the traditional IoT
to effectively conduct analysis and control functions for intelligent decision making.
Through synthesizing and analyzing data collected via IoT infrastructure in unprece-
dented volumes and rates, AIoT can achieve a lot of functionality, such as automating
2 Future Directions 261
great potential in shortening the project duration from weeks or months to a matter of
hours or days. Secondly, the 3D printer is more flexible in creating curved walls, roofs,
and other unstandardized shapes, endowing the printer objects with the characteristic
of great geometric freedom. Thus, more sophisticated designs and structures can be
easily built and customized with no restriction of the usual fabrication. Thirdly, the
materials for 3D printing possess special features, such as great tensile strength,
corrosion resistance, high-temperature resistance. This kind of optimized material
with controlled mechanical behavior helps to raise the durability of buildings. Lastly,
most adopted materials are recycled, organic, and eco-friendly, contributing to great
sustainability. It is possible to estimate materials costs and quantities more accurately,
which is helpful to avoid unnecessary waste and decrease material costs. However,
some problems do exist to impede the promotion of 3D printing, such as how to
control structural safety, how to conduct architectural paradigm shift, and how to
develop a rational and digital design workflow, which need to be solved at first.
Currently, 3D printing is moving towards 4D printing under the consideration of time
dimension and intelligent behavior, aiming to make breakthroughs in transforming
configurations for self-assembly, multi-functionality, and self-repair. The research
status of 4D printing in construction is still at its experimental stage, and thus some
new challenges will arise, such as the great demand for digitally savvy engineers,
advanced computational analysis, and structure verification in new designs [3].
(6) Blockchain
A nascent technology called blockchain is a powerful shared global infrastructure,
whose original goal is to simplify and secure transactions among parties [16]. To
make blockchain easier to understand, its basic concept can be explained as a veri-
fied chain with blocks of information, where each block saves process-related data
in a trusted environment. All historical data along with modifications can be there-
fore stored across a network and protected by cryptographic technology. Due to the
distributed ledger built in the blockchain, the stored digital information can be acces-
sible to users of the network simultaneously. Once a block is entered and verified, any
modifications are forbidden in the information. When the application of blockchain
is extended to construction, it provides two value opportunities: one is to gather
and consolidate data in a shared dashboard, while the other is to convert the project
management systems into a more transparent and secure practice with the following
potential applications [18]. For instance, blockchain can work together with BIM to
support life-cycle project management. It allows for collecting large data from various
stages of the project and sharing data securely among stakeholders. Also, the BIM
model can possibly be updated in time when it receives the next block of information.
As expected, such integration of blockchain and BIM will drive the project delivery
to be more automated and streamlined along with improving productivity, trustwor-
thiness, and cost. Besides, another key application of blockchain lies in creating a
smart contract, aiming to enforce and monitor the expected behavior by itself. Only
when certain criteria are satisfied, the process will be executed. This contributes
to high accuracy, compliance, transparency, cost-effectiveness, and collaboration in
activities, like payment, contract administration, and others. It is also important to
2 Future Directions 263
notice that blockchain can provide open and trustworthy information for improving
the supply chain. Therefore, all participants are allowed to make audits on issues
from the downstream supply chain.
References