Probabilistic Seismic Hazard Analysis (PSHA) Training Manual
Probabilistic Seismic Hazard Analysis (PSHA) Training Manual
Probabilistic Seismic
Hazard Analysis (PSHA)
Training Manual
Version No. 1.0.0
Global
Earthquake
Model (GEM)
Foundation
1
GEM Foundation
via Ferrata, 1
20133 Pavia
Italy
Citation
Please cite this document as:
GEM Foundation (2021). Probabilistic Seismic Hazard Analysis (PSHA) Training Manual.
doi: 10.13117/OQ_HAZARD_TRAINING_MANUAL
Disclaimer
The GEM Foundation’s Probabilistic Seismic Hazard Analysis (PSHA) Training Manual is
distributed in the hope that it will be useful, but without any warranty: without even the
implied warranty of merchantability or fitness for a particular purpose. While every pre-
caution has been taken in the preparation of this document, in no event shall the authors
of the Manual and the GEM Foundation be liable to any party for direct, indirect, special,
incidental, or consequential damages, including lost profits, arising out of the use of infor-
mation contained in this document or from the use of programs and source code that may
accompany it, even if the authors and GEM Foundation have been advised of the possibility
of such damage. The Manual provided hereunder is on as “as is” basis, and the authors and
GEM Foundation have no obligations to provide maintenance, support, updates, enhance-
ments, or modifications.
This manual is made possible by the support of the American People through the United
States Agency for International Development (USAID.) The contents of this study are the
sole responsibility of the GEM Foundation and do not necessarily reflect the views of USAID
or the United States Government.
License
This Manual is distributed under the Creative Commons License Attribution- ShareAlike 4.0
International (CC BY-SA 4.0). You can download this Manual and share it with others as
long as you provide proper credit.
The example material that complements this Manual is the result of an effort carried out
jointly by the Information Technology and Scientific teams working at the Global Earthquake
Model Foundation (GEM) Secretariat. It is freely distributed under an Affero GPL license
(http://www.gnu.org/licenses/agpl-3.0.html)
By using this training package you agree to the terms of this license.
"OpenQuake" and "GEM Global Earthquake Model" are registered trademarks of the GEM
Foundation. Unauthorized use of these terms represents an infringement of trademark law.
If you have any questions please contact the GEM Foundation at:
[email protected]
Preface . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
Acronyms . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
I Introduction 9
1.3.1 HMTK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1.3.2 OQ-MBTK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
1.3.3 GMPE-SMTK . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 12
3 Basic data . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 21
3.3 Geodesy . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
3.3.1 GPS . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 27
4.1.4 Declustering . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
4.2.1 Faults . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
5.2.1 vs30 . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 54
6.3 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
V PSHA results 75
9.0.1 Hazard results . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 77
10.0.4 Disaggregation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 80
10.1 Examples . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
11 Bibliography . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
9
Preface
This document is an example-based instructional manual that teaches how to construct a
simple seismic hazard model. The intended audience is the novice hazard modeler at uni-
versity level or higher, and thus instructs on the basic state-of-practice. The methodologies
presented within are meant to be accessible to a wide range of users, and so an emphasis is
given to openly available data sets and tools. The provided examples use public data sets,
and toolkits developed by the GEM Foundation.
The users of this manual will benefit from introductory level prerequisite knowledge of
the following topics:
• probability theory
• earthquakes and tectonics
• geospatial information system (GIS)
• beginner- to intermediate-level Python
The instruction provided herein is complemented by several available resources, which
we aim to build on but not replace. In particular, we refer to earlier articles and manuals
that explain in greater detail many of the concepts or algorithms included in this training
material. These resources are:
The development of this training material was funded by USAID through the Training
and Communication for Earthquake Risk Assessment (TREQ) project. The development
efforts were led by the Hazard Team of the Global Earthquake Model Foundation, a public-
private partnership initiated by the Global Science Forum of the Organisation for Economic
Co-operation and Development (OECD)1 .
1
A short description of the process promoted by OECD is available here:
http://www.oecd.org/science/sci-tech/theglobalearthquakemodelgem.htm
10 Acronyms
Acronyms
rrup
closest distance between the site and rupture.
AFE
annual frequency of exceedence.
GAF-DB
Global Active Faults Database.
GEM
Global Earthquake Model Foundation.
GHEA
Global Historical Earthquake Archive.
GMC
Ground Motion Characterization.
GMM
Ground Motion Model.
GMPE
Ground Motion Prediction Equation.
GMPE-SMTK
GMPE Strong Motion Modeller’s Toolkit.
HMTK
Hazard Modeller’s Toolkit.
IASPEI
International Association of Seismology and Physics of the Earth’s Interior.
ISC
International Seismological Center.
ISF
IASPEI Seismic Format.
MFD
Magnitude-Frequency Distribution.
oq-engine
OpenQuake-engine.
OQ-MBTK
OpenQuake Model Building Toolkit.
Acronyms 11
PGA
Peak Ground Acceleration.
PGV
Peak Ground Velocity.
PSHA
Classical PSHA.
SA
Spectral Acceleration.
SSC
Seismic Source Characterization.
SSM
Seismic Source Model.
TRT
tectonic region type.
UHS
uniform hazard spectrum.
Part I
Introduction
15
Seismic hazard analysis is a procedure used to determine the level of danger that may occur
due to earthquakes within an area of interest over a given time frame. The two major types
are determinstic seismic hazard analysis, which is interested in the worst case event that can
be expected at a site, and probabilistic seismic hazard analysis, which calculates the rates
at which each shaking level will occur or the probability that each level will be exceeded
during a fixed time frame.
This instructional manual teaches classical probabilistic seismic hazard analysis (PSHA),
and how to construct the underlying input models. Classical PSHA (Cornell, 1968) uses the
hazard integral, which considers all possible seismic sources expected to affect a site and
their rates (or, alternatively, probabilities) of occurrence, and the expected ground motions
that will occur for earthquakes produced by these sources. The hazard integral evaluates
the probability that a specified ground motion intensity level will be exceeded over an in-
vestigation period or the frequency at which the intensity level is exceeded.
A number of software options are available for developing seismic hazard models and
executing Classical PSHA calculations. This manual uses the OpenQuake Engine (Pagani
et al., 2014) and its accompanying toolkits.
The OpenQuake Engine is a seismic hazard and risk calculation software that’s develope-
ment and maintainence is led by the Global Earthquake Model Foundation (GEM) (Pagani
et al., 2014; Silva et al., 2014). The Engine is written in Python and supported on Linux,
MacOS, Windows, and cloud platforms. The software emerged from a 2010 project aiming
to develop an open and transparent hazard and risk calculation engine with community-
driven development, and includes contributions from many of GEM’s collaborators. The
Engine’s development emphasises reproducibility, and thus is founded on several testing
approaches including unit tests, comparisons against other PSHA codes and benchmarks
(e.g. PEER, Hale et al., 2018), and comparisons of outputs from other PSHA codes using
end-to-end calculations.
The Engine is hosted on GitHub (https://github.com/gem/oq-engine).
Additional reading: Pagani et al. (2014) and Silva et al. (2014), the OpenQuake Engine
User Instruction Manual, and the OpenQuake Underlying Hazard Science
16 Chapter 1. Introduction to seismic hazard analysis
The examples included in this training manual use three toolkits: the Hazard Modeller’s
Toolkit (HMTK), the OpenQuake Model Building Toolkit (OQ-MBTK), and the GMPE Strong
Motion Modeller’s Toolkit (GMPE-SMTK). Like the OpenQuake Engine, all three toolkits are
Python-based. The HMTK is built into the Engine, while the other two toolkits must be
installed separately.
1.3.1 HMTK
The HMTK is a Python library that provides tools that assist in developing seismic hazard
models, and in particular seismic source models, in the OpenQuake Engine format. The
toolkit does not eliminate the requirement of modeller judgement in the hazard modelling
process, but provides modellers with implmentations of many state-of-practice algorithms
used to develop seismic source models.
The HMTK is the most established of the toolkits used in this training manual. As a
component of the OpenQuake Engine, it is thoroughly tested. It is accompanied by a users
manual available here.
1.3.2 OQ-MBTK
The OQ-MBTK is a container for a number of separate packages that are useful for building
seismic hazard models and analysing their performance. The two OQ-MBTK pacakges most
relevant for this training manual are the catalogue toolkit (CAT) that assists modellers in
preparing the earthquake catalogue to be used in the model-building process and the model-
building toolkit (MBT) that streamlines the use of the HMTK for preparing seismic source
models inputs in the OpenQuake Engine format.
The OQ-MBTK can be downloaded at https://github.com/GEMScienceTools/oq-mbtk.
1.3.3 GMPE-SMTK
The GMPE-SMTK is a suite of Python tools that helps hazard modellers to select GMPEs
(ground motion prediction equations, also called ground motion models) to be used in a
seismic hazard model. The toolkit is based upon the GMPEs included in the OpenQuake
Engine.
The GMPE-SMTK can be downloaded at https://github.com/GEMScienceTools/gmpe-
smtk.
17
(1) Seismic source model (2) Ground motion model (3) Logic trees
Figure 2.1 – The components of a PSHA input model. (1) An example of a seismic source
model: a collection of seismic sources that together generate all the seismicity affecting an area
of interest. (2) An example of a ground-motion model, showing that the ground-motion value
attenuates with distance. (3) Example of a logic tree structure used to incorporate uncertainty
in components (1) and (2).
Epistemic uncertainty in the SSC can be described by using a number of SSMs, each one
18 Chapter 2. Components of a seismic hazard model
with an associated weight, or by some initial SSMs and a set of modifications that describe
the various epistemic unceratainties of source parameters.
Annual rate
Magnitude
Annual rate
Magnitude
Fault source
Smoothed seismicity
Annual occurrence rate, M>m (point sources)
Figure 2.2 – A simplified example of a seismic source model that shows the position, source
typology, and magnitude frequency distributions of sources covering a region of interest. The
left panel shows two source typologies: faults, with surface traces denoted by red lines, and
point sources colored by the cumulative annual occurrence rate of earthquakes with magnitude
M > m, where m is a minimum magnitude threshold. The top of the right panel shows two
examples of magnitude frequency distributions, which indicate the occurrence rates of sources
denoted by the dashed lines. (NB: this is a purely hypothetical source model.)
Hazard modellers use the following information and considerations to complete seismic
source characterization.
• Recorded seismicity: do instrumental and historical earthquake catalogues contain
events in this region? How frequent are recorded earthquakes, and what are their
sizes?
• Distribution of observed seismicity: are patterns observed in the occurring seismicity?
Do observed earthquakes tend to concentrate in delineable areas/volumes (in terms
2.2 Ground motion characterization 19
of both lateral and depth distribution), or are hypocenters more sparsely distributed?
How are the earthquakes distributed throughout time? What is the ratio of large to
small earthquakes?
• Tectonic setting/geology: what is known about the tectonic activity in the area covered
by the model? For example, is it close to a plate boundary, or in the stable continental
interior? If it’s close to a plate boundary, what kind of earthquakes does that boundary
produce? What are the relative plate motions?
• Active faults: other than plate boundary faults, does the region include other known
faults, and are they thought to be seismically active? Are they covered by paleoseismic
or geodetic data that could reveal their activity rates?
Figure 2.3 – Components of the ground motion characterization workflow. Top: Locations
of earthquakes near Mexico for which strong motion data are available, colored by tectonic
region type and scaled by magnitude. Bottom: Plots showing how the computed peak ground
acceleration (PGA) changes with distance for three different magnitudes using seven existing
ground motion models for subduction interfaces.
2.3 Logic trees 21
Figure 2.4 – An example of a logic tree, with labels corresponding to the terminology of the
OpenQuake Engine (Pagani et al., 2014). Each branching level introduces an additional epis-
temic uncertainty to the input model.
declustering). In the case of the source characterization, epistemic uncertainties include, for
example, the seismic source delineation and geometry, maximum magnitude admitted, the
parameters of the magnitude-frequency distribution, and any other parameter that cannot
be reduced to a single hypothesis. In the case of the ground motion characterization, the
epistemic uncertainties include the regional inter- and intra-event variability of ground mo-
tions for a given magnitude-distance pair, which prohibits us from fitting recorded seismic
recordings for a tectonic region using a just single ground motion model, and data gaps (or
general lack of data), such as for high magnitudes, or at long and short site-source distances.
Part II
Seismic source characterization is the definition of one or more seismic source models
that represent all earthquake sources that could produce seismicity within the model cover-
age area during an investigation period and the associated epistemic uncertainties. In this
chapter, we introduce:
1. types of data that are used in seismic source characterization, including datasets thats
that are open-source and commonly used for PSHA
2. the source typologies that are most frequently used to represent seismic sources
3. a seismic source characterisation workflow, from catalogue processing to defining seis-
mic sources
3 Basic data
This chapter introduces three types of data that are commonly used to build seismic hazard
models:
• Earthquake catalogues
• Fault data
• Geodetic data or information
Figure 3.1 – Example of an earthquake catalogue: the ISC-GEM extended catalogue coverage
of the Aegean Sea region Weatherill et al., 2016. Hypocenters are colored by depth and size is
scaled by magnitude.
28 Chapter 3. Basic data
All earthquakes occur on faults, discontinuities within the Earth’s crust where rock has
been crushed or fractured, separating the surrounding earth into separate blocks. Faults
range in area from very small (mm2 -scale) to 100s or 1000s of km2 . Some faults are known
to scientists, while others are too small to detect; are buried deep underground without
surface expressions; or have not ruptured historically. Faults that are thought to be active
should be included in seismic hazard models. While this is mostly straightforward in the
case of plate boundaries, crustal faults can be more ambiguous, often due to a shortage of
data. For this reason, faults are not always included in seismic hazard assessments, even
if we know that they are active. This section introduces the fault characteristics that are
critical to constructing fault sources, and the ways of collecting these data.
Geological surveys from around the world typically depict the state of known faulting on
geologic maps for their regions of oversight, sometimes collecting the faults into a separate
database that includes the information available for each. In these databases, each fault
includes at minimum a fault trace, which is the curve representing the intersection of the
fault surface with the topographic surface. In order to convert a fault into a seismic source
for PSHA, we must also assign it a geometry and activity rate. These parameters can be
assigned by studying historical earthquakes, and by conducting remote sensing and field
investigations.
Historical earthquakes that are large enough to be linked to a specific fault (or faults), and in
particular those that produce surface ruptures (e.g., offset at the ground surface), indicate
currently active faults. Sometimes these earthquakes occur on faults that are known to be
active, but other times are the first indicator of ongoing fault activity, or the first evidence
that the fault exists. Recent earthquakes that appear in instrumental catalogues are typically
studied with a variety of methods – field mapping, seismology, remote sensing, GPS – that
provide a 3D image (the geometry) of the causative faults, as well as a robust value of
moment magnitude; many of these topics are discussed in the next sections. Studies of
earthquakes that occurred earlier in history, before the instrumental period but still observed
by humans (those appearing in historical catalogues) and that can be linked to a known
fault, can still yield magnitude estimates, both using the surface rupture qualities and by
inverting the observed damage field.
Independently, the magnitudes of past earthquakes can only be used to constrain the
maximum magnitude of a source, serving as a known minimum value of the largest earth-
quake that a fault can accommodate. The timing of the event can be used to make very
coarse estimates of the activity rate, e.g., using the time that has elapsed since the earth-
quake, and the time that passed between the arrival of the first humans and the earthquake.
Typically, these time scales can only indicate the minimum recurrence interval, the period
between consecutive events of the given magnitude; the inverse of the recurrence interval
3.2 Fault data 29
is the activity rate. These parameters can be computed more robustly by using data col-
lected from targeted fault investigations using approaches of geology, paleoseismology, and
geodesy.
Maximum magnitude
The maximum magnitude potential of a fault can be estimated using scaling relationships
that convert from fault length, surface area, or slip to magnitude (discussed more in Chapter
4.4.3).
Fault length is the total length of its surface trace. Importantly, surface rupturing earth-
quakes do not always rupture the full length of the fault. Fault surface traces can be mapped
in the field, or from aerial and satellite data (topography as well as various image types).
The applicability of aerial and satellite data to a fault, and the ease of mapping fault surface
traces in general, depends on the qualities of the fault site. The easiest faults to map are
those that occur in arid, unvegetated landscapes, where fault morphology is preserved and
distinguishable even in optical imagery. On the other hand, forested landscapes cover the
surface expressions of faults, and in these cases, methods that “see through” vegetation are
required to identify the full length of the fault.
Fault surface area is the contact area between the two blocks that comprise a fault. This
parameter can be coarsely estimated using basic geometric relations: fault length, dip, and
an assumption of the seismogenic zone thickness, the part of the brittle crust that can fail
elastically in an earthquake. Dip can be measured from paleoseismic trenches or geologic
outcrops that are cut by the fault, especially those that reveal more than a few meters of
stratigraphy; or geophysical data such as potential fields methods (gravimetry and magne-
tometry) and active seismic.
Fault slip is the distance that the fault blocks move relative to each other during an
earthquake. In magnitude scaling relationships, slip from a single earthquake event is used
with a length-to-slip ratio assumption, and so is most useful as an indicator of magnitude
potential for assessing pre-historic fault behavior. One way to measure slip is to identify
landform features (e.g., alluvial fans, stream channels, ridges) that span both blocks of
a fault, and quantify how they have deformed by earthquakes. For example, a channel
crossing a strike-slip fault may be offset during an earthquake such that the upstream part
of the channel has moved laterally compared to the downstream channel. Importantly, the
slip preserved in a landform does not necessarily indicate the slip produced by a single event.
Slip measurements are most straightforward when the rupture reaches the surface. Slip
on ruptures that are preserved in geologic outcrops can be measured from the separation of
stratigraphic layers that have been faulted.
One challenge in collecting fault slip data is identifying the number of earthquakes that
have occurred since the measured feature was formed. This is addressed with earthquake
timing.
30 Chapter 3. Basic data
Earthquake timing
Knowledge of earthquake timing is important for understanding the rate at which faults
produce their maximum magnitude events. The most effective way to determine the earth-
quake ages is to acquire ages of the stratigraphic layers that were deposited before and after
faulting events. Usually, this is done by digging a paleoseismic trench, an elongate hole ex-
cavated across a fault, and studying the rock and sediment layers that the trench exposes
to determine which layers were present when each preserved earthquake occurred. Then,
samples of material that can be dated are collected from the layers most closely bracketing
earthquakes, (e.g., organic material or minerals using radiometric, cosmogenic, or lumines-
cence dating). These paleoseismic approaches can also be applied to study faulting that does
not occur in the shallow crust (e.g., interface) by instead dating formations that preserve
the effects of earthquakes, such as coastal uplift (e.g. terraces or coral reefs), and turbidites.
When landforms that preserve and bracket earthquake timing are not available, more
approximative methods are used. For example, the age of a scarp – a step in topography
created by an earthquake – preserved in sediment can be approximated by the comparing
the shape of the scarp to local rate of erosion, considering both sediment composition and
climate (e.g. rainfall).
One goal of collecting earthquake timing and magnitude (or slip) data is to calculate
slip rates – the slip per unit time – for active faults. Slip rates are indicators of how fast the
two blocks of a fault are moving relative to each other, and indicate the potential seismic
moment release of that fault, that is, the amount of energy stored in a seismogenic fault that
may be released by earthquake. In lieu of – or to complement – paleoseismic and geologic
slip rate data, GPS stations placed on opposite sides of faults record the velocity of relative
block movements; this is discussed more with Geodesy.
In many cases, the data on characteristics of a given fault are insufficient, however,
some parameters can be reasonably estimated. This is discussed more in Chapter 4.3.2 on
building fault sources.
3.3 Geodesy
3.3.1 GPS
The type of geodetic data with the highest immediate utility to seismic hazard analysis is
GPS velocities. From these data, we can infer how fast the Earth’s crust is deforming, and
therefore indicate the distribution of earthquake potential across the globe, or a region.
modelling as one of its primary anticipated utilities. Thus, the catalogue is available in a
pre-processed format that expedites its immediate utility for seismic source modelling. In
particular, the catalogue reports a single instance of each event with homogenous mag-
nitudes. On the contrary, the catalogue does not include small magnitude events, so has
limitations where seismicity rates are low.
The current version (6.0, as of January 2020) of the ISC-GEM catalogue dates back to
1904 and includes events Mw >5.5, with additional Mw >5.0 earthquakes on the continents.
Read more about the ISC-GEM catalogue and download it here:
http://www.isc.ac.uk/iscgem/.
Since 2017, the GEM Global Active Faults Database (GAF-DB) has aimed to merge these
databases into a single open-source database which includes the basic level of information
for each fault that is needed to produce seismic sources to be used in hazard analyses. At
minimum, this includes:
• fault surface trace coordinates,
• dip direction and sense of motion, and
• slip rate
As of January 2020, GAF-DB has near-global coverage, and can be downloaded in GeoJ-
SON, GeoPackage, ESRI Shapefile, and KML formats; the OpenQuake tools mostly make use
of the the GeoJSON format. The database is kept up-to-date using Git versioning, and so
users can easily update as the database evolves to include more faults, and can more easily
contribute to the database.
Read more about GAF-DB and download the data here:
https://github.com/GEMScienceTools/gem-global-active-faults.
The effort was led by the University of Nevada Reno, University of California Los Angeles,
Chinese Earthquake Administration, and UNAVCO. Read more about the model here:
https://gsrm2.unavco.org/intro/intro.html.
according to the precision of the catalogues in question. For example, older entries may only
provide time stamps to the day or hour, whereas newer entries are reported to sub-second.
The spatial accuracy may depend on the number of and proximity to recording stations of
the earthquakes, and the accuracy of the velocity model used by the agency to record to
locate the event.
Once the duplicate events are identified, criteria must be applied to choose which event
to include in the final catalogue. For example, the criteria might include considerations on
the reliability of one network compared to the others, or the primary area covered by a
monitoring network. In some cases, the origin may be taken from one catalogue, while the
magnitude is taken from another.
Subduction sources
Subduction zones produce the largest earthquakes globally, and account for the vast ma-
jority of seismic moment release. Therefore, it is very important to correctly classify the
seismicity that is produced by these sources, and significant scientific effort has focused on
understanding subduction zone structures and productivity.
At a basic level, subduction sources are divisible into two tectonic region types: interface
and intraslab. Interface earthquakes occur on the contact surface between the two tectonic
plates that are moving toward each other; the earthquakes accommodate the plate conver-
gence. Intraslab earthquakes occur within the plate that is being subducted, e.g., within
the slab. The majority of intraslab earthquakes occur at greater depths than the subduction
interface, which permits the simplest level of classifying subduction earthquakes: using a
depth cutoff that considers shallower earthquakes to be interface, and deeper earthquakes
to be intraslab.
There are many resources available to modellers that help to choose a cutoff depth be-
tween interface and intraslab events. The works of Heuret et al. (2011), Christopherson
4.1 Catalogue processing 37
et al. (2015), and Hayes et al. (2012, 2018) include compilations of subduction zone pa-
rameters, including the depth ranges of locked interfaces for individual subduction zones.
Among other data, these models assign depths by inspecting the focal mechanisms, assum-
ing that most reverse faulting occurs on the interface, while most normal faulting occurs
within the slab.
Subduction earthquake classification by depth is easy to implement, but causes many
crustal earthquakes and those intraslab earthquakes that occur directly beneath the interface
close to the trench to be classified as subduction interface. A more advanced approach is
to define a volume within which most interface earthquakes occur, establishing a “buffer
zone” around a surface that represents the interface. In the past, this was often done using
a planar surface with a dip defined by the observed focal mechanisms, but recent work has
produced high resolution models of subduction interfaces that consider many types of data.
This approach is used in the methodologies of Garcia et al. (2012), Zhao et al. (2015), and
Pagani et al. (2020).
In some cases, a modeler may also want to also divide a subduction zone along strike,
or in other words, to impose segmentation. In this case, each interface or intraslab segment
would be treated as a separate source, and so the seismicity should also be separated into
sub-classes that correspond to the segments.
4.1.4 Declustering
Declustering is the procedure by which the mainshock events in an earthquake catalogue
are separated from aftershock events. PSHA usually uses declustered earthquake catalogues
in order to evaluate only the seismic hazard that is due to independent earthquakes, as
opposed to those that are “triggered" by other earthquakes (aftershocks). The resulting
mainshock catalogue is then treated as Poissonian or "memory-less", and thus events of a
given magnitude occur at rates which are homogenous in time.
The most common algorithmic procedure is to compare earthquakes within magnitude-
dependent space-time “windows" (e.g. Gardner and Knopoff, 1974), treating the largest
earthquake within a window as the mainshock for the encompassed events; this method-
ology is used in the examples that accompany this training material. Before proceeding to
the examples, please read HMTK User Guide Section 2.2.1, which explains three windowing
options and an additional declustering algorithm included in the HMTK.
Read more about declustering methodologies in Stiphout et al. (2010).
several reasons, for example, some earthquake do not produce strong enough shaking to
be detected by the present seismic network (either because the earthquakes have such low
magnitudes, or because the network itself is sparse). Small earthquakes can also get “lost"
in the signals of larger earthquakes, or the agencies compiling the catalogues may not be
interested in the smallest events.
Knowledge of the catalogue completeness is important in PSHA (and, more generally,
in applications of statistical seismology) because the occurrence rates assigned to sources
are often constrained by earthquake observations. The occurrence rates of sources that are
modelled using a Gutenberg-Richter magnitude-frequency distribution (MFD; introduced in
Section 4.4) depend on the relative proportion of large to small earthquakes; if too low of
a magnitude is used as MC , the resolved proportion is distorted.
Global catalogues, such as the ISC and ISC-GEM catalogues, typically have a magnitude
completeness that varies in both time and space, since global network coverage is not spa-
tially homogenous and station locations change with time. Therefore, in order for these
catalogues to be used in PSHA, the completeness must be analysed for the proximity of the
model coverage. Use of the most conservative magnitude threshold would eliminate a large
number of valuable observations. Instead, a time-dependent MC can be used, allowing for
more robust occurrence rates of higher magnitudes that are based on longer observations
periods.
There are many analytical ways to find MC for a catalogue, including both determin-
istic and probabilistic methods, and those that are catalogue- or network-based. Read
about these methodologies, and more about MC in this review by CORSSA (Mignan et al.,
2010). The included methods are developed to find a single value of MC . In cases where a
time-dependent completeness magnitude is desired, the analytical methods in Mignan et al.
(2010) can be used with windowing.
The HMTK provides two means of defining time-dependent MC . The first is the Stepp
(1971) algorithm, which is described in the HMTK Users Manual. The second is to manually
assign a completeness table by observing time-magnitude density plots, such as the example
in Figure 4.1.
The OpenQuake Engine and Hazard Science Manuals describe how each of the various
source typologies are implemented in and treated by the OpenQuake Engine.
4.2.1 Faults
Geometrically, fault sources (sometimes called line sources) are modelled as finite planes
or surfaces. Typically, these source types are comprised of discrete faults with a given sense
of motion, and during PSHA calculation, ruptures are generated that correspond to indi-
vidual earthquakes that could occur on the fault surface. Each fault source is prescribed its
own magnitude-frequency distribution, which is represented in the OpenQuake Engine as
magnitude and occurrence rate pairs.
Within PSHA, there are sub-typologies of fault sources which differ based on the com-
plexity of the source geometry and restrictions to the rupture geometries or magnitudes that
the fault sources can produce. Faults that have been studied more thoroughly can some-
times be modelled with more detailed geometry. For example, it is common practice to use
faults with 3D geometries, termed “complex faults" in the OpenQuake Engine terminology,
to model subduction interfaces. On the other hand, faults for which only the surface trace
40 Chapter 4. Seismic source model
Figure 4.2 – The two main categories of source typologies used in PSHA. Top: Fault sources are
3D surfaces that delineate where ruptures can occur, and correspond to sources for which the
geometry is well-defined. The dark gray rectangle indicates a single rupture, while the white
surface indicates the fault. Bottom: Distributed seismicity is used where earthquakes are known
to occur, but cannot be easily attributed to known structures. Instead, hypothetical ruptures are
modelled at many positions and with many geometries and orientations. The figure shows an
area source, for which a grid of points inside a perimeter is used to define rupture positions,
and other source attributes (described in the text) are used to create ruptures around the grid
points.
is known are usually represented by a simpler geometry (“simple faults" in OpenQuake ter-
minology).
All earthquakes occur on faults, but not all faults are known to the scientific community.
Fault source typologies are reserved for faults that we do know about – cases where we feel
confident about the position and orientation of ruptures. However, knowing the position
and shape of a fault is not enough to model it using a fault source, we must also have enough
information to constrain the occurrence rate of earthquakes on the fault. The methodologies
for constraining occurrence rates are covered in Section 4.4.
of an earthquake. During hazard calculation in the OpenQuake Engine, area sources are
discretized into point sources, each with an equal proportion of the seismic moment rate and
identical source attributes. Thus, area sources are appropriate for representing seismicity
within a boundary that is thought to have consistent (or similar) characteristics, including
occurrence rate. For example, an area source could be used to model seismicity in the
outer rise of a subduction zone, or a stable crustal area with no known faults but occasional
earthquakes that do not conform to a spatial pattern.
In some cases, area sources are difficult to use effectively. For example, where earth-
quake rates are spatially variable, use of many area sources would lead to low numbers
of earthquakes inside each, which will cause difficulty in constraining occurrences. An al-
ternative approach used for modelling distributed seismicity is called smoothed seismicity,
which removes the subjective component associated with the need to define the geometry
of an area source polygon. Smoothed seismicity is similar to area sources in that, at first,
regions with internally consistent attributes (here called “source zones” ) are defined and
treated as a single source in order to determine occurrence parameters; the source zones are
then discretized into a grid of point sources. However, instead of imposing uniform moment
rates on the point sources as is done for area sources, smoothed seismicity sources allow
for variable occurrence rates across the grid. A fraction of the source zone rate is assigned
to each point source by applying a smoothing kernel to the catalogue occurrences. Often, a
Gaussian smoothing filter is used.
In both examples of distributed seismicity, seismicity is ultimately modeled as point
sources. Despite their name, many calculation engines (including the OpenQuake Engine)
do not treat the sources as points during hazard calculations; instead, they use the point
source attributes (e.g., hypocenter depth, nodal plane geometry) as a basis for generating
ruptures, and the ruptures are used during calculation. This is important because of the
impact of rupture finiteness on distance metrics used by ground motion models (GMMs,
covered in more detail in Chapter 6.1). For example, many GMMs measure distance to
the rupture, and not distance to the hypocenter. The dimensions of ruptures produced at
each point source are based on magnitude-scaling relationships (discussed more somewhere
else), and so point sources can still be used for distributed seismicity with larger magnitudes.
of the ways that hazard modelers assign these attributes. In addition to the text in this
section, please read the OpenQuake Engine User Instruction Manual, which explains the
specific attributes required for OpenQuake source typologies.
Point sources
Although termed “point source” , many PSHA softwares actually use a rupture that is built
based on the point source and its attributes. Intuitively point sources must correspond to a
hypocenter: a latitude, longitude, and depth. The point is converted into a rupture based on
a given nodal plane (the rupture’s strike, dip, and rake); a magnitude-scaling relationship,
which converts between the magnitude and area of a rupture; and rupture aspect ratio,
which controls the ratio between the down-dip width and along-strike length of the rupture.
Typically, seismogenic depth limits are also assigned to ensure that ruptures remain within
the brittle crust.
Area sources
Because an area source is modelled as a collection of uniform point sources, most of the
geometric attributes of the point source persist for area sources. However, instead of defin-
ing single latitude-longitude pairs that position the hypocenter, a polygon of coordinates
defines the perimeter of the area source. Recall that smoothed seismicity is also initialized
with a source zone perimeter.
In order to constrain the source geometry attributes for distributed seismicity, we return
to the considerations initially posed in Chapter 2.1. Is there recorded seismicity? How is it
distributed? What is the tectonic setting? And are there known active faults?
It is logical to first constrain the larger scale geometric attributes: the perimeters of area
sources or source zones. As a pre-modeling step, the basic datasets described in Chapter 3
should be interpreted together for the region of intended coverage. Importantly, the sources
that contribute to hazard extend beyond the region of interest (ROI), and on this premise
we define the first step in seismic source characterization:
1. Define the lateral extent of the seismic source model/s: the region of interest, plus
an additional buffer zone around the perimeter.
The width of the buffer zone depends on the region, and should be based on the follow-
ing consideration:
• How far away are the seismic sources that contribute to seismic hazard in the
ROI? Larger magnitude earthquakes are felt over longer distances. Earthquakes in
older stable crust are felt over longer distances than those in younger active crust.
4.3 Defining source geometry and postion 43
These latter points also serve as guidelines for selecting an integration distance during
hazard calculations.
Once the perimeter of the source model is defined, it can be subdivided into sources.
2. Define the perimeters of area sources/source zones that will model distributed seis-
micity.
Note that while we call this a perimeter, the sources will actually occupy a volume. A source
model can consist of many depth layers. Recall that area sources and source zones used
for smoothed gridded seismicity should encompass space with internally consistent seismic
properties. These properties include:
• Source of deformation
• Faulting style
• Deformation rate
• (for area sources only) spatial distribution of seismicity
There are a multitude of tools that can be used to delineate source perimeters, and the
best tool will depend upon the modeler. However, we suggest using QGIS (http://www.qgis.org),
since this is an open-source geographic information system (GIS) program that can be used
to display numerous datasets and models together and interactively, and to initiate shape-
files (or other vector dataset) that store the source parameters.
Likewise, there are no firm rules that a modeler can use to define the source perime-
ters, but rather the guidelines above.Seismic source characterization is typically an iterative
process, and the source perimeters defined in the first undertaking of Step 2 are liable to
change.
Once the source perimeters are defined, the seismicity and tectonic processes occur-
ring inside each can be analyzed in order to assign the other parameters. The OpenQuake
Hazard Modeler’s Toolkit (HMTK) includes tools to derive these parameters from source
perimeters (shapefile format) and catalogues. Before continuing, please learn about the
HMTK catalogue format in Section 2.1.1 of the HMTK User Guide!
Many of the next steps to source characterization are completed using subcatalogues
that are “clipped” to the source zone perimeter, as well as a depth range if appropriate, as
is done in tectonic regionalization. The examples included in this manual use the HMTK to
filter events into sub-catalogues, but again, many tools and strategies are available.
Additional reading: QGIS tutorial for making vector datasets, HMTK User Guide
The hypocentral depths assigned to a source should correspond to the depths of events in
the subcatalogue for the area source or source zone. In most seismic hazard softwares (in-
cluding the OQ Engine) it is possible to assign a hypocentral distribution (e.g., pair of depths
and weights summing to 1.0), rather than a single value, thus capturing the aleatory uncer-
tainty in hypocentral depth. Aleatory uncertainty is used to model the range of possibilities
that are expected given an infinite observation period, in contrast to epistemic uncertainty,
which reflects our incomplete knowledge.
Visual approaches can be helpful for identifying depth concentrations within a subcata-
logue, including histograms that count events within a depth range, and time-depth density
plots. In addition to revealing where events are occurring, these plots can also reveal char-
acteristics of the earthquake catalogue. For example, hypocentral depths of moderate and
small earthquakes may be assigned a fixed depth value (e.g., 10 km, 33 km), and so these
precise values, in addition to their encompassing depth ranges, may have excessively high
counts. We caution modelers to make themselves aware of these catalogue characteristics.
Additionally, we note that earthquake depths are usually uncertain on the scale of – at
least – kilometers. Thus, it is important not to over-interpret the depths of a subcatalogue by
assigning a hypocentral distribution that is too finely discretized. Instead, for example, the
distribution can use a handful of depths that each represent a range (e.g., a depth of 7.5 km
representing hypocenters that span 5-10 kms depth. There are two major reasons for this.
First, using a large number of precise hypocenters (e.g., 1.0, 2.0, 3.0, . . . n km) does not
necessarily improve the model, even if all depths are known with small errors. One reason
for this is that ground motion models (GMMs) compute intensity measures for a mean value
with a standard deviation that typically allows for larger variation than would be caused by
a small shift in hypocenter depth; actually, some GMMs do not consider hypocenetral depth.
A second reason is that each additional hypocentral value increases the computational load
of the model. These points should be considered while assigning the hypocentral depth.
In some cases, little or no seismicity is available to constrain hypocentral depth. In these
cases, a common solution is to apply “rules of thumb” , or generalizations based on the tec-
tonic region type. One simple approach is to assume that all seismicity occurs at the halfway
point of the seismogenic thickness.
Focal mechanisms — which are comprised of two nodal planes, each with a strike, dip,
and rake –are used in two ways during hazard calculation. First, for each source included
in a seismic source model, earthquake ruptures are defined to use as the basis for GMMs to
calculate the intensity measure levels produced by earthquakes at each site. The rupture is
oriented according to the source nodal plane strike and dip. Some GMMs also consider the
focal mechanism, either applying a scaling term based on the simplified classification (e.g.,
normal, reverse, or strike-slip), or by the rupture’s rake.
4.3 Defining source geometry and postion 45
Like hypocentral depths, nodal planes can usually be assigned singularly, or as a weighted
distribution. Also like hypocentral depths, a distribution of nodal planes should include up
to a few possibilities that represent the faulting styles and geometries in the source zone,
aiming to find a balance between what is observed in earthquakes and computational ex-
pense.
A focal mechanism distribution can be assigned based on the focal mechanisms of earth-
quakes that have occurred within the source perimeter, and assigned based on a subcata-
logue of the GCMT database, or any other catalogue of focal mechanisms. Often, as in the
GCMT project, focal mechanisms correspond to an earthquake centroid, not to the hypocen-
ter; this is important to keep in mind while preparing a subcatalogue of focal mechanisms.
Using only focal mechanism catalogues, assigning nodal plane distributions to source
zones is not as straightforward as is assigning hypocentral depth distributions. Raw nodal
planes cannot be plotted as histograms, since they include three parameters (strike, dip, and
rake); however, they can be converted to simplified classifications (e.g. strike-slip, normal,
reverse, and the oblique senses of motion in between), and binned by these classifications.
Map-view plots of focal mechanism “beachballs” and strike-vs-dip plots – along with the
judgement of the modeler – can be used to look more closely at the nodal plane orientation.
Focal mechanisms include two nodal planes: one on which the earthquake occurred,
and an auxiliary plane which would result in the same patterns of energy radiation and
first arrival polarities. From the focal mechanism alone, it is impossible to know which of
the nodal planes hosted the earthquake. This is important for modelers to consider when
using strike-vs-dip plots to assign nodal planes to sources. If both nodal planes are plotted,
clusters of data will form at the true nodal planes as well as the auxiliary planes. Sometimes,
both nodal planes are likely candidates for seismicity within a source zone, such as for
normal faulting mechanisms in basin and range-style extensional tectonic settings, whereas
for vertically dipping, purely strike-slip mechanisms, one nodal plane may be more likely.
Alongside focal mechanisms, some other data types are useful for assigning nodal planes
to distributed sources. GPS and strain rate data will reveal the expected faulting types, and
active faults encompassed by the source zone can reveal the orientation of the seismogenic
structures. In lieu of other data, Andersonian mechanics can be used to assign dip values to
nodal planes.
The final nodal plane distribution for a source or source zone should include few enough
nodal plane-weight pairings that their coexistence within a single source can be justified. If
there are too many, then the modeler should consider whether the source should be further
divided. Typically, this constraint is considered before the source perimeters are delineated.
However, observing seismicity characteristics on a smaller scale often reveals patterns that
are harder to see in large-scale datasets, and thus source delineation often includes multiple
iterations.
Most hazard calculation engines use magnitude-scaling relationships and rupture aspect
ratios to build a set of ruptures from the hypocenters and nodal planes, which should be
based on the judgement of the modeler.
The most important consideration regarding magnitude-scaling relationships(MSRs) is
that they must be compatible with the tectonic region type (TRT); while the earliest MSRs
did not have enough data to divide the available data into TRTs or by faulting style, the
more recent MSRs tend to include multiple relationships that each correspond to a category
of earthquakes. MSR selection for PSHA is discussed in more detail by Stirling et al. (2013).
Although the depths of hypocenters are assigned to each source, the rupture scaling
attributes may create ruptures that extend unrealistically deep, or that extend beyond the
Earth’s free surface. To prevent this, many calculation engines also require that the lower
and upper seismogenic depths are defined, and adjust the rupture dimensions to comply
with the intended seismogenic layer (each engine uses its own approach to do this).
The way that fault source surface geometry is defined can be summarised as by the fol-
lowing approaches:
• Define the 3D coordinates of the fault surface, covering its full extent
• Define the surface trace of the fault, as well as a fault dip and seismogenic depth
limits; together, these attributes can be used to compute the 3D coordinates of the
fault surface
Fault databases such as the GEM Global Active Faults Database expedite the latter ap-
proach, providing both surface traces and fault dips. The former approach is most commonly
used if a 3D model of a fault already exists or when a wealth of geophysical data spans the
fault (e.g. subduction interfaces, for which Hayes et al. (2018) have released Slab 2.0, a
model with global coverage).
When constructing fault sources of any surface complexity, the modeller faces decisions
about the extents of faults that should be modelled as a single source. In other words,
modellers must decide whether the limits of a fault strand imposed in a fault database
correspond to barriers along a fault or fault system that earthquakes will not cross; the
fault segmentation used to construct a fault database may not correspond to the supposed
4.3 Defining source geometry and postion 47
are variable, and the modeller must consider the specific software when defining source
geometry and rupture-scaling attributes.
Gutenberg-Richter MFDs
Gutenberg and Richter (1944) were the first to resolve a relationship between magnitude
and occurrence rate for earthquakes in a region of interest:
l o g N = a − bm (4.1)
where N is the annual rate of earthquakes with magnitude M ≥ m, a is the rate of all
earthquakes, and b is the relative distribution of earthquakes among magnitudes. Equa-
tion 4.1, called the Gutenberg-Richter occurrence relationship, has a negative exponential
functional form, in which occurrence rates decrease with increasing magnitude. Higher b-
values release a larger proportion of the seismic moment in small earthquakes, while higher
a-values indicate an overall higher rate of seismicity.
While the Gutenberg-Richter occurrence relationship allows computation of the occur-
rence rate of earthquakes with any magnitude, some sources may only be capable of produc-
ing earthquakes up to some maximum magnitude, Mmax (discussed in Section 4.4.3). To ac-
count for these constraints, a truncated GR MFD is used to cut the MFD at Mmax magnitude.
4.4 Defining source occurrence 49
For calculation purposes, MFDs are also cut at a minimum magnitude (“double-truncated")
that signifies a threshold below which earthquakes are not expected to cause damage.
Figure 4.3 shows an example of a truncated GR MFD where a=4.0 and b=1.0, plotting
both an incrememntal (or discrete) and a cumulative MFD, and a truncated instance of
the cumulative MFD for which Mmax has been assigned. The cumulative MFD shows the
MFD computed from the Gutenberg-Richter recurrence relationship, which indicates the
annual occurrence rate of earthquakes with a magnitude greater than or equal to the x-axis
magnitude, while the incremental MFD includes only the occurrences within the respective
bin, here using a bin width of 0.1. The discrepancy between these two representations of the
MFD is significant in terms of hazard calculation, and the modeller must check which MFD
is assumed by the utilized calculation engine. Please read the OpenQuake Engine User’s
Manual to see how to assign MFDs in the case of the OpenQuake Engine.
Catalogue observations classified to a seismic source or source zone (review Section
4.1.3 about catalogue classification) are the most widely used data for constraining GR
MFDs. There are several algorithms used to resolve a GR MFD from observed seismicity;
one of the most common is the maximum likelihood method of Weichert (1980). A major
advantage of this method is that it is possible to include earthquake magnitudes with un-
equal observation periods, allowing the modeller to include small earthquakes that were
detected more recently, while utilizing a longer duration—and thus computing more robust
rates—for larger magnitudes. The Weichert (1980) method, as well as other maximum like-
lihood methods, is described in Chapter 2.4 of the Hazard Modeller’s Toolkit User’s Guide.
In PSHA, common practice is to constrain MFDs to model only independently occurring
earthquakes (e.g., mainshocks; review Section 4.1.4 about declustering earthquake cata-
logues). Thus, when using catalogue observations to constrain a GR MFD, the declustered
catalogue should be used. We emphasize that the decision to include only mainshocks is a
topic of debate in the hazard community, however, we present the standard approach as of
this writing.
Characteristic MFDs
The characteristic MFD, first proposed by (Schwartz and Coppersmith, 1984), hypothesizes
that seismic moment for individual earthquake sources (i.e. faults) is predominantly re-
leased through large earthquakes that occur cyclically, contrasting the GR MFD by concen-
trating earthquake occurrences around upper magnitudes. The characteristic MFD is pre-
dominantly reserved for seismic sources that are geometrically well constrained and thought
to produce large earthquakes, e.g., fault sources or subduction interfaces.
The simplest characteristic MFD assigns all occurrences of a source to a single magnitude—
the magnitude that corresponds to full-fault rupture—but MFDs of this type more commonly
use boxcar or truncated Gaussian functions centered on the characteristic magnitude, e.g.
Figure 4.4. Other characteristic MFDs allocate the majority of seismic moment release to the
50 Chapter 4. Seismic source model
105
Incremental GR MFD
Cumulative GR MFD
103
Annual occurrence rate
101
10 1
10 3
0 2 4 6 8
Magnitude
Figure 4.3 – Example of a GR MFD.
A is the fault surface area, µ is the shear modulus, c the seismogenic coupling coefficient,
and ṡ is the fault slip rate. The moment balance distributes Ṁo over the range of permitted
magnitudes following Equation 4.3.
4.4 Defining source occurrence 51
Z Mmax
Ṁ0 = λ(m)M0 (m)d m (4.3)
Mmin
10-2
Characteristic
10-3
Annual Rate
10-4
10-5
For some source typologies, a single MFD per source sufficiently models the occurrences;
this is the case for fault sources and distributed seismicity that is modelled by area sources.
Alternatively, distributed seismicity can be modeled using smoothed gridded seismicity,
which heterogenously distributes the MFD occurrence rates across a source zone using a
smoothing kernel and the distribution of observed seismicity. If a GR MFD is used, the b-
value remains constant, while the a-value is distributed among of grid of point sources that
discretize the source zone. In PSHA, smoothed seismicity is often used to account for the
spatial variability in intraslab seismicity or background seismicity in shallow crustal tectonic
regions.
One of the most common smoothing algorithms used in modern PSHA is the method of
Frankel (1995). This method is described in detail in the HMTK Tutorial Section 2.6.
52 Chapter 4. Seismic source model
100
YoungsCoppersmithExponential
YoungsCoppersmithCharacteristic
10-1
Annual Rate
10-2
10-3
0.3 or 0.5).
An algorithm by Makropoulos and Burton (1983) considers the amount of seismic mo-
ment that has accumulated in a source zone or on a fault compared to the moment that has
been released in earthquakes, thus resolving the seismic moment budget.
Lastly, when fault sources are used to model seismicity, the dimensions of the fault sur-
face can be used to constrain Mma x using magnitude scaling relationships, the same equa-
tions that are used to generate ruptures from sources. The magnitude scaling relationships
included in the OpenQuake Engine are summarised here.
The algorithms of Kijko (2004) and Makropoulos and Burton (1983) are explained in
Section 2.5 of the HMTK Tutorial.
4.5 Examples
• 1_Catalogue_Basics.ipynb: introduces common catalogue formats, and the format
used by the catalogue module (cat) of the OpenQuake MBTK
• 2_Declustering.ipynb: uses the HMTK to demonstrate catalogue declustering
• 3_Completeness.ipynb: uses the HMTK to demonstrate completeness analysis
• 4_Magnitude_homogenization.ipynb: uses the cat module of the MBTK to demon-
strate magnitude homogenization
• 5_Maximum_magnitude.ipynb: shows methods from the HMTK that are used to
determine the maximum magnitude of a source
• 6_MFD_from_seismicity.ipynb: demonstrates how to use catalogue seismicity to de-
rive a source magnitude frequency distribution using the HMTK and MBTK
• 7_hypocentral_depth_distribution.ipynb: uses the HMTK and MBTK to compute
the probability mass function of depths for seismicity produced by a source zone
• 8_nodal_plane_distribution.ipynb: uses the HMTK and MBTK to show the relative
proportion of recorded earthquakes with each simplified cateogry of focal mechanisms
• 9_MFD_from_slip_rate.ipynb: uses the HMTK and MBTK to use fault slip rate and
surface area to derive various types of MFDs
Part III
with the corresponding period. The spectral acceleration as a function of period is called
the response spectra. PGA is a valuable parameter for evaluating or designing shorter build-
ings, while for taller buildings, SA of a period corresponding to that of the building is more
useful.
Strong motion stations usually record data on three components: one oriented to record
each of north-south, east-west, and vertical motion. In most cases, PSHA modellers are
interested in the horizontal components of ground motion, since these components include
the seismic waves that are most hazardous to the built environment (e.g. Figure 5.1).
Strong motion data can be retreived in a few ways. In the optimal case, databases pro-
vide PGA, SA, and other parameters from accelerograms that have already been processed.
Datasets of this type are usually only available for the largest or most impacting earthquakes.
More commonly, the accelerograms corresponding to the same event have been grouped and
clipped to the duration of that event, simplifying the data compilation for users. For earth-
quakes that are less widely studied, the time-series of raw accelerograms must be collected.
Strong motion data sources are discussed more in Chapter 5.3.1.
Utility in PSHA
Strong ground motions are the shaking levels that modellers aim to replicate in the ground
motion characterization. One way that strong motion data are used in PSHA is to derive
a Ground Motion Prediction Equation (GMPE) for an area of interest; this is a significant
and time-consuming procedure in itself, and is beyond the scope of this training material.
We refer interested readers to a review by Douglas and Edwards (2016). More commonly,
strong motion data are used in PSHA as a way to choose a GMPE from the existing database
to use for a model. In summary, this is done by comparing the observed ground motions to
the intensity levels predicted by each GMPE, and choosing that which performs best. One
such workflow is presented in more detail in Chapter 6.2.
5.2.1 vs30
Shear wave velocity is a mechanical parameter linked to the rigidity of the medium. Within
the Earth, on average, the shear wave velocity decreases as depth decreases (i.e. the shear
5.2 Site information 59
80
EW
60
Acceleration (cm/s/s)
40
20
0
20
40
60
80
0 2 4 6 8 10
Time (s)
80
NS
60
Acceleration (cm/s/s)
40
20
0
20
40
60
80
0 2 4 6 8 10
Time (s)
40
Vertical
30
Acceleration (cm/s/s)
20
10
0
10
20
30
40
0 2 4 6 8 10
Time (s)
60 Chapter 5. Basic data and parameters
wave velocity is lower close to the surface than at great depth). A common parameter
used for site characterization in PSHA is VS,30 (Borcherdt, 1994), the average shear wave
velocity of the materials in the uppermost 30 m of the soil column, inferred by the travel
time of a shear wave from the surface to 30 m depth. In general, VS,30 increases for more
consolidated and harder rock. VS,30 first emerged as a means of divising site classes that
correlate with shallow rock or soil type, ranging from hard rock to soft soils, as part of the
US National Earthquake Hazards Reduction Program (NEHRP) building code (Borcherdt,
1994; Borcherdt and Glassmoyer, 1992), and other building codes have since incorporated
similar VS,30 -based site classifications.
Most GMPEs include VS,30 as a variable that accounts for the site class, and as a result,
computed hazard corresponds to only that site class. Regional or national hazard maps are
usually computed for a single site class either to provide reference conditions for a wide
area, or because detailed site characteristics are not available.
Measured VS,30 values can be obtained via both invasive and non-invasive shallow geo-
physical methods, including borehole analysis and surface wave inversions techniques. Where
these are not available, proxies based on geology or topography must be used; the topogra-
phy proxy developed by Allen and Wald (2007) is described in Chapter 5.3.2.
VS,30 , while a useful proxy for site characterisation and its impacts on ground motion ampli-
fication, is most effective for shorter spectral periods. At longer spectral periods (> ∼ 1.0 s,
e.g. Zhu et al. (2019)), the subsurface beyond 30 m may have greater significance. To ac-
count for this, some recent GMPEs have added terms that serve as proxies for basin structure,
utilizing shear wave isosurfaces that indicate the local depth to a fixed shear wave velocity:
the depth to shear wave velocity Vs = 1.0 and 2.5 km/s. These values can be obtained using
similar methods to those used for VS,30 : borehole and seismic inversion techniques, and—in
their absence—VS,30 -based conversion equations. Vs = 1.0 km/s is usually on the order of
tens to hundreds of meters, while Vs = 2.5 km/s ranges from hundreds of meters to a few
kilometers.
The OpenQuake Engine includes more than 250 GMPEs. The site parameters required by
each GMPE are listed under REQUIRES_SITES_PARAMETERS in the OpenQuake GSIM
Package Documentation.
62 Chapter 5. Basic data and parameters
Figure 5.3 – An example strong motion data summary from CESMD (truncated) for the 30 July
2020 Pacoima earthquake in southern California.
summarised strong motion data for select earthquakes, e.g. CESMD provides tables of peak
ground motion at each station (example in Figure 5.3).
We encourage readers to explore the data and data products provided by each of the
strong motion data agencies and organizations.
While some openly available site characterisation data exists, only small geographic regions
are covered, and in general there is not a global resource for these data. Instead, several
proxies have been developed for estimating these parameters based on more readily avail-
able parameters, e.g., parameters availble from remote sensing.
One of the most common proxies for estimating VS,30 is using the Allen and Wald (2007)
method that is based on topography, or - more specifically - slope. This method works
on the assumption that flatter topography more likely has a surface cover composed of
unconsolidated sediment than of rock. Thus, steeper slopes correspond to harder rock and
lesser amplification of strong ground motions due to resonance effects.
The most common algorithms for estimating the depths of isosurfaces at which shear
waves travel 1.0 and 2.5 km/s are based on VS,30 (measured or computed using Allen and
Wald (2007)), while some proxies for the 2.5 km/s isosurface are based on the 1.0 km/s
64 Chapter 6. Ground motion characterisation
(e.g. subduction interface, active shallow crustal, volcanic, etc.) as a means of categoriz-
ing sources with common properties. Some GMPEs go beyond this, adding geographical
constraints to their applicability. These categorical factors are imposed by the data used to
develop the GMPE; e.g. only seismic recordings for subduction interface earthquakes are
used to constrain the respective GMPEs.
A comprehensive summary by John Douglas of the epirical GMPEs developed since 1964
is available http://www.gmpe.org.uk/gmpereport2014.html.
Figure 6.1 – Distance metrics that are commonly used in GMPEs; see Table 6.1.
The mean value predicted by a GMPE results from the regression of many strong mo-
tion data to a functional form. The data maintain a scattered distribution about the mean
result and are the basis of the statistical standard deviation. The standard deviation permits
variability in the ground motions for the same rupture parametrization due to factors that
cannot reasonably be accounted for by GMPE variables, e.g. rupture velocity, slip distribu-
tion, source-to-site path, etc.
The standard deviation of a GMPE can be attributed to two types of variability, intra-
and inter-event, which together capture the total standard deviation (see Atik et al. (2010)).
Intra-event standard deviation is the variation in intensity measure level that occurs in a sin-
gle earthquake for equal source-to-site distances. For example, during a single earthquake,
the PGA recorded for events 50 km from the earthquake hypocenter may have scatter that
depends on the geology between the rupture and the station, the strike of the fault com-
pared to the station azimuth, and the direction of rupture propagation. On the other hand,
inter-event standard deviation is the change in intensity measure levels recorded at the
same station during different earthquakes. For example, the state of stress in the crust may
have changed between two consecutive co-located earthquakes of comparable magnitude,
impacting the energy in the seismic waves that propagate away from the rupture.
6.2 Choosing ground motion models for PSHA 67
At least one GMPE must be selected for each large-scale tectonic region included in the
seismic source characterisation of a PSHA input model. Given the large number of available
GMPEs, selecting them is not a simple task. To date, there is not a rigorous and uniform
methodology used by experts to choose GMPEs, which is in part due to high variability
of strong motion data availability globally. However, some "best practice" guidelines and
quantitative approaches are widely used. In this chapter, we introduce the most common
ways that hazard modellers select GMPEs. We discuss the following approaches as if the
aim were to select a single GMPE for each tectonic region, however, PSHA ground motion
characterizations usually incorporate epistemic uncertainty by weighting several GMPEs per
tectonic region using a logic tree; this is discussed more in Part IV.
capture local or regional characteristics that are not included as GMPE variables. GMPEs
derived for specific areas should not be used outside of their intended region without careful
consideration and a defensible explanation.
The GMPE qualities required to apply these criteria are summarised in a GMPE com-
pendium by John Douglas, which can be found at http://www.gmpe.org.uk.
Figure 6.2 – An example of trellis plots showing how ground motion levels predicted by five
GMPEs scale with magnitude for four spectral periods for an earthquake ∼20 km away from a
site on rock.
Trellis plots can be used to compare both the mean ground motions predicted by GMPEs
and the associated sigma (standard deviation). Again, there is no firm "recipe" for selecting
a GMPE from a set of candidates by using trellis plots, however, there are a number of
characteristics of the comparative plots to keep in mind:
1. Are any GMPEs distinct outliers for the majority of trellis plots?
2. Which GMPEs best represent the median and upper and lower bounds for the distances
6.2 Choosing ground motion models for PSHA 69
6.3 Examples
• 10_GMPEs_Introduction.ipynb: uses the OpenQuake Engine and GMPE-SMTK to
overview the basic properties of GMPEs
• 11_Trellis_Plots.ipynb: uses the GMPE-SMTK to demonstrate aspects of comparative
scaling
Part IV
Epistemic uncertainty pertaining to seismic hazard analysis exists for several reasons.
While the last decades have yielded major advances in the scientific understanding of earth-
quake occurence and strong ground motions, there is still a general scarcity of data relative
to the duration of the seismic cycle and global variability in tectonic processes; this challenge
was also discussed in Parts II and III. For example, strong-motion stations have only existed
for a few decades, and so in many cases they have not have recorded ground shaking during
the largest and most impactful earthquakes at a site (i.e. the most significant earthquake
occurred during the pre-instrumental period), and thus the expected ground motions are
unknown, and must be predicted by ground motion models that extrapolate or use large
magnitude earthquakes from other areas. Even in cases where the largest observed earth-
quake was recorded, it is impossible to assert whether it represents the largest magnitude
earthquake that the causative source can produce. This latter point corresponds to another
source of epistemic uncertainty: theoretically, an infinite data colletion period is required
to reject alternative hypotheses.
While the examples listed so far represent data challenges that lead to epistemic uncer-
tainty, another cause is disagreement among experts, either due to differing assumptions,
interpretations, methodologies, or theoretical bases regarding seismic processes. These fac-
tors can result in input models that yield a range of projected hazard levels, even when
using overlapping or identical datasets, as we demonstrated using the examples in Parts II
and III of this tutorial.
Figure 7.1 – An example of a logic tree from the OpenQuake User’s Manual, labelled using the
OpenQuake terminology. The logic tree is divided into hierarchical branching levels, each of
which represents a specific epistemic. The model begins at Branching level 1. Then, Branching
Level 2 introduces an epistemic uncertainty with two branches, the green ones, each of which
have two alternatives in branching Level 3, the ones colored blue, for a total of four branches.
The schematic shows how logic trees can easily grow to have a large number of end branches
The complexity and comprehensiveness of logic trees used in PSHA varies based on the
utility of the model, its maturity (e.g., its version), and the agencies and experts involved
in costructing the input model. For example, site-specific hazard studies, and especially
those produced for critical facilities, are typically more comprehensive in their logic tree
structures than national or regional PSHA. The most comprehensive logic trees aim to be
mutually exclusive and collectively exhaustive (e.g Bommer and Scherbaum, 2008; Jaynes,
2003) for each included epistemic uncertainty (in OpenQuake Engine terminology, for the
branches in a branchset), meaning that the alternative parametrizations inherently reject
each other, and together cover the range of possible solutions.
In this tutorial, we introduce the basics of the two logic trees used in the OpenQuake
Engine: the seismic source logic tree and the ground-motion logic tree. We additionally
recommend the following resources:
• source typology. For example, one model could use faults plus smoothed seismic-
ity, while the alternative model uses area sources with internally homogenous activ-
ity rates. In this case, the alternative models make different interpretations of how
strongly the known faults and recorded seismicity will correspond the positions of
seismicity within the investigation time.
• resolution. For example, one source models may comprise a more sources/source
zones with smaller dimensions. Smaller or more discrete sources and source zones
define the positions of seismicity more precisely, but suffer from lesser data available
to characterise their rates (i.e. they might have larger epistemic uncertainty affecting
the parameters describing their earthquake occurrence).
• basis for source delineation. For example, one source model may be based on
geodesy, while another is based on geology or seismicity.
• when modelling assumptions largely impact source parameter values, and ultimately
the hazard results, or
• when imprecise or sparse data are used to resolve source parameters.
While there are certain state-of-practice methodologies used to construct a seismic source
model, many of these methodologies still depend on some modeller input, for example, the
selection of magnitude scaling relationships, the type of MFD to assign to a source, and
the maximum magnitude. The examples in Part II demonstrated that choices made during
catalogue processing—such as the completeness table and the declustering algorithm—can
produce significantly different subcatalogues, and subsequently very different MFDs. Thus,
a modeller may choose to include two or more alternative MFDs for one seismic source.
(NB: this would result in numerous alternative initial SSMs.)
Uncertainty is inherent in many data types used to build seismic hazard models, often
due to short observation periods relative to the processes observed, sub-optimal measure-
ment positioning, or very low numbers of data points. For example, fault slip rates derived
from geodesy (e.g. GPS), paleoseismic trenching, or other geochronology often consist of a
minimum, maximum, and preferred value that together span a few mm/yr. For slow moving
faults, the range of slip rates may be larger than the preferred slip rate itself. As in the for-
mer example, a modeller may choose to include multiple alternative MFDs for a fault source.
The OpenQuake Engine seismic source logic tree supports several epistemic uncertain-
ties that are commonly incorporated into seismic source characterisations, including seismic
source models, Gutenberg-Richter a- and b-values, and maximum magnitude. Please read
the OpenQuake Engine User Instruction Manual and the OpenQuake Engine Underlying
Hazard Science for a complete list of supported uncertainties and further information on
constructing a seismic source logic tree for the OpenQuake Engine.
There is no universally used method for weighting ground motion models, in particular
when strong-motion data is unavailable. However, some quantitative approaches exist for
data-driven ground motion model selection. The Scherbaum et al. (2009) method, which
uses likelihood priniciples, and the Kale and Akkar (2013), which uses Euclidean distance-
based ranking, are implemented in the GMPE-SMTK.
Part V
PSHA results
81
Together, the components described in Parts II–IV comprise the input required for PSHA.
Here, we introduce the major types of hazard results produced by Classical PSHA:
1. hazard curves
2. hazard maps
3. uniform hazard spectra
This manual focuses on understanding the different result types. Please read the Open-
Quake Engine User Instruction Manual to learn how to set up and execute OpenQuake Clas-
sical PSHA jobs and the OpenQuake Underlying Hazard Science Chapter 2 to understand
the mathematics used by the OpenQuake calculators.
The state-of-practice PSHA softwares compute hazard results either in terms of annual rates
of exceedance, also called annual frequency of exceedence (AFE), or probabilities of ex-
ceedance in a given time span. They differ in that the former is independent of the investi-
gation time, while probabilities correspond to a specific time period. AFE is the inverse of a
return period. For an investigation time of 1 year, probabilities and rates of exceedance are
nearly equal, and thus the analyst must be attentive to the spefications of their software.
The OpenQuake Engine computes results in terms of probabilities of exceedance.
Equation 9.1 shows the conversion from rate to probability, where P(t) is the probability
given investigation time t and λ is the rate of occurrence. Table 9.1 lists some of the numbers
that are of interest to engineering applications in terms of annual probability of exceedance,
rate of exceedance, and return period.
Table 9.1 – Common recurrences in terms of annual probability of exceedance (POE), rate of
exceedance, and return period.
Figure 10.1 – An example of hazard curve in terms of probabilities of exceedance (POEs) for
PGA levels with a 1 year investigation period (e.g. annual POEs). Common return periods of
interest are indicated.
Hazard curve
10 1
Annual probability of exceedance
10 2
10 3
10 4
10 5
10 6
10 2 10 1 100
PGA (g)
logic tree end branches and N is the number of ground motion logic tree end branches.
Together, all realizations represent the range of probabilities for a site’s ground motion
parameter levels that are permissible by the PSHA input model, thus reflecting parameter
uncertainties. However, the applications of PSHA are typically more interested in statistical
parameters of the computed results, and in particular the mean hazard and some quantiles.
The mean hazard curve is computed using the individual realization probabilities and
the weights of their associated end branches as defined by the logic tree. Quantiles are
usually computed by interpolating between the values of the realizations, and sampling
accordingly. Please refer to the OpenQuake Underlying Hazard Science Chapter 5.3.3 to
see the mathematics of hazard statistics as computed by the OpenQuake Engine.
Figure 10.2 shows an example of the mean hazard curve as well as the 0.15, 0.50, and
0.85 quantiles hazard results for a model with numerous realizations.
10% in 50 years
2% in 50 years
Figure 10.3 – Example hazard maps covering Colombia showing mean PGA with (left) 10%
POE and (right) 2% POE in 50 years (NB: the maps are for demonstration purposes only).
10.0.4 Disaggregation
The PSHA outputs described so far have reflected the seismic hazard contributions from
all the sources and their possible ruptures and ground motions that may impact a site.
10.1 Examples 85
Figure 10.4 – An example UHS with 10% POE in 50 years. Left: Hazard curves for several
spectral periods used to construct a UHS, with filled circles indicating where the curve intersects
10% POE in 50 years. Right: the UHS constructed from these (and more) hazard curves.
However, for some applications, i.e. in engineering or risk analysis, it may be useful to
focus on independent earthquake ruptures, those that are contriubting most to the computed
intensity measure level for a given annual POE and spectral period. For example, one may
want to know which rupture distances and magnitudes are contributing most, and which
tectonic region types produce those ruptures.
The method used to identify which sources and ruptures contribute most significantly
is called seismic hazard disaggregation (sometimes called deaggregation). Disaggregation
reveals the relative contributions to a given frequency of exceedance of the different in-
dependent variables considered in the hazard integrals, such as magnitude, distance, or
epsilon (the coefficient multiplied by the GMPE sigma) or other categorical parameter, e.g.
tectonic region type. (Bazzurro and Cornell, 1999) and (Pagani and Marcellini, 2007) de-
scribe some of the methodologies used to disaggregate the seismic hazard. The approach
used by the OpenQuake Engine is described in the OpenQuake Engine Underlying Hazard
Science and the OpenQuake Engine User Instruction Manual.
Figure 10.5 shows two possible configurations of a disaggregation calculation: mag-
nitude, distance, and the GMPE standard deviation "epsilon" (the coefficent multiplied by
GMPE sigma); and latitude, longitude, and tectonic region type.
10.1 Examples
• 12_simple_PSHA_calculations.ipynb: uses the OpenQuake Engine as a library for
performing simple hazard calculations, and reviews common hazard outputs
Figure 10.5 – Two disaggregation configurations. Left: magnitude, distance, and the GMPE
standard deviation "epsilon". Right: latitude, longitude, and tectonic region type. In each, the
height of the column indicates the contribution from that parameter bin.
87
11 Bibliography
Books
Campbell, K. W. and Y. Bozorgnia (2007). Campbell-Bozorgnia NGA ground motion rela-
tions for the geometric mean horizontal component of peak and spectral ground motion
parameters. Pacific Earthquake Engineering Research Center (cited on page 60).
Jaynes, E. T. (2003). Probability theory: The logic of science. Cambridge university press
(cited on pages 70, 73).
McGuire, R. K. (2004). Seismic Hazard and Risk Analysis. EERI (cited on page 5).
Articles
Abrahamson, N. A. and W. Silva (2008). “Summary of the Abrahamson & Silva NGA Ground-
Motion Relations”. In: Earthquake Spectra 24.1, pages 67–97 (cited on page 60).
Atik, L. A., N. Abrahamson, J. J. Bommer, F. Scherbaum, F. Cotton, and N. Kuehn (2010).
“The variability of ground-motion prediction models and its components”. In: Seismo-
logical Research Letters 81.5, pages 794–801 (cited on page 62).
Bazzurro, P. and C. A. Cornell (1999). “Disaggregation of Seismic Hazard”. In: Bulletin of
the Seismological Society of America 89.2, pages 501–520 (cited on page 81).
Bommer, J. J. and F. Scherbaum (2008). “The Use and Misuse of Logic Trees in Probabilis-
tic Seismic Hazard Analysis”. In: Earthquake Spectra 24.4, pages 997–1009 (cited on
pages 66, 69, 70, 73).
Bommer, J. J., J. Douglas, F. Scherbaum, F. Cotton, H. Bungum, and D. Fah (Sept. 2010). “On
the Selection of Ground-Motion Prediction Equations for Seismic Hazard Analysis”. In:
Seismological Research Letters 81.5, pages 783–793. ISSN: 0895-0695. DOI: 10.1785/
gssrl.81.5.783. eprint: https://pubs.geoscienceworld.org/srl/article-
pdf/81/5/783/2762378/783.pdf. URL: https://doi.org/10.1785/gssrl.
81.5.783 (cited on page 63).
Boore, D. M., J. Watson-Lamprey, and N. A. Abrahamson (2006). “Orientation-independent
measures of ground motion”. In: Bulletin of the seismological Society of America 96.4A,
pages 1502–1511 (cited on page 53).
Borcherdt, R. D. (1994). “Estimates of site-dependent response spectra for design (method-
ology and justification)”. In: Earthquake spectra 10.4, pages 617–653 (cited on page 57).
Borcherdt, R. D. and G. Glassmoyer (1992). “On the characteristics of local geology and
their influence on ground motions generated by the Loma Prieta earthquake in the San
Francisco Bay region, California”. In: Bulletin of the seismological society of America
82.2, pages 603–641 (cited on page 57).
88 Chapter 11. Bibliography
Campbell, K. W. and Y. Bozorgnia (2014). “NGA-West2 ground motion model for the aver-
age horizontal components of PGA, PGV, and 5% damped linear acceleration response
spectra”. In: Earthquake Spectra 30.3, pages 1087–1115 (cited on page 60).
Chiou, B. S.-J. and R. R. Youngs (2008). “An NGA Model for the Average Horizontal Compo-
nent of Peak Ground Motion and Response Spectra”. In: Earthquake Spectra 24, pages 173–
215 (cited on page 60).
Chiou, B. S.-J. and R. R. Youngs (2014). “Update of the Chiou and Youngs NGA model for
the average horizontal component of peak ground motion and response spectra”. In:
Earthquake Spectra 30.3, pages 1117–1153 (cited on page 60).
Cornell, C. A. (1968). “Engineering Seismic Risk Analysis”. In: Bulletin of the Seismological
Society of America 58, pages 1583–1606 (cited on page 11).
Cotton, F., F. Scherbaum, J. Bommer, and H. Bungum (Apr. 2006). “Criteria for Selecting
and Adjusting Ground-Motion Models for Specific Target Regions: Application to Central
Europe and Rock Sites”. In: Journal of Seismology 10, pages 137–156. DOI: 10.1007/
s10950-005-9006-7 (cited on page 63).
Douglas, J. and B. Edwards (2016). “Recent and future developments in earthquake ground
motion estimation”. In: Earth-Science Reviews 160, pages 203–219 (cited on page 54).
Field, E. H., R. J. Arrowsmith, G. P. Biasi, P. Bird, T. E. Dawson, K. R. Felzer, D. D. Jack-
son, K. M. Johnson, T. H. Jordan, C. Madden, A. J. Michael, K. R. Milner, M. T. Page, T.
Parsons, P. M. Powers, B. E. Shaw, W. R. Thatcher, R. J. Weldon, and Y. Zeng (June
2014). “Uniform California Earthquake Rupture Forecast, Version 3 (UCERF3)—The
Time-Independent Model”. en. In: Bulletin of the Seismological Society of America 104.3,
pages 1122–1180. ISSN: 0037-1106, 1943-3573. DOI: 10.1785/0120130164. URL:
http://bssa.geoscienceworld.org/content/104/3/1122 (visited on 05/05/2016)
(cited on page 43).
Frankel, A. (1995). “Mapping Seismic Hazard in the Central and Eastern United States”. In:
Seismological Research Letters 66.4, pages 8–21 (cited on pages 47, 89).
Garcia, D., D. J. Wald, and M. Hearne (2012). “A global earthquake discrimination scheme to
optimize ground-motion prediction equation selection”. In: Bulletin of the Seismological
Society of America 102.1, pages 185–203 (cited on page 32).
Gardner, J. K. and L. Knopoff (1974). “Is the sequence of earthquakes in Southern Califor-
nia, with aftershocks removed, Poissonian?” In: Bulletin of the Seismological Society of
America 64.5, pages 1363 –1367 (cited on page 33).
Gutenberg, B. and C. F. Richter (1944). “Frequency of Earthquakes in California”. In: Bulletin
of the Seismological Society of America 34, pages 185 –188 (cited on page 44).
Hale, C, N Abrahamson, and Y Bozorgnia (2018). “Probabilistic seismic hazard analysis
code verification”. In: Pacific Earthquake Engineering Research Center, Report 3 (cited
on page 11).
89
Silva, V., H. Crowley, M. Pagani, D. Monelli, and R. Pinho (2014). “Development of the
OpenQuake engine, the Global Earthquake Model’s open-source software for seismic
risk assessment”. In: Natural Hazards 72.3, pages 1409–1427 (cited on page 11).
Stewart, J. P., J. Douglas, C. Di Alessandro, Y. Bozorgnia, N. A. Abrahamson, D. M. Boore,
K. W. Campbell, E. Delavaud, M. Erdik, and P. J. Stafford (2012). “Selection of a global
set of GMPEs for the GEM-PEER Global GMPEs Project”. In: (cited on page 65).
Stirling, M., T. Goded, K. Berryman, and N. Litchfield (2013). “Selection of earthquake scal-
ing relationships for seismic-hazard analysis”. In: Bulletin of the Seismological Society
of America 103.6, pages 2993–3011 (cited on page 42).
Weatherill, G. A., M. Pagani, and J. Garcia (Sept. 2016). “Exploring earthquake databases for
the creation of magnitude-homogeneous catalogues: tools for application on a regional
and global scale”. en. In: Geophysical Journal International 206.3, pages 1652–1676.
ISSN: 0956-540X. DOI: 10.1093/gji/ggw232. URL: https://academic.oup.com/
gji/article/206/3/1652/2583518 (visited on 03/22/2019) (cited on pages 22,
23, 27, 30).
Weichert, D. H. (1980). “Estimation of the Earthquake Recurrance Parameters for Unequal
Observation Periods for Different Magnitudes”. In: Bulletin of the Seismological Society
of America 70.4, pages 1337 –1346 (cited on page 45).
Youngs, R. R. and K. J. Coppersmith (Apr. 1985). “Implications of fault slip rates and earth-
quake recurrence models to probabilistic seismic hazard estimates”. In: Bulletin of the
Seismological Society of America 75, pages 939–964 (cited on pages 44, 46, 48).
Zhao, J. X., S. Zhou, P. Gao, T. Long, Y. Zhang, H. K. Thio, M. Lu, and D. A. Rhoades (2015).
“An earthquake classification scheme adapted for Japan determined by the goodness of
fit for ground-motion prediction equations”. In: Bulletin of the Seismological Society of
America 105.5, pages 2750–2763 (cited on page 32).
Zhu, C., F. Cotton, and M. Pilz (2019). “Testing the depths to 1.0 and 2.5 km/s velocity
isosurfaces in a velocity model for Japan and implications for ground-motion modeling”.
In: Bulletin of the Seismological Society of America 109.6, pages 2710–2721 (cited on
page 57).
Other Sources
Allen, T. I. and D. J. Wald (2007). Topographic slope as a proxy for seismic site-conditions
(VS30) and amplification around the globe. Technical report. Geological Survey (US)
(cited on pages 57, 59).
Budnitz, R., G Apostolakis, and D. Boore (1997). Recommendations for probabilistic seismic
hazard analysis: guidance on uncertainty and use of experts. Technical report. Nuclear
Regulatory Commission, Washington, DC (United States). Div. of . . . (cited on page 70).
Johnston, A., L. Kanter, K. Coppersmith, and C. Cornell (1994). The earthquakes of stable
continental regions. Volume 1, Assessment of large earthquake potential, Final report.
91
Technical report. Electric Power Research Inst., Palo Alto, CA (United States); Memphis
State . . . (cited on page 48).
Kulkarni, R., R. Youngs, and K. Coppersmith (1984). “Assessment of confidence intervals
for results of seismic hazard analysis”. In: Proceedings of the eighth world conference
on earthquake engineering. Volume 1, pages 263–270 (cited on page 69).
Mignan, A., J. Woessner, and D. Scherlemmer (2010). Theme IV - Understanding Seismicity
Catalogs and their Problems. Technical report doi: 10.5078/corssa-00180805. Commu-
nity Online Resource for Statistical Seismicity Analysis (cited on page 34).
Stepp, J. C. (1971). “An investigation of earthquake risk in the Puget Sound area by the use
of the type I distribution of largest extreme”. PhD thesis. Pennsylvania State University
(cited on page 34).
Stiphout, T. van, J. Zhuang, and D. Marsan (2010). Theme V - Basic Techiniques/Models
for Analysing Seismicity. Technical report. Community Online Resource for Statistical
Seismicity Analysis. URL: http://www.corssa.org (cited on page 33).
92 Glossary
Glossary
activity rate
See magnitude-frequency distribution.
aleatory uncertainty
Uncertainty in a phenomenon that is due to true variability..
annual frequency of exceedence
The frequency in 1/yr at which a ground motion level is expected to be exceeded..
annual rate of occurrence
The average number of occurences in a one year period, e.g. earthquakes of a given
magnitude by a source..
area source
A source type usually adopted to model distributed seismicity. In an area source the
seismicity occurrence rate is assumed uniform over the source area; this produces an
hazard pattern with a plateau of constant hazard inside the polygon delimiting the
area source and values of hazard that tend to decrease as we move away from the
border of the source.
centroid
the center of the energy release of an earthquake.
characteristic MFD
An magnitude-frequency distribution in which all or most moment release occurs at
one magnitude or over a small range of magnitudes..
classified
grouped based on a category such as tectonic region, i.e. a classified catalogue..
completeness magnitude
The lower magnitude threshold above which all earthquakes occurring within the
region covered by a catalogue are thought to be included..
cumulative MFD
An magnitude-frequency distribution in which the given annual occurrence rate cor-
responds to magnitudes greater than or equal to the respective magnitude..
declustering
The procedure by which independent earthquakes (mainshocks) in a catalogue are
identified and separated from other occurrences (foreshocks and aftershocks)..
distance
i.e. the distance between an earthquake and a site. There are several ways to measure
distance, e.g. see rrup .
Glossary 93
distributed
i.e. distributed seismicity. a collection of possible seismic source with less definitive
geometry.
earthquake catalogue
A database of earthquakes, where each entry include the basic information about an
earthquake on record, at minimum including the earthquake time, magnitude, and
hypocentral location.
earthquake rupture
A 3D surface - representing a portion or the entire fault surface - over which a slip
event (i.e. an earthquake) occurs.
epistemic uncertainty
The type of uncertainty that results from limited knowledge or data about a phe-
nomenon.
event
In seismology, a term sometimes used to refer to one earthquake.
fault
Discontinuities within the earth’s crust where rock has been crushed or fractured,
separating the surrounding earth into separate blocks. See also the OpenSHA website.
fault trace
A curve representing the intersection between the surface containing the fault surface
(or its prolongation) and the topographic surface .
focal mechanism
Geometrical or mathematical representations of the faulting that occurs during an
earthquake Earthquake, Focal Mechanism.
global catalogue
Earthquake catalogues with global coverage..
grid source
A source typology usually adopted to model distributed seismicity. It is routinely pro-
duced by a seismicity smoothing algorithm (one of the most famous algorithm is the
one proposed by Frankel (1995)).
ground-motion logic tree
A method used to systematically describe the epistemic uncertainties related to the
ground motion models used in the computation of hazard using a specific PSHA input
model.
ground-motion model
An object that given a rupture with specific properties computes the expected ground
motion at the given site. In simplest case a ground motion model corresponds to a
94 Glossary
ground-motion prediction equation. In case of complex PSHA input models, the pro-
duced ground motion models contains a set of Ground Motion Prediction Equations
(GMPEs), one for each tectonic region considered.
ground-motion prediction equation
An equation that - given some fundamental parameters characterizing the source, the
propagation path and the site (in the simplest case magnitude, distance and VS,30 ) -
computes the value G M of a (scalar) ground motion intensity parameter.
hazard curve
A set of intensity measure levels and their corresponding probabilities of exceedance
for a single site, site condition, and investigation time..
hazard map
The spatial distribution of intensity measure levels for a fixed probability of exceedance
and site condition..
historical catalogue
A compilation of past earthquakes and their parameters comprising events that oc-
curred before the time of seismometers, but that were felt by humans..
hypocenter
The 3D nucleation point of an earthquake rupture.
incremental MFD
An magnitude-frequency distribution in which the given annual occurrence rate cor-
responds to magnitudes only of the specified magnitude. In practice, e.g. in the
OpenQuake Engine, the specified magnitude is a magnitude range defined by some
bin width and center or edge magnitude value..
initial seismic source input model
It is the ensemble of information needed to fully describe the seismic sources compos-
ing a seismic source input model. The initial seismic source input model is included
in the first branching level of a seismic source logic tree.
input model
An object containing the information necessary to describe the seismic source and the
ground motion models - plus the related epistemic uncertainties.
instrumental catalogue
A compilation of past earthquakes and their parameters comprising events for which
the event information is obtained by processing seismometer data..
length
i.e. Fault length; the total length of the fault trace..
local catalogue
Earthquake catalogues with higher density station coverage but covering only sections
of the globe (<100 km-scale)..
Glossary 95
logic tree
Data structure used to systematically describe uncertainties on parameters and models
used in a PSHA study.
magnitude
The size of an earthquake.
magnitude homogenization
The process by which the magnitudes for a set of earthquakes are converted to use a
single measure of magnitude..
magnitude-frequency distribution
A distribution describing the frequency of earthquakes with a specific magnitude.
It can be continuous or discrete. One frequency- magnitude distribution frequently
adopted in Classical PSHA (PSHA) is the double truncated Gutenberg-Richter distri-
bution.
magnitude-scaling relationship
An empirical relationship linking the magnitude with a parameter describing the size
of the corresponding rupture (e.g. the area of the rupture or the rupture length).
nodal plane
In seismology, the two great circles that separate the extensional and compressional
quadrants of a focal mechanism..
paleoseismic trench
An elongate hole excavated across strata offset by a fault and used to reveal a sequence
of depositional and offsetting events.
peak ground acceleration
The highest acceleration experienced by a particle on the ground during an earth-
quake. See also the USGS webpage: Earthquake Hazards 201.
point source
The elemental source typology used in the OpenQuake-engine to model distributed
seismicity.
pre-historic catalogue
A compilation of past earthquakes and their parameters comprising events that oc-
curred before humans, which are revealed through geological investigations..
prime
The preferred earthquake origin parameters as determined by the organization re-
porting the data..
probabilistic seismic hazard analysis
A methodology to compute seismic hazard by taking into account the potential con-
tributions coming from all the sources of engineering importance for a specified site.
96 Glossary
probability of exceedance
In PSHA, the probability that a ground-motion value for a given intensity measure
type will be exceeded during an investigation time..
rake
The rake is the direction in which a hanging wall block moves during a rupture, mea-
sured relative to fault strike on the plane of the fault.
rate of exceedance
The number of times that an intensity measure level (e.g. a ground-motion value
generated by an earthquake) is exceeded..
realization
The hazard results computed from a single end branch of the logic tree for a input
model..
recurrence interval
The average time between consecutive earthquakes on a fault.
regional catalogue
Earthquake catalogues with higher density station coverage but covering only sections
of the globe (>100 km-scale)..
residual analysis
A quantitative method of comparing strong motion data to GMPE predicted values in
order to evaluate or select GMPEs for a PSHA model.
response spectra
The magnitude of a parameter describing strong ground motion (e.g. acceleration,
velocity, displacement) as a function of the natural vibrational period.
rupture aspect ratio
The ratio between the lenght and the width of an earthquake rupture.
scarp
The surface expression of an earthquake, often manifesting as a step in topography
across a fault..
seismic hazard analysis
A procedure used to determine the level of danger that may occur due to earthquakes
within an area of interest over a given time frame. See also the OpenSHA website.
seismic hazard disaggregation
A methodology to investigate the contributions to a specific level of hazard in terms
of fundamental variables commonly used to characterize seismic sources and ground
motion models (e.g. magnitude, source-site distance, epsilon.
seismic source
An object that can generate earthquake ruptures.
Glossary 97
tectonic region
A area on the topographic surface that can be considered homogeneous in terms of
tectonic properties such as the prevalent seismogenic properties and/or the seismic
wave propagation properties.
time-independent
A modelling assumption that indicates a model is valid for any time frame.
98 Glossary
truncated GR MFD
A Gutenberg-Richter MFD that is “cut" at an upper magnitude that is considered the
maximum credible earthquake that can be produced by the corresponding seismic
source..
VS,30
Average shear wave velocity of the materials in the uppermost 30m of the soil column.