IMechE Alarp-Technical-Safety-Guide - 2021
IMechE Alarp-Technical-Safety-Guide - 2021
IMechE Alarp-Technical-Safety-Guide - 2021
ENGINEERS:
A TECHNICAL
SAFETY GUIDE
What we do:
• Provide best practice guides and information to engineers associated with designing,
manufacturing, operating or maintaining products.
• Increase the public understanding of risk, while continuing to work on the assessment
of Hazardous events, dissemination of lessons learned and the promotion of risk
reduction strategies.
• Represent the Institution on the Hazards Forum which provides an interdisciplinary focus
for the study of disasters.
• Organise events, such as the annual ALARP Seminar.
• Support Proceedings of the Institution of Mechanical Engineers, Part O: Journal of Risk
and Reliability.
The SRG is a voluntary group, which meets quarterly. We welcome new members who have
experience in, and a passion for, safety. Please send CV to [email protected].
Acknowledgements
This document has been developed by Working Group 2 of the IMechE Safety and Reliability
Group, with support from the other SRG working groups.
We would like to thank the many external reviewers from legal firms, industry, consultancies,
regulators, engineering institutions and societies, whose contributions have been invaluable.
Glossary of Terms 06
Acronyms 08
1. Introduction and Scope 09
1.1 Why this Document is Needed 10
1.1.1 Process Safety 10
1.1.2 Systems Approach to Safety 10
1.1.3 Reasonable Foreseeability 10
1.1.4 Probabilistic Assessment 11
1.1.5 Evidential Admissibility 11
1.1.6 Risk Management Contiguity 11
2. UK Safety Legislation 12
2.1 Regulatory Control and Prosecution 14
2.2 Reasonable Foreseeability and the ‘Incidence of everyday life’ 16
2.3 Reasonable Practicability and Gross Disproportion 18
2.4 Risk Transfer and Risk Trade-offs 20
2.5 Target Safety Levels (TSLs) 21
2.6 The Legal Duty Holder 22
2.7 Reverse Burden of Proof 22
2.8 Retrospective Application of ALARP Principles 22
3. Overview of the Safety Risk Management Process 23
3.1 The Hazardous System 25
3.2 Proportionality 25
3.3 The Proportionality Matrix 27
3.4 Engineering Management Systems 27
4. Identification 30
4.1 The Hierarchy of Controls 30
4.2 Basic HAZID 31
4.3 Comprehensive HAZID 35
4.3.1 Hierarchical Guideword Expansion 35
4.3.2 Contextual Guideword Expansion 36
4.3.3 The Degree of Guideword Expansion 38
4.3.4 Guidewords for Identifying RRMs and Actions for Further Study 40
4.4 HAZID Conclusions 42
5. Evaluating Reasonably Foreseeable Consequences 43
5.1 Consequence Reduction Measures 44
6. Analysis 45
6.1 RRM Effectiveness Review 47
6.2 Human Error, Human Factors, and Ergonomics 49
6.3 Task Risk Analysis (TRA) and Procedure HAZOP 51
6.3.1 Task Risk Analysis 51
6.3.2 Procedure HAZOP 52
6.4 Systems Theoretic Process Analysis (STPA) 53
6.5 Failure Modes and Effects Analysis (FMEA/FMECA) 55
Barrier Another term for a Risk Reduction Measure (RRM), generally used in the context of
Bow Ties.
Cause Anything whose absence may prevent a loss, e.g. errors in design, manufacture,
control, human factors, ergonomics, management systems, equipment unreliability,
RRM flaws, safety culture, conflicts of interest, cognitive bias. (NB. Terms such as
root, direct and indirect Causes may be ambiguous, so they are avoided.)
Failure Mode A type of failure, e.g. corrosion, fatigue, rupture, unrevealed control failure.
Good Practice Standards, practices, methods, guidance, and procedures conforming to the law and
the degree of skill and care, diligence, prudence and foresight which would
reasonably and ordinarily be expected from a skilled and experienced person or
body engaged in a similar type of undertaking under the same or similar
circumstances.
NB. The HSE have also defined Good Practice as the generic term for those
standards for controlling Risk which have been judged and recognised by HSE as
satisfying the law when applied to a particular relevant case in an appropriate
manner.
Gross An RRM that is not considered to be Reasonably Practicable, because the sacrifice
Disproportion (in time, trouble, or money) is disproportionate to the Risk being mitigated.
See Section 2.3 for a full definition.
Hazard A useful means of expressing a condition or material that presents the ‘possibility
of danger’ in a specified situation, e.g. cooking oil fire in busy kitchen, pressurised
release of toxic material from relief valve upwind of an office building, aircraft stalls
whilst turning finals at one thousand feet. A defined Hazard is useful when the
analysis of it would not benefit further if it were contextualised anymore.
NB. Depending on the analytical objectives the hazard could be either
i. poor control panel layout
ii. operator presses wrong button
iii. reactor goes unstable
iv. reactor melts down
v. radiation reaches domestic housing
vi. people exposed to radiation or
vii. people die from radiation sickness.
Any of these could be described as a Hazard or a Cause, depending on the analytical
context, so Hazards may be described as mutually exclusive Risk scenarios that
generally represent several Causes. The differences between Hazards, Risks and
Causes are not always clear.
Legal Duty The organisation or individual upon which the legal obligation lies.
Holder
Lifecycle Safety Critical data that needs to be established at the design stage and maintained
Criteria or updated throughout the lifecycle of the product.
Reasonable A legal concept to define which Hazards fall within legislation. See Section 2.2.
Foreseeability
Regulator The UK HSE, ONR, ORR, BSR, CAA, local authorities, and the police.
Risk A combination of severity and likelihood, expressed in case law as ‘the possibility
of danger’.
Risk Reduction Any measure that removes a Hazard or prevents harm to people, reduces its
Measure (RRM) likelihood or mitigates the consequences.
Safety Critical Any component, function, activity, process, or procedure whose omission, failure or
incorrect operation could increase Risks associated with the System.
Scenario A specific set of conditions that may be physical, sociological and/or environmental,
which could create a Hazardous situation.
Well-Reasoned A qualitative, rational explanation of how all Reasonably Foreseeable Hazards have
Argument been systematically identified and all Reasonably Practicable RRMs have been
(WRA) implemented – see Section 8.1 and 8.2. It may be supported by quantitative
arguments provided they are based on Robust Statistics, or modifications thereof,
and comply with the Royal Statistical Society, Practitioner Guides No’s. 1 to 4,
Guidance for Judges, Lawyers, Forensic Scientists and Expert Witnesses, and are
free of any of the errors in Appendices A and B of this document.
The document applies to engineering decisions • Explain the Systems approach to risk
for all Risks, minor or major, throughout the assessment.
lifecycle of any product, activity, process, or • Highlight some of the common
installation, including its design, construction, misconceptions about legal obligations.
commissioning, operation, modification, integrity
• Communicate legal precedent with respect to
management, decommissioning and disposal.
Reasonable Foreseeability.
NB. There are many examples of UK Legislation • Explain the arguments for and against
which are prescriptive as to the safeguards qualitative versus quantitative analytical
which must be taken, regardless of the techniques.
SFAIRP approach, for example the Pressure • Explain the criteria for probabilistic evidence to
Systems Safety Regulations, Water Supply be admissible in a UK law court.
(Water Fittings) Regulations, COSHH, to • Explain the common probabilistic error types.
name only a few. • Describe acceptable methods of
demonstrating gross disproportion.
However, whilst recognising that certain
• Describe how to build a Well-Reasoned
industries have dedicated legislation and have
Argument (WRA) to demonstrate legal
developed specific guidance for their types of
compliance.
Hazards and Risk Reduction Measures (RRMs),
this guidance is intended to establish the generic • Outline a risk management process which is
principles and encourage cross-pollination of logical, objective, contiguous and systematic.
best practice across all industries and products.
The document has been reviewed by lawyers,
There have been several developments since regulators, and safety experts from various
much of the existing guidance was written, which industries. It is primarily aimed at engineers,
has rendered some aspects misleading. These but may also be useful to managers, students,
include changes in the law, legal precedent, auditors, and accident investigators. Lawyers
evidential admissibility, lessons from accidents, wanting to know more about how engineers can
new analytical techniques, and the discovery manage Risk may also find it informative.
of fundamental errors in some existing
methodologies. The need for a single source of
updated guidance to help engineers comply with
the law, improve safety, and avoid unnecessary
costs and studies is therefore clear. The specific
objectives of this document are therefore to:
• Explain the risk management processes
involved.
• Explain lifecycle considerations and lessons
from process safety.
The principles and methodologies
discussed here could be regarded
as sound practice, regardless of a
country’s legislative regime.
Key Messages
The UK legal regime for health and safety is based on the Health And Safety At Work Act (HASAWA),
which places an obligation on the Legal Duty Holder to identify all Reasonably Foreseeable Hazards
and ensure, So Far As Is Reasonably Practicable (SFAIRP), that neither employees nor the public are
exposed to Risks to their health or safety. This is often expressed as an obligation to reduce Risks to
As Low As Reasonably Practicable (ALARP).
These legal obligations give discretion to the Legal Duty Holders, as to how they should best
minimise Risks. They cannot be delegated or transferred, but this does not mean that subordinates
or contractors are absolved of their responsibilities either (Section 2.6).
There are also prescriptive laws that apply to certain types of activity and industry. Furthermore, any
standards, guidance, Approved Codes of Practice may be regarded as Good Practice, which should
be complied with, unless a WRA has been provided to justify deviating from them.
The analysis must be Proportionate to the nature and complexity of the situation and the severity of
its effects if it is to identify all Reasonably Foreseeable Hazards.
The analysis must recognise the social and technical aspects of the System(s) involved, and this
may include materials, potential failure modes, activities, scenarios, escalations, human factors,
environmental aspects, and any other factor that might influence Risk (Section 6.0).
There is no legal requirement to quantify Risk, nor does a Risk figure demonstrate SFAIRP or ALARP.
‘Tolerable’ Risk is not a legal concept.
The argument for excluding a Risk Reduction Measure (RRM) on the grounds that it would be Grossly
Disproportionate would need to be convincing to a reasonable person (Section 2.3).
The inability of a Legal Duty Holder to pay for an RRM is not a justification.
The Reverse Burden of Proof puts the onus on the Legal Duty Holder to demonstrate that all
Reasonably Practicable RRMs have been implemented (Section 2.7). It is therefore prudent to record
all significant decisions with a WRA, regardless of whether a Safety Case is legally required.
Disclaimer: This Section is intended to provide engineers with a brief description of the key principles of UK goal setting
safety law, to assist with their day to day professional responsibilities. However, there may be exceptions, nuances,
special requirements and prescriptive laws that may apply to specific industries, activities, or products, that have not
been covered. Some case law examples have been summarised, to assist in this understanding but these cannot
constitute a rigorous explanation; only generalities that may require further research. Wherever uncertainty exists the
reader is strongly advised to seek legal advice.
ALARP has become common parlance to describe SFAIRP, (which may be because it simply rolls
of the tongue more easily). Although this document refers to SFAIRP and ALARP, both should be
interpreted as answering the same basic question, ‘Have all reasonably practicable RRMs been
implemented, in the Systems we design and build, to ensure that people are safe from the Hazards
to which they will be exposed?’
1 Also known as Common Law, as opposed to Statutory Law, which is made by parliament.
The following case law examples are helpful to clarify these issues:
Foreseeability with
Baker v Quantum Clothing Group, 2011.
respect to standards,
The court found the employer not liable because, at the time of the employees hearing
or generally available
damage, there was no recognised guidance on acceptable noise levels in the workplace,
understanding of,
so it was therefore deemed not Foreseeable.
the Risk:
The last example above, R v Tangerine & Veolia, Logically, any one of these questions is a valid
is particularly relevant because it imposes a duty determinant of whether the legal obligations have
to undertake appropriate studies to identify the been met.
Hazards and the means by which they could be
liberated, prevented or mitigated, i.e. establish In this context, a reasonable definition of ‘not
a detailed understanding of the technical and/ Foreseeable’ could be compliance with the
or human factor issues. An alternative way of following criteria:
stating this could be: ‘The analysis must be • It is not required by any standard, guidance,
Proportionate to the nature and complexity of the code of practice or accepted Good Practice.
situation and the severity of its effects if it is to
• Appropriate studies (such as HAZID, HAZOP,
identify all Reasonably Foreseeable Hazards’.
FMEA) would not identify the Hazard,
consequence, scenario, failure mode or Cause.
The guidance to the Offshore Safety Case
• There is no history of this or similar events.
Regulations states that it is Foreseeable that a
helicopter could crash into an offshore oil and • It is not plausible that it will occur in future
gas installation, but not an airliner. This implies (where plausibility is justified by a reasonable
that the airliner scenario is not plausible, rather Qualitative Argument, e.g. it can be
than not Foreseeable, as it has been foreseen. demonstrated that are several RRMs in place,
The premise here, is that a reasonable person which are sufficiently effective, diverse, reliable
would agree that, although the airliner scenario and redundant that total failure would not be
is possible, the combination of conditions that plausible), OR by a Quantitative Argument,
could lead to it may be dismissed as not plausible which is based on Robust Statistical evidence
or, in the accepted legal terminology, not of integrity, for that item, operating under
Foreseeable. The rationale would appear to be those conditions.
that there are enough natural or man-made RRMs
to make the collision implausible, e.g. aviation If an event is Reasonably Foreseeable, then the
law, pilot competence, the remote vicinity of an Legal Duty Holder must demonstrate that all Risk
offshore platform, radar and visual flight rules. controls have been implemented.
It therefore follows that any event that is provided
with enough suitable RRMs could be argued to be
not Foreseeable. This conclusion is supported by
Case Law, R v HTM (Section 2.3).
R v HTM, 2006.
HTM claimed that two employees who had been trained, coupled with warnings on the
machinery, nevertheless broke the rules. It was not Foreseeable that they could have done
anything more to prevent their deaths. The company were found not liable and this was
The relationship upheld by the court of appeal, who made the several points, including the following:
between • It was correct that evidence of foreseeability should be allowed because that evidence
Foreseeability was potentially relevant to the issue of reasonable practicability.
and Reasonable
Practicability • Foreseeability was merely a tool with which to assess the likelihood of a risk
eventuating.
• A defendant to a charge under sections 2, 3 or 4 of the HSWA, when asking a jury to
consider whether it had done everything that was reasonably practicable, could not be
prevented from bringing evidence as to the likelihood of risk occurring.
1. Expert witnesses must have appropriate 1. It should demonstrate that the analysis
competence in statistics if they are to give has been Proportionate (Section 3.1), i.e.
probabilistic evidence. This must therefore be undertaken in enough detail to understand
true for anyone providing such evidence. the Hazards and any relevant failure modes,
Causes, activities or scenarios that could
2. Methods of modifying statistical base
lead to an accident and how they may be
rates are limited to a form of deductive
prevented or mitigated.
Bayesian inference, discouraging the use
of mathematical formulae. The guides only 2. It should demonstrate a sufficient
refer to one stage of modification, (from prior understanding of the technical and social
probabilities, known as base rates, to inferred issues associated with the Risk.
posteriors), thereby indicating that multiple 3. If appropriate, it should show that any relevant
modifications, such as the complicated standards, approved codes of practice,
algorithms used in computer models, are too guidance and relevant Good Practice are not
complex to be admissible. applicable to the unique set of circumstances
3. Evidence must be relevant to the case, (which being considered.
is, in any case, a requirement of the courts).
2 The only known exception involving the use of ‘broadly acceptable’ comes from EU railways regulation:
COMMISSION IMPLEMENTING REGULATION (EU) No 402/2013 of 30 April 2013 on the common safety
method for risk evaluation and assessment and repealing Regulation (EC) No 352/2009, quote:
2.2.2. To focus the risk assessment efforts upon the most important risks, the hazards shall be classified
according to the estimated risk arising from them. Based on expert judgement, hazards associated with
a broadly acceptable risk need not be analysed further but shall be registered in the hazard record. Their
classification shall be justified in order to allow independent assessment by an assessment body.
2.2.3. As a criterion, risks resulting from hazards may be classified as broadly acceptable when the risk is so
small that it is not reasonable to implement any additional safety measure. The expert judgement shall
take into account that the contribution of all the broadly acceptable risks does not exceed a defined
proportion of the overall risk.
In practice 2.2.3 interprets ‘broadly acceptable’ in a similar manner to ALARP.
The identification and analysis of safety Risks needs to be as objective, systematic, and scientific as
practicable, only using judgement where it can be shown to be accurate enough.
The principles involved in demonstrating that all Reasonably Practicable Risk controls are in place
are essentially the same for all industries, although the methodologies may vary.
The Proportionality principle is key to determining the proper level of analysis for any Hazard.
Risk management applies throughout the lifecycle of the subject matter, from design to disposal.
The basic elements of the Risk management • Reducing the likelihood - There are many
process are summarised in Figure 1. Whilst the methodologies for analysing the problem to
elements of Figure 1 are applicable across all identify further RRMs. The objective of these
industries, the process can be relatively simple studies is generally to identify ways to prevent
and sequential, or complex and iterative, so the an incident by exposing inadequate or absent
figure indicates potential feedback loops where controls.
one stage may reveal new information relevant • Recording and communicating safety critical
to a previous one. However, the process will messages - It is important to document the
generally include the following elements: demonstration of ALARP, together with safety
• Determining Proportionality - For the goals and how they are to be achieved. This
analysis to be Proportionate, the worst also ensures that these lessons are not lost,
Reasonably Foreseeable consequences so that integrity can be maintained throughout
need to be established, together with a broad the product’s lifecycle.
understanding of the type of System and its • Maintaining ALARP throughout the Lifecycle
complexity. The Proportionality Matrix (Section - ALARP applies throughout the product
3.2) is then used to establish the Risk analysis lifecycle, so systems need to be in place to
strategy. ensure that changes, such as modifications,
• Identifying Hazards - This could be described wear and tear and aging do not compromise
as systematic brainstorming to establish the integrity.
the high level Hazards for further analysis. • Identifying RRMs - RRMs can be identified at
This stage may also identify Risk Reduction almost any of the above stages, as this is the
Measures (RRMs). fundamental purpose of risk management.
• Reducing consequences - This may be the
simplest and most effective Risk reduction
strategy if it is practicable. However, it
may require physical effects modelling to
understand the potential escalations and
effects on people. If the consequences can be
reduced significantly, or eliminated, there may
be no need for further analysis.
IDENTIFY HAZARDS
Systematic brainstorming, hazardous materials,
equipment and activities, Hierarchy of Controls, HAZID,
good practice, standards, accident history (Section 4)
Each Hazard will relate to a System, which A key principle in risk analysis is that of
comprises anything that can trigger or influence Proportionality, which dictates the amount
the likelihood or consequences of that Hazard of effort justified for a particular Hazard.
causing an accident over the lifecycle, including Unfortunately, there is little explanation of what
equipment, software, people, management this means or how it can be measured. Most
systems, procedures, activities, and natural guidance on this uses ambiguous wording, such
Causes. It may also be necessary to define the a ‘low risk’, without any explanation of what
System limits, or the design envelope. this means. HSE COMAH guidance states ‘The
depth depth of the analysis in the operator’s
The System may need to be considered over risk assessment should be proportionate to
some or all stages of its lifecycle, as follows: (a) the scale and nature of the major accident
• Construction/Manufacture Hazards (MAHs) presented by the establishment
and the installations and activities on it, and (b)
• Commissioning, set-up and/or trials
the risks posed to neighbouring populations
• Storage/Mothballing/Transportation and the environment’ i.e. the assessment has
• Modes of Operation to be site specific. ‘The depth of analysis that
• Maintenance and Inspection needs to be present depends on the level of
• Modification risk predicted before the identified measures
are applied.’ However, Risk cannot be known
• Dismantlement and Disposal
until the analysis is complete, and even then, it
may not be possible. Appendices A & B show
The relevant stages could be incorporated into
that Risk calculations and predictions may be
the hazard identification guidewords, or even
prone to elusive errors of thousands, millions or
have dedicated columns (see Table 2).
even billions of times, with no means of sense
checking or validating them.
Example 3.1 System Definition -
As stated in Section 2, the law courts are unlikely
The Grenfell Tower Disaster
to accept any Risk prediction without robust
This accident illustrated how the System evidence of its veracity, so an objective measure
boundaries can be set too tightly, as the true of Proportionality is necessary. This could
System was much more than the cladding include:
on the outside of the building that ultimately 1. The worst Reasonably Foreseeable
caught fire; it included its structure, its contents, consequences.
escape routes, ventilation systems, the
2. The potential for RRMs.
occupants, emergency procedures, drills, and
the emergency services, all of which influenced 3. The System complexity.
either the likelihood or consequences of the
outcome, so analysis of the cladding alone
would be incomplete.
A B C D
Single injury or 1
health effects.
Single fatality or 2
chronic health
effect, or multiple
injuries.
Multiple fatalities 3
or chronic health
effects.
NB. This bears similarity with Risk Matrices, policies, quality assurance, audit, review, change
which are discouraged for the reasons given in control, procedures, roles, responsibilities,
Section 6.12 and Appendix A6. competence requirements, and processes
for capturing lessons from experience and
The Proportionality Matrix therefore provides history. This is especially true for permissioned
the engineer with a more objective and legally industries that are obliged to produce a safety
defensible method for determining the depth report/case for acceptance by the Regulator.
of analysis than most guidance and/or risk
matrices would. It would be prudent to record One useful and well established framework for
the basis for rating the Hazards, especially a management system is the Plan, Do, Check,
within the amber zones. This also provides a Act (PDCA) principle, as shown in Figure 2,
basis for a more objective WRA. although other models may be as effective.
This can be applied at many levels, to test the
effectiveness of individual tasks through to
3.4 Engineering Management the high level policies of an organisation. It
Systems also illustrates the principles behind Figure
1, showing how the feedback loops facilitate
In order to demonstrate that all Reasonably iterative problem solving.
Practicable Risk controls are in place it may be
necessary for the whole engineering process The PDCA is most useful for determining how
to be governed by an appropriate safety well the key elements have been integrated
management system, which may comprise into the organisation, to ensure that it has the
Key Messages
A HAZID will normally be required for any significant Hazard, and this should draw upon guidewords,
experience and past incidents and accidents.
The HAZID should consider the whole System relevant to the Hazard, which may include hardware,
software, competence, procedures, emergency response etc.
For each Hazard, the HAZID will either conclude that the Risks are ALARP or that further study is
required.
Example 4.2 -
Brent Bravo Offshore Platform Fatalities
In 2003 two persons working inside a leg of the Brent Bravo platform in the North Sea were
engulfed by hydrocarbon gases from a leak. They were unaware of the anaesthetic effect of these
gases, and they collapsed and died. Hydrocarbons are highly flammable and potentially explosive,
so there were many RRMs to prevent ignition of the gas but, perhaps because of the severity of
these threats, no one had considered the anaesthetic Hazards. Nevertheless, the Hazard was well
understood by divers who work in the oil and gas industry. It is possible that a HAZID could have
identified the Hazard.
Guideword Hazard Worst Reasonably Causes Safety RRMs ALARP justification Action Party
Foreseeable Goals or action for further and date
Consequences and study
Escalations
Corrosion
Coating
Galvanic Surface Loss Intergranular Inspection
Failure
Stainless
Water Material UV Visual Procedures
bolting
Insulation
Oxygen Welding Abrasion NDT Frequency
Failure
Set-up/ Lifecycle
Activity Item Location Proximity Environment Function
Condition stage
Maintenance Lift Basement N/A Main road Day/Night N/A Sleeping
Cleaning Cladding Kitchens - Park Heatwave - Cooking
Removing Electricity
- Stairwell - - - Storage
refuse substation
4.3.3. The Degree of Guideword Expansion every item of equipment in the HAZID if
this could be done more effectively by
Whenever the Foreseeable consequences a few specialists in an FMEA. However,
of a hazard are significant enough, the scope of the FMEA may not be as
and there is potential to identify more broad as the HAZID, so care should also
RRMs, then further study or guideword be taken that scenarios are not getting
expansion may be justified. The level overlooked if this course of action is taken
of analysis should be Proportionate too early.
to the consequences of each Hazard,
some of which would be dealt with by a Nevertheless, the HAZID cannot be
single guideword, as in a basic HAZID, expected to identify all Causes, which
(e.g., electric shock or slips, trips, and may be too numerous and subtle.
falls, which would be unlikely to affect Example 4.8 is a good illustration of
more than one person), whereas more these difficulties and why further study
guidewords may be necessary for fire, may be necessary. The guidewords in
which may have the potential for multiple this case could have been ‘explosion’
fatalities. and ‘fuel tanks’, but this is too superficial
to identify such things as swarf from
For some Hazards more advanced drilling and riveting causing short-circuit
methodologies, such as FMEA, HAZOP between aging high and low voltage
or STPA, may be preferable, so it may cables which would ignite fuel vapours
be better to raise actions for them to (the industry subsequently changed the
be analysed that way. There is no point procedures for drilling and riveting on
in working through the failure modes of aircraft). Nevertheless, although drilling
Whilst a basic HAZID might employ In this incident a tank level alarm failed during filling and the tank
around eight specialists for a day or overflowed into a bund, which was designed to take the tank
two, a comprehensive one could take capacity. However, the incident happened on a Sunday, when no
these people away from their jobs for one was around, so it went unnoticed, the filling continued, and
significantly longer, which may not be the bund overflowed. A large gas cloud formed, resulting in an
practicable. Team brainstorming has explosion that damaged the plant offices, local businesses, and
come to be regarded as Good Practice, residential properties. Fortunately, there were no fatalities.
so any deviation from this may need
The criminal prosecution resulted in approximately £10 million in
to be justified. However, other
fines and costs. Civil liabilities amounted to around £700 million.
methodologies, such as STPA (Section
6.4), use a core team of two persons
If the subject matter is characterised by Automation may not be able to deal with
activities, it may be more appropriate to all situations, as it may need to be over-
substitute or supplement the HAZID with ridden with manual intervention in some
a Procedure HAZOP (Section 6.3.2) or situations.
STPA (Section 6.4).
These options may create a need for
4.3.4. Guidewords for Identifying RRMs and further study, as suggested in the
Actions for Further Study last column. For example, placing of a
guard on a machine may be all that is
Other than the RRMs from Good Practice Reasonably Practicable without significant
and standards, it will be necessary to effort or analysis, but if there are
consider bespoke or novel measures. Reasonably Foreseeable situations where
These may relate to prevention or the guard would not be effective, such as
mitigation, either reducing the likelihood during maintenance, then a more detailed
or consequences of the loss. Table causal analysis may be required.
4 categorises some of the potential
types of Risk reduction to illustrate how An RRM may also create unexpected new
guidewords could be developed for this Hazards or adversely affect other Risks,
purpose. The priorities should still be in so they should always be checked to
line with the Hierarchy of Controls (Figure make sure these have been identified and
3) but the assumption here is that inherent addressed.
safety could not be achieved, so more
avenues will need to be explored.
Error Recovery Control system design, built in time to TRA/Procedure HAZID (6.3), STPA (6.4),
Human error
correct/recover. HE/HFA (6.2)
Consequences Quantification of immediate health/
Initial Severity Consequence analysis (3.0)
mortality effects.
Key Messages
A conservative judgement of the Reasonably Foreseeable consequences, which may have been
made during the HAZID, might not always be sufficient to identify suitable RRMs.
The simplest means of mitigating a Risk may The Reasonably Foreseeable consequences will
be to address the consequences, as the not necessarily be the worst possible (Section
Causes may be manifold and relatively elusive. 2.2), provided a WRA can be made to dismiss
In some cases, where the Hazards are known, the latter on the grounds of sufficiently effective
such as fire, this could be the starting point of and reliable RRMs or Robust Statistics (Appendix
the evaluation. A2) that the likelihood of the conditions required
for it to occur would not be plausible. The
The consequences may have been evaluated consequences may also be evaluated for each
conservatively using professional judgement Foreseeable Failure Mode, if applicable, e.g. fires
during the HAZID process, but it may be due to a small corrosion leak and a pipe rupture
necessary to undertake simulation or testing may have quite different consequences.
to understand its severity and whether it could
escalate to something more significant. The validity of potential escalations may need to
be checked with empirical testing, calculation,
The Reasonably Foreseeable consequences or simulation (typically Physical Effects
could be a function of the following factors: Models, which may be utilise Computational
The nature of the Hazard or loss of control. Fluid Dynamics). CFD models are popular for
• The Failure Mode, which may influence the simulating fires, explosions, and the dispersion
extent of the Hazard. of toxic or flammable gases. Where escalations
are Reasonably Foreseeable, the consequences
• Whether and how it could escalate to a larger
could increase substantially, so consequence
or different Hazard.
reduction measures should be considered to
• Its effects on health, injury, or mortality, prevent these or mitigate them.
including who will be affected and how many,
e.g. constructors/manufacturers, operators/
users, maintainers, other staff, customers,
public and emergency responders.
The Reasonably Foreseeable consequences
will not necessarily be the worst possible.
Key Messages
The analytical strategy needs to consider the available methodologies, their strengths, and
weaknesses, and optimise or modify them to suit the subject matter.
Assumptions and omissions should be eliminated wherever possible, or minimised and stated.
Predictive methods, such as QRA and risk matrices are not recommended analytical techniques.
If the previous stages have not demonstrated systematically assess all possible System
that all Reasonably Practicable RRMs are conditions to identify Hazardous scenarios,
in place using simple solutions, then it may such as HAZOP.
be necessary to reduce their likelihood by
identifying further Causes or flaws in RRMs. A common framework for analysis is to
Accidents and their outcomes are either the describe loss as a linear chain of events which
result of component failure or interactions are independent of each other, such as Event
between the elements of a System, so each and Fault Tree Analyses, Functional Safety
of these may require a different analytical Assessments, Layers of Protection Analysis
approach. History shows that virtually all or Bow Ties. Whilst these may be effective for
accidents have multiple Causes, and typically component failures, they may otherwise be an
involve sociotechnical Systems, which can be oversimplification, because each event may
influenced by such things as actions, omissions, be influenced by the others or by separate
latent conditions, human factors, ergonomics, common factors, for which more complex
unclear objectives, errors in the design, models are appropriate.
software or procedures, environmental factors
and any vulnerabilities or limitations in the RRMs. Recent developments, such as STPA, view the
It is therefore rarely possible to attribute a risk to System as hierarchical control frameworks
a single Cause. The analysis may therefore need with feedback loops, which may be more
to identify multiple Causes that could influence effective at identifying interactions, especially
the likelihood of any Hazard becoming a loss. in sociotechnical Systems. Any System has
goals and constraints, which may need to be
The strategy for analysing the Causes may clearly identified. Processes, such as Goal
vary significantly. Some methodologies Setting Notation may help to understand the
take a blanket approach, such as human objectives, define the constraints, and identify
factors checklists or management system their weaknesses. The framework for analysing
reviews, to identify anything that might have the Causes will therefore be a major influence on
a causal relationship. Alternatively, others will the effectiveness of any study.
Part of Action Extra Action Other Action Out of Sequence More Time
Abnormal
Clarity Training Maintenance Ergonomics
Conditions
STPA (6) views the System as a set of functions, Policy, Rules, Procedures
rather than equipment items. Many accident
models, such as FTA, FMEA, ETA and Bow Ties,
assume accidents to be caused by a linear
succession of discrete, equipment failure-based Human Controller
events, which omit any feedback or interactional
aspects. This limits the causal detail that is Mental Model (beliefs)
practical with these methodologies and omits
human error, regulatory and management
constraints on the System. STPA requires a
different mindset, as it treats safety as a control
Programmed Controller
problem in which accidents arise from complex
processes that may operate concurrently
Process Model Control Algorithm
and interact to create unsafe situations. It
therefore has the benefit that human error and
management systems can be addressed by a
bottom-up approach, rather than by the more
generic methodologies discussed elsewhere in
Monitoring/Feedback Signal/
this section.
for Hazards, Failures and Comms/
System Status Actuators
All accidents involve lack of control over the
System Hazards, so the control view of safety
defines a safe System as one that enforces
safety constraints on the behaviour of the
System. STPA is the Hazard analysis arm External Conditions and Controlled
of System Theoretic Accident Model and Other Hazards Activity(ies) or Process
Processes (STAMP), which was created to find
more Causes, including social, organisational,
Figure 5: Example STPA Control Structure
human error, design and requirements flaws,
and dysfunctional interactions among non-failed
components. It originated in the aerospace
industries to review software systems, has and actions. The hierarchy can be extended up
become widely used in autonomous vehicles to regulatory and governmental influences if
and rail but is now finding wider applications, necessary.
such as petrochemical and medical.
The STPA process comprises four stages:
By breaking down any System functionality
1. Define the purpose of the analysis.
into decisions, communications, actions, and
feedback, it facilitates their representation in 2. Model the control structure.
schematic form, like an instrumented control 3. Identify Unsafe Control Actions.
loop, as shown in Figure 5. This enables the 4. Identify the loss scenarios.
hardware and sociotechnical aspects of the
System, such as equipment items, human error,
STPA has a wide range of applications and may
management systems and regulatory controls,
to be presented diagrammatically. be the most comprehensive risk management
process currently available, as is evident from
The structure may vary from a single control Table 4 (Section 4.3.4.). Example 6.6 below
loop to multiple loops representing different shows how STPA can be used in a nuclear power
equipment items, feedback, and control plant. The methodology is detailed in the STPA
paths for various data/commands, as well as Handbook (7).
automated or manually controlled decisions
Losses:
L1: Loss of life/injury, L2: Asset damage, L3 Environmental, L3 Loss of production
Control Structure:
Pump 1
on/off Digital Control System
Vessel with
Exothermic
Pump 2
Reaction
Cooler
Solution:
Exothermic reaction needs to be controllable with one pump or two. Recalibrate flow transmitters to cope with
both pumps. However, the effects of doing so would need to be checked to ensure that this does not introduce
another problem to production or safety, e.g. sensitivity reduction effects performance monitoring or detection
of leaks.
However, it may also be used as a method of 6.9 Functional Safety Analysis (FSA)
brainstorming causal mechanisms by working
backwards from the Hazardous situation to the FSAs are intended to determine the appropriate
scenarios and routes that may lead up to it. reliabilities for Safety Instrumented Functions
(SIFs). The resulting Safety Integrity Levels (SILs)
FTA can suffer similar problems to ETA, determine the Risk reduction, as described in BS
especially for Systems involving manual input. EN 61508, which is supported by BS EN 61511
for the process industry, 61513 for nuclear,
6.7 Event Tree Analysis (ETA) 62061 for machinery and 26262 for automotive.
These may be regarded as Good Practice with
These are useful for identifying the possible certain caveats.
courses of development/escalation of a
Hazardous situation with many potential There are two assumptions that need to be
outcomes. In the same way that FTA looks at considered, namely:
the events leading up to a Hazardous condition, • that there is a tolerable risk target that needs
ETA plots the events thereafter. Although it was to be met and
originally developed for probabilistic analysis, it • where relevant, the unmitigated risk is either
may be used as a qualitative analytical tool or as known or can be calculated.
a graphical method of presenting the full set of
potential outcomes.
The main problems with ETA are to overlook
Consequences
unrealistic judgements.
Hazards/Causes
Event
Bow Ties have become an increasingly popular
Emergency Response,
means of graphically representing Hazards
Escalation Prevention
and their RRMs, as shown in Figure 6. They are
Consequence
by James Reason (10) and shown in Figure 7.
Mitigation
The cheeses represent the RRMs, which may
Prevention/
reduce the likelihood or consequence of an
Mitigation
event. The holes in the cheeses represent flaws
or weaknesses in the RRMs, which can lead to
an accident when they align. One problem with Figure 6: The Bow Tie
the Bow Tie is that it shows the cheeses but not
the holes, which may create an unduly optimistic
Hazard
picture of Risk reduction.
This accident was also discussed in Example 6.4 (Section 6.1), where it was shown that most of the problems
contravene the criteria involved in an RRM Effectiveness Review.
However, a Bow Tie would have presented an extremely optimistic view of the water washing Risks because it
would inevitably show many barriers/RRMs without showing their effectiveness, common mode failures and any
other factors discussed in Section 6.1.
Isolation Valve(s)
Nitrogen Blanket
Instrumentation
Competence
Water Spray
Procedures
Tank Ullage
Evacuation
Slip Flange
Scrubber
Alarms
Flare
Top
Event
Nevertheless, a set of Bow Ties could provide a starting point for the RRM Effectiveness Review, if there is one for
each identified scenario.
Hazard: A B C D E F G H J K
Modify
existing
System
Build
new
System
6.14 Specialist Materials Studies and FMEA. It will be critical to identify any
weaknesses in RRMs, so an RRM Effectiveness
Failure of materials may be a significant Cause Review may be necessary, especially for non-
of accidents, which requires detailed analysis instrumented systems.
by specialists or rigorous testing in all relevant
environments and load conditions, especially On the other hand, for some high consequence
when new materials are introduced. Factors such events there may be no practical mitigation
as corrosion, fatigue, creep, flammability, toxicity, RRMs, such as aircraft crashes, so much greater
ultraviolet deterioration, impact, fretting, extreme effort will be necessary to identify all Reasonably
temperature operation, chemical reactions may Foreseeable Causes. This may involve a
combination of TRA, Procedure Reviews, HAZOP,
need to be tested, sometimes in combinations.
FTA, FMEA/FMECA, FSA, HFA, STPA, or any other
industry specific methodologies. Complex and/
or sociotechnical Systems with internal and/or
6.15 Conclusions external interfaces or interdependencies and
significant consequences may necessitate STPA.
Each of the above methodologies has strengths
and weaknesses and may only work in certain Any methodology should be regarded as a
contexts. The best approach to selecting the flexible process that can be modified to suit
types and degrees of study required should be the subject matter. The ideal approach may be
based on the Proportionality Matrix (Section a combination of different ones, employing the
3.3) and the Potential RRM Categories (Table 5, best aspects of each. This is especially true for
Section 4.3), together with the pros and cons HAZID, TRA and tailor made matrices, which
described in this section. can be effective when designed to focus on
recognisable characteristics of the problems to
In some cases, the Causes are reasonably be solved.
generic, at least at a high level, such as in fires
in buildings. The HAZID may identify prevention Although this section is not a comprehensive list
measures, like the use of non-flammable of methodologies, it is intended to illustrate the
materials, but the Causes may be too numerous principles involved and how they may be applied.
to be identified, so significant effort may The preparation for any analysis may therefore be
be required to identify and deliver effective key to the success of the work. This may require
mitigation RRMs to prevent escalations and experience and some imagination but, based
serious consequences. These may be physical on the principles set out, it should be possible
RRMs or single function instrumented systems, to create bespoke methodologies that are both
which explains the use of processes like FSA cost effective and produce better results.
Key Messages
Risk control throughout the lifecycle requires certain criteria to be established at the early stages of
feasibility and design and maintained throughout the lifecycle.
These criteria need to be formulated and communicated in a manner that ensures the necessary
actions will be taken throughout the lifecycle.
It will be necessary to demonstrate that any Safety Critical element would perform its function(s)
effectively.
This demonstration may be based on trials, standards, certification, engineering calculations, and/or
a Well-Reasoned Argument.
An ALARP demonstration is not complete until the Hazards and their controls are effectively
communicated to all stakeholders, whether management, workforce, users and/or the public.
This may take many forms, including safety cases, Safety Critical records, training, warnings, and
competency requirements.
For some installations, buildings or products there may be a legal requirement to provide a formal
safety report/case, for acceptance by the Regulator.
A safety report/case should comprise a WRA for each System to communicate in a clear,
comprehensive, and defensible manner that the Risks will be ALARP throughout the lifecycle.
(NB. A safety case should also include certain operational and managerial issues which may
not be fully covered in this document.)
To be effective, the WRA may need to be developed throughout the risk management process, by
an accountable individual, who is supported by a team of the relevant disciplines. This may reveal
flaws or omissions in the analysis, which may need correction or further study work.
Regardless of the safety case obligations, appropriate WRAs and Safety Critical data should be
recorded and maintained throughout the lifecycle to satisfy the legal obligations.
Given the Reverse Burden of Proof (Section 2.7) • Training and competency.
it will be prudent to document all decisions that • Procedures (applying to Safety Critical
could influence the Risks. The documentation activities throughout the lifecycle).
can take many forms, ranging from studies, • Hazard Register
reports, checklists (Appendix C), ALARP
• Requirements for activity control and risk
Worksheet (Table 7), through to full Safety
assessment,
Reports/Cases (Table 8). It should provide a
e.g. permit systems.
fully auditable record of the risk management
process for each Hazard, summarising the • Lifecycle Criteria included in relevant
conclusions and how they were reached in an documentation,
accessible style, i.e. demonstrate that every e.g. performance standards.
Foreseeable Hazard has been identified and • Safety philosophies.
subjected to suitable and sufficient analysis • WRAs for Systems.
and that all reasonable measures to mitigate • A safety report/case.
the Risks have been implemented. Some
justification for the type and level of analysis
adopted for each Hazard may also be necessary. 8.1 Building a WRA
Appendix C suggests a framework for a checklist The basis of any safety demonstration should
(which is a summary of the key points throughout be a WRA, which may be necessary for each
this document) and may act as a final check, or System and Hazard. Its content should be as
for legal compliance, audit, and review purposes. factual, qualitative, and concise as practicable,
employing graphical, diagrammatic, and tabular
Effective communication of safety messages representations wherever desirable. It should
may require diverse means, such as: preferably be based on recognised methods
• Publicity campaigns. such as Claims, Argument, Evidence and/or Goal
• Localised warning signs. Structured Notation (Section 8.2).
• What are the safety objectives, and Relevant Good Practice, Standards, ACoPs and
Guidance
have they been effectively recorded and
communicated? Assumptions and Uncertainties
• If an RRM has been rejected, what is Assessment
Options considered
the justification, and does it fulfil the
requirements stated in Section 2.3? Detailed Safety Objective(s)
In its simplest form, the WRA for each Hazard Option rejections and reasons (e.g. Gross
may be captured in an ALARP proforma, Disproportion, inapplicability)
as in Table 8, which is, in any case, a good
Hardware control measures
communications tool for summarising the
Systems and their RRMs. The proformas can be Risk
Software controls (competence, procedural,
Reduction
contained in a single document, or incorporated management systems)
Measures
into the safety report/case, so that it is available
Lifecycle Criteria (where appropriate)
for engineers, maintenance and inspection
personnel, operators, and regulators. Table 8: Suggested ALARP Communications Proforma
Problem: Identify which, if any, tunnel systems are not reasonably practicable.
The following is an outline of how to develop a WRA to select a tunnel system and demonstrate whether specific
designs would be grossly disproportionate.
RRMs:
• Options of five tunnel systems, as shown in the diagram.
• Allowable train speeds. Single Double
• Train and track type Track Tunnel
• Frequency of trains. plus service tunnel
• Tunnel firefighting systems.
• Rescue provision.
• Remote monitoring.
• Operational controls, e.g. opposing train limits. Double Single
Track Tunnel
Example variables:
• Train speed.
• Train design (safety standards). Double Single
• Size, straightness, and length of tunnel. Track Tunnel
• Number and type of carriages. plus service tunnel
• Track type.
• Diesel or electric and other flammable materials.
Double Single
• Passengers or cargo.
Track Tunnel
• Train frequencies, especially passing.
plus service tunnel
• Debris potential.
• Fire-fighting, ventilation and escape.
• Rescue potential.
It is immediately apparent that there are many variables that can affect both the Causes and consequences of
accidents in the tunnel. Whilst there may be some statistical data, this would be too limited to reflect all these
variables.
The consequences for each Hazard could vary enormously depending on the tunnel system and many of the
variables listed. This may involve significant judgement, as there may not be enough evidence to model specific
crash scenarios, or their relationship with the above variables. The constraints on the system will be key to the
Risks, e.g. train speed limits, oncoming trains, train types and safety standards. These constraints may also be
different for each tunnel system.
The Causes could also be manifold, e.g. potential for debris, maintenance tasks, train speeds, flammable materials.
The analyses of these would need to be rigorous enough to demonstrate that all Reasonably Foreseeable Causes
had been identified.
The final decision would need to compare the remaining differences in the Causes and consequences to the
costs for each tunnel system. If the former are not significant but the costs are, then an argument can be
made that the more expensive options do not provide sufficient benefits to warrant their adoption, i.e. such an
argument would necessarily include Causes and consequences, and should make sense to a ‘reasonable person’
as well as the other criteria for Gross Disproportion (Section 2.3).
Problem: Identify and evaluate RRMs for fires and explosions on the platform.
The following summarises the key arguments used in the WRA (some aspects, such as safe refuges, alternative
muster locations, lifeboats, and escape equipment have been omitted).
Hazards:
H1. Gas jet fire radiation affecting personnel at the time of ignition (escape to safe place).
H2. Gas jet fire radiation preventing rescue of injured parties.
H3. Gas jet fire escalating to other hydrocarbon equipment.
H4. Explosions due to delayed ignition or extinguishing and re-igniting gas cloud.
H5. Smoke impairment of muster locations.
H6. Liquid hydrocarbon fires in drip trays under pumps escalating to equipment above.
H7. Heli-fuel fire on helideck preventing rescue of passengers after crash on the helideck.
H8. Heli-fuel fire on helideck preventing rescue of passenger after spillage on the deck.
RRMs:
RRM A. Fire pump(s) with deluge and/or fire monitors for jet fires.
RRM B. Foam system for helideck.
RRM C. Fire extinguishers.
RRM D. Emergency Shut Down (ESD) and Depressurisation (EDP) systems.
Radiation fall off follows an inverse square law. Three zones around flame i) not survivable ii) survivable for limited
time iii) not a threat. Survivable zone is a thin annulus, approximately 2m thick. Deluge has little effect on the
radius or thickness of the survivable annulus and may fill it with superheated steam, thus increasing the Risks to
individuals.
If rescue teams require deluge protection, then victim would have succumbed before arrival.
It cannot prevent escalations as water cannot penetrate the flame envelope and it cannot achieve 100%
coverage of target equipment.
Testing causes corrosion of pressurised equipment, further increasing Risks and maintenance.
It causes significant maintenance and testing, which requires extra manning on platform, thus increasing the
number of people exposed to all offshore Risks.
It cannot extinguish a jet fire and would create potential for explosions if it did.
Conclusions:
A pumped fire water system with deluge and monitors was rejected. Fire extinguishers were incorporated
into procedures and permit system for work on liquids systems. A one shot foam system was adopted for the
helideck.
Although fire-fighting systems on offshore installations were considered to be Good Practice, they were shown to
be not applicable in this case.
This can be likened to the claims, argument, Mitigate gas jet All Process
fire effects Releases
evidence approach discussed in Section 8.1.
This is especially suitable for control Systems
where options are binary, but the principles are
sound for any demonstration that all reasonably
practicable RRMs have been identified. The Prevent
Prevent Enable Enable
escalations
System must first be defined in terms of structural escape rescue of
to process
objectives, or goals, each of which must then be collapse from area survivors
equipment
justified by structured arguments.
Emergency Facilities (e.g. muster points, fire and rescue) and procedures.
Response External support and communications.
Table 9: Suggested Structure and Content for a Major Accident Safety Case/Report
A demonstration of ALARP is not a one off exercise but continues throughout the lifecycle.
A formalised Management of Change (MoC) process may need to be established. This may also
need to be backed up by a formalised revalidation process where complex systems may change
for reasons beyond the control of the MoC, e.g. COMAH 5 year revalidation of Safety Report.
A formalised means of monitoring creeping change is recommended, whether by audit, review and/
or inspection or monitoring leading indicators.
The management system should specify accountabilities, roles and responsibilities and
competence for those undertaking this activity.
It is recommended that contingencies are set out for Reasonably Foreseeable outages and
changes of status, to avoid shutting down Systems or product recalls.
UK Worldwide
The CCHAZID (14) is a variant of the standard • Change of use, additional uses, process
HAZID technique (Section 4), with a similar changes.
structure and process but different guidewords, • Hazardous materials and environmental
to help identify new or increased Risks occurring changes.
over time.
• Equipment or infrastructure changes.
(e.g. electrical, mechanical, instrumentation,
The primary guidewords in this case could
relate to: structural and process).
SCE #2 R R G G G R
SCE #4 R R Rule C R G G
Other Pitfalls:
1. Baker, J. The Report of the BP U.S. 13. Richard J. Goff and Justin Holroyd,
Refineries Independent Safety Review UK Health and Safety Laboratory.
Panel. 2007. Development of a Creeping Change
HAZID Methodology. IChemE. [Online]
2. Health and Safety Executive. Reducing
2017. https://www.icheme.org/
Risks, Protecting People (R2P2). 2001.
media/11897/paper-61.pdf.
ISBN 0 7176 2151 0.
14. Health and Safety Executive. Good
3. Royal Statistical Society. Practitioner
Practice and pitfalls in risk assessment,
Guides No’s. 1 to 4, Guidance for Judges,
RR151. 2003.
Lawyers, Forensic Scientists and Expert
Witnesses, Royal Statistical Society. s.l. : 15. —. HSG238, Out of control: Why control
Royal Society of Statistics, 2009 to 2014. systems go wrong and how to prevent
failure, 2003.
4. Perrow, C. Normal Accidents: Living
with High Risk Technologies. s.l. : ISBN: 16. Confidential Enquiry into Sudden Death
9780691004129, 1984. in Infancy” (or “CESDI”), entitled “Sudden
Unexpected Deaths in Infancy. s.l. : BMJ.
5. Health and Safety Executive. HSG65,
Managing for Health and Safety, 3rd 17. Hill, Pr. R. Cot Death or Murder? -
Edition. [Online] 2013. https://www.hse. Weighing the Probabilities. s.l. : Salford
gov.uk/pubns/books/HSG65.htm. University, 2002.
6. Leveson, N. Engineering a Safer World: 18. The fabrication of facts: The lure of the
Systems Thinking Applied to Safety. 2011. incredible coincidence. Derksen, Ton. s.l. :
ISBN: 9780471846802. Neuroreport, 2009.
7. Leveson, N., Thomas J. https:// 19. Kahneman, Daniel. Thinking Fast and Slow.
psas.scripts.mit.edu/home/get_file. 2011.
php?name=STPA_handbook.pdf. STPA
20. Tetlock, P. E. and Gardner, D.
Handbook. [Online] 2018.
Superforecasting: The Art and Science of
8. Kurt Lauridsen, Igor Kozine, Frank Prediction. 2015.
Markert, Aniello Amendola, Michalis
21. Robson, D. The Intelligence Trap: Why
Cristou, Monica Fiori. Assessment of
Smart People Do Stupid Things and How
Uncertianties in Risk Analysis of Chemical
to Make Wiser Decisions. s.l. : Hodder &
Establishments. Roskilde : Riso National
Stoughton, 2019.
Laboratory, 2002. Riso-R-1344(EN).
22. Cox Jr., L.A. What’s Wrong with Risk
9. Reason, J.,. Managing the Risks of
Matrices? 2008.
Organizational Accidents. Aldershot, UK :
Ashgate, 1997. 23. Thomas, P., Bratvold, RB, Eric Bickel JR.
The Risk of Using Risk Matrices. 2013.
10. Health and Safety Executive. A Review
of Layers of Protection Analysis (LOPA) 24. Miller, K. Quantifying Risk and How It All
analyses of overfill of fuel storage tanks, Goes Wrong,. s.l. : IChemE, 2018.
RR716. [Online] 2009. https://www.hse.
25. Health and Safety Executive. RR672
gov.uk/research/rrhtm/rr716.htm.
Offshore hydrocarbon releases 2001 to
11. Tinsley C. H., Dillon R. L., Madsen P. M. 2008. s.l. : HSE Books, 2008.
How to Avoid Catastrophe. [Online] 2011.
26. Ashwanden, C. You Can’t Trust What You
https://hbr.org/2011/04/how-to-avoid-
Read About Nutrition. FiveThirtyEight.
catastrophe.
[Online] 2016. FiveThirtyEight.com.
12. The Assurance Working Group. The GSN
27. Reason, Pr. James. The Human
Community Standard . [Online] 2018.
Contribution.
https://scsc.uk/r141B:1?t=1.
28. J., Thomas. 2020. 2020 MIT STAMP
Workshop.
Measured Risk is a statistic, an average, which may not be indicative of the Risks in a unique System
and may not provide any indication of why the Risk exists. Robust Statistics are only possible for
mass produced items tested under identical conditions for sufficient time.
Calculated Risks must be based on logic or Robust Statistics that are modified in accordance with
Bayesian theory. This is rarely possible for engineered or sociotechnical systems (Appendix B6).
Predicted Risk is an epistemic belief, a knowledge related opinion, but accident history shows that
the unknowns can be as important as the knowns, if not more so. It may be undermined by multiple
cognitive biases, conflicts of interest, errors due to poor understanding of probability theory and
counter-intuitive mathematical relationships
Predicted Risk is not real, it is an opinion, generally unique to an individual, which may be subject to
almost unlimited errors that can be extraordinarily difficult to identify, even by experts.
Risk assumes randomness, which is an absence of control, either due to inability, ignorance, or
choice. Ignorance cannot be quantified. Choice must be justified. Accidents are due to ineffective
or absent controls, which cannot be quantified but may be rectified.
The Risks associated with engineered and sociotechnical Systems are nuanced with potentially
chaotic aspects. Predictions, based on experience, appearance, or comparison with similar
Systems, may therefore be highly deceptive. ‘Expert judgement’ or ‘sound engineering judgement’
can only relate to Foreseeability, not to Risk quantification.
Accidents only happen because someone thinks the Risk is low, so dismissing low Risks is illogical.
This appendix deals with the legal definition A1 Legal Definition of Risk and
of Risk and the feasibility of measuring, Risk Assessment
calculating, or predicting it for the purposes of
demonstrating whether it has been reduced to The concept of risk was described in the case
ALARP. of Regina V Board Of Trustees Of The Science
Museum, 1993 as ‘a possibility of danger’.
A controlled study by the Risø National Institute (9) employed six specialist consultancies to assess
the Risks on a chemical installation. Figure A1 shows the results, with three of the eleven equipment
items having more than four orders of magnitude Risk variation, which is 10 times larger than the
ALARP region for workers, and 100 times larger than that for the public (2).
10-2
Worker Intolerable
10-3
Public Intolerable
10-4
10-5
Risk/Year
Broadly Acceptable
10-6
10-7
10-8
It should be noted that this trial measured variation not error (which could be larger). The report
concluded that the differences were necessary assumptions caused by ‘analysts guessing’. The
process is therefore random over a potential range of at least four orders of magnitude.
Because the Risk predictions for rare events are unverifiable errors cannot be known, but they
could be even larger.
The law courts provide the best evidence of how large the errors could be, because some gross
miscarriages of justice have warranted academic challenge to the figures. The Sally Clark trial (17)
(18) infamously had a critical expert witness error of one billion times (Appendix B6, Error #6) and
the case of Lucia de Berk in the Dutch courts (19) erred by ten orders of magnitude. Both errors
resulted in life sentences for the defendants, who were later released when this came to light. The
Sally Clark case was probably the most researched example but nevertheless had a UCC error
(Appendix B7, Error #7) of one hundred times that went unnoticed even by the academics.
Cause #1 – The fuel rods had graphite tips, which made the reaction unstable at low levels of power.
Cause #2 – The operators did not understand this.
Cause #3 – The test procedures contained ambiguities and omissions.
Cause #3 – The test required standard operating criteria to be contravened.
Cause #4 – The test was rushed because of pressures to get it done.
Cause #5 – The run down test was to prove that power to the pumps could be maintained during
a reactor shut down, but this should have been proved by design and safe simulations during
commissioning.
These are just a few of the many reasons that resulted in the disaster, but they illustrate how
changing any one of them could have either prevented the accident from happening, or
at least changed the probability enormously.
Any one of these factors can become dominant, because the Risks are epistemic and partly
or combinations of them could change the controllable, and may be reduced to ALARP,
probability from negligible to highly likely, or but no one could have sufficient knowledge to
even certain. This is too complex to model estimate them in any meaningful way.
probabilistically or to make mental estimates of
their likelihood. The Chernobyl disaster (Example The tossing of a coin, or the rolling of a die, are
A4.4) illustrates these principles quite well. therefore deceptive examples of Risk because
they have unique problem characteristics – i)
Chernobyl is one example of these sensitivities, they are based on a logical argument (that
but similar conclusions could be drawn for of symmetry) and ii) they involve a complete
virtually all major accidents. The level of detail inability to control the outcome. Engineering
required to meet the knowledge requirements problems are inevitably much more complex,
above, would therefore be impracticable, making with multiple variables, which may often be
any meaningful predictions totally unrealistic. controlled, if only partially. There is no logical
argument for Risk, as was evident for the coin.
This illustrates how the unknowns can become The Russian Roulette example shows how the
more important than the knowns and, even if introduction of more variables immediately
they are known, it may be impossible to evaluate makes the problem insoluble.
or quantify them in any meaningful way.
Accident inquiries rarely, if ever, state that the
Engineers can sense check most engineering event was random, as there was inevitably good
calculations within an order of magnitude (e.g. reason for its occurrence. Example A4.6 lists
a pump or a bridge that is either ten times 30 well-known major accidents, of which only
too large or small would be apparent to a one could be considered entirely random, the
competent engineer). This may be the reason Space Shuttle Columbia, which was doomed
why they believe that they can do the same the moment the tile fell off and hit the wing.
with Risk, because they fail to recognise the (NB. Some would argue that even the Shuttle
different nature and sensitivities inherent in the accident was not random because there may
probabilistic domain (20), (21). have been more that NASA could have done to
prevent it.) The other accidents were initiated
The sensitivities can become almost unlimited, by identifiable errors in design or operation,
as illustrated in Example A4.5. – Russian most of which could have been rectified with an
Roulette. This example is analogous to many effective Risk management process.
industrial applications of ALARP legislation
Russian Roulette illustrates some of the key points here because it has epistemic, controllable
and random aspects, which are analogous to many industrial situations. The Risk of the gun firing
is governed by three variables:
i) weight of the bullet
ii) friction of the cylinder
iii) orientation of the gun, (which may either increase or decrease the Risk).
A simple logical argument might predict the Risk as 1/6 because there is one bullet in six
chambers, but this is incorrect because the weight of the bullet tends to make the cylinder stop
at the bottom of the spin. In Bayesian terms, the 1/6 is the prior probability, which ignores the
weight effect, whereas the actual probability is known as the posterior. It may not be reasonably
practicable to assess the effects of friction, so the posterior probability would be incalculable.
If friction is the dominant factor, the weight of the bullet could be a minor influence only. On the
other hand, a frictionless cylinder would make the weight of the bullet dominant, and therefore
almost inconceivable that it would stop at the top. Depending on the orientation of the gun, the
range of possible probabilities is virtually infinite and not something that anyone could judge,
unless they had Robust Statistical data from trials under identical conditions. In the absence of
these, the posterior probability cannot be known, and it would clearly be unacceptable to quote
the prior. Any attempt to assess the posterior probability would be nothing more than guesswork,
which would not be legally admissible evidence.
If the gun’s orientation were known, it may be possible to say whether the probability is greater or
less than 1/6, but nothing more. If the gun’s orientation is fixed, then the Risks are uncontrollable
and random, but with no known probability. If it is not fixed, and the operator can choose the
orientation, then it is at least partially controllable, because holding the gun upright (with the
barrel at the top of the spin) will reduce the probability, thereby achieving ALARP, even though the
benefit cannot be quantified. The demonstration of ALARP can therefore only be based on
a WRA.
UK Worldwide
Ladbroke
Aberfan Flixborough Bhopal Guadalajara Seveso
Grove
Clapham
Hillsborough Titanic Feyzin Mumbai High Texas City
Junction
Three Mile
Comet Airliner Kegworth Windscale Fukushima Pasadena
Island
Virtually all major accidents happen because The above limitations are only made worse
the Risk is underestimated at some point in by the many cognitive biases that affect Risk
the product lifecycle, i.e. a failure to recognise estimation (21), some of which are:
either the Hazard, a flaw in its controls, or the
likelihood of failure (12). 1. Confirmation Bias (aka Wilful Blindness)
(20)
The difference between the ability to judge - Probably the best known cognitive bias, as
plausibility (Foreseeability) and probability, virtually everyone tends to select evidence
is recognised in the RSS legal guidance (3). that favours their own beliefs and disregard
Although an expert may be able to judge evidence to the contrary.
Foreseeability, probability is a different matter,
e.g. a metallurgist may be able to say whether 2. Availability Heuristic (aka Outcome Bias)
a fatigue crack is Foreseeable, but he/she (20)
does not have the competence to judge the - Personal experience dominates judgement,
probability and, in the absence of Robust although it may differ from reality.
Statistical data on that item operating under
those precise conditions, any judgement would 3. Anchoring (20)
have no sound basis. Expertise has boundaries,
- Where any initial suggestion of likelihood
which do not encompass probabilistic
is liable to influence the final assessment.
quantification.
This can occur in QRA, where the results are
compared to previous studies, even though
The primary objective in risk management is to
those studies may be flawed and cannot be
identify controls and determine whether they are
verified.
effective, not to dismiss the Hazard as random.
Prediction attempts to quantify unknowns,
4. Bounded Rationality (20)
although identifying them should enable them
to be addressed. If analysis concludes that they - This is where the argument is limited
cannot be controlled, then that provides a WRA to only those factors that can be
for demonstrating that all reasonable RRMs evaluated, ignoring nuanced, complex, or
have been implemented. If the conclusion is that unquantifiable variables.
control would be too expensive, then Appendix
A7 explains why prediction is not accurate 5. Question Substitution (20)
enough to demonstrate Gross Disproportion and - When a question cannot be answered
a WRA may be the only viable means of rejecting directly, it has been shown that people tend
the RRM. to substitute a similar but different question
to the one they are given when it is not
understood (20) and (Appendix B6, Error #6).
The figure shows a commonly used layout for a safety risk matrix.
Chronic Multiple
First Aid Minor Injury Fatality
Injury Fatalies
Very likely Red
Likely
Possible Amber
Unlikely
Very Unlikely Green
used, and it is common for none of these to be 4. Which consequences does it refer to?
stated. These are as follows: - The expected outcome?
1. What is its purpose? - The worst Foreseeable outcome?
- Measuring tolerability of Risk? - The worst possible outcome?
- Ranking Risks for different items or 5. What type of Risk metric is in use?
activities? - Relative Risk? (If so, to what? High and low
- Determining Proportionality, i.e. the level are meaningless without a reference.)
of Risk assessment required? - Absolute Risk?
2. What is it measuring? 6. Which Risk it is measuring?
- One Risk to an individual? - The unmitigated Hazard Risks?
- One Risk to all people? - The Top Event Risk (Failure Mode)?
- All Risks to one individual? - The outcome?
- All Risks to all people?
7. What timescale is the Risk is measured
3. What level of granularity is it focussing on? over?
- Average Risks for that industry? - A single action?
- Specific Risks for that product, location, - The whole activity?
or plant? - Per Year (assuming the activity repeats
- Average or specific Risks for a particular continuously for a year)?
system? - Per year (for the average number of times
- Average or specific Risks for a given that activity is likely to occur)?
function or component? - Over the product Lifecycle?
The matrices are sometimes used to determine A Risk thought to be low may be because it has
the acceptability of Risk, with green cells not been experienced, or because the Hazards,
regarded as acceptable, amber worthy of Causes or flaws in RRMs are not appreciated,
consideration for Risk reduction, and red being so it may be much higher than estimated.
unacceptable. The inherent assumption is that Conversely, a Risk believed to be high may be
the matrix adds some value, which is greater because of past accidents and a knowledge
than simply estimating whether something is of causal mechanisms, but that would mean
acceptable or not. In practice, it simply splits the they have probably been addressed already
judgement into two parts, i.e. the consequence, in standards, Good Practice or by common
which can normally be judged within acceptable sense and may therefore be much lower than
limits, and the frequency, which cannot. The estimated. Risk management seeks to identify
danger is that the assessor starts with a unknowns so, on this basis, these would be
preconceived idea of the acceptability of a expected to be where the Risks are believed to
Risk and subconsciously ‘reverse engineers’ be low. So, although focussing efforts on what
it into the desired cell, with all the cognitive we perceive to be high Risks may be intuitive, it
biases discussed previously. It is not possible may be illogical, and it could lead to a situation
to scientifically prove the accuracy of Risk where the largest Risks are deliberately being
judgements for rare events, but the Riso excluded.
institute study (9) showed that the variability
alone can regularly exceed a factor of 10,000
times (Example A4.3), so the errors could be
even larger. Risk simply cannot be judged in
this way.
The reasons for using risk matrices for business decisions is quite different to their use in
safety. For example, a decision whether to invest in Project A or Project B may involve much
more tangible and quantifiable variables, such as exchange rates, materials and labour
availability, interest rates and potential for change. Each of these may be estimated with
reasonable accuracy compared to safety decisions because they tend to be mid-range
numbers, say between 10 and 90%. Uncertainties in this range are generally much more
limited, whereas safety deals with exceedingly small numbers, which people are extremely poor
at estimating.
In business the matrix can be a useful means of communicating opinions as to which option may
have the least Risk (both in terms of maximum loss and its likelihood). Unlike safety decisions, these
are binary ones that have no legal obligation to be transparent or robust.
Risk matrices have been heavily criticised in A7 Cost Benefit Analysis (CBA)
scientific papers. Thomas (24) stated, ‘our
literature search found more than 100 papers Cost Benefit Analysis (CBA) is calculated as
in the OnePetro database that document the follows:
application of RMs in a risk-management context.
However, we are not aware of any published Cost Cost
=
empirical evidence showing that they actually Benefits (Risk reduction) * (Value of Preventing a Fatality)
help in managing risk or that they improve
decision outcomes’. It went on to say ‘A tool that where Risk = Probability *Number of Fatalities
produces arbitrary recommendations in an area
as important as risk management in O&G should Gross Disproportion is demonstrated when cost/
not be considered an industry best practice’. Cox benefits > 1 x (a Gross Disproportion factor),
(23) stated, ‘Typical risk matrices can correctly where the factor could range from approximately
and unambiguously compare only a small fraction 3 to 10 according to guidance on the HSE
(e.g., less than 10%) of randomly selected pairs website, but there are no known legal precedents
of Hazards…. Risk matrices can mistakenly
assign higher qualitative ratings to quantitatively However, caution must be exercised, to ensure
smaller risks. For risks with negatively correlated that the quantification of Risks and costs
frequencies and severities, they can be “worse are scientifically derived, error potential and
than useless,” leading to worse‐than‐random uncertainties are taken into consideration, and
decisions’. the value of life, injury or health issues are in line
with societal expectations.
The matrices may therefore constitute little more
than easily manipulated, ambiguous heuristics, CBA cannot demonstrate Gross Disproportion
based on intuition rather than logic and subject if the uncertainties are more than an order of
to multiple cognitive biases, with enormous magnitude because reliance on the mean is
error potential, and all of which are applied to mathematically unsound in such circumstances.
an undefined notion. In a legal context it is not The true moment of a probability distribution
unreasonable that any probabilistic argument should use the integral of moments across it,
should be transparent and reasoned deductively i.e. ∫y.f.df, as shown in Graph A1, with the point
(plausibility) or inductively (based on Robust weightings (y.f.δf) shown in Graph A2. With Robust
Statistics with no more than one level of Bayesian Statistical data y.f.δf peaks close to the mean,
modification). Their use in risk management but as the uncertainty or errors increase, (e.g. the
is therefore strongly discouraged and it is 90% confidence figure is more than an order of
recommended that the Proportionality Matrix magnitude above the mean), then y.f.δf continues
(Section 3.3) is used instead. to increase and tends towards infinity, as shown
Risk models are necessarily based on unverifiable hypotheses and assumptions, so they constitute
beliefs, not science.
They are susceptible to numerous errors, caused by non-representative data, false assumptions,
unfalsifiable chance correlations, omitted variables and simplified or omitted interdependencies.
The resulting errors can exceed many orders of magnitude.
Equipment failure is often the primary variable in Risk models, but it is rarely the direct Cause of
accidents and, even when it is, those failures can often be traced back to deeper, resolvable reasons
that cannot be modelled or quantified.
Probabilistic errors can be notoriously difficult to find, even by experts, so a sufficiently rigorous
quality assurance process, (which interrogates the data collection, its interpretation, the algorithms,
and the model’s architecture), may not be a realistic proposition.
An important distinction that often goes assumptions and/or predictions that cannot be
un-noticed in probabilistic evaluation is the verified. Common errors are (25):
difference between error and uncertainty. 1. Non-representative data.
Uncertainty relates to the shape and dimensions 2. Causal fallacy.
of the Probability Distribution Function (PDF)
3. Omission.
and it can normally be quantified, especially if
the PDF complies with a mathematical form, 4. Null Hypothesis.
such as the Normal/Gaussian or Bell Curve 5. Ludic Fallacy – Independence.
distributions. This means that confidence 6. Illegitimate Transposition of the Determinant.
bounds can be determined, e.g. the 95% 7. Unfalsifiable Chance Correlations.
confidence interval, which could be much larger
than the 50% mean value, and therefore more The following sections explain these errors in
conservative. more detail.
Brittle Fracture
Wear and tear on valve stem seals
(Longford)
Overpressure rupture
Poorly fitting instrument connection
(Grangemouth, Ocean Odyssey)
Process upset
Wear and tear on door seals
(Texas City, Buncefield, Seveso, Bhopal)
Isolation/reconnection error
Sampling
(Piper Alpha, Pasadena)
The only solution is to measure something more known major accidents, which are typically
frequent that has a quantifiable relationship large, normally full bore, failures, greater than
(random or Bayesian) with the rare event and 10 kg/s. However, the right-hand column looks
calculate its frequency using this knowledge. It at the Causes of the more frequent, smaller
will be necessary to provide suitable evidence leaks, commonly known as weeps and seeps.
of any such relationship, whether by logical There are few, if any, major accidents that have
argument or by some form of statistical proof. been caused by these failure modes, yet they
dominate the datasets (26), despite having no
It is Good Practice to state the study objectives quantifiable random relationship with the rare
and define the relevant variables prior to the events that they are used to predict.
collection of any statistical data. However, when
used for predicting rare events, such as major Because the larger leaks are not statistically
accidents, this will not be practicable because significant it is normal practice to plot all leak data
the data population is generally worldwide. on a frequency vs. size graph and find the best fit
In practice, it will be necessary to draw upon curve, in the erroneous belief that this overcomes
whatever industry data is available, even the paucity of large leak data. This is therefore
though it may have been collected for different an attempt to overcome contravention of the
objectives to the study purpose, e.g. reliability, statistical significance criterion by contravening
rather than safety. representativeness instead. The problem is
that leak sizes are intuitively the right thing to
A good example of this error occurs in the measure, provided the Causes are ignored.
process industries where data is collected by Conversely, the idea that data on the common
size of leak, even though very few leaks lead cold could be used to predict cancer rates
to major accidents. However, the small leaks would be quickly dismissed, because it is well-
have different Failure Modes to large ones and known that one is a virus, whilst the other is not.
may have little or no potential to Cause major Statistical sampling must therefore be subject to
accidents. In Example B1 the left-hand column rigorous scrutiny to ensure that it represents that
is a list of these Failure Modes for some well- which the study is attempting to calculate.
The ratio of accidents to Hazards is known as potential is almost unlimited and that even
the Likelihood Ratio, but this still requires the expert judgement cannot be relied upon to
Hazard probability to be known, so there may be notice it or to sense check Risk figures. Even
no credible solution. This is a common problem the appeal court judges failed to understand
with Bayesian analysis, which is often fudged by the difference when it was explained to them
simply guessing the denominator, on the basis eighteen months later, stating that it was “a
that this at least gets closer to the true answer. straight mathematical calculation to anyone who
However, if the original error is four orders of knew the birth-rate over England, Scotland and
magnitude and Bayesian analysis reduces this to Wales”. It was not until the case became a Cause
three, then the result is still one thousand times celebre and academics wrote papers on it that
out. With rare events, such as major accidents, the defence were able to get the judgement
the ratio could be many orders of magnitude, overturned on a second appeal; by which time
giving unacceptable errors. she had spent three years in jail.
One of the most researched examples of this All data, together with its application, should
mistake was the trial of Sally Clark, for the therefore be thoroughly reviewed for possible
murders of her two children, who were later Illegitimate Transposition of the Determinant
concluded to have died of cot deaths. There was errors.
no evidence of foul play, except that an expert
witness, Professor Sir Roy Meadows, stated that
the probability of natural deaths was only 1 in B7 Error #7
73,000,000 (17), (18). Ms. Clark was found guilty
Unfalsifiable Chance Correlations
and given a life sentence.
Correlations between any two variables, e.g.
In practice, he stated the probability of two cot
fatigue and equipment type, should have proven
deaths given that she was innocent:
causal relationships, free from Unfalsifiable
Chance Correlations (UCCs), where multiple
Pr(2 cot deaths│innocence) = 1:73,000,000 tests looking at different variables may
eventually find correlations due to chance only,
However, this is not the question that the trial e.g. testing at a P value of 0.05 (95% confidence)
sought to establish, which was, “What was the would create one false correlation in every 20
probability that she was innocent given two cot trials. A good explanation of UCCs is given in
deaths?” to which the answer was: a study by Ashwanden (27) that demonstrated
that if people cut the fat off their meat there is
a 99.7% probability that they are atheists. The
Pr(innocence│2 cot deaths) = 15:1
point was deliberately provocative to illustrate
that mathematical rigour alone is not enough.
The fact that the error, which was over a billion 3,200 equally absurd hypotheses were tested,
times (25), went unnoticed, shows that the most of which showed little or no correlation,
IDENTIFICATION STAGE:
The System is defined and understood
Formal identification/brainstorming exercise
Appropriate disciplines/personnel involved
Lessons from incidents, accidents, and precursors
Health, injury and mortality effects assessed
Lifecycle – construction/manufacture
Lifecycle – commissioning/set-up/trials
Lifecycle – storage/mothballing
Lifecycle – operations
Lifecycle – maintenance and inspection
Lifecycle – decommission, dismantle, dispose
Reliability issues
Control issues
Activities and procedures
Exceeding the design envelope
Natural environment
Deterioration, ageing
ANALYTICAL STAGE:
Procedural analysis
Human factors analysis
Software analysis
Functionality analysis
RRM analysis – Effectiveness
RRM analysis – Failure to Safety
RRM analysis – Reliability
RRM analysis – Independence/Redundancy
RRM analysis – Diversity
RRM analysis – Self-Revealing Failures
imeche.org