IMPERFECT & Partial Valve Stroke Testing Impact On SIF
IMPERFECT & Partial Valve Stroke Testing Impact On SIF
IMPERFECT & Partial Valve Stroke Testing Impact On SIF
Master Thesis
MASTEROPPGAVE
Vrsemester 2007
ERKLRING
Jeg erklrer herved p re og samvittighet at jeg har utfrt ovennevnte hovedoppgave selv
og uten noen som helst ulovlig hjelp
____________________________________________________________
Signatur
Besvarelsen med tegninger m.v. blir i henhold til Forskrifter om studier ved 20, NTNU's
eiendom. Arbeidene - eller resultater fra disse - kan derfor ikke utnyttes til andre forml uten
etter avtale med de interesserte parter.
PREFACE
This master thesis was written during 20 weeks throughout spring 2007 as the final work
performed by Hanne Roln at Norwegian University of Science and Technology (NTNU). The
thesis is written within the Department of Industrial Economics and Management, study
field Health, Environment and Safety. It is also closely related to the Department of
Industrial Production and Quality as the thesis is within the field of technical safety. The
thesis is written in close cooperation with Aker Kvrner Subsea.
Intended audience are those with knowledge of reliability theory, and it is recommended that
the reader is familiar with the concepts described in the book System Reliability Theory by
Rausand and Hyland (second edition, 2004).
I would like to thank my colleagues at Aker Kvrner Subsea for the support throughout the
semester, and especially Thor Ketil Hallan as the supervisor. Further on I would like to thank
Ring-O, Lars Bak (Lilleaker Consulting), and Luciano Sanguineti and Enrico Sanguineti at
ATV for giving me the necessary practical understanding of valves. And finally thanks to
Mary Ann Lundteigen (NTNU) for good discussions and Marvin Rausand (supervisor at
NTNU) for important feedback and input throughout the thesis.
Hanne Roln
Introduction
SUMMARY
In order to avoid substantial hardware costs of building platforms, moving petroleum
production facilities subsea is becoming a popular solution. Fields can be remotely operated
and stand-alone fields that would not be profitable to develop separately can now be tied
together to one pipeline/riser and hence save expenses. Safety instrumented systems are
implemented to reduce or eliminate unacceptable risk associated with such production, and
the safety integrity level is a common requirement describing the safety availability of the
equipment. When performing analysis of the consistency of the safety functions to perform
when needed, it is important to evaluate the assumptions that form the basis for the
calculations. The author has in particular assessed the assumption that a component is as
good as new after each proof test, meaning that the unavailability is reduced to zero.
The reasons for imperfect tests may be related to the five M-factors; method, machine,
milieu, man-power and material. Through different case studies the potential effects of
imperfect tests have been analyzed. SINTEF has proposed a method for including the
systematic failures in the calculations of the probability of failure on demand (PFD) by
adding a constant value called PTIF (test independent failures) in the PDS method. A method
for quantifying the PFD impact of an imperfect test due to non-testable random hardware
failures have been proposed by the author. Case results indicate that the PFD impact is far
more significant for imperfect testing of hardware failures than the PDS approach for
systematic failures.
Implementing partial stroke testing enables to reveal failure modes only before possible
through tests that require process shutdown. A successful implementation may improve the
safety integrity level rating of the system. The use of partial stroke testing in subsea
petroleum production has so far not been common and several of the arguments for and
against implementing partial stroke testing are assessed.
It has been argued that partial stroke testing leads to an increase of the spurious trip rate, as
it is likely that if it starts to move it will continue to closed position. The likely reasons for
such an event were placed in a Bayesian belief network and proved the need for the right
equipment to be implemented. New devices such as smart positioners and digital valve
controllers have been introduced for the purpose of partial stroke testing, reducing the
human interference in partial stroke testing and thus reducing the causes for spurious trips.
Partial stroke testing may be implemented in order to justify extended proof test intervals. As
common cause failures are those failures that happen within the same proof test interval, an
extension of the interval could imply that more failures are classified as common cause
failures (Rausand, 2007). In such situations, it should be discussed whether the -factor
should be incremented to reflect the PFD impact this may have.
Another argument for implementing partial stroke testing has been the opportunity to reduce
the hardware fault tolerance (implies cost saving) since the safe failure fraction is
11
Introduction
12
Introduction
INDEX
PREFACE
9
11
13
15
15
16
SUMMARY
INDEX
LIST OF TABLES
LIST OF FIGURES
TERMS AND ABBREVIATIONS
1
INTRODUCTION
1.1
1.2
1.3
1.4
1.5
2
BACKGROUND
OBJECTIVES
DELIMITATIONS
SCIENTIFIC APPROACH
STRUCTURE OF THE REPORT
17
17
17
18
19
THEORETICAL FRAMEWORK
2.1
2.1.1
2.1.2
2.2
2.2.1
2.2.2
2.2.3
2.3
2.3.1
2.3.2
2.3.3
2.4
2.5
3
17
20
IMPERFECT TESTING
21
21
23
23
23
23
24
25
25
28
31
34
37
39
13
39
41
43
46
52
54
4.1
4.2
4.3
4.4
4.5
4.6
5
DISCUSSION
55
57
58
62
63
65
69
6.1
6.2
6.3
6.3.1
6.4
7
55
5.1
5.2
5.3
6
Introduction
69
71
71
72
72
72
74
75
76
CONCLUDING REMARKS
79
REFERENCES
ANNEX A, XMT
81
84
14
Introduction
LIST OF TABLES
TABLE 1, SIL FOR LOW AND HIGH DEMAND MODE OF OPERATION (IEC 61508-1, 2002)............................................................26
TABLE 2, SIL FOR TYPE A SUBSYSTEM (IEC 61508-2) ..........................................................................................................32
TABLE 3, SIL FOR TYPE B SUBSYSTEM (IEC 61508-2)...........................................................................................................32
TABLE 4, DATA FOR THE SYSTEM TEST EXAMPLE ..................................................................................................................41
TABLE 5, UNAVAILABILITY AT TIME T OF A SINGLE COMPONENT UNDER IMPERFECT TEST CONDITIONS .......................................48
TABLE 6, PFD AVERAGE DIFFERENCES BETWEEN PERFECT AND IMPERFECT TESTS ....................................................................49
TABLE 7, MATRIX FOR SIL RATING SENSITIVITY DUE TO IMPERFECT TESTING ..........................................................................50
TABLE 8, DANGEROUS FAILURE MODES AND TEST STRATEGY FOR A SAFETY GATE VALVE ( ADAPTED FROM SUMMERS & ZACHARY
2000A, MCCREA-STEELE 2006, KOP 2002, BAK 2007 AND ATV 2007) ......................................................................59
TABLE 9, RELIABILITY DATA AS BASIS FOR PST COVERAGE ESTIMATION (ADAPTED FROM LUNDTEIGEN & RAUSAND, 2007).......61
TABLE 10, PFD RELATED TO DIVERSE PST COVERAGES, TEST INTERVALS AND (IM)PERFECT TESTING ........................................68
TABLE 11, MORVIN HIPPS REQUIREMENTS (STATOIL, 2007A)...............................................................................................73
TABLE 12, MORVIN HIPPS CASE DATA ...............................................................................................................................76
LIST OF FIGURES
FIGURE 1, IEC 61508 SAFETY LIFECYCLE (IEC 61508) .........................................................................................................22
FIGURE 2, SKETCH OF A SIMPLE SIS (RAUSAND & HYLAND, 2004).......................................................................................25
FIGURE 3, ALLOCATION OF SIL (IEC 61508-1).....................................................................................................................27
FIGURE 4, RISK REDUCTION (IEC 61508) .............................................................................................................................27
FIGURE 5, FAILURE MODE CLASSIFICATION (IEC 61508) .......................................................................................................28
FIGURE 6, FRACTIONS OF DIFFERENT TYPES OF FAILURES FOR A SYSTEM WITH TWO COMPONENTS .............................................30
FIGURE 7, SIS DESIGN REQUIREMENTS .................................................................................................................................31
FIGURE 8, WELLHEAD AND XMT (OREDA, 2002)...............................................................................................................34
FIGURE 9, HORIZONTAL XMT (AKS, 2007).........................................................................................................................35
FIGURE 10, GATE VALVE WITH ACTUATOR (RING-O, 2007) ...................................................................................................36
FIGURE 11, UP AND DOWN TIME RELATED TO TESTS ..............................................................................................................38
FIGURE 12, CAUSES FOR IMPERFECT TESTING OF SUBSEA SAFETY VALVES ...............................................................................40
FIGURE 13, RELIABILITY BLOCK DIAGRAM OF A SIMPLE SIS..................................................................................................41
FIGURE 14, CONTRIBUTION TO UNAVAILABILITY (PDS METHOD, 2006) ..................................................................................43
FIGURE 15, SKETCH OF THE PFD IMPACT WITH PTIF ADDITION (CASE A) .................................................................................44
FIGURE 16, UNAVAILABILITY UNDER IMPERFECT TEST CONDITION CASE A ..............................................................................44
FIGURE 17, SKETCH OF THE PFD IMPACT WITH IMPERFECT TEST ADDITION (CASE B)................................................................46
FIGURE 18, SERIES STRUCTURE WHEN IMPERFECT TEST OF A COMPONENT ...............................................................................46
FIGURE 19, UNAVAILABILITY UNDER IMPERFECT TEST CONDITION CASE B ..............................................................................47
FIGURE 20, UNAVAILABILITY FOR DIFFERENT FAILURE RATES UNDER IMPERFECT TESTING ........................................................49
FIGURE 21, THE M-FACTORS CONTRIBUTION TO THE IMPERFECT TEST ADDITION ....................................................................50
FIGURE 22, UNAVAILABILITY WITH DECREASING PTIF ADDITION (CASE C1) .............................................................................52
FIGURE 23, UNAVAILABILITY WITH DECREASING IMPERFECT TEST ADDITION (CASE C2) ...........................................................53
FIGURE 24, PFD RESULTS FROM CASE STUDIES ON IMPERFECT TESTING ...................................................................................54
FIGURE 25, PST IMPACT ON THE PFD (LUNDTEIGEN & RAUSAND, 2007)................................................................................55
FIGURE 26, SIMPLE SIS WITH PST IMPLEMENTATION (ADAPTED FROM MCCREA-STEELE, 2006) ..............................................56
FIGURE 27, OVERVIEW OF RELEVANT FAILURE RATES (LUNDTEIGEN & RAUSAND, 2007) .........................................................58
FIGURE 28, BAYESIAN BELIEF NETWORK FOR ST DURING PST ...............................................................................................62
FIGURE 29, UNAVAILABILITY WITH PST ..............................................................................................................................65
FIGURE 30, UNAVAILABILITY WITH PST AND PTIF ADDITION ..................................................................................................66
FIGURE 31, UNAVAILABILITY WITH PST AND IMPERFECT TESTING ..........................................................................................67
FIGURE 32, HIPPS SCHEMATIC (KOP, 2004)........................................................................................................................74
FIGURE 33, HIPPS RELIABILITY BLOCK DIAGRAM FOR M ORVIN FIELD DEVELOPMENT ..............................................................76
FIGURE 34, PFD RESULTS FOR THE DIFFERENT CALCULATION APPROACHES .............................................................................78
15
Introduction
16
Introduction
1 Introduction
1.1 Background
As safety instrumented systems are getting increasingly important within petroleum
production, there is a need to have a good understanding of the assumptions and
simplifications that form the basis for the assessment. Reliability calculations of subsea
production systems are usually based on the IEC 61508 (2002) approach and utilizing
OREDA (2002) data. A basic assumption is that after test and repair, the system is as good
as new, meaning that the system unavailability is reduced to zero. As the demand for
continuous production extends the test intervals further every time, it is increasingly
important to study these aspects more profoundly. Are the safety levels as good as claimed?
Is it required to change the calculation-method in order to reflect the reality?
The topic of this thesis has been developed in cooperation with Marvin Rausand at NTNU
and Thor Ketil Hallan at Aker Kvrner Subsea, as this is of interest for both parties.
Intended audience are those with knowledge of reliability theory, and it is recommended that
the reader is familiar with the concepts described in the book System Reliability Theory by
Rausand and Hyland (second edition, 2004). Basic knowledge of subsea petroleum
production is also an advantage.
1.2 Objectives
The main objectives have been to;
Study the IEC 61508 (2002) and IEC 61511 (2004) standards and OLF 070 guideline
(2004)
Perform a case study from one of Aker Kvrner Subseas actual projects
1.3 Delimitations
The time scope to carry out the master thesis is set to 20 weeks; hence it has been a need to
choose more specific topics to assess. Since the author has been present at Aker Kvrner
Subsea during the thesis period, the focus is upon reliability in subsea petroleum production.
Because of the authors special interest in safety topics, the reliability assessment is limited to
the safety reliability (availability), excluding the production reliability.
The purpose of the case study is to relate the results to an actual field. The weight of the
thesis is to be found in assessment of PST and imperfect testing.
17
Introduction
18
Introduction
Chapter 2
Theoretical framework
Chapter 3
Imperfect testing
Chapter 4
Partial stroke testing
Chapter 5
Discussion
Chapter 6
Case
The theories from the former chapters are used for a real life
system; the Morvin field development.
Chapter 7
Concluding remarks
19
Theoretical framework
2 Theoretical framework
Subsea equipment is becoming the typical solution for the petroleum offshore industry as the
production is moved to deeper and more demanding areas. It is expected to be a substantial
increment of the subsea petroleum and gas production in the years to come (Subseazone,
2007). At the same time it can be claimed that there has been an increased concern regarding
health, safety and environment (HSE) among the general public and governments the last
years, which has lead to a more strict legislation within the field. Together with the high costs
for subsea intervention this gives incentives for the oil companies to achieve a high level of
safety and reliability of their systems.
Safety instrumented systems (SIS) have become ever more common as a measure for
reducing risk. A SIS is designed to prevent, or mitigate, the hazardous event that could harm
the system in which it is implemented to protect. Examples of hazards related to subsea
production are blowouts topside (possible personnel fatalities and material damage) and
leakage to water (environmental danger). One SIS can perform one or several safety
instrumented functions (SIF).
With this increased dependency of SIS to mitigate risk, it is crucial to be aware of the
assumptions and simplifications that form basis for the reliability calculations. In the
following short introductions to the important standards and guidelines within the field are
given, as well as a more thorough description of SIF.
20
Theoretical framework
21
Theoretical framework
The IEC 61508 has a lifecycle approach for the SIS as presented in Figure 1. Following the
steps in the lifecycle is to ensure that the SIF is achieved through a systematic approach to all
the necessary activities.
The system design complies with the IEC 61508 and IEC 61511 standards when the company
accomplishes the requirements related to:
SIS decommissioning
The calculation of the reliability of SIF is only a small part of the IEC 61508 compliance.
Some of the assumptions and simplifications done in the IEC standard related to reliability
are assessed in this thesis. Lundteigen & Rausand (2006) stated; The standards are not
prescriptive, which gives room for different interpretations, and hence opens up for new
methods, approaches and technology.
The implications the IEC 61508 standard has for SIS is described more in detail throughout
chapter 2.
22
Theoretical framework
2.2.2 PDS
PDS is the Norwegian abbreviation for reliability of computer-based safety systems.
SINTEF is the author of both a PDS Method Handbook and a PDS Data Handbook. The PDS
approach is described in the first, while the latter contains a data dossier for different
components. These are based on OREDA, but the project group have done some adjustments
and expert judgements upon the figures. The PDS method is in line with the main principles
in the IEC 61508 standard, except of a somewhat different approach regarding failure
classification, modelling of common cause failures and the treatment of systematic failures
(PDS Method, 2006). Of special relevance for this thesis are the quantification of systematic
failures called PTIF (test independent failures, TIF), and the concept of Critical Safety
Unavailability (CSU) which in addition to the IEC 61508 approach to PFD calculation also
includes downtime due to test and repair.
23
Theoretical framework
2.2.3 Exida
Exida (excellence in dependable automation) provides reliability data for use in process and
machinery industries. The numbers are based on FMEDA data or exida comprehensive
analysis with data from OREDA, PDS etc. The main reliability concepts and failure
classifications correspond to a great extent those described in IEC 61508.
24
Theoretical framework
Examples of SIS are among others the emergency shut-down system in a hazardous chemical
process plant, automobile indicator lights, anti-lock braking and engine-management
systems, and remote monitoring, operation or programming of a network-enabled process
plant (adapted from Rausand & Hyland, 2004).
The safety integrity level (SIL) specifies the safety integrity requirements of the SIF to be
allocated to the SIS. It states the probability of the SIS to fail to perform the requested SIF
upon demand, often referred to as the PFD (probability of failure on demand). The PFD may
be interpreted in two ways; the probability that the system will be in a dangerous failure
mode upon demand, and the fraction of time the system will be in a dangerous failure mode
and not work as a SIF. In order to attain the requested SIL it is also required to avoid and
25
Theoretical framework
control systematic failures and to select the hardware configuration within the architectural
constraints. These requirements are further described in section 2.3.3.
The IEC 61508 standard divides the SIL into 4 levels, where the highest SIL rating states the
lowest probability that the SIS will fail to perform the required SIF. Depending if it is a low or
high/continuous demand mode of operation, the range of the levels differ as shown in Table
1. Low demand mode embrace systems where the frequency of demands for operation made
on a safety related system is no greater than one per year and no greater than twice the prooftest frequency (IEC 61508, 2002), otherwise it is classified as a high demand system. An
example of a low demand application in subsea production is the safety valve where the valve
remains static until a demand occurs. An application in high demand mode can for example
be the brake system in a car.
Table 1, SIL for low and high demand mode of operation (IEC 61508-1, 2002)
Safety
integrity
level
10 5 to < 10 4
10 9 to < 10 8
10 4 to < 10 3
10 8 to < 10 7
10 3 to < 10 2
10 7 to < 10 6
10 2 to < 10 1
10 6 to < 10 5
Note that one SIS may perform several SIFs, and that the reliability assessments are done for
each SIF and not for the SIS. The SIL rating is often required to be within the midpoint of
each level to be considered good enough, meaning that the PFD has to be less than
26
Theoretical framework
The IEC 61508 standard, the Norwegian Activities regulations 1 -2 and the Haddon
energy model (1973) all have the same philosophy regarding barriers. In order to reduce risk,
the priority should be to apply measures in the design, as this is the best way to eliminate the
hazards. If this does not reduce the risk to the tolerable region (reference ALARP principle as
described in IEC 61508-5 and the Norwegian Framework regulations 9). Barriers are
introduced in order to prevent or mitigate impact on people, environment and/or material
assets. A SIS can be such a barrier as shown in Figure 4. Further reduction of risk than
strictly necessary in order to come within the acceptable area should be done as long as this is
economically reasonable.
Note that introducing a SIS is only one measure among others. It is equally important to
introduce other barriers.
27
Theoretical framework
Examples of dangerous detected failures are those revealed by diagnostic testing. On the
other hand, a dangerous undetected failure is a failure not revealed before a proof test or by a
demand. These failures are important to discover as soon as possible. Representative for a
safe failure is a spurious trip, which is for example that the safety valve closes without a real
demand. Note that to classify a failure according to these classes is not always
straightforward and can easily be interpreted differently among the users.
Random hardware failures and systematic failures
Another way of classifying the failures is to differentiate between random hardware
(physical) failures due to aging and stress, and systematic (non-physical) failures due to
design and interaction (adapted from IEC 61508);
Systematic failures: Failures that are related in a deterministic way to a certain cause,
which can only be eliminated by modification of the design, manufacturing process,
operational procedures, documentation or other relevant factors. Design faults and
maintenance procedure deficiencies are example of causes that may lead to systematic
failures. System failure due to systematic failures can not be easily predicted.
28
Theoretical framework
It is of great interest to assess the unavailability of the safety system. It is only the dangerous
undetected failures that form the basis for the PFD calculation. With as the proof test
interval and F(t) the distribution function, the safety unavailability is expressed by PFD as
given below;
(t) = Pr(a DU failure has occurred at, or before, time t) = Pr(Tt) =F(t)
PFD =
1
1
(t )dt = F (t )dt
0
0
29
Theoretical framework
Common cause failures (CCF) may bring down both safety valves at the same time, reducing
the potential positive effects by introducing redundancy to the system. By IEC 61508
definition a common cause failure is a failure, which is the result of one or more events,
causing coincident failures of two or more separate channels in a multiple channel system,
leading to system failure. The -factor is one way of describing common cause failures
quantitatively, where the -factor gives the fraction of common cause failures among all
failures of a component;
Pr (Common cause failure | Failure) =
This is illustrated in Figure 6. Rausand & Hyland (2004) give more details about the factor model and other alternative models.
Component 1 Component 2
1-
1-
Figure 6, Fractions of different types of failures for a system with two components
Goble (2003) state that three principles should be followed in order to avoid CCF;
1. Reduce the chance of a common stress physical separation and electrical separation
in redundant units.
2. Respond differently to a common stress redundant units should use diverse
technology/mechanisms.
3. Increased strength against all failures.
A method for estimating the common cause beta factor is provided in IEC 61508-6, or the
maximum values given in table D.4 in IEC 61508-6 can be used directly.
30
Theoretical framework
SFF =
DD
Tot
+
+ +
S
DD
DD
DU
Theoretical framework
A. A safety valve is normally defined as a type A subsystem (Lundteigen and Rausand, 2007).
Table 2 gives the attainable SIL rating under these constraints for type A subsystems.
Table 2, SIL for type A subsystem (IEC 61508-2)
Safe failure
fraction
< 60%
60% - < 90%
90% - < 99%
99%
Safe failure
fraction
< 60%
60% - < 90%
90% - < 99%
99%
Because of the difficulties of achieving and maintaining a SIL 4 throughout the safety
lifecycle, applications which require the use of a single SIF with SIL 4 should be avoided
where reasonably practicable (IEC 61511-1).
Related to hardware safety integrity there are also requirements for the PFD. In the
calculations it should be taken into account the system architecture, dangerous failures
undetected/detected by diagnostic tests, susceptibility to common cause failures, diagnostic
coverage, test intervals and repair times etc. The calculations should be done for each sub
element and gives the following formula for the SIS in Figure 2;
1
1
Where the PFD is given by PFD = (t )dt = F (t )dt as presented in last section.
0
0
32
Theoretical framework
33
Theoretical framework
34
Theoretical framework
Two common types of XMT are the conventional (dual bore) and horizontal. One of the main
advantages for choosing a conventional XMT is that the tree can be retrieved without
removing the tubing hanger and the tubing (Sangesland, 2007). In a horizontal XMT the
diameter of the tubing can be larger and the tubing can be replaced without retrieving the
tree. The stack-up height of the XMT including BOP may otherwise be difficult to handle on a
conventional drilling vessel. On the other hand, retrieval of the tree implies retrieval of the
tubing. Illustrations of both the horizontal and conventional tree are in Annex A. In Figure 9
is an example of a horizontal XMT.
The gate valve is normally preferred as safety valve compared to a ball valve. The ball valve
requires a rotation force in addition to vertical movement. Gate valves normally have a lower
internal leakage, and a shorter stem travel. It is assisted by pressure in the bore cavity
35
Theoretical framework
pushing the stem out of the valve cavity when closing. The actuator spring is also designed to
close also without pressure in the system. In Figure 10 a gate valve is shown in closed (left)
and open (right) position. Note that this is only one of many solutions.
Note that for fail safe gate valves the hole in the gate is on the upper part, thus when the valve
is de-energized it will shift upwards and close. The valve will close whenever loss of electric
and/or hydraulic power is detected.
36
Theoretical framework
37
Theoretical framework
in question and the chosen test approach. The proof test itself may however also be
incomplete and may be considered as an imperfect test. This is further elaborated in the next
chapter. In Figure 11 the relationship between the time concepts and tests is given. Rausand
& Hyland (2004) give a thorough description of these concepts.
C- Diagnostic or partial
testing reveals the failure
before the scheduled proof
test, thus reducing the
time of the undetected
dangerous failure.
Figure 11, Up and down time related to tests
It has been developed a high level of diagnostic coverage for the sensors and logic and with
redundancy it has been succeeded to reduce the contribution to the PFD (Metso automation,
2002), leaving the greatest contributor to be the actuating items/final elements.
Because of the disturbances testing impose on the production, the risks associated with the
testing itself and restarting after the test is finalized, it is preferred to test as seldom as
possible. Hence it is a need to optimize the test intervals to maintain both the safety and
production interests. That is to assure as high safety availability as possible without
introducing additional production downtime. The causes and consequences of imperfect
testing, and partial stroke testing is further discussed in chapter 3 and 4.
38
Imperfect testing
Imperfect testing
In the IEC 61508 standard it is assumed that after a proof test the component is as good as
new. For the proof test to be fully effective this means that it is necessary to detect 100 % of
all dangerous failures, reducing the unavailability to zero. This may not be feasible. An
imperfect test situation may be defined as a situation where the test does not discover all
dangerous failures and subsequently component unavailability remains. It can be claimed
that there are two possible classifications of an imperfect test situation;
1. The test does not cover all possible failures inadequate test method.
2. The test does not detect all the failures unsuccessful test.
Hence the function test, the PST and the diagnostic test can all be classified as an imperfect
test since they do not cover all failure modes while a proof test may be imperfect due to
unsuccessful testing.
Since the focus in this thesis is upon testing it is assumed that as long as all failures are
discovered they can be repaired to an as good as new condition. Analogue to the definition
of imperfect testing in this thesis, imperfect repair can be defined as the situations where the
fault is not repaired perfectly or that the failure is chosen not to be repaired, as well as lack of
an adequate method for repairing the component. This can be the case when for example the
leakage is considered minimal, or the repair of a somewhat delayed operational time (DOP) is
postponed until it is more significant. Rausand & Hyland (2004) give an introduction to
imperfect repair processes.
This uncertainty related to the test quality is not included in the reliability calculations, and is
neither discussed much in the IEC 61508/61511 standards nor literature in general. IEC
61508-6 mention briefly the effects of a non-perfect proof test in annex A (informative only).
This topic is elaborated in the following, both the possible causes for imperfect testing and its
impact on the PFD.
39
Imperfect testing
create a static well. The fishbone diagram in Figure 12 gives additional causes for imperfect
testing of safety valves subsea.
As described in Figure 12 the reasons for imperfect testing can be related to the attributes;
methods, materials, machines, milieu and manpower. The attribute materials cover the test
equipment, methods the procedure and formalities around the testing, machines the
subsea-system itself, milieu describes the context of the system and manpower the
managers and workers conducting the tests. As illustrated it is obvious that the human
interference is an important reason for imperfect testing.
There are no data collected for the proportion of tests that can be claimed to be imperfect. A
possible method for estimating the contribution of each of the Mfactors described above is
proposed in 3.2.2.
40
Imperfect testing
Failure rate DU
Component
Test interval
hours
3.0 10 7
3%
8760
3.94 10 5
1.0 10 7
2%
8760
9.01 10 6
1.0 10 7
2%
8760
9.01 10 6
( hours )
(OLF 070)
Sensors
(2oo3)
Logic
(1oo2)
Valve
(1oo2)
Common cause
failure; - factor
(PDS Data, 2006)
System
( hours 1 )
5.74 10 5
This result corresponds to a SIL 4 classification for the system. No diagnostic testing is
assumed for the sensors or logic in this example. The time required to test and repair the
items is considered negligible. This may make the result higher than genuine field results.
41
Imperfect testing
Note that HFT and SFF requirements are not considered in the case examples, only the PFD
requirement for the SIL rating is assessed.
To determine the PFD impact of an imperfect test is dependent of the type of failure that
remains undetected. It is only the dangerous undetected failures that should be included in
the calculations. Using the simplified equations for the PFDs, the PFD for the system is:
=
=
DU =
DU [(1 ) DU ]
[(1 ) DU ]
+
+ DU +
+ DU
2
3
2
3
2
2
These simplified equations can be used when DU is small (<0.1), and is often used in
practical calculations. The approximation is conservative, which means that the
approximated value is always greater than the correct value (Rausand & Hyland, 2004).
In the calculations of the system test example in the next sections, the PFD is chosen to be
calculated for 20 years because the XMTs are often required to have a life span of at least 20
years. It is assumed that the XMT is not retrieved and overhauled during this period. In order
to estimate the impact of PFD imperfect testing of the valves to the test-system, it is
necessary to assess the valves PFD first and then use the result in the system test example at
the end of each section.
42
Imperfect testing
The definition of PTIF is The probability that the module/system will fail to carry out its
intended function due to a (latent) systematic failure not detectable by functional testing
(therefore the name test independent failure). The PTIF is assumed to be constant
throughout the lifetime, and for extended testing (proof testing) of the valves, a value of
PTIF= 1.0 10 5 hours 1 is suggested. PDS does not give any details for this choice of value.
The difficulty with detecting systematic failures is an example of an imperfect test due to
inadequate test methods. This gives the following equation:
43
Imperfect testing
Figure 15, Sketch of the PFD impact with PTIF addition (case A)
With failure rate DU = 1.0 10 7 hours 1 and PTIF= 1.0 10 5 , the unavailability for one
component for a lifecycle of 20 years is illustrated in Figure 16.
Unavailability Case A
0,0012
0,001
(t)
0,0008
Unavailability
SIL 2
0,0006
PFD average
PFD avg perfect test
0,0004
0,0002
16 0
30
32 0
60
48 0
90
65 0
20
81 0
50
97 0
8
11 00
41
13 00
04
14 00
67
16 00
30
00
Hours
For this exact example the PFD for the imperfect test situation is PFD = 4.48 10 4 while a
perfect test yields the result PFD = 4.38 10 4 and a difference of 1.0 10 5 , which is the PTIF
addition. This addition does not lead to a change of the SIL rating for the component as the
PTIF is small.
44
Imperfect testing
[(1 ) DU ]
[(1 ) DU ]
+ DU +
+ DU +
2
3
2
3
2
DU
+ C1oo 2 PTIF
2
45
Imperfect testing
Figure 17, Sketch of the PFD impact with imperfect test addition (case B)
As a basis for calculating the PFD impact of imperfect testing, a component is divided into a
series structure where one part is non-testable and the remaining part testable. In order
to function, both parts of the component have to function.
The dangerous undetected failures rate is split into two parts dependent of the imperfect test
fraction, here named .
DU = DU NT + DU T
DU NT = DU
DU T = (1 ) DU
When the test interval for the testable part is =8760h, and the NT =175200h corresponds to
the component life span of 20 years, the PFD for the component is described by:
46
PFD =
Imperfect testing
(1 ) DU DU NT
+
2
2
As the relation between PFD and the test interval is a linear function, it is reasonably that a
shorter test interval leads to a smaller PFD. Because of the implications with proof testing,
shorten the test intervals is not a desired solution for achieving the required SIL. For this
reason the test interval is held fixed throughout the calculations.
In Figure 19 a sketch is drawn for the failure rate DU = 1.0 10 7 hours 1 over 20 years. It is
assumed that the non-testable part is =20%. This means that 80% of the failure rate is set to
zero after every proof test while the remaining 20% continues over the whole interval. As
shown in Figure 19 the non-testable part increasingly override the testable part of the system.
Unavailability Case B
0,0045
0,004
Unavailability
0,0035
SIL 2
(t)
0,003
0,0025
Unavail. Perfect
testing
0,002
PFD average
0,0015
0,001
0,0005
15 0
90
31 0
80
47 0
70
63 0
60
79 0
50
95 0
4
11 0 0
13
12 00
72
14 00
31
15 00
90
17 00
49
00
Hours
PFD
DU NT
+ (1 ) DU PT
2
2
PFD 0.2
1.0 10 7 20 8760
1.0 10 7 8760
+ (1 0.2)
= 2.10 10 3
2
2
Note that the PFD average is actually an average of average in order to illustrate the possible
change in the SIL rating. For this exact example the PFD for the imperfect test situation is
PFD = 2.10 10 3 while a perfect test yields the result PFD = 4.38 10 4 which gives a
difference of 1.66 10 3 .
The PFD impact by different combinations of the failure rate and percentage non-testable are
given in Table 5. The unavailability is calculated for different failure rates, and the range from
DU = 1.0 10 8 hours 1 till DU = 1.0 10 6 hours 1 is chosen as this interval reflects the
failure rate of a PMV in subsea XMT. The non-testable part ranges from 10% till 90% in the
calculations. For convenience it is assumed that the same failures remain undetected during
47
Imperfect testing
the whole lifetime. For illustrating the PFD development the accumulated PFD is shown for
year 1, 10 and 20.
Table 5, Unavailability at time t of a single component under imperfect test conditions
Years
1
10
20
10%
1
10
20
20%
1
10
20
30%
1
10
20
40%
1
10
20
50%
1
10
20
60%
1
10
20
70%
1
10
20
80%
1
10
20
90%
1
10
20
DU = 1.0 10 8
DU = 1.0 10 7
DU = 1.0 10 6
8.80 10 5
8.70 10 5
8.70 10 5
8.80 10 5
1.66 10 4
2.54 10 4
8.80 10 5
2.46 10 4
4.22 10 4
8.80 10 5
3.25 10 4
5.89 10 4
8.80 10 5
4.05 10 4
7.56 10 4
8.80 10 5
4.84 10 4
9.24 10 4
8.80 10 5
5.63 10 4
1.09 10 3
8.80 10 5
6.43 10 4
1.26 10 3
8.80 10 5
7.22 10 4
1.43 10 3
8.80 10 5
8.01 10 4
1.59 10 3
8.80 10 4
8.70 10 4
8.70 10 4
8.80 10 4
1.66 10 3
2.54 10 3
8.80 10 4
2.46 10 3
4.21 10 3
8.80 10 4
3.25 10 3
5.88 10 3
8.80 10 4
4.04 10 3
7.54 10 3
8.80 10 4
4.83 10 3
9.20 10 3
8.80 10 4
5.62 10 3
1.09 10 2
8.80 10 4
6.41 10 3
1.25 10 2
8.80 10 4
7.20 10 3
1.42 10 2
8.80 10 4
7.99 10 3
1.58 10 2
8.76 10 3
8.66 10 3
8.66 10 3
9.70 10 4
1.66 10 2
2.53 10 2
1.84 10 3
2.44 10 2
4.15 10 2
2.70 10 3
3.21 10 2
5.75 10 2
3.57 10 3
3.98 10 2
7.32 10 2
4.44 10 3
4.74 10 2
8.86 10 2
8.78 10 3
5.50 10 2
1.04 10 1
8.78 10 3
6.24 10 2
1.19 10 1
8.77 10 3
6.98 10 2
1.33 10 1
8.77 10 3
7.71 10 2
1.47 10 1
In Table 6 the average differences between perfect testing results and the imperfect test
situation is given. The imperfect test situation yields higher average PFDs than with perfect
testing.
48
Imperfect testing
8.32 10 5
1.66 10 4
2.50 10 4
3.33 10 4
4.16 10 4
4.99 10 4
5.83 10 4
6.66 10 4
7.49 10 4
1.0 10 7
1.0 10 6
8.32 10 4 8.32 10 3
1.66 10 3 1.66 10 2
2.50 10 3 2.50 10 2
3.33 10 3 3.33 10 2
4.16 10 3 4.16 10 2
4.99 10 3 4.99 10 2
5.83 10 3 5.83 10 2
6.66 10 3 6.66 10 2
7.49 10 3 7.49 10 2
4.16 10 4 4.16 10 3
4.16 10 2
As illustrated in Figure 20 the impact is greater when the failure rate is getting higher. For a
component with failure rate of DU = 1.0 10 6 hours 1 and a high percentage of non-testable
failures could potentially lead to a change of the SIL rating as the result would tend to go to
the outer limit of the classification. Often the SIL is required by the client to be in the
midpoint of the range.
Unavailability
0,06000
0,05000
PFD diff
0,04000
10^-6
0,03000
10^-7
10^-8
0,02000
0,01000
0,00000
0
10
20
30
40
50
60
70
80
90
Figure 20, Unavailability for different failure rates under imperfect testing
Based on the PFD average differences given in Table 6, special care should be shown in cases
with high failure rates for the valves, while for the lower failure rates the impact is not
considered as critical if the SIL requirement is low. If imperfect tests are to be included in the
calculations, refer to Table 7 for knowing when this topic should be given attention.
49
Imperfect testing
4.16 10 4
4.16 10 3
4.16 10 2
DU = 1.0 10 8
DU = 1.0 10 7
DU = 1.0 10 6
Failure
rate valves
SIL
( hours 1 )
4
3
2
1
Red
Yellow
Green
The inclusion of imperfect test addition will not have an impact on the SILrating
As discussed at the beginning of this chapter it may be hard to assess the exact percentage
that remains untested after a proof test, hence using the imperfect test addition as proposed
in Table 7 ensures that conservative estimates are made.
Note that these imperfect test additions are done for one component only, and for different
architectures they need to be modified. For the system in question where the valves have a
1oo2 configuration, the PFDdiff shouled be modified analogous to the PDS method (2006);
50
Imperfect testing
0.5 = 25 % non-testable
1.0 = 50 % non-testable
1.5 = 75 % non-testable
2.0 = 100 % non-testable
Pimp, the proability of imperfect tests, is then calculated by the average of the contributors;
Pimp =
(M 1 + M 2 + M 3 + M 4 + M 5 )
PFDdiff
5
Pimp =
(1 + 0.5 + 1 + 1 + 1)
4.16 10 3 = 3.74 10 3
5
Using the simplified equations for calculating the PFD as described in the introduction of this
chapter, and adding the imperfect test contribution as calculated, the PFD for the system is:
[(1 ) DU ]
[(1 ) DU ]
+ DU +
+ DU +
2
3
2
3
2
DU
+ Pimp
2
51
Imperfect testing
(t)
0,0008
Unavailability
SIL 2
0,0006
PFD average
PFD avg perfect test
0,0004
0,0002
16 0
30
32 0
60
48 0
90
65 0
20
81 0
50
97 0
8
11 00
41
1 3 00
04
1 4 00
67
1 6 00
30
00
Hours
The impact on the PFD for a single component is hardly traceable. The PFD for the imperfect
test situation is PFD = 4.41 10 4 while a perfect test yields the result PFD = 4.38 10 4 and
a difference of 3.0 10 6 .
Result system test example
Using the same approach as in case B, but with a decreased PTIF value gives the following
result for the test system example:
[(1 ) DU ]
[(1 ) DU ]
+ DU +
+ DU +
2
3
2
3
2
DU
+ PTIF
2
52
Imperfect testing
Unavailability Case C2
0,002
0,0018
0,0016
0,0014
Unavailability
(t)
0,0012
SIL2
0,001
PFD average
0,0008
0,0006
0,0004
0,0002
16 0
30
32 0
60
48 0
90
65 0
20
81 0
50
97 0
8
11 00
41
1 3 00
04
1 4 00
67
1 6 00
30
00
Hours
Figure 23, Unavailability with decreasing imperfect test addition (case C2)
The equation for one component for this situation is given below;
PFD
DU NT
+ (1 ) DU PT
2
2
PFD 2 0.2
1.0 10 7 10 8760
1.0 10 7 8760
+ (1 0.2)
= 3.85 10 3
2
2
For this exact example the PFD for the imperfect test situation is PFD = 3.85 10 3 while a
perfect test yields the result PFD = 4.38 10 4 which gives a difference of 3.41 10 4 for the
component.
[(1 ) DU ]
[(1 ) DU ]
+ DU +
+ DU +
2
3
2
3
2
53
Imperfect testing
DU
+ Pimp
2
PFD
In this chapter several interpretations of imperfect testing have been proposed. The PDS
(2006) perspective on systematic failures was described in case A, while the influence of
undetectable random hardware failures was assessed in case B. Case C gave an alternative
approach to the prior ones, where the additions were assumed to reduce over time. The PFD
results from the cases are gathered in Figure 24.
1,40E-04
1,20E-04
1,00E-04
8,00E-05
6,00E-05
4,00E-05
2,00E-05
0,00E+00
Base Case
Case A
Case B
Case C1
Case C2
Case
The case B and case C2 PFD values are considerably higher than the base case, and are also
the only ones that lead to a shift of the SIL rating. The PDS approach to systematic failures in
case A and case C1 hardly has an impact on the PFD. The results illustrate the importance of
making a testable system and the impact this may have if it is not taken into consideration.
There may be several difficulties applying the methods described in this chapter. One topic
that has been debated is the challenges with quantifying the PTIF-value correctly to reflect the
hidden systematic failures. Further on should it should be developed a detailed approach for
deciding the contribution of each of the M-factors proposed in case B.
Another aspect that should be taken into consideration is that normally the PFD is based on
the average values. For increasing unavailability as shown in case B, it is a substantial
difference between the PFD for the first and the last proof test interval as illustrated by the
unavailability sketches. This makes it interesting to assess the PFD max values in addition to
the traditional PFD average. For high failure rates the imperfect test addition could
potentially lead to another decrease of the SIL rating for the system at the end of its life span.
54
Figure 25, PST impact on the PFD (Lundteigen & Rausand, 2007)
There are three basic types of partial stroke test equipment; mechanical limiting, position
control and solenoids (Summers & Zachary, 2000a):
Mechanical limiting: Requires manual interaction and visual inspection of valve
movement which is obviously not practical to incorporate in subsea systems.
55
Position control: Enables to detect how far the valve has moved. This requires
additional hardware to be installed, and a system for collecting the test information, making
the cost a major drawback.
Solenoids: The test is conducted by pulsing the solenoid, and the preset valve travel is
confirmed by a limit switch or position transmitter, allowing for automatic documentation of
test status.
The solenoid may either be integrated with the SIS, or it may be a separate PST package
(Lundteigen and Rausand, 2007). The SIS sketch in Figure 26 is an illustration of a solution
with both position control and solenoid (adapted from Beurden & Amkreutz, 2001).
Figure 26, Simple SIS with PST implementation (adapted from McCrea-Steele, 2006)
In subsea petroleum production PST has been implemented in the Kristin field for testing the
High Integrity Pressure Protection System (HIPPS) (Lundteigen and Rausand, 2007).
56
In addition to this there are some problems related to estimate the coverage when using PST.
The measuring devices used to confirm that the test was successful may introduce failures
themselves. Besides, there are not many methods that can measure with certainty that the
PST actually moved the valve, very often it is only assumed on basis of for example that the
hydraulics were bled off. These topics are discussed throughout the next chapters.
57
Figure 27, Overview of relevant failure rates (Lundteigen & Rausand, 2007)
It is necessary to estimate the coverage of PST in order to optimize the proof test intervals or
to determine if a higher SIL rating can be obtained. The PST coverage is defined as the
fraction of dangerous undetected failures detected by PST relative to the total number of
dangerous undetected failures by Lundteigen & Rausand (2007).
PST =
DU , PST
DU
DD = DC D
DU , PST = (1 DC ) PST D
DU , PT = (1 DC ) (1 PST ) D
Where DC is the diagnostic coverage (as explained in section 2.5):
DC =
DD
D
Summers (1998) put emphasis on the importance of being plant specific when the PST
coverage is assessed, as the valve and exposure environment may differ greatly from case to
case. She also claims that credit for partial stroking in the quantitative verification of a SIL
should be considered only when the process service is clean and tight shutoff is not required.
58
In Table 8 the reasons for the different failure modes are described and connected to the test
strategy they are assumed to be revealed by.
Table 8, Dangerous failure modes and test strategy for a safety gate valve (adapted from
Summers & Zachary 2000a, McCrea-Steele 2006, KOP 2002, Bak 2007 and ATV 2007)
Failure Descriptor
Failure Mode
Test
Strategy
FT/PT
PST/FT/PT
FT/PT
PST/FT/PT
PST/FT/PT
FT/PT
PST/FT/PT
FT/PT
PST/FT/PT
FT/PT
PT
FT or PST
with speed
of travel test
FT or PST
with speed
of travel test
PT (PST in
low
pressure
fields)
PT
PT
PT
It is the safety gate valve that forms the basis for the assessment of the failure modes in Table
8. The table does not cover all possible failure descriptors for all different valve designs, but
reflects the design of the valve producer ATV to a high degree (ATV, 2007). Note that it isnt
59
any reason why moving the valve 20% should lead to a build up of corrosion/scaling that
eventually would lead to a fail to close failure (Bak, 2007), hence it is not put as a failure
descriptor.
Note that as a basis for the PST coverage estimation several assumptions have been done:
Both the critical and degraded failure rates from OREDA are included in the
calculations as the practical distinction between the two may be vague (adapted from
Lundteigen & Rausand, 2007).
Whether it is FTC or FTO that is safety critical for the system is not possible to read
from the OREDA data, but for avoiding calculations with twice the contribution the
FTC rate was chosen to be used throughout the calculations.
Only the failure modes DOP, ELP, FTC and LCP are considered dangerous and
relevant for testing of a safety system, hence are only these failure modes included in
the calculations.
Since it is not much reliability data for subsea systems in particular, the available data
for topside and subsea is merged. This may not be accurate as there are different
requirements related to subsea equipment than topside. Even so, the inner
environment is considered equal for both subsea and topside valves, and the mix may
thus be justified.
Only the latest OREDA edition (2002) is utilized. It may be discussed whether the
earlier OREDA edition should be used or not since it is a continuously improvement
of the design as more operational experience is attained. Hence the old versions may
not reflect the failure rates of new equipment realistically.
Leakage in closed position can not be detected by PST as the valve needs to be fully closed.
Neither is external leakage in closed position assumed to be discovered by PST as the
pressure difference over the valve may be minimal for a high pressure field. The failure mode
itself is very unlikely to occur for valves with backseat, as the leakage is only possible during
transaction. Consequently both of these coverages are set to 0 %. It is likely that DOP can be
discovered by PST if a speed of travel detector is installed. For the failure descriptor due to
high friction the valve may delay/be prevented from fully close the PST will probably not
detect problems with closing the very last part. Regarding FTC, it may be discussed if the
assumption that the valve will continue to a closed position if it can start to move is realistic.
The coverage is estimated on the failure descriptors in Table 8 only and on basis of the
assumptions described above, a tentative estimation of the PST coverage for each failure
mode may be set to;
FTC- 80%
LCP- 0%
DOP- 90%
ELP- 0%
The PST coverage factor is estimated by collecting the relevant failure modes from OREDA
(2002) as shown in Table 9. Only the valves used for ESD, control and safety purposes is
60
chosen. Then the fraction of failures that can be discovered by PST is calculated as shown
below:
PST =
Table 9, Reliability data as basis for PST coverage estimation (adapted from Lundteigen
& Rausand, 2007)
Failure data
(OREDA 2002)
FTC
LCP
DOP
ELP
Total
1
134
2
40
0
30
1
44
4
248
61
21
10
93
34
22
69
42
49
21
23
11
20
317
67
71
71
526
Total
When only the data from the latest OREDA edition (2002) is used and with the PST coverage
for the different failure modes as estimated above, the PST coverage is 62%. With more
optimistic PST dangerous failures coverages, e.g. both FTC and DOP set to 95%, the PST
coverage is estimated to be 72%. On the contrary, a more pessimistic approach, with both
FTC and DOP set to 50%, the PST coverage is estimated to be 38%.
Summers & Zachary (2000a) proposed a PST coverage of 70% and Lundteigen & Rausand
(2007) estimated a PST coverage for the Kristin HIPPS valves of 62%, hence the result is in
about the same range as former research.
61
From the figure it can be concluded that by implementing an automatic system for
conducting the PST, many of the potential reasons for ST can be designed away.
62
DU PT
+ PST DU PST + DD DT
2
2
2
The proportion of failures that is not detected through PST is left to be discovered through
proof testing which is assumed to be perfect. This may not be a realistic assumption as
discussed in chapter 3 and is further assessed in section 4.6. A system which has
implemented PST may be influenced by several factors than normally included in PFD
calculations. Some of the simplifications and assumptions are briefly discussed in the
following.
Increase of -factor
It is likely that when one of the two valves in the same line fails due to corrosion, there is a
high probability that given the same body materials and process conditions, the other valve
will fail as well (metso, 2002).
To be considered a CCF, the failures have to be within a short time interval. To be classified
as such, the failures have to occur within the same proof test interval. Hence an extension of
the test interval might imply that the -factor should increment over time (Rausand, 2007).
Reasons for this are;
1. Longer time intervals that several components can fail within.
2. Preventive maintenance initiated because of findings on one component will not be
performed as often as with shorter proof test intervals.
If PST is used to extend the proof test intervals, this should lead to an assessment of the factor to reflect the PFD impact this may have.
SFF calculation
An argument for implementing PST is the potential improvement of the SFF and
consequently the possibility to reduce the hardware fault tolerance requirement. This is
obtained by the PST converting part of the dangerous undetected failures from the
denominator, till dangerous detected in the numerator;
SFF =
SD + SU + DD
SD + SU + DD + DU
63
64
DU = 3.0 10 7 hours 1 is chosen. The PST coverage is set to PST=60% and the test interval
PST=1 week.
Unavailability
0,0014
0,0012
0,001
(t)
Unavailability
0,0008
0,0006
PFD average
SIL 2
0,0004
0,0002
13 0
70
27 0
40
41 0
10
54 0
80
68 0
50
82 0
20
95 0
9
1 0 00
96
1 2 00
33
1 3 00
70
1 5 00
07
1 6 00
44
00
Hours
DU PT
+ PST DU PST
2
2
3.0 10 7 8760
3.0 10 7 182.5
PFD (1 0.6)
+ 0.6
= 5.42 10 4
2
2
The implementation of PST would for this case lead to a change of SIL rating from SIL 2 to
SIL 3 for the single component since the PFD is reduced from the previous PFD = 1.31 10 3
till PFD = 5.4 10 4 .
In the following PST is applied onto the case A and case B described in chapter 3.
65
Unavailability
0,0016
0,0014
0,0012
Unavailability
(t)
0,001
0,0008
PFD average
0,0006
SIL 2
0,0004
0,0002
13 0
70
27 0
40
41 0
10
54 0
80
68 0
50
82 0
20
95 0
9
10 00
96
12 00
33
13 00
70
15 00
07
16 00
44
00
Hours
DU PT
+ PST DU PST + PTIF
2
2
1.0 10 7 8760
1.0 10 7 182.5
PFD (1 0.6)
+ 0.6
+ 1.0 10 5 + 1.0 10 3 = 1.19 10 3
2
2
As the PFD when proof testing the component is PFD = 4.38 10 4 , implementing PST leads
to a worse PFD result when the PTIF-additions proposed by PDS is used. This is the case when
the PTIF is significantly greater than the failure rate. Hence, for a safety valve with adequately
low failure rate, the PDS method does not support the implementation of PST. But with two
valves in series, the PST contribution will be positive since Addition = PTIF , which also
implies that for small -values the PTIF-addition becomes very small.
66
of
component
with
failure
rate
DU = 1.0 10 7 hours 1 . The PST coverage is set to PST=60% and the test interval PST=1
week. The non-testable part is assumed to be =20% and it is assumed that the non-testable
part is equal for both the PT and PST.
Unavailability
0,004
0,0035
0,003
Unavailability
(t)
0,0025
0,002
PFD average
0,0015
SIL 2
0,001
0,0005
13 0
70
27 0
40
41 0
10
54 0
80
68 0
50
82 0
20
95 0
9
10 00
96
12 00
33
13 00
70
15 00
07
16 00
44
00
Hours
DU NT
PFD 0.2
1.0 10 7 20 8760
2
1.0 10 7 8760
1.0 10 7 182.5
= 1.90 10 3
+ (1 0.2) (1 0.6)
+ 0.6
2
2
The insertion of the non-testable part and PST leads to a change of SIL rating from SIL 3 to
SIL 2 rating since the PFD increases from PFD = 4.38 10 4 (only proof testing conducted)
till PFD = 1.90 10 3 .
The impact on the PFD of a component with failure rate DU = 1.0 10 7 hours 1 and where
20% is assumed non-testable is assessed in Table 10. The table shows the results for both the
imperfect test situation and perfect testing for diverse PST coverages and intervals.
67
Table 10, PFD related to diverse PST coverages, test intervals and (im)perfect testing
PST coverage 50 %
Without
PST
With
PST
PST = 1
week
PST = 1
month
PST = 3
months
4.36 10 4
Imperfect
test
Perfect
test
Imperfect
test
Perfect test
1.94 10 3
2.23 10 4
1.91 10 3
1.80 10 4
1.95 10 3
2.38 10 4
1.92 10 3
1.98 10 4
1.98 10 3
2.71 10 4
1.95 10 3
2.38 10 4
PST coverage 70 %
With
PST
PST = 1
week
PST = 1
month
PST = 3
months
PST coverage 60 %
PST coverage 80 %
Imperfect
test
Perfect
test
Imperfect
test
Perfect
test
1.87 10 3
1.37 10 4
1.84 10 3
9.50 10 5
1.88 10 3
1.58 10 4
1.86 10 3
1.19 10 4
1.93 10 3
2.05 10 4
1.90 10 3
1.72 10 4
Considering only the perfect test situation it can be claimed that there are relatively small
differences between the diverse PST coverages and for the different test intervals. A change of
the failure rate to DU = 1.0 10 6 hours 1 showed that it more important to conduct the PST
often, than assess the exact PST coverage. A lower test interval could potentially lead to a
change of the SIL rating of the component while a higher PST coverage hardly has an impact.
Dependent of each situation, the improved PFD by PST implementation can be obtained by
either improving the PST coverage, or shorten the PST test interval (or both).
It is clear that an imperfect test gives a greater negative impact to the PFD than the positive
PST contribution; hence the priority should be upon reducing the non-testable part. A
reduction of the non-testable part with 10% gives a greater improvement of the PFD than for
a component with 80% PST coverage and a test interval of 1 week.
68
Discussion
5 Discussion
DU = 1.0 10 6 hours 1 and above, it is more important to conduct the PST often than to
achieve a higher PST coverage. The reason is that the positive PFD impact is greater if the
tests are carried out more often than to improve the coverage by additional 10%. This makes
the difficulty deciding the exact PST coverage somewhat less important as the assumptions
taken doesnt give a great impact.
On the contrary, a reduction of the non-testable part with 10% gives a greater improvement
of the PFD than obtaining both higher PST coverage and shorter test intervals. Hence the
focus should be upon diminishing the factors that leads to unsuccessful tests.
If PST is implemented in order to extend the proof test interval, it might be necessary to
change the -factor (Rausand, 2007). As CCF are those failures that happen within the same
proof test interval, an extension of the interval could lead to more failures are classified as
CCF. Analogous for the imperfect test situation, the period for the non-testable part is larger
and it is likely that several components would fail within the same period. Hence it should be
discussed whether the -factor should be incremented to realistically reflect the PFD impact.
69
Discussion
The SIL rating is not a static measure. The PFD is greatly influenced by the operator
companies in the production lifetime. The equipment can be delivered in excellent condition
and with the opportunity to check and validate the system, but the operational philosophy is
significant. As mentioned, the operational philosophy may be to minimize the stress on the
system by not conducting the tests in a realistic manner, e.g. conducting function tests
instead of proof tests (which implies leakage test). A leakage test may lead to degradation of
the components which results in higher failure rates and consequently higher PFD. This
makes it interesting to optimize the function tests, PST and proof tests interval with the
possible degradation of the system if proof tests were conducted. For some situations it may
be worth including the imperfect test addition instead of degrade the components by proof
testing.
70
Discussion
The PFD impact of imperfect testing should be assessed also for the logic and sensors.
For illustrative reasons the focus has been only upon the safety valves in this thesis.
71
Case study
6 Case study
The topics discussed throughout the thesis have in this section been applied on a genuine
field development. The Morvin field have been chosen for this purpose. Special attention has
been given to examine the HIPPS. Up to this date, the contract for Morvin has not yet been
awarded hence has this case study been based on the initial concept studies done by AKS and
the first drafts done by Statoil.
72
Case study
Topic
Definition of safety function
Definition of functional limits
Equipment under control
(EUC)
Safe state of the function
SIL requirement
Requirement or description
Closing valves upon high pressure in production header
The function includes pressure transmitter subsea, logic,
and valves.
The EUC is defined as the flowline and riser.
Safe state of the function is when one of the valves is
closed.
SIL 3. PFD shall be less than 5.0 10 4 .
PFD allocation:
- Initiator < 35 %
- Logic < 15 %
- Final element < 50 %
Closing time shall be less than 13 seconds, including signal
polling time.
Internal leak shall be 0 kg/s at FAT
The status/position of all safety critical components in the
function shall be available at any time.
Along with the requirements directed to the physical structures, there are several other
requirements related to activities, testing and documentation. Regarding safety and
reliability analyses, the following is required as a minimum:
HAZIDs; including hazard register, according to ISO17776
HAZOPs; Hazard and Operability analyses
SAZOPs; Safety and Operability analyses
FMECAs; Failure Mode Effect and Criticality Analyses
The documentation requirements for compliance with the IEC 61508 standard are specified
as follows (Statoil, 2007b);
73
Case study
6.3 HIPPS
The HIPPS may be installed topside or subsea on a X-mas tree, manifold or pipeline end
terminal. HIPPS provides a pressure break between the subsea systems rated to full shut-in
pressure and the flowline and riser rated to a lower pressure (Patni & Davalath, 2005). An
example of a HIPPS schematic is given in Figure 32 (KOP, 2004) and as shown the HIPPS
basically consists of two safety valves in series and redundant pressure transmitters. One
HIPPS configuration is normally placed on each header on the manifold securing redundancy
for the function.
There are several advantages of implementing HIPPS subsea. Among others (adapted from
Patni & Davalath, 2005);
74
Case study
75
Case study
Sensor 1
Sensor 2
Sensor 1
Sensor 3
Sensor 1
Sensor 4
Logic unit 1
Valve 1
Sensor 2
Sensor 3
Logic unit 2
Valve 2
Sensor 2
Sensor 4
Sensor 3
Sensor 4
Figure 33, HIPPS reliability block diagram for Morvin field development
When assuming the same results of the FMEA analysis performed for the Kristin project
(AKS, 2002) the data for the HIPPS valve is as described in Table 12. The failure rate includes
all the final elements as actuator and solenoid. The sensor and logic failure data is achieved
from a topside HIPPS example in the PDS data handbook (2006).
Table 12, Morvin HIPPS Case data
Failure rate DU
Component
Sensors
(2oo4)
Logic
(1oo2)
Valve
(1oo2)
( hours 1 )
*PDS data, 2006
- factor
(PDS Data, 2006)
SFF
Test interval
(hours)
3.0 10 7 *
3%
< 60%
8760
1.0 10 7 *
2%
99%
8760
1.01 10 6
2%
60%
8760
It is assumed that the equipment complies with the SFF requirement given by Statoil.
Considering the HFT and SFF requirements in Table 2, the 2oo4 voting on the sensors
enables a SIL 3 rating, the logic enables a SIL4 rating and the final elements allow a SIL 3
rating. Hence the system complies with the SIL 3 requirements.
76
Case study
PFD SYS
PFD SYS
[(1 ) DU ]
[(1 ) DU ]
+ DU +
+ DU +
2
3
2
3
2
DU
+ PTIF
2
[(1 ) DU ]
[(1 ) DU ]
+ DU +
+ DU +
2
3
2
3
2
DU
+ Pimp
2
DU PT [(1 ) DU PT ]
DU PT
+
+
2
3
2
2
[(1 ) DU PT ]2 DU PT
+ (1 PST )
+
+ PST
3
2
5
6
5
PFD SYS = 3.94 10 + 9.01 10 + 4.56 + 4.53 6
77
3
2
5
= 9.85 10
Case study
DU PT [(1 ) DU PT ]
DU PT
+
+
+
2
3
2
2
[(1 ) DU PT ]2 DU PT
+ (1 PST )
+
3
2
0,001
0,0009
0,0008
0,0007
0,0006
0,0005
0,0004
0,0003
0,0002
0,0001
0
1. PT
2. P TIF
3. P imp
4. PST
5. All
Calculation approach
It is evident that special attention should be put upon discovering the non-testable part of the
system as this has a great impact on the PFD. Introducing PST as a manner to decrease the
PFD has practically no impact when PTIF and Pimp is included in the calculations.
78
Concluding remarks
7 Concluding remarks
SIL rating is a common requirement for subsea petroleum systems, making it interesting to
evaluate the assumptions that form basis for the calculations. The assumption that a
component is as good as new after each proof test, meaning that the unavailability for the
component is reduced to zero, has been subject for assessment. The effect of a test that is
imperfect, meaning that the unavailability is not reduced to zero, has not been discussed to a
great extent in the literature. Hence the author has aimed to define and analyze the effect of
imperfect testing.
An imperfect test was classified according to two dimensions;
1. The test does not cover all possible failures inadequate test method.
2. The test does not detect all the failures unsuccessful test.
The reasons for imperfect tests were related to the five M-factors; method, machine, milieu,
man-power and material. It has been proven that the PFD impact by imperfect tests can be
significant. While the PDS proposed PTIF value hardly makes any impact, an imperfect test
with a high proportion of non-testable failures proved the potential of changing the SIL
rating of the system. As it may be difficult to decide the exact percentage that is non-testable
for a system, a method based on the M-factors facilitating such estimation was proposed.
PST has been introduced in order to reveal failure modes which before only has been feasible
through tests that require process shutdown. A successful implementation may improve the
SIL rating of the system. The use of PST in subsea petroleum production has so far not been
common. Several of the arguments for and against implementing PST in subsea equipment
have been assessed.
A tentative PST coverage factor was set to 62 %, based on a failure mode assessment of gate
valves and OREDA data. The result is in accordance with former research. The PST coverage
for the dangerous failure modes FTC, LCP, DOP and ELP could not be justified quantitatively
as the production companies do not give such detailed information. The coverage may differ
dependent on the valve type in question, design and production environment.
It has been argued that PST leads to an increase of the ST rate, assuming that if the valve
starts to move it will continue to closed position. The likely reasons for such an event were
assessed in a Bayesian belief network and proved the need for the right equipment. New
devices such as smart positioners and digital valve controllers have been introduced for the
purpose of PST, reducing the human interference in PST and thus the reasons for ST.
PST is by some implemented in order to justify extended proof testing intervals. As CCF are
those failures that happen within the same proof test interval, an extension of the interval
could lead to more failures being classified as CCF. In such situations, it should be discussed
79
Concluding remarks
whether the -factor should be incremented to realistically reflect the PFD impact this may
have.
Another argument for implementing PST has been the opportunity to reduce the HFT since
the SFF is incremented by detecting more dangerous undetected failures and convert them to
dangerous detected failures. As PST is not fulfilling the criteria for being a diagnostic test, it
is argued that the PST should not be used to affect the SFF, and hence can not be an
argument for a reduction in the HFT (McCrea-Steele, 2006).
Especially for components with higher failure rates, from DU = 1.0 10 6 hours 1 and above,
investing in PST can be recommended. The case studies showed that achieving the exact PST
coverage was less important than the test frequency. The positive PFD impact was greater if
the tests were carried out often than if the coverage was improved by additional 10%.
On the contrary, a reduction of the non-testable part with 10% gave a greater improvement of
the PFD than if both higher PST coverage and shorter test intervals were obtained. Hence the
focus should be upon diminishing reasons for why the test should be unsuccessful.
Throughout the thesis it has become obvious that it is necessary to assess the assumptions
regarding reliability calculations more closely. The imperfect test case proved that ignoring
the estimation of non-testable failures could lead to an inaccurate PFD result. As the use of
SIS develops as one of the standard method for reducing risk in the petroleum production, it
should be highly relevant to improve the quality of these calculations.
80
REFERENCES
Articles and textbooks
Ali, R. 2004. Problems, concerns and possible solutions for testing (and diagnostic
coverage) of final control element of SIF loops. FIELDVUE Business, USA.
Bak, L. 2007. Personal communication 2nd of May 2007. Sandvika, Norway.
Beurden, I. Amkreutz, R. 2003. The effects of partial valve stroke tesing on SIL level.
exida.com.
Goble, W.M. 2003. Estimating the Common Cause Beta Factor. Exida.com, USA.
Haddon, W. Jr. 1973. Energy damage and the ten countermeasure strategies. Human
Factors, 15(4):355-66.
Hovden, J., 2003. Theory formations about the Risk Society. NoFS XV, Karlstad, Sweden.
Lundteigen, M. and Rausand, M., 2006. Assessment of Hardware Safety Integrity
Requirements. Proceedings of the 30th ESReDA Seminar. NTNU,Trondheim-Norway.
Lundteigen, M. and Rausand, M. 2007. The effect of partial stroke testing on the reliability
of safety valves. NTNU, Trondheim-Norway.
McCrea_Steele, R. 2005. Partial Stroke Testing. Implementing for the Right Reasons. Paper
at ISA EXPO 2005, Chicago.
McCrea_Steele, R. 2006. Partial Stroke Testing. The Good, the Bad and the Ugly. Premier
Consulting Services, USA.
Metso automation, 2002. Comparison between testing methodologies to achieve the
required SIL level. Application report 2726/01/02.
Rausand, M. 2007. Personal communication 10th of April 2007. Trondheim, Norway.
Rausand, M. and Hyland, A., 2004. System Reliability Theory. Second edition. John Wiley
& Sons, Inc., Hoboken, New Jersey.
Sangesland, S. 2007. Drilling and completion of subsea wells. Course compendia, NTNU,
Trondheim.
Sanguineti, L. & Sanguineti, E. 2007. Personal communication 4th of May 2007. ATV,
Colico, Italy.
81
82
OLF 070, 2004. Application of IEC 61508 and IEC 61511 in the Norwegian Petroleum
Industry. OLF, Rev. 02, 10.29.2004.
OREDA, 2002. Offshore Reliability Data. 4th Edition. SINTEF, Trondheim-Norway.
OREDA, 2007. OREDA homepage. Internet: http://www.sintef.no/static/tl/projects/oreda/
PDS Data Handbook, 2006. Reliability Data for Safety Instrumented Systems. 2006
Edition. SINTEF, Trondheim-Norway.
PDS Method Handbook, 2006. Reliability Prediction Method for Safety Instrumented
Systems. 2006 Edition. SINTEF, Trondheim-Norway.
Other documents
AKS, 2007. Dalia X-mas Tree. Internal document Aker Kvrner Subsea.
KOP, 2005. IEC 61508 / IEC 61511 - SIL kurs, modul 1. Presentation for training purposes.
KOP, 2004. There is something about Kristin. Presentation at Society for Underwater
Technology.
KOP, 2003. Safety requirement specification, system 18, HIPPS. Doc. Number 22-KC000502
KOP, 2002. FMEA report HIPPS valve and actuator. Doc. Number C074-KOP-S-RA-0002
Ring-O, 2007. Single acting FSC actuator for 51/8 gate valve. Internal document Ring-O
Valves, Colico - Italy.
Statoil, 2007a. Morvin Safety Requirement Specification. Doc. number TR2250.
Statoil, 2007b. IEC61508/61511 Compliance. Doc.number TR2249
Statoil, 2007c. Delivery of subsea production system. Scope of work. Frame agreement
no.4600004645.
83
ANNEX A, XMT
Horizontal XMT
(Sangesland,2007)
Conventional XMT
(Sangesland,2007)
84