A Systematic Approach For The Design of Safety Critical Systems
A Systematic Approach For The Design of Safety Critical Systems
A Systematic Approach For The Design of Safety Critical Systems
com
Volume 3, Issue 4, July-August 2014 ISSN 2278-6856
Volume 3 Issue 4 July-August, 2014 Page 27
Abstract: A brief overview of the fields that must be
considered when designing safety-critical systems is presented.
The design of safety critical systems has been adopted static
techniques to minimize error detection and fault tolerance.
This paper specifies basic design approach by identifying the
basic components of a safety critical computer system mishap
causes and risk factors. Examines the design approach that
implements safety and reliability. This paper also deals with
some implementation issues.
Keywords Safety Critical System, Risk, Safety,
Mishap, Risk Mitigation, Fault tolerance .
1. INTRODUCTION
Critical systems are systems in which defects could have a
dramatic impact on human life, the environment or
significant assets. Such systems are expected to satisfy a
variety of specific qualities including reliability,
availability, security and safety [1]. A real-time system is
safety critical when its incorrect behavior can directly or
indirectly lead to a state hazardous to human life.
A safety critical system is a system where human safety is
dependent upon the correct operation of the system [6]
[10]. However, safety must always be considered with
respect to the whole system including software, computer
hardware, other electronic and electrical hardware,
mechanical hardware and operators or users not just the
software element.
Defining Safe:
The notion of safety comes when we drive a car, fly on an
airliner, or take an elevator ride [11]. In each case, we are
concerned with the threat of a mishap, which the US
Department of Defense defines as an unplanned event or
series of events that result in death, injury, occupational
illness, Damage to or loss of equipment or property, or
damage to the environment.
The mishap risk assesses the impact of a mishap in terms
of two primary concerns: its potential severity and the
probability of its occurrence [13]. For example, an airliner
crash would affect an individual more severely than an
automobile fender-bender, but it rarely happens. This
assessment captures the important principle that systems
such as cars, airliners, and nuclear plants are never
absolutely safe. It also provides a design principle: Given
our current knowledge, we can never eliminate the
possibility of a mishap in a safety-critical system; we can
only reduce the risk that it will occur [7].
Risk reduction increases the system cost. In some
applications such as in nuclear energy, safety dominates
the total system cost. When creating a safe system by
minimizing cost forces us to compromise to the extent that
Weexpend resources to reduce mishap risk, but only to a
level considered generally acceptable.
2. BASIC SAFETY DESIGN APPROACH
Typically, Any computer systemwhether its a fly-by-
wire aircraft controller, an industrial robot, a radiation
therapy machine, or an automotive antiskid system
contains five primary components [13]:
The application is the physical entity that the system
monitors and controls. Sometimes developers refer to an
application as a process. Typical applications include an
aircraft in flight, a robotic arm, a human patient, and an
automobile brake.The sensor converts an applications
measured physical property into a corresponding
electrical signal for input into the computer. Developers
sometimes refer to sensors as field instrumentation.
Typical sensors include accelerometers, pressure
transducers, and strain gauges.The effector converts an
electrical signal from the computers output to a
corresponding physical action that controls an
applications function. Developers sometimes call an effect
as an actuator or final element. Typical effectors include
motors, valves, brake mechanisms and pumps. The
operator is thehuman or humans who monitor and activate
the computer system in real time [13]. Typical operators
include an airplane pilot, plant operator, and medical
technician. The computer consists of the hardware and
software that use sensors and effectors to monitor and
control the application in real time. The computer comes
in many forms, such as a single board controller,
programmable logic controller, airborne flight computer,
or system on a chip.Many computer systems, such as those
used for industrial supervisory control and data
acquisition; consist of complex networks built from these
basic components.
Mishap Causes:
In the basic computer system, developers fully define the
application, including all hardware, software and operator
functions that are not safety related [9]. Because the basic
computer system employs no safety features, it probably
will exhibit an unacceptably high level of mishap risk.
When this occurs, solving the design problem requires
modifying the operator, computer, sensor, and effector
components to create a new system that will meet an
A Systematic Approach for the Design of Safety
Critical Systems
Ch.Dinesh
1
1
Sr.Asst.Professor,
Department of CSE,
Dadi Institute of Engineering & Technology,NH-5,
Anakapalle-531002,VISAKHAPATNAM
Web Site: www.ijettcs.org Email: [email protected], [email protected]
Volume 3, Issue 4, July-August 2014 ISSN 2278-6856
Volume 3 Issue 4 July-August, 2014 Page 28
acceptable level of mishap risk.The design solution begins
with the question, how can this basic computer system fail
and precipitate a mishap? The key element connecting a
failure in the basic system to a subsequent mishap is the
hazard, defined as any real or potential condition that can
cause
injury, illness, or death to personnel;
damage to or loss of a system, equipment, or property;
damage to the environment.
Hazard examples include loss of flight control, nuclear
core cooling, or the presence of toxic material or natural
gas. All such hazards reside in the application. Thus,
system design focuses first on the application component
of the system to identify its hazards. Then designers must
have their attention to the operator, sensor, computer, and
effector components.To determine how these components
can fail and cause a mishap, the designers perform a
failure-modes analysis to discover all possible failure
sources in each component. These include random
hardware failures, manufacturing defects, programming
faults, environmental stresses, design errors, and
maintenance mistakes. These analyses provide information
for use in establishing a connection between all possible
component failure modes and mishaps, as Figure 1shows.
With this analytical background in place ,actual design
can begin.
Mishap Risk Mitigation Measures
For any given the system having a high risk of mishap,
design attention turns to modifying it to mitigatethis risk.
We can do this in three ways:
1) Improve component reliability and quality
2) Incorporate internal safety and warning devices
3) Incorporate external safety devices
Figure 2 shows how and where applying these mishap-
risk-mitigation measures can alleviate the computer system
mishap causes shown in Figure 1.
Improving reliability and quality involves two measures:
improving component reliability and exercising quality
measures. Reliability improvement seeks to reduce the
probability of component failure, which in turn will
Reduce mishap probability [8]. A widely used and effective
approach for improving reliability employs redundant
hardware and software components. Redesign can remove
component reliability problems. Other sources of
component failure such as procedural deficiencies ,
personnel error are difficult to find.
Although reliability and quality measures can reduce
mishap risk, they normally will not lower it to an
acceptable level because component failures will still
occur. If the project requires additional risk mitigation
steps ,internal safety devices used as defense. Even after
designers have taken these measures, system failure still
continues, resulting in mishaps. Finally external safety
devices are used as last line of defense against these
residual failures [6]. External safety devices range from
simple physical containment through computer based
safety instrumented systems. To achieve effective mishap
risk mitigation ,developers usually apply all three of these
mitigation measures to create a layered approach to system
protection.In addition ,risk mitigation efforts must be
distributed evenly across the system s sensor, effector,
computer and operator components because single failure
in any of the part of the system can make the aggregate
mishap risk totally unacceptable.
3. EVALUATING SAFETY CRITICAL
COMPUTER SYSTEMS
The design of any safety critical system must be as simple
as possible ,taking no unnecessary risks.Software point of
view ,this usually involves minimizing the use of interrupts
and minimizing the use of concurrency within the
software.Ideally,a safety critical system requiring a high
integrity level would have no interrupts and only one
task.However,this is not achievable in practice.There are
two distinct philosophies for the specification and design
of safety critical systems[2].
To specify and design a "perfect" system, which cannot go
wrong because there are no faults in it, and to prove that
there are no faults in it.