Understanding Mil STD 882 - System Safety
Understanding Mil STD 882 - System Safety
Kanchan Biswas
Former Director (Aircraft), CEMILAC, DRDO
(+ 91 9448376835, [email protected])
Abstract
A system may be defined as a composite entity consisting of hardware, software, integrated through
some interfaces to be used following a defined operating procedure by the operators in a supporting
environment to achieve any defined task or complete some specific mission. The system safety
studies are primarily aimed at eliminating system malfunctions or failures to prevent loss of human
life and properties. While it is not possible to eliminate system failures totally, the system safety
engineering aims at identifying the level of risk and trying to mitigate the risk to an acceptable level.
Mil – STD 882 indicates the procedure for classifying the severity of hazard as well as the
acceptable probability of occurrences for each category of hazard. System safety engineers cannot
wait for an accident or defect investigation report to improve safety, rather all safety provisions are
to be complied with during the design conceptual stages. System safety engineering is an
application of engineering and management principles, criteria and techniques to optimize safety
within the constraint of operational effectiveness, time and cost throughout all phases of the system
life cycles.
Key Words: System Safety, Mil 882, Hazard, Risk, Probability of occurrences, Risk
Mitigations, RTCA DO 178, RTCA DO 278.
a) Identify the risks using hazard analysis techniques as early as possible in the system
life cycle
b) Develop options to mitigate – Eliminate, control or avoid the hazards
c) Provide for timely resolution of the hazard
d) Implement the best strategy
e) Control the hazard through a closed-loop system.
System safety is not only a function of engineering but also an integral part of the
management activates. Participation of management is required to ensure timely
identification and resolution of the hazards. The management as well as the engineering
tasks are identified in the Mil standard 882 [2].
Risk = ∑ f x s;
t =t 1
= Product of f and s, summed for all the hazards during the defined time interval
(t 2- t 1).
Further, we may define an Accident as a Hazard triggered into an unsafe condition which
may result in a Mishap – an unpleasant event that may result into death, injury,
occupational illness or damage to equipment or property.
Preliminary Hazard Analysis (PHA) is done at the design concept stage. This is done early
so that initial risks assessment and their mitigation makes the design concept sound. Lot of
time and money may be wasted if the design concepts are to be changed later. The
hazards may be associated with hardware, software, operational procedures, system
interfaces, environmental and health hazards.
There is no clear-cut procedure laid down in literature on how to identify hazards. However,
PHA is basically a brainstorming technique, hence some kind of organized approach helps
in conducting this analysis.
Mil – STD – 882 [2] defines the basis for classification of Hazards based on the extent of
loss or damage (see later). For every class or criticality of Hazard, Mil 882 defines the
acceptance level of probability of occurrences. The probability of occurrence should be
inverse of the criticality level of the Hazard; in the sense that more critical hazard should
have lower probability of occurrences.
In figure 3 (b), if the reliability of each of the components F x 1, F x 2.., F x 4 is 0.9. The reliability
of the sub-systems at level 2, can be calculated as:
a) Sub system A and B are in ‘AND’ with their elements F x 1, F x 2.., F x 4 . Elements F x 1
and F x 2 are in ‘And’ gate, components are in parallel, therefore,
b) Sub sytems A and B are in ‘Or’ gate. They are in series, therefpre,
Remember that the defintion of ‘And’ and ‘OR’ gates in reliability analysis and FTA are
opposite. Reliability values of each components, sub system and overall system reliability is
shown in the figure 3(b).
Hazard The The Combin- Measures Post Probability If the cell After
description severity probability ation of taken to mitigation of is red, mitigation all
Study may of the of severity reduce the Measures occurrenc further cells should
include worst occurrence and risk to severity e of failure elimination be green
-Source of credible of the probability public level post all of
potential effect hazard or to (reducing mitigation mitigation
harm without failure determine either the measures must be
-Mechanism any mode qualitative severity or taken done to
by which the mitigation without the risk to the probability). reduce the
harm may be measures mitigation public. Typically risk
caused measures Mark the through
-worst cell red to design
credible indicate changes,
outcome unaccept- safety
assuming no able risk. devices,
measure warning
employed devices,
procedures
and
training.
Risk is a function of the severity of a failure (event) and its probability of occurrence. The
hazards are assigned priorities so that the catastrophic and critical ones are prevented.
The system safety standard practice for military systems as identified in the DoD System
Engineering approach, is to eliminate the hazards, where possible, ‘And’ minimize the risks,
where it is not possible to eliminate the risks. The detail procedure is identified in Mil STD
882 E, 11 May 2011 [2].
The system safety process consists of eight activity elements. The logical sequence of the
eight elements are shown in figure 4.
A hazard is defined as a real or potential condition that could lead to an unpleasant event or
series of events (i.e. mishaps) resulting in death, injury, occupational illness, damage or
loss of equipment or property, or damage to the environment.
The assessed risks are expressed as a Risk Assessment Code (RAC) which is a
combination of Risk severity category and the probability of its occurrence level. The Risk
Assessment Matrix is shown in table 6. For example, in table 6, RAC in cell 1A is High,
which has risk severity as “Catastrophic” and probability of occurrence is “Frequent”. The
risks are assigned a risk levels of High, Serious, Medium or Low for each RAC.
Note: The definitions in table 2, 3 and 4 (RAC) shall be used, unless tailored alternative definitions are
formally approved by procurement executives.
The potential risk mitigation(s) shall be identified, and the expected risk reduction(s) of the
alternative(s) shall be estimated and documented in the Hazard Tracking System (HTS).
The goal should always be to eliminate the risk, however, when a hazard cannot be
eliminated, the risk should be reduced to the lowest acceptable level within the constraints
of cost schedule and performance by applying the system design order of precedence Mil
882 standard has gone through several revisions. Released in 1960, it was revised as 882A
on 15 Aug 1979, 882B on 30 Mar 1984, 882C on 19 Jan 1993 and 882D was made on 10
Feb 2000 and last revision 882E came on 11 May 2011. There have been a few changes
happened during the revisions, notable among them are:
a) The Hazard Risk index and suggested mitigation criteria have been changed.
b) In 882E, risk mitigation goal has been redefined as to eliminate the hazard if
possible. When a hazard cannot be eliminated, the associated risk should be
reduced to the lowest acceptable level within the constraints of cost, schedule, and
performance by applying the system safety design order of precedence. However,
no quantitative value has been indicated.
c) Mil 882C had considered software as a critical item. However, no specific direction
for software safety analysis was indicated. Mil 882 E has introduced specific
Software Safety analysis.
Frequent 1 3 7 13
Probable 2 5 9 16
Occasional 4 6 11 18
Remote 8 10 14 19
Improbable 12 15 17 20
Mil STD 882D, 10 Feb 2000 uses the risk assessment matrix as shown in table 7b and
uses the same acceptance criteria as shown in table 7c.
Table 7c. Mishap risk categories and mishap risk acceptance levels (882D, App–
A).
18 – 20 Low As Directed
The assessment of risk for software controlled or software intensive systems cannot rely
solely on the risk severity and probability of occurrence like in the case of hardware. A
software (s/w) is generally application-specific, and the reliability parameters associated
with a s/w cannot be estimated in the same manner as a hardware. Therefore, another
approach will have to be used for the assessment of s/w’s contributions to system risk that
considers the potential risk severity and the degree of control the software exercises over
the hardware.
The system safety approach of Mil 882E generates a Software Safety Criticality Matrix
(SSCM) using the same severity categories as catastrophic, critical, marginal and
negligible. However, in place of the probability of occurrences, we place the Software
Control Categories (SCC) levels. The SCC defines the degree of control the software
exercises on the system. The levels are defined from 1 to 5 for the ‘Autonomous’ to ‘No
Safety Impact (NSI)’ software (see table 8). The SSCM index obtained depending on the
SCC and severity categories indicates the Level of Rigor (LOR) task category. The ‘LOR’ is
a specification which defines the depth and breadth of software analysis and verification
activities necessary to provide a significant level of confidence that a safety critical or a
safety-related software function will perform as required.
Level of Rigor
SwCI 1 Program shall perform analysis of requirements, architecture, design, and code; and
conduct in-depth safety-specific Testing.
SwCI 2 Program shall perform analysis of requirements, architecture, and design; and conduct
in-depth safety-specific testing
SwCI 3 Program shall perform analysis of requirements and architecture; and conduct in-depth
safety-specific testing.
SwCI 5 Once assessed by safety engineering as Not Safety, then no safety specific analysis or
verification is required.
The risks associated with system hazards that have software causes and controls may be
acceptable based on evidence that hazards, causes, and mitigations have been identified,
implemented, and verified in accordance with DoD customer requirements. The evidence
supports the conclusion that hazard controls provide the required level of mitigation and the
resultant risks can be accepted by the appropriate risk acceptance authority. In this regard,
software is no different from hardware and operators. If the software design does not meet
safety requirements, then there is a contribution to risk associated with inadequately
verified software hazard causes and controls. Generally, risk assessment is based on
quantitative and qualitative judgment and evidence.
Table 11 shows risk levels and how the risk can affect the system functioning.
Table 11. Software Hazard Causal Factor Risk Assessment Criteria
Risk Description of Risk Criteria
Levels A software implementation or software design defect that upon occurring during normal or
credible off-nominal operations or tests:
Medium Influences a marginal or negligible mishap, reducing the system to a single point of
failure, or
Places the system in a condition where two independent functioning interlocks or
human actions remain to preclude the potential occurrence of a catastrophic or critical
hazard
Low Influences a catastrophic or critical mishap, but where three independent functioning
interlocks or human actions remain, or
Would be a causal factor for a marginal or negligible mishap, but two independent
functioning interlocks or human actions remain.
A software degradation of a safety critical function that is not categorized as high,
serious, or medium safety risk.
A requirement that, if implemented, would negatively impact safety; however code is
implemented safely
In RTCA DO 178B [4], a software level is decided based upon its contributions to potential
failure conditions as determined by the system safety assessment process. The software
level implies that the level of effort required to show compliance with certification
requirements varies with the failure condition category. The software level definitions are:
Software life cycle data are assigned to one of two categories: Control Category 1 (CC1)
and Control Category 2 (CC2). These categories are related to the Software Configuration
Management (SCM) controls placed on the data. SCM introduces more severe control for
the category CC1 data control during each activity of the SCM process.
The SCM is involved with 13 different process objectives verifications. While CC1 data
control verifies all 13 objectives, CC2 data controls verifies only 6 of the process control
objectives.
For certification of aircraft and engine software, the certification authority considers the
software as part of the airborne system or equipment installed on the aircraft or engine; that
is, the certification authority does not approve the software as a unique, stand-alone
product. The certification authority establishes the certification basis for the aircraft or
engine in consultation with the applicant. The certification basis defines the regulations
together with any special conditions which may supplement the published regulations.
For modified aircraft or engines, the certification authority considers the impact the
modification has on the certification basis originally established for the aircraft or engine. In
some cases, the certification basis for the modification may not change from the original
certification basis; however, the original means of compliance may not be applicable for
showing that the modification complies with the certification basis and may need to be
changed.
The certification authority assesses the Plan for Software Aspects of Certification for
completeness and consistency with the means of compliance that was agreed upon to
satisfy the certification basis. The certification authority satisfies itself that the software
level(s) proposed by the applicant is consistent with the outputs of the system safety
assessment process and other system life cycle data. The certification authority informs the
applicant of issues with the proposed software plans that need to be satisfied prior to
certification authority agreement.
Prior to certification, the certification authority determines that the aircraft or engine
(including the software aspects of its systems or equipment) complies with the certification
basis. For the software, this is accomplished by reviewing the Software Accomplishment
Summary and evidence of compliance. The certification authority uses the Software
Accomplishment Summary as an overview for the software aspects of certification.
The Software Verification Plan is a description of the verification procedures to satisfy the
software verification process objectives. These procedures may vary by software level as
defined in Annex A-1 of DO 178B. The plan should include:
d) Verification environment – Equipment for testing, analysis tools and testing guidelines.
The verification methods for objectives for A, B, C and D levels of Software and their
outputs are identified in the annexure Tables A-1 to A-10. The data control category (CC1
and CC2) is also indicated for each objective verification process (including the sub-
objectives) for different software levels.
These annexures provide guidelines for the software life cycle process objectives and
outputs described in this document by software level. These tables reference the objectives
and outputs of the software life cycle processes previously described in this document.
b. The independence by software level of the software life cycle process activities
applicable to satisfy that process's objectives.
c. The control category by software level for the software life cycle data produced by
the software life cycle process activities (subsection 7.3).
DO 178 Table No. Applicable Verification Method, Output and Data Control Category.
A-1: S/w Planning The table has 7 objectives, and the outputs are various plans like design,
Process (7 objectives) development, certification plans, coding, SCM, SQA etc. Verification – V for A, B,
and C and AD for D. Data control CC1 and CC2 based on the data.
A-2: S/w Development There are 7 objectives, and the outputs are S/w design requirement, design
Process (7 objectives) description, source and executable codes. All verifications are V category and
Data control - CC1 & CC2
A-3: Verification of OP There are 7 objectives, and the outputs are S/w verification results. IV (for
of S/w Requirement Accuracy & compliance of high-level S/w, and Accuracy of Algorithm – A & B level
Process (7 objectives) S/w). V for C & D level with some D levels are AD. Data control mostly CC2.
A-4: Verification of OP There are 13 objectives, and the outputs are S/w verification results. IV for A & B
of S/W Design level for S/w Compliance, Accuracy, architecture and portioning integrity. C level
Process (13 s/w are all V and D level s/w are of AD category. Data controls are CC2.
objectives)
A-5: Verification of OP There are 7 objectives, and the outputs are S/w verification results. IV for A level
of S/w Coding & for Source Code Compliance to low level requirement, s/w architecture, accuracy
Integration Processes & Consistency, for rest of the objectives V category. IV for source code accuracy
(7 objectives) and consistency of B level s/w, rest V category. C levels are V category, and all D
level s/w are AD category. Data control is CC2
A-6: Testing of There are 5 objectives, and the outputs are S/w verification cases and Procedures
Outputs of Integration and S/w verification results. IV for A level S/w for two objectives – object code
Process (5 objectives) Compliance to low level requirement, and robustness. All other objectives are V
category. IV for B level s/w for objective code compliance to low level requirement
and rest of the objectives V category. All C level S/w are V category and D levels
are V and AD category. Data control is CC1 for requirements and CC2 for rest.
A-7: Verification of There are 8 objectives, and the outputs are S/w verification cases and Procedures
Verification Process and S/w verification results. IV for all objectives for A level S/w. IV for 3 objectives
Results (8 objective of B level s/w – Test coverage for s/w structures, and V for rest. C and D level S/w
cases) are V and AD category. Data control is CC2.
A-8: S/W There are 6 objectives, and the outputs are SCM Records and S/w Life cycle
Configuration environment Configuration Index. All 6 objectives for A, B, C and D level s/w are
4.7 Comparison of Software Safety – Mil 882, RTCA DO 178 and DO 278
In the aviation industry the Unmanned Aerial System (UAS) contains Ground Control
Centre and the Datalink which are essential part of the UAS. These being ground based
components; the software elements of these components may be designed as per DO 278
standard.
DO-278 provides guidelines to produce software for ground-based avionics systems and
equipment that performs its intended function with a level of confidence in safety. The
guidelines are in the form of:
Objectives of software life cycle processes
Description of activities and design considerations for achieving these objectives
Description of the evidence that indicate that the objectives have been satisfied.
The document discusses those aspects of certification that pertain to the production of
software for ground-based avionics systems and used in CNS or ATM equipment.
A comparison of the Software Safety/Assurance Levels amongst Mil 882, DO 178 and DO-
278 is shown in table 12.
Table 12. Comparison of Software Assurance levels: Mil 882, DO 178 and DO-278
5. MiL-STD-882E Tasks
Mil-STD-882 defines system safety tasks to be performed. The 100-series tasks apply to
management. The 200-series tasks apply to analysis. The 300-series tasks apply to
evaluation. The 400-series tasks apply to verification. These tasks can be selectively
applied to fit a tailored system safety effort. Each desired task shall be specifically called
out in a contract because the task descriptions do not include requirements for any other
tasks.
b. The task description describes the work a contractor shall perform if the task is placed
on contract. When preparing proposals, the contractor may recommend inclusion of
additional tasks or deletion of specified tasks with supporting rationale for each
addition/deletion.
c. The details to be specified in each task description lists specific information, additions,
modifications, deletions, or options to the requirements of the task that should be
considered when requiring a task.
a) Task 101 is to integrate “Hazard Identification and Mitigation” into the Department
of Defense Acquisition Systems Engineering process using the system safety
methodology.
d) Task 104 is to support reviews, certifications, boards, and audits performed by or for
the Government.
e) Task 105 is to provide support to designated program office Integrated Product Teams
(IPTs) or Working Groups (WGs).
f) Task 106 is to establish and maintain a closed-loop Hazard Tracking System (HTS).
g) Task 107 is to submit periodic progress reports summarizing the pertinent hazard
management and engineering activities that occurred during the reporting period.
f) Task 206 is to perform and document an “Operating and Support Hazard Analysis”
(O&SHA) to identify and assess hazards introduced by operational and support
activities and procedures; and to evaluate the adequacy of operational and support
g) Task 207 is to perform and document a “Health Hazard Analysis” (HHA) to identify
human health hazards, to evaluate proposed hazardous materials and processes
using such materials, and to propose measures to eliminate the hazards or reduce the
associated risks when the hazards cannot be eliminated.
c) Task 303 is to participate in the “Test and Evaluation” (T&E) process to evaluate the
system, verify and validate risk mitigation measures, and to manage risks for test
events.
d) Task 304 is to perform and document the application of the system safety process
described in Section 4 of this Standard to “Engineering Change Proposals” (ECPs);
change notices; deficiency reports; mishaps; and requests for deviations, waivers and
related change documentation.
a) Task 401 is to define and perform tests and demonstrations or use other verification
methods on safety-significant hardware, software, and procedures to “verify
compliance with safety requirements”.
6. Conclusion
Defining and following a process for assessing risk associated with hazards is critical to the
success of a program, particularly as systems are combined into more complex System of
Systems (SoS). These SoS often involve systems developed under disparate development
and safety programs and may require interfaces with other Service (Army, Navy/Marines,
and Air Force) or DoD agency systems. These other SoS stakeholders may have their own
safety processes for determining the acceptability of systems to interface with theirs.
References
1. Dev G Raheja, ‘Assurance Technologies – Principles and Practices’, McGraw Hill
Publication.