Guideline Systems-Safety-Assurance
Guideline Systems-Safety-Assurance
Guideline
Systems Safety Assurance
Notice to users
This RISSB product has been developed using input from rail experts from across the rail
industry and represents good practice for the industry. The reliance upon or manner of use of
this RISSB product is the sole responsibility of the user who is to assess whether it meets their
organisation’s operational environment and risk profile.
Document control
Identification
Document title Version Date
Document history
Publication version Effective date Page(s) affected Reason for and extent of change(s)
Copyright
© RISSB
All rights are reserved. No part of this work may be reproduced or copied in any form or by any means, electronic or
mechanical, including photocopying, without the written permission of RISSB, unless otherwise permitted under the
Copyright Act 1968.
Table of contents
1 Introduction.................................................................................................................... 3
1.1 Introduction .............................................................................................................. 3
1.2 Aim and purpose ...................................................................................................... 3
1.3 Scope ...................................................................................................................... 3
1.4 Who this guideline applies to .................................................................................... 3
1.5 How to use this guidance ......................................................................................... 4
1.6 Definitions and abbreviations ................................................................................... 4
2 System safety assurance: Requirements..................................................................... 10
2.1 Why SSA? ............................................................................................................. 10
2.2 SSA considerations ................................................................................................ 11
2.3 Systems safety assurance process outline ............................................................. 11
3 System safety assurance: Organisational matters ....................................................... 14
3.1 Organisational legal status ..................................................................................... 14
3.2 Organisational responsibilities................................................................................ 14
3.3 SSA stakeholders .................................................................................................. 16
3.4 Organisational structure ......................................................................................... 17
3.5 Personnel with safety duties: Qualifications and competence ................................ 17
4 Systems safety assurance: Process details ................................................................. 19
4.1 Impact assessment ................................................................................................ 19
4.2 System safety assurance: Hazard management .................................................... 24
4.3 Independent safety assessment ............................................................................. 41
4.4 Human factors integration ...................................................................................... 43
4.5 Applying SIL methods in a SFAIRP environment ................................................... 44
4.6 The safety case...................................................................................................... 45
4.7 Assurance and acceptance requirements .............................................................. 47
4.8 Accreditation and variation ..................................................................................... 47
5 References .................................................................................................................. 49
Appendix A Systems lifecycle model ............................................................................... 52
Appendix B Development lifecycle model ........................................................................ 53
Appendix C SSA task allocation: Sample scoping table ................................................... 56
Appendix D Functional failure analysis (FFA) .................................................................. 57
1 Introduction
1.1 Introduction
System safety assurance (SSA) provides the necessary governance, processes and objective
evidence by which all interested parties satisfy themselves that a given product, service, system
or organisational change can be safely integrated, operated and maintained into the transport
network, so far as is reasonably practicable (SFAIRP).
1.3 Scope
This document applies to organisational, operational and asset change and provides guidance
on:
• why do SSA?
• key SSA considerations;
• important organisational matters relevant to SSA; and
• the SSA process.
This guideline outlines high-level, structured safety assurance processes that:
• can be applied throughout the change;
• can be tailored to fit the size and complexity of the change;
• ensure regulatory and legal requirements are met; and
• ensure existing standards may be applied.
The guideline provides a SSA lifecycle model to safely design, deliver, construct, commission,
operate, maintain, modify, and dispose of railway assets, systems and operations. The
guideline applies to new and modified railway infrastructure and equipment, including rolling
stock, electrical, telecom, signalling and civil infrastructure. It applies to significant changes to
operation and maintenance of existing systems. While specifically concerned with safety, it is
also relevant to assuring prevention of environmental and asset damage, cybersecurity and
reliability, availability and maintainability (RAM).
The guideline does not include the daily management of workplace safety which is covered by
WHS standards, including during construction.
• executives and senior managers in order for them to understand the requirements
of SSA management and the duty of care that which applies to an organisation; and
• designers, engineers, safety and assurance managers, project managers, contractors
and suppliers and procurement authorities who need a detailed understanding of
SSA principles in the Australian rail context.
The guideline is applicable to all sectors in the rail industry including light rail and heritage
operators.
Design (verb): a process to define the architecture, system elements, interfaces, and other
characteristics of a system or system element (ISO 15288:2015).
Design (noun): result of the process of design (ISO 15288:2015).
ETA: event tree analysis.
FFA: functional failure analysis.
FMEA: failure modes and effects analysis.
FMECA: failure modes, effects and criticality analysis.
FTA: fault tree analysis.
GSN: goal structuring notation.
HAR: hazard analysis report.
Hazard: a condition that could lead to an accident (EN 50126-1:2017).
Hazard analysis: the process of identifying hazards and analysing their causes, and the
derivation of requirements to limit the likelihood and consequences of hazards to a tolerable
level (EN 60050-821:2017 and adopted by EN 50126-1:2017).
Hazard log: a document in which hazards identified, decisions made, solutions adopted, and
their implementation status are recorded or referenced (EN 60050-821:2017 and adopted by
EN 50126-1:2017).
Hazard management: process typically involves:
• Hazard identification - Identification of reasonably foreseeable hazards.
• Assessing the hazard risks - Assessment of the level of risk associated with exposure
to, or unintentional release of, the hazard.
• Hazard mitigation and controls - Development of risk control measures and their
implementation.
• Hazard monitoring and review - Evaluation of the control measures and their
modification, if necessary, to ensure effective control of risk (Hazard Management
Procedure; Government of South Australia; April 2011 referenced in SAI Global
Effective hazard identification and management 2012).
HAZops: hazard and operability studies.
HF: human factors.
HFI: human factors integration.
IHA: interface hazard analysis.
Independent safety assessment: the independent process to determine whether the
system/product meets the specified safety requirements and to form a judgement as to
whether the system/product is fit for its intended purpose in relation to safety (EN 50126-
1:2017).
ISA: independent safety assessor.
Risk assessment: the overall process comprising a risk analysis and a risk evaluation (IEC Guide
73:2014, IEC 60050-903:2013 and adopted by EN 50126-1:2017).
Risk evaluation: the procedure based on the risk analysis to determine whether the risk SFAIRP
has been achieved (IEC Guide 53:2014, IEC 60050-903:2013 and adopted by EN 50126-1:2017).
Risk management: the systematic application of management policies, procedures and
practices to the tasks of analysing, evaluating and controlling risk (EN 50126-1:2017).
Risk tolerability: risk which is accepted [by an entity] in a given context based on the current
values of society (ISO Guide 51:1999 and adopted by EN 61508.4:2018). Note that where there
are perceived to be other reasonably practicable risk controls available, Australia’s RSNL may
require a higher level of risk reduction than such an entity’s risk tolerability.
RSNL: Rail Safety National Law.
RSO: rolling stock operator.
RTO: rail transport operator.
Safety: a measure of the degree of freedom from risk or conditions that can cause death,
physical harm (adapted from Stephenson, 1991).
Safety assurance report: forms part of the safety case which should be supported by an
independent assessment where required, a hazard log and control output data (evidence all the
chosen safety control measures/treatments have been implemented). These in turn are
supported by a wide range of evidence which the organisation uses to demonstrate safety of
rail assets.
Safety case: a documented demonstration that the product (e.g. a system, subsystem or
equipment) complies with the specified safety requirements (IEC 60050-821:2017 and adopted
by EN 50126-1:2017).
Safety critical system: a safety-related system of high criticality (Storey, 1996).
Safety integrity: probability of an E/E/PE safety-related system satisfactorily performing the
specified safety functions under all the stated conditions within a stated period of time (EN
61508.4: 2018).
Safety integrity level: discrete level (one out of a possible four), corresponding to a range of
safety integrity values, where safety integrity level 4 has the highest level of safety integrity and
safety integrity level 1 has the lowest (EN 50126-1:2017 and EN 61508.4: 2018
Safety-related system: a system by which the safety of equipment or plant is assured (50126-
1:2017 and Storey, 1996).
SAR: safety assurance report.
SDD: safety design documents
SE: systems engineering.
SiD: safety in design.
SFAIRP: so far as is reasonably practicable.
SSA is most effective to influence the safety outcomes when applied throughout the change
lifecycle. In principle, this requires system safety activities to be linked to safety requirements
development early in the change through to verification and validation of those requirements.
For asset changes, SSA links to systems engineering (SE) standards to achieve this purpose. SSA
needs to integrate with other activities through the project lifecycle such as design
management, SE, RAM and human factors integration (HFI).
• Safety case: The structured argument and supporting evidence to demonstrate that
a particular system is deemed to be safe SFAIRP.
How the
organization How the project How safety is
manages safety manages safety demonstrated
Safety
SSA guidance or
SSAMP Case and
SSA procedure
evidence
Concept design
Impact assessment
Process
Safety
argument and
evidence
To achieve these duties whenever new railway assets or systems are introduced, or existing
railway assets or systems are modified, upgraded or removed RTOs need to ensure the
following:
• Operational safety risks are identified, assessed and managed with new or modified
systems or assets, when operating as an integrated part of the transport network.
• Sufficient evidence is provided to demonstrate a safety argument that the new or
altered system or asset has achieved the following:
o Designed to ensure safety SFAIRP during its operation.
o Manufactured or constructed and transitioned into the transport network in a
manner which ensures safety SFAIRP.
o All reasonably foreseeable safety risks are eliminated or minimised SFAIRP.
• Disposal: arranging for the safe disposal of assets.
its appointed suppliers or for the principal contractor by the RTO. It also includes
independent safety assessment, independent design checks, independent
professional review and the independent validator required under EN50126:2017
clause 7.
• Level 3 (accept): At this level are the activities undertaken by the RTO or asset
accepting organisation. This includes the Tier 1 configuration management process
and a configuration change board (CCB) (see section 4.7) as well as all due diligence
activities undertaken by the RTO/asset acceptor1. Due diligence activities would
normally be conducted in a risk-based manner focussing on the highest levels of risk
to delivery, cost and safety.
1
Refer to section 3.4 for detail about configuration management.
Is the project:
• Highly complex, novel, large
or with significant
conseqeunces?
No Yes
Allocate task:
• SE developer;
• SS practitioner support;
• SE manager reviewer; and
• Chief engineer approver.
Criticality assessment:
• Define risks of project
failure; and
• Assess risks as per matrix.
Are any rated as critical?
No Yes
Non-critical Critical
The following is a list of sources of information that can be used in the impact assessment
process:
• Similar or related railway assets or systems.
• Safety.
• Reliability.
• Availability.
• Maintainability.
• Economy.
Possible May occur in the event of failure Note: The example is suitable
for an organisation undertaking
Likely Will occur in most circumstances
non-complex tasks. Where the
Consequence organisation is undertaking a
wide variety of highly critical
Insignificant Minor injury not requiring treatment
tasks it may be beneficial to
Major Non-life threatening injury classify a criticality rating (for
example, (C1 – C4) to provide
Catastrophic Life threatening injury or death
better direction to engineering
Consequence tasks.
Insignificant Major Catastrophic
• Approach to allocating and demonstrating safety integrity levels (SIL) /functional safety
integrity requirements.
• Monitoring and control of the safety program (including audit program with
internal/external auditing).
• Liaison with the ONRSR and other key stakeholders.
• The format of the safety assurance report suite of documents that will be prepared and
the authorities who will authorise the relevant supporting documents.
• Configuration management.
• Quality management.
• Safety management of subsystems.
• Type approval process.
• ISA tasking and management.
• Plan for certification, accreditation and acceptance.
SSA tasks identified within the SSAMP should be aligned to systems engineering tasks
undertaken within the project, enabling deliverables to support project assurance tasks.
The SSAMP should be a ‘living’ document throughout the SSA program and is therefore open to
update and review.
Assess Re-assess
Identify Refine Propose SFAIRP
likelihood & likelihood & Document
hazard causes controls assessment
consequence consequence
A hazard is “a condition that could lead to an accident”, (EN 50126-1:2017). It is not an internal
fault, nor an external event, which may be potential causes, nor it the accident itself, nor is it
the measure of risk. Hazards should be described with respect to the system under
investigation.
Assess Re-assess
Identify Refine Propose SFAIRP
likelihood & likelihood & Document
hazard causes controls assessment
consequence consequence
Preliminary hazard
identification (PHI)
Derived
PHL
requriements
PHI Output
completeness of the resultant hazard log. While it is possible for the PHI to be blended with the
PHA (see the following section) using system techniques such as hazard and operability study
(HAZop), it is recommended that these processes are separate.
PHI Output: The following documents should be retained as supporting evidence to
development of SARs and for auditing/control of the SSA process:
• Preliminary hazard list.
• Initial hazard log.
• PHI records including an attendance list, meeting minutes and any subsequent
action items.
Assess Re-assess
Identify Refine Propose SFAIRP
likelihood & likelihood & Document
hazard causes controls assessment
consequence consequence
Preliminary hazard assessment (PHA)
Derived
Hazard log PHAR
requriements
PHA Output
The PHA should include identification of hazards which relate to complete system behaviour,
considering hazards associated with:
• all operational modes and circumstances;
• degraded function;
• loss, inappropriate activation and incorrect application of each system function; and
Assess Re-assess
Identify Refine Propose SFAIRP
likelihood & likelihood & Document
hazard causes controls assessment
consequence consequence
System hazard assessment (SHA)
System design
SFAIRP
Hazard log SHAR control
statements
identification
SHA Output
The SHA should always be undertaken via workshop with all applicable stakeholders and SMEs
present in order to support the SFAIRP assessment, with all decisions documented to provide
evidence of the assessment process.
For SHA a systematic method should be used to analyse systematic failures, such as hazard and
operability studies (HAZops, refer ISO 61882:2016 - Hazard and operability studies (HAZOP
studies). Application guide), fault tree analysis (FTA, refer ISO 61025:2007 Fault tree analysis
(FTA) and failure mode effects analysis (FMEA, refer ISO 60812:2006 Analysis techniques for
system reliability. Procedure for failure mode and effects analysis [FMEA]) sometimes
combined the consideration of criticality – FME(C)A. HAZops can also be adapted computer
applications by considering additional flows associated with data flows and interconnections
and in this case the acronym becomes CHAZops.
It can often be difficult to decide what sort of techniques to use. Table 2 shows some of the
indicative techniques that can be used. The choice of technique(s) should be based on the
purpose of the analysis - known/unknown causes and consequences.
Table 2 - How to determine the technique to use based on purpose of the analysis
There are four different types of analysis and a number of tools, some sit across each
combination of known/unknown cause and consequence:
• When there are unknown causes and unknown consequences, the approach needs
to be exploratory in nature and HAZop is a good tool, HAZop is suitable for PHI and
PHA, but it works best in SHA.
• When there are known causes and known consequences, PHA would have already
been done, and so FME(C)A and ETA are good tools to use in this verification area.
ETA is a good tool in PHA.
• When there are known consequences and unknown causes, some PHA would have
already been done to identify consequences, but the causes are unknown in terms
of what components are going to lead to those hazards. In this causal analysis area
FTA is a good tool.
• When there are unknown consequences and known causes, experience with the
component may tell if they will fail in a certain way, e.g. a switch has two failure
modes – stuck on and stuck off. In other more complex components, there is a set
of failure rates associated with them. If the causes are known but not the
consequences, then in the consequence analysis area FMEA and ETA are good tools.
Reliability block diagrams (RBD) can be used to assess the reliability of the system. The RBDs of
the function of the system can be developed in either parallel or series:
• Series means that both of them have to be operating correctly for the system to be
operating correctly.
• Parallel means only one of them needs to be operating correctly for the system to
be operating correctly.
Having developed the system as an RBD a reliability figure can be given to each component in
the system, then using probability theory the reliability of the system as a whole can be
calculated.
SILs have been discussed where they were studies on high-level functions. The process is
repeated at a low-level component level.
SHA Output: The following documents should be retained as supporting evidence to
development of SARs and for auditing/control of the SSA process:
• Updated hazard log.
• SHA records including an attendance list, meeting minutes and any subsequent
action items.
• System hazard analysis report (SHAR).
The SHAR should include:
• identification of hazard causes of the system, including what the components are
that fail or even the combination of failing components;
• further justification of the preliminary and system risk and integrity level
assignments made;
• details of risk assessments carried out to date;
• an updated set of safety requirements as component level and safety requirements
are looked at;
• SILs assignation in the same way as discussed for high-level functions, e.g. how
much integrity these components are going to need to have; and
• risk apportionment to various components so riskiest and those that have to be
most reliable can be identified.
It is important to note that even if the SHA process determines that the project is not one of
highest safety criticality, a record should be kept as evidence for due diligence process.
Note: the risk assessment techniques mentioned here are the most common used for PHA and
SHA, however there are many more techniques available in:
• SA/SNZ HB 436:2013 Risk management guidelines – Companion to ISO 3100:2009,
and
• ISO 31010:2009 Risk management – Risk assessment techniques.
While the following sections provide less detail on these activities, the process and output are
similar to the PHA and SHA discussed above.
Assess Re-assess
Identify Refine Propose SFAIRP
likelihood & likelihood & Document
hazard causes controls assessment
consequence consequence
Subsystem hazard assessment (SSHA)
Subsystem
Hazard log SSHAR design control
identification
SSHA Output
To avoid repetition, where it is identified that the hazard affects the system as a whole, these
hazards should be managed within the SHA process.
Assess Re-assess
Identify Refine Propose SFAIRP
likelihood & likelihood & Document
hazard causes controls assessment
consequence consequence
Interface hazard assessment (IHA)
Identification of controls
Updated SFAIRP
Hazard log IHAR provided by other
statements
systems / entities
IHA Output
As the system may not be able to affect the design of other systems or entities, the IHA
typically allocates responsibilities to other parties and provides recommendations/SRACs
necessary to ensure that the system is safe SFAIRP.
Assess Re-assess
Identify Refine Propose SFAIRP
likelihood & likelihood & Document
hazard causes controls assessment
consequence consequence
Operating and support hazard assessment (OSHA)
Administrative
Updated SFAIRP
Hazard log OSHAR control
statements
identification
OSHA Output
Change of
SSA plan
project scope
External events
Safety review
Audit
SSA tasks Safety review
deliverables
Qualitative Quantitative
• Relies mainly upon expert judgment and past • Relies upon failure.
experience.
• Requires less effort/time to complete; and • Can provide greater precision; and
• Does not require significant data to undertake. • Clear demonstration that hazards meet the tolerable
hazard rate (THR).
During the PHI and PHA the system is typically conceptual, due to the lack of data. Qualitative
methods are better used when identifying the PHL. During later design stages (such SSHA, SHA
and IHA) there is more detailed data available and quantitative assessments may be completed.
When considering the use of qualitative methods, considerations should include:
• hazards which may lead to a major or catastrophic consequence may require a
greater level of analysis;
• their appropriateness to highly novel systems or products; these may require a
quantitative approach due to a lack of prior experience/knowledge; and
• SILs required and that the approach is capable of arriving at THRs.
The quantitative approach should only be considered if a higher level of confidence is required,
it is generally not justified unless:
• the significance of the decision to be made is substantial;
• the data to support the analysis is available and reliable; and
• the decision makers are able to use quantitative results.
If these requirements are not satisfied then qualitative analysis is typically more appropriate,
noting that where critical risks exist these may be assessed qualitatively to support the hazard
assessment.
adapted from IEC 60050-192:2015). The inter-linking of railway RAMS elements, reliability,
availability, and maintainability and safety is shown in Figure 14.
Attainment of in-service safety and availability targets can only be achieved by meeting all
reliability and maintainability requirements and controlling the ongoing, long-term,
maintenance and operational activities and the system environment.
Failures in a system, operating within the bounds of an application and environment, will have
some effect on the behaviour of the system. All failures adversely affect the system reliability
whereas only some specific failures will have an adverse effect on safety within the particular
application. These links are shown in Figure 15.
firmware), processes, people, information, techniques, facilities, services and other support
elements (INCOSE, 2015).
Before undertaking any SSA work, the system boundaries need to be defined because safety is
a whole of system issue. Then thought needs to go into possible failures, i.e. the cessation of
required function or performance of undesired function/s, within the system boundary. This
can cause the system to enter into a hazardous state, which when interfacing with co-effectors
(i.e. an external event within the environment in which a system operates) can cause an
accident.
4.2.6.5 Software
Railway technology employs software-intensive systems, where many system functions are
implemented in software, for applications at all levels of safety criticality.
Writing software for safety-critical and safety-related applications is different from business
applications. With software, it is easy to introduce unmanageable complexity and
unpredictable component behaviour. The behaviour of software is not continuous; behaviour
can change radically with small changes in input values. Failures are systematic, rather than
random, due to latent errors in the design. Multiple redundant systems are vulnerable to
common failure mode.
Safety development is about improving the correctness of the software design, rather than the
reliability of the equipment. The objective is to produce software with zero faults. The more
rigorous the methods used, the more likely the goal will be achieved. Standards for safety-
related systems recommend or require methods for software at different levels of criticality.
For example, EN 50128:2011, Annex A has tables of recommended practises addressing these
issues, at different SILs.
RTOs and application developers should determine, justify and uphold an appropriate software
development standard.
Given the stringent requirements and additional costs of safety-related software, developers
should consider restricting safety functions to smaller, more readily verified components, while
other components undertake the bulk of functionality.
Commercial, off-the shelf software (COTS) should not be used without due certification and
consideration of the safety functions of the application. This includes, for example, regular
business operating systems which are not suitable for real-time systems at the highest levels of
criticality. Similar considerations apply to firmware, custom digital electronic hardware etc,
which also encapsulate complex design logic.
Detail
Risk Identification Formal risk identification activities are discussed in detail at section 4.2 and 4.10,
however hazards may be identified at any stage throughout the system lifecycle.
Regardless of their source, hazards are to be assessed, controlled and accepted in
the same manner. Potential risk sources may be:
• SME recommendation;
• User/operator recommendation;
• Incident reports; and
• Design process.
Refine Causes Once risks have been identified it is important to review the causal events to validate
that the risk exists and to identify other means by which the risk may eventuate.
Prior to continuing to risk assessment, it is recommended that the list of risks is
reviewed to ensure that each risk is unique to avoid duplication of effort.
Assess Likelihood & Risks are to be assessed prior to identification of any control actions to provide a
Consequences baseline of risks.
Propose Controls Controls should be identified and assessed through hazard assessment working
groups, further detail of control assessment is included below.
Re-assess Likelihood Once controls have been identified the likelihood and consequences are to be re-
& Consequences assessed to define the status of the hazard.
SFAIRP Assessment Each hazard is to be assessed within working groups to make the claim that they
are minimised SFAIRP. Further detail of SFAIRP assessment is included below
• Hazard control.
• Damage limitation.
There needs to be differing control strategies depending on whether the failures are random or
systematic, i.e.
• Random failures (associated with hardware component failures – gather right data
may be able to predict probability of failure in a given time) can be mitigated by:
o using the most reliable components available, noting that manufacturers will
often supply known failure rates in the product specification;
o applying redundancy to achieve overall reliability;
o performing preventative maintenance to replace components before faults
arise; and
o executing on-line or off-line diagnostic checks.
In theory systematic failures (associated with software faults – not random so not easy to
predict their impact of system reliability) can be eliminated, but in practice it is too costly,
therefore, instead, focus effort on the most critical areas by:
• identifying safety requirements using hazard analysis.
• assessing risk in system and operational context, then look to:
o elimination or reduction of errors using quality development processes by
verifying compliance with safety requirements and
o integrating and test against safety requirements.
hazard
Administrative
Change the way people
Least effective
Controls work
Figure 16 – The hierarchy of controls (National Institute for Occupational Safety and Health, DART, USA May 2018)
All controls should be identified and discussed through design review and hazard assessment
workshops. The hazard log should contain sufficient information to:
• identify the control;
• outline how the control affects the hazard;
• identify the responsible party for that control;
• demonstrate if the control is proposed, implemented or rejected; and
• where applicable justify why a control is rejected.
Controls need to be applied until there are no more reasonably practicable controls available
that will reduce the risk any further, with the rationale recorded for rejecting any controls not
implemented in the hazard log. The assertion that a hazard has been mitigated SFAIRP should
be validated during hazard assessment activities and recorded within the hazard log.
There is no prescriptive rule that can be used to determine whether or not a control is
reasonably practicable. Instead, there are a number of principles that should be considered,
including what can be done, should be done, unless it is reasonable in the circumstances to do
less. Specific reference is made to the ONRSR Guideline: Meaning of duty to ensure safety so
far as is reasonably practicable – SFAIRP and RISSB Guideline – Safe Decisions.
A potential control may generally be considered to be reasonably practicable if any one of the
following applies:
• It is required by law.
• It is readily commercially available.
• It is recommended by reputable standards, codes of practice, or guidance.
• It has been implemented in similar situations elsewhere.
To ensure this requirement is met, a documented ISA brief and plan appropriate to the scale of
the change being assessed should be created and implemented. Independent safety
assessment is an investigative activity, aimed at producing an informed judgment on the level
of safety achieved, rather than an audit of activities undertaken.
An ISA is appointed to assess that:
• safety requirements are appropriate and adequate for the planned application;
• safety management is appropriate and sufficient to ensure that SSA activities are
adequately undertaken;
• all safety related issues have been identified and are appropriately managed;
• SSA activities are being undertaken in accordance with documented processes; and
• the safety argument and evidence are complete and robust and demonstrates that
the system meets its safety objectives and is safe SFAIRP.
Independent safety assessment is a significant body of work in its own right. The ISA should be
engaged early in the system development lifecycle, with agreed objectives and assessment
plan.
The level of independence of the ISA should be determined by impact assessment. The ISA will
generally be fully independent of the supplier. The ISA should also be independent of the
procuring RTO, or at least the RTO’s project management line. This provides a buffer against
pressure to accept an unready product. To be effective, access and a good working
relationship between the ISA and supplier should be established.
This independent judgement provides an additional level of assurance to give the asset and
assurance acceptor confidence in the validity of the safety argument presented and the
development of the progressive assurance. The ISA will generally present their judgement to
the CCB of other acceptance authority at relevant assurance gates to support the decision-
making process.
An independent safety assessment is a team-based activity. It is led by a lead assessor that is
suitably qualified and experienced in system safety supported by a team of SMA with
competence in disciplines relevant to the assessment scope. For example, an ISA for a
signalling project would require a signalling SME and potentially a testing and commissioning
SME in order to ensure an appropriate assessment of the engineering arrangements and
activities are conducted. Quality and competence of staffing is important in getting best value
from an ISA.
Independent safety assessment should be conducted in a risk-based manner with the majority
of effort focussed in the areas of highest risk.
Independent safety assessment activities include safety management and other audits, reviews
of safety management and safety engineering arrangements, document assessments,
witnessing of project activities, technical interviews and vertical slice analysis.
Independent safety assessment outcomes should be reported in independent safety
assessment reports making a clear recommendation based on the judgement reached linked to
the evidence witnessed by the assessment activities.
The assessment may undertake product and process reviews as part of the assessment:
• Process-focused reviews to check how things are being done, looking for objective
evidence that evidence that the plans for safety are being followed.
• Product- focused reviews to check what is being produced, looking for objective
evidence that the safety requirements are being met.
The frequency and rigour for each type of review, and the degree of independence of the
reviewer, will depend on the extent of the risk and novelty and on how complicated the project
is.
The assessment activities should include:
• interviews with project personnel;
• examination of project documents;
• observation of normal working practices, project activities and conditions;
• re-work of parts of the safety analysis work to check accuracy, concentrating on
particularly critical areas or where the assessor has reason to suspect a problem (for
assessment only); and
• demonstrations arranged at the assessor’s request.
The output of the independent safety assessment will be an independent safety assessment
report (ISAR). The plan should be brief and should include:
• a statement of the assessment requirements, according to the assessment remit, but
taking into account any agreed amendments;
• identification of any dependencies on the project or others, such as access to project
personnel or documents;
• identification of the assessor or assessment team, including qualifications, experience
and level of independence;
• identification of individuals to be interviewed;
• management arrangements for reporting findings and reviewing, endorsing and
distributing;
• assessment reports;
• assessment timescales, including the expected date of issue of assessment reports; and
• communication channels, and the access rights of key stakeholders to ISA
documentation and personnel.
In the case of an audit the ISAR should record the evidence for compliance or non-compliance
with the plans for the SSA program activities.
HFI can have a direct impact on SSA so it is critical that relevant HFI is integrated and issues and
safety requirements are transferred into the SSA process. Evidence of HFI will form part of the
safety case.
It important that HF is integrated with SSA through the whole lifecycle, including requirements
development, hazard management, design development and verification and validation
activities. HF requirements are an input the design development stage and are typically
associated with ensuring a design meets the physical, functional or performance needs of its
identified end users. HF requirements will be verified and validated as part of the HF activities
in line with other systems engineering processes.
HFI deliverables may include:
• HF integration plan;
• HF issues register;
• HF assurance report; and
• verification/validation of derived requirements.
More detailed guidance is provided by international standards and guidelines and the following
RISSB products:
• AS/RISSB 7470:2016 Human factors integration in engineering design - General
requirements;
• Guideline - Integration of human factors in engineering design; and
• Guideline – Human factors integration to the project lifecycle (currently under
development).
HF incorporates a number of different domains and resource areas including cognitive
psychology, physical sciences and ergonomics, organisation design, human-machine interface
design/useability and human reliability assessments. When HF advice is required it is important
to use a person with an appropriate competency for the required task.
the lack of harmonisation between functional safety standards and ensure all definitions are
clearly defined within the SSAMP and approved prior to use.
Due to the potential sensitivity associated with establishing THRs amongst some stakeholders,
and the potential for significant cost and schedule impacts if they are not agreed when required
by the project life cycle, the SSAMP should either clearly articulate any THRs to be applied in
the project, or else clearly identify the authority, procedure and schedule for their
determination.
The SAR is a rigorous argument of why the system is safe SFAIRP and should communicate a
clear, comprehensive and defensible argument that a system is acceptably safe to operate in a
particular context. The SAR should be seen as a roadmap whereby the contextual and
supporting evidence is analysed, and meeting the safety requirements is demonstrated, i.e. it
provides traceability to all the relevant safety-related information. Preparation of the SAR is
started early in the process and many other documented outcomes flow into it. It should
include:
• the safety argument – a high level summary with appropriate sign-off that controls are
in place that make risks SFAIRP;
• safety requirements and objectives – records overall goals and strategies;
• contextual information - this is helpful for all the intended audiences, but in particular
for auditors and evaluators. It should include:
o a specification of the safety requirements;
o standards, codes and other criteria;
o project and development constraints; and
o supporting documents – these provide the building blocks upon which the
SAR is built and contains the outputs of processes such as PHI, PHL and SHA.
A SAR will normally include a vast amount of data that can be easily lost if not structured
appropriately. Most SARs are developed using word processing and spreadsheets, and the
documents and the audit trail become cumbersome. More recently computer tools have been
developed to manage the process though visibility and consistency of information.
Goal Structuring Notation (GSN) is a graphical method which may be used to document and
communicate the safety argument within a SAR. GSN has a number of advantages over
traditional text-based argument formats, in that it can be readily understood by stakeholders
who are not safety professionals and provides an explicit structure to the safety argument
which can make it easier to establish the ‘completeness’ of the argument. Further guidance on
GSN is available in the GSN Community Standard V1, (November 2011).
Regardless of the above, the SAR still needs to be structured to be able to manage the
information that gives the safety justification. A compounding issue is the input of multiple
disciplines, i.e. software, hardware, analogue electronics, electrical engineering, mechanical
engineering, pneumatics, hydraulics, HF and psychology. As such gathering and representing
views on interactions and safety issues is complex and difficult.
At each stage of the SAR development the ‘proof’ in a safety assurance report must be carefully
justified and assumptions made explicit, without this it is flawed. The production of the safety
assurance report in a coherent argument is one of the most difficult and demanding aspects of
creating a safety-critical system.
Change is inevitable and so the safety assurance report should be seen as a ‘living-breathing’
document and changes need to be managed in the safety assurance report for, but not limited
to: user requirements, design solutions, acceptance of risk. The challenge is to efficiently
maintain consistency, completeness and correctness of the safety assurance report in the
presence of such change.
It is important that underlying assumptions, constraints, and dependencies relating to safety
cases and risk assessments are explicitly articulated, as they provide critical contextual
information that needs to be considered when evaluating the applicability of any assessment,
or the acceptability of a reported risk.
These three elements can be described as follows:
Assumptions:
• a statement that can’t be proven but is believed to be true (and is expected to remain
true for the life of the system without further management); and
• the associated risk assessment cannot be relied upon unless the assumption is actually
true.
Constraints:
• a limiting condition that is accepted on the operation of the system;
• it is the boundary condition for the safety argument. If we operate the system outside
of a constraint, the safety argument may no longer be valid; and
• constraints are something the system operator will need to be mindful of for the life of
the system; but
• constraints are not really suitable for conditions that require active management,
monitoring and ongoing maintenance – those conditions should be implemented as
controls in the hazard log.
Dependencies:
• an action or activity that has not yet been finished but will generally be completed
before the system enters service;
• closure of the dependency may result in further amendments or additions to the
risk assessment; and
• the risk assessment therefore can’t be considered finalised until all dependencies
have been closed out.
All assumptions, constraints, and dependencies should be clearly described in the SAR, and
formally accepted by the relevant RTO prior to a system commencing operation.
The purpose of accreditation or approval for a variation by the ONRSR is to demonstrate that a
RTO has the competence and capacity to manage safety risks associated with its railway
operations by implementing its SMS and to safely manage changes to its operations.
Although accreditation comes at the end of the project, planning for it is important, including
robust verification and validation iterations, need to be commenced at an early stage. The
process involves convincing an independent body of the matters identified in this guideline and
it would be wise to ensure early liaison with that body to ensure communication channels are
opened to foster a common understanding of all parties in the process.
This early liaison will define what standards are to be used as criteria documents, any variations
and any internal criteria or other good practice that can be considered to demonstrate
appropriate safety and integrity levels moving into operation and maintenance phase of the
project and beyond.
As part of the SSAMP a verification plan needs to be developed and approved by any regulatory
bodies who will be involved in the accreditation process. The verification plan should contain
elements from the SSAMP to give enough context and then also the agreed criteria (including
standards), development methods to be used and the documents to be provides as part of the
certification. Should there be any deviations from standards then a justification also needs to
be documented in the verification plan. This plan needs to be developed early in the process
and signed by all parties and doing such should alleviate any unnecessary and expensive
rework. However, that’s not to say that as the project unfolds the verification plan will not be
updated as change is generally inevitable in large complex system development that could
impact on certification.
Having previously said accreditation comes at the end of the project, i.e. a signed certificate,
the process is continuous throughout the project with data and documents provided at key
stage gates. This will then trigger reviews with the accrediting body with advice given on levels
of satisfaction given by the certifying body at those stage gates.
It should be noted that accreditation may be conditional, i.e. it will impose certain operational
constraints.
5 References
The reference material below has been drawn on to create this guideline or is a handy further
reference on the subject of systems safety assurance.
A Guide to Hazard and Operability Studies, Chemical Industries Association, 1992
AS 61508.4: 2018, Functional safety of electrical/electronic/programmable electronic safety-
related systems - Definitions and abbreviations
AS/RISSB 7470:2016 Human factors integration in engineering design - General requirements
AS/RISSB 7659:2012 Rail equipment type approval
BS 5760: Part 5 1991, Reliability of systems, equipment and components: Part 5 Guide to failure
modes, effects and criticality analysis
Def(Aust)5679, Issue 2, 2008, Safety engineering for defence systems
Design and safety assessment of critical systems, Bozzano and Villafiorita, 2011
EN 50126-1:2017, Railway Applications - The specification and demonstration of reliability,
availability, maintainability and safety (RAMS) - Part 1: Generic RAMS process
EN 50128: 2011, Railway applications. Communication, signalling and processing systems.
Software for railway control and protection systems
EN 50128:2011, Railway applications. Communication, signalling and processing systems.
Software for railway control and protection systems
EN 50129:2003, Railway applications - Communication, signalling and processing systems -
Safety related electronic systems for signalling
EN 60812:2006, Analysis techniques for system reliability. Procedure for failure mode and
effects analysis (FMEA)
EN 61882:2016, Hazard and operability studies (HAZOP studies). Application guide
EN 62502:2011, Analysis techniques for dependability. Event tree analysis (ETA
Engineering a safer world, Leveson, 2009
Engineering requirements, Hull et al, 2002
Engineering Safety Management, issue 4, ‘Yellow Book 4’, note withdrawn
GSN Community Standard V1, November 2011
Hazard Analysis Techniques for System Safety, Ericson, 2005
Hazard Management Procedure; Government of South Australia; April 2011 referenced in SAI
Global Effective hazard identification and management 2012
HB 436:2013: risk management guidelines - Companion to AS/NZS ISO 31000:2009
Human factors in systems engineering, Chapanis, 1996
IEC 60050-192:2015, International electrotechnical vocabulary - Part 192: Dependability
Figure 18 is a systems lifecycle model from EN 50126 that provides a structure for planning,
managing, controlling and monitoring all aspects of a system from concept to decommissioning.
The top-down branch (left side) is generally called ‘development’ and is a refining process
ending with the manufacturing of system components. The bottom-up branch (right side) is
related to the assembly, the installation, the hand-over and then the operation and
maintenance of the whole system.
As can be seen verification is required at each stage to verify that “demonstrate that the
requirements of each lifecycle phase have been fulfilled” (EN 50126-1:2017), and the inputs are
then used to the validation process. The validation process at stage 4 assures “that system
requirements (including RAMS requirements) have been properly specified applying the
requirements defined in this standard and any additional specific requirements defined by
applicable legal framework” (EN 50126-1: 2017). The validation process at stage 9 assures
“that the system under consideration meets the specified requirements for the intended use or
application” (EN 50126-1: 2017).
Other lifecycle models are widespread in the use of systems safety assurance; however, a
simplified development lifecycle model is most commonly used. That said, the reliability,
availability, maintainability and safety (RAMS) process should be noted as being part of the
overall process. RAMS is a process that is used to reduce the incidence of failures and/or the
consequences throughout the lifecycle, and thus minimise the residual risk resulting from these
errors.
Figure 19 is an adapted development lifecycle model and depicts the overarching system and
lifecycle process in the front square-corner boxes and the deliverables in the rounded-corner
boxes behind. There is a slight change in terminology when compared to the system lifecycle,
but these are the models accepted by and used in industry.
Requirements Commisioning
definition
User Operational
requirements capability
System
Completed
requirements
system
Systems Integration
architectural
design
Architectural
design Assemblies
Implementation
Component
development Components
Figure 19 is a simplified model and omits the traditional elements normally represented on top
of the ‘V’ and elsewhere:
• Verification and validation processes, because they are usually depicted in the
model as being a linear step associated with one leg of the model or towards the
end of the lifecycle process. In practice the validation processes are continuous,
and the verification process is iterative all the way through the lifecycle process,
and
support, test, manufacturing and disposal. SE considers both the business and the technical
needs of all customers with the goal of providing a quality product that meets the user needs”.
The technical process and outputs of the INCOSE 2015 approach map closely to the
development lifecycle model. There are also other aspects such as enterprise processes,
agreement processes and project-enabling processes. All these processes are collectively
known as systems engineering management. EN 15288 is also a handy reference for the SE
process.
The enterprise processes “are used to establish and evolve plans, to execute plans, to assess
actual achievement and progress against the plan and control execution through to fulfillment”,
INCOSE, 2015.
The agreement processes “define the activities necessary to achieve an agreement between two
parties”, INCOSE, 2015.
The project-enabling processes “help ensure the organisation’s capability to acquire and supply
products or services throughout the initiation, support and control projects”, INCOSE, 2015.
Author:
Reviewer:
Approver:
Note: The example above is suitable for an organisation undertaking non-complex tasks. The
SSA tasks should be modified to represent the range of SSA tasks deemed applicable for the
scope of works undertaken.
There is no reference guide for FFA although it is a well-accepted multi-industry process for
PHA.
A FFA is a system safety and reliability analysis technique. Its purpose is to identify the
consequences of system functional failures, and so identify those functional failures that are
hazardous or adversely impact availability.
FFA is very similar to FMEA, another technique used in SHA. However, FFA is done on functions
rather than components, so it is useful in very early stages of in the system development
lifecycle when the operational concept is being explored and before system architecture has
been decided. So, to complete an FFA, you need a high-level model of the functions of the
system.
Because components, technologies, etc. may not have been decided yet, detailed failure modes
are not known, but functional failure modes can (and should) be postulated. For each system
function, consider cause and effects of:
• loss of function (omission failure);
• provision of function when not required (commission failure), and
• incorrect value of function (error).
Environmental and operational contributing factors also are important to take into
consideration before appropriate hazard controls but into place.
The advantages of FFA include it:
• is a simple concept.
o can be applied at system level, before selection of specific technologies, and
o considers hypothetical failure modes, which can’t be done in HAZops dues to
the system design and guidewords.
The disadvantage is that in software, single functions may be too complex to have confidence in
the completeness of analysis, i.e. it does not have the same granularity as HAZops.
The outcomes often influence the choice of safety features, guide further risk assessment and
assist with SIL determination
If the system architecture has been defined it would generally be more appropriate to move
straight into a SHA using a failure modes and effects analysis.
Table 7 is a template for a FFA process.
Melbourne Office
Level 4, 580 Collins Street,
Melbourne, Vic 3000
Brisbane Office
Level 4, 15 Astor Terrace
Brisbane, QLD, 4000
PO Box 518
Spring Hill, QLD, 4004