Project Risk Management
Project Risk Management
Author:
Jørn Vatn
1 Introduction 5
1.1 About this compendium . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
1.2 Definitions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5
2 Risk Management 8
2.1 Project objectives and criteria . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 8
2.2 Risk identification . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.3 Structuring and modelling of risk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.3.1 Model for project execution time/schedule modelling . . . . . . . . . . . . . . 11
2.3.2 Cost modelling . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.3.3 Uncertainty in schedule and cost modelling . . . . . . . . . . . . . . . . . . . 11
2.4 Risk elements for follow up: Risk and opportunity register . . . . . . . . . . . . . . . 12
2.5 Correction and control . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.6 Collection and analysis of experience - learning . . . . . . . . . . . . . . . . . . . . . 14
3 Probability theory 22
3.1 Basic probability notation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.1.1 Event . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.1.2 Probability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 22
3.1.3 Probability and Kolmogorov’s axioms . . . . . . . . . . . . . . . . . . . . . . 23
3.1.4 The law of total probability . . . . . . . . . . . . . . . . . . . . . . . . . . . . 24
3.1.5 Bayes theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 25
3.1.6 Stochastic variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 26
3.2 Common probability distributions . . . . . . . . . . . . . . . . . . . . . . . . . . . . 29
3.2.1 The normal distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 30
3.2.2 The exponential distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.2.3 The Weibull distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 31
3.2.4 The gamma distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 32
3.2.5 The inverted gamma distribution . . . . . . . . . . . . . . . . . . . . . . . . . 32
3.2.6 The lognormal distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3.2.7 The binomial distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3.2.8 The Poisson distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 33
3.2.9 The inverse-Gauss distribution . . . . . . . . . . . . . . . . . . . . . . . . . . 34
3.2.10 The triangular distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
3.2.11 The PERT distribution . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 35
3.3 Assessment of parameters in parametric distributions . . . . . . . . . . . . . . . . . . 36
3.4 Distribution of sums, products and maximum values . . . . . . . . . . . . . . . . . . 37
1
3.4.1 Distribution of sums . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 37
3.4.2 Distribution of a product . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 38
3.4.3 Distribution of maximum values . . . . . . . . . . . . . . . . . . . . . . . . . 38
4 Schedule 41
4.1 Critical Path Method (CPM) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
4.2 Program Evaluation and Review Technique (PERT) . . . . . . . . . . . . . . . . . . 43
4.3 Successive schedule planning (SSP) . . . . . . . . . . . . . . . . . . . . . . . . . . . . 43
4.4 Monte Carlo simulation (MCS) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 46
4.5 Penalty for default . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 48
4.6 Event uncertainty in the schedule model . . . . . . . . . . . . . . . . . . . . . . . . . 49
4.7 Updating the model as we get more information . . . . . . . . . . . . . . . . . . . . . 49
4.8 Examples of advanced schedule modelling . . . . . . . . . . . . . . . . . . . . . . . . 50
7 Parameter estimation 74
7.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
7.2 The MLE principle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 74
7.3 Method of moments – PERT distribution . . . . . . . . . . . . . . . . . . . . . . . . 75
7.4 The LS principle . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 76
7.5 Bayesian meothds . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 79
8 Expert judgements 81
8.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.1.1 Purpose . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.1.2 Extent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.1.3 Use . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 81
8.2 General theory . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
8.2.1 History . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 83
8.2.2 Which types of evaluations/estimates do we consider as expert judgements? . 84
8.2.3 Why is there a need for expert judgements? . . . . . . . . . . . . . . . . . . . 86
8.2.4 Who are the experts? . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 88
8.2.5 How should the expert judgement be carried out? . . . . . . . . . . . . . . . 89
8.2.6 Terms and notions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 89
2
8.3 Check list for expert judgement exercises . . . . . . . . . . . . . . . . . . . . . . . . . 92
8.3.1 Checklist for phase I: The preparation phase . . . . . . . . . . . . . . . . . . 92
8.3.2 Checklist for phase II: The elicitation phase . . . . . . . . . . . . . . . . . . . 93
8.3.3 Checklist for phase III: The calculation phase . . . . . . . . . . . . . . . . . . 94
8.4 Calculation aspects . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 95
8.4.1 Criterion for performing calibration . . . . . . . . . . . . . . . . . . . . . . . 95
8.4.2 Method for performing calibration . . . . . . . . . . . . . . . . . . . . . . . . 96
8.4.3 Weighting of experts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 96
8.4.4 Standard weighting model - Experts only . . . . . . . . . . . . . . . . . . . . 98
8.4.5 Transforming percentiles . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 98
8.4.6 Standard weighting model - Experts and data . . . . . . . . . . . . . . . . . . 99
8.5 Worked example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 100
3
11.3.1 Worked example . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 136
11.4 Stochastic programming . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
11.4.1 Introduction . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 138
11.4.2 Discretization . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
11.4.3 The Value of the Stochastic Solution . . . . . . . . . . . . . . . . . . . . . . . 141
11.4.4 Expected value of perfect information . . . . . . . . . . . . . . . . . . . . . . 143
11.4.5 Scenario building . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 143
11.4.6 How to perform discretization? . . . . . . . . . . . . . . . . . . . . . . . . . . 144
Bibliography 144
Index 149
4
Chapter 1
Introduction
• Risk identification
• Risk structuring
• Risk follow up
We will also discuss elements related to decision analysis where risk is involved, use of life cycle
cost and life cycle profit models and methods to assess the numerical values of parameters used in
the risk models.
The course compendium comprises a large number of exercises, and it is recommended to do
most of the exercises in order to get a good understanding of the topics and methods described. A
separate MS Excel program, pRisk.xlsm has been developed in order to assist numerical calculations
and to conduct Monte Carlo simulations.
1.2 Definitions
Aleatory uncertainty
Variation of quantities in a population. We sometimes use the word variability rather than aleatory
uncertainty.
Epistemic uncertainty
Lack of knowledge about the “world”, and observable quantities in particular.
Dependency
The relation between the sequence of the activities in a project.
Observable quantity
A quantity expressing a state of the “world”, i.e., a quantity of the physical reality or nature, that is
unknown at the time of the analysis but will, if the system being analysed is actually implemented,
5
take some value in the future, and possibly become known.
Parameter
We use the term parameter in two ways in this report. The main use of a parameter is that it
is a quantity that is a part of the risk analysis models, and for which we assign numerical values.
The more academic definition of a parameter used in a probability statement about an observable
quantity, X, is that a parameter is a construct where the value of the parameter is the limiting value
where we are not able to saturate our understanding about the observable quantity X whatsoever
new information we could get hold of.
Parameter estimate
The numeric value we assess to a parameter.
Probability
A measure of uncertainty regarding the occurrence of an event.
Risk
Risk is uncertainty regarding occurrence and severity of future events. In project risk management
focus is on undesired events, whereas economists also include opportunities as part of risk. To
quantify risk three elements are introduced: < e, p, S >. p is used as a probability measure of the
occurrence of an event, say e. S represents the severity of the event. Note that S is a multidimen-
sional random quantity, covering several dimensions like personnel safety, environmental impacts,
material damages, project delays, extra costs, etc. Since there is more than one event to treat, i is
used as an index to run through all relevant events. An operational definition of risk is thus the
set of all relevant triplets: R = {< ei , pi , Si >}.
Risk picture
A set of undesired events, the causes and factors that may contribute to the event, the possible
consequences of the event with corresponding influencing factors, and uncertainties related to all
these issues.
Risk acceptance
A decision to accept a risk.
Schedule
A plan which specifies the start and finalisation point of times for the activities in a project.
Stochastic dependency
Two or more stochastic variables are (stochastically) dependent if the expectation of one stochastic
variable depends on the value of one or more of the other stochastic variables.
Stochastic variable
A stochastic variable, or random quantity, is a quantity for which we do not know the value it will
take. However, we could state statistical properties of the variable or make probability statement
about the value of the quantity.
6
Uncertainty
Lack of knowledge about the performance of a system, and observable quantities in particular.
7
Chapter 2
Risk Management
– Risk analysis (“Systematic use of available information to identify hazards and to esti-
mate the risk to individuals or populations, property or the environment”), and
– Risk evaluation (“Process in which judgements are made on the tolerability of the risk
on the basis of risk analysis and taking into account factors such as socio-economic and
environmental aspects”)
8
some severe consequences are defined, e.g. accident with fatalities. Then we try to set an upper
limit for the probability of these consequences that could be accepted, i.e., we could not accept
higher probabilities in any situations. Further these probabilities could only be accepted if risk
reduction is not possible, or the cost of risk reduction is very high.
In recent years it has been a discussion in the risk analysis society whether it is fruitful or not
to use risk acceptance criteria according to the principles above. It is argued that very often risk
acceptance criteria are set arbitrary, and these do not necessarily support the overall best solutions.
Therefore, it could be more fruitful to use some kind of risk evaluation criteria, rather than strict
acceptance criteria.
In project risk management we could establish acceptance criteria related to two types of events:
• Events with severe consequences related to project costs, project quality, project duration, or
even termination of the project.
In this course we will have main focus on the project costs and the duration of the project. Note
that both project cost and project duration are stochastic variables and not events. Thus it is not
possible to establish acceptance criteria to project cost or duration directly. Basically, there are
three types of numeric values we could introduce in relation to such stochastic variables describing
the project:
1. Target. The target expresses our ambitions in the project. The target shall be something
we are striving at, and it should be possible to reach the target. It is possible to introduce
(internal) bonuses, or other rewards in order to reach the targets in a project.
2. Expectation. The expectations are the value the stochastic variables will achieve in the long
run, or our expectation about the outcome. The expectation is less ambitious than the target.
The expectation will in a realistic way account for hazards, and threats and conditions which
often contribute to the fact that the targets are not met.
3. Commitment. The commitments are values related to the stochastic variables which are
regulated in agreements and contracts. For example it could be stated in the contract that a
new bridge shall be completed within a given date. If we are not able to fulfil the commitments,
this will usually result in economical consequences, for example penalties for defaults, or in
the worst case cancelling of the contract.
Problem 2.1
Discuss targets, expectations and commitments related to a new railway track between two big
cities in Norway. ♢
We sometimes also want to discuss the uncertainty in e.g. the project costs. In Section 2.3.3 we
have discussed the uncertainty concept in relation to project duration and costs.
9
3. And if it goes wrong, how serious is it (the consequences)?
With respect to risk identification it is the first question we will answer. Since risk addresses all three
questions, it would have been better to use terms like ‘hazards-’ and ‘threats identification’ rather
than the term ‘risk identification’, but the common practice is to use the term ‘risk identification
when it comes to identification of what could go wrong. The risk identification could be addressed
from different angles:
An undesired event is an event which might occur, e.g., a large water leakage in a tunnel. A scenario
is a description of a imagined sequence or chain of events, e.g., we have a water leakage, and we are
not able to stop this leakage with ordinary tightening medium due to the possible environmental
aspects which is not clarified at the moment. Further the green movement is also likely to enter
the scene in this case. A hazard is typically related to energies, poisonous media etc, and if they
are released this will result in an accident or a severe event. A threat is a wider term than hazard,
and we include also aspects as “wrong” method applied, “lack of competence and experience”. The
term threat is also very often used in connection with security problems, e.g., sabotage, terrorism,
and vandalism.
Problem 2.2
List examples of “undesired events”, “scenarios”, “hazards” and “threats” in relation to building a
new railway track between two major cities in Norway. ♢
There exist several methods that could be used in order to identify undesired events and threats,
e.g.:
• Preliminary Hazard Analysis (PHA). PHA is used to establish threats in an early phase of
a project. The method will usually require some project breakdown, e.g Work Breakdown
Structure (WBS) or Cost Breakdown Structure (CBS), project phases or similar. A detailed
project description is usually not available at this moment.
• Task analysis (TA) and Hazard and Operability Study (HAZOP) are used on a more detailed
level where we have knowledge about the various tasks.
• Use of experience data means that we try to identify events and threats based on systematic
analysis of experience from the past, i.e., what have gone wrong in earlier projects.
• Checklists. Checklists exist on different levels, and could either be used in a separate analysis,
or as an aid in another method, e.g. in a PHA. The checklists should, however, be put a way
initially since introducing checklists early will often prevent the process of revealing project
specific conditions. A checklist is primarily a list to be used in the end of the process to
ensure that the “obvious” elements have not been overlooked. In Table 2.1 a such generic list
of risk factors is provided. For each risk factor in the list, also some “cues” are listed which
could be used to assess the significance of the risk factor in a given project. Note that this
list is on a very general level, and not specific to e.g. a construction project, a tunnel project
and so on.
10
2.3 Structuring and modelling of risk
In Section 2.2 we have identified methods to identify events and threats. We now want to relate
these events and threats to the explicit models we have for project costs and project duration.
11
we split this uncertainty into three different categories, i) aleatory uncertainty (variability due to
e.g. weather conditions, labour conflicts, breakdown of machines etc.), ii) parameter or epistemic
uncertainty due to lack of knowledge about “true” parameter values, and iii) model uncertainty
due to lack of detailed, or wrong modelling. Under such a thinking, the aleatory uncertainty
could not be reduced, it is believed to be the result of the variability in the world which we
cannot control. Uncertainty in the parameters is, however, believed to be reducible by collecting
more information. Also uncertainty in the models is believed to be reducible by more detailed
modelling, and decomposition of the various elements that go into the model. It is appealing to
have a mental model where the uncertainty could be split into one part which we might not reduce
(variability), and one part which we might reduce by thorough analysis and more investigation
(increased knowledge). If we are able to demonstrate that the part of the uncertainty related
to lack of knowledge and understanding has been reduced to a sufficient degree, we could then
claim high confidence in the analysis. In some situation the owner, or the authorities put forward
requirements which could be interpreted as confidence regarding the quality of the analysis. It is
though not always clear what is meant by such a confidence level. As an example, let E(C) be the
expected cost of a project. A confidence statement could now be formulated as “The probability
that the actual project cost is within an interval E(C)±10% should at least be 70%”. It is, however,
not straight forward to document such a confidence level in a real analysis. The “Successive process
(trinnvisprosessen)” [4] is an attempt to demonstrate how to reduce the “uncertainty” in the result
to a certain level of confidence.
We also mention that Aven [12] has recently questioned such an approach where there exist
model uncertainty and parameter uncertainty, and emphasises that we in the analysis should focus
on the observable quantities which will become evident for us if the project is executed, e.g. the
costs, and that uncertainty in these quantities represent the lack of knowledge about which values
they will take in the future. This discussion is not pursuit any more in this presentation.
Problem 2.3 Discuss different type of uncertainties in a tunnel project. Propose a classification,
and identify uncertainty elements that could be reduced by i) further physical investigation, and
ii) by further analysis. Also list uncertainty elements that could not be reduced before the project
is actually executed. ♢
2.4 Risk elements for follow up: Risk and opportunity register
As risk elements and threats are identified in Section 2.2 these have to be controlled as far as
possible. It is not sufficient to identify these conditions and model them in the schedule and cost
models, we also have to mitigate the risk elements and threats. In order to ensure a systematic
follow up of risk elements and threats it is recommended to establish a so-called threat log. The
terms ‘Risk Register’ and ‘Risk & Opportunity Register’ (R&OR) is sometimes used rather than
the term ‘threat log’.
A R&OR is best managed by a database solution, for example an MS-Access database. Each
row in the database represents one risk element or threat. The fields in such a database could vary,
but the following fields seems reasonable:
• ID. An identifier is required in order to keep track of the threat in relation to the quantitative
risk models, to follow up actions et.
• Description. A description of the threat is necessary in order to understand the content of the
the problem. It could be necessary to state the immediate consequences (e.g. occupational
12
accident), but also consequences in terms of the main objectives of the project, e.g., time and
costs.
• Likelihood or probability. A judgement regarding how probable it is that the threat or the
risk condition will be released in terms of e.g. undesired or critical events.
• Impact. If possible, give a direct impact on cost and schedule if the event occurs, either by
an expected impact, or by L, M and H values.
• References to cost and schedule. In order to update the schedule and cost models it is
convenient to give an explicit reference from the R&OR into the schedule and cost models.
• Manageability. Here it is descried how the threat could be influenced, either by implementing
measures to eliminate the threat prior to it reveals it self, or measures in order to reduce the
consequences in case of the threat will materialize.
• Alert information. It is important to be aware of information that could indicate the devel-
opment of the threat before it eventually will materialize. If such information is available
we could implement relevant measures if necessary. For example it could be possible to take
ground samples at a certain cost, but utilising the information from such samples could enable
us to choose appropriate methods for tunnel penetration.
• Deadline and responsible. Identification of who is responsible for implementing and follow up
of the measure or threat, and any deadlines.
• Status. Both with respect to the threat and any measure it is valuable to specify the devel-
opment, i.e., did the treat reveal it self into undesired events with unwanted consequences,
did the measure play any positive effect etc.?
Problem 2.4
Consider threats and risk conditions in Problem 2.2 and discuss the possibilities to mitigate the
threats, and if any prior information could be available. ♢
13
2.6 Collection and analysis of experience - learning
After project execution it is valuable to systematise the information and knowledge we have achieved
during the project. The most important data sources will be
Problem 2.5
Identify other sources of information that could be relevant. ♢
It is especially two types of analyses we will conduct:
• Systematising of what went wrong, and which measures that proved to be efficient.
• Estimation of parameters which we could include in later probabilistic schedule and cost
models.
The main elements of the project risk management process is shown in Figure 2.1.
Project objectives,
requirements etc.
Learning
Experience
Risk identification
data
Project execution,
correction and controll
14
Table 2.1: Risk factors, adapted from State of
Texas: Department of Information Resources,
http://www.dir.state.tx.us/eod/qa/risk/index.htm
Risk Factors Low Risk Cues Medium Risk Cues High Risk Cues
Project Fit to Provider directly supports indirectly impacts does not support or re-
Organization provider organization one or more goals of late to provider organi-
mission and/or goals provider zation mission or goals
Customer Perception customer expects this organization is working project is mismatch
organization to provide on project in area not with prior products
this product expected by customer or services of this
organization
Work Flow little or no change to will change some aspect significantly changes
work flow or have small affect on the work flow or
work flow method of organization
Goals Conflict goals of projects within goals of projects do not goals of projects are in
the program are sup- conflict, but provide lit- conflict, either directly
portive of or compli- tle direct support or indirectly
mentary to each other
Resource Conflict projects within the pro- projects within the pro- projects within the pro-
gram share resources gram schedule resources gram often need the
without any conflict carefully to avoid con- same resources at the
flict same time (or compete
for the same budget)
Customer Conflict multiple customers of multiple customers of multiple customers of
the program have com- the program have dif- the program are trying
mon needs ferent needs, but do not to drive it in very dif-
conflict ferent directions
Leadership program has active pro- program has person or program has no leader,
gram manager who co- team responsible for or program manager
ordinates projects program, but unable to concept is not in use
spend enough time to
lead effectively
Program Manager Ex- program manager has program manager has program manager is
perience deep experience in the some experience in do- new to the domain
domain main, is able to leverage
subject matter experts
Definition of the Pro- program is well-defined, program is well-defined, program is not well-
gram with a scope that is but unlikely to be han- defined or carries con-
manageable by this or- dled by this organiza- flicting objectives in the
ganization tion scope
Continued on next page
15
Table 2.1 – Continued from previous page
Risk Factors Low Risk Cues Medium Risk Cues High Risk Cues
Political Influences no particular project has several po- project has a variety
politically-driven litically motivated de- of political influences or
choices being made cisions, such as using most decisions are made
a vendor selected for behind closed doors
political reasons, rather
than qualifications
Convenient Date date for delivery has date is being partially date is being totally
been set by reason- driven by need to meet driven by need to meet
able project commit- marketing demo, trade marketing demo, trade
ment process show, or other mandate show, or other man-
not related to technical date; little considera-
estimate tion of project team es-
timates
Use of Attractive Tech- technology selected has project is being done project is being done as
nology been in use for some in a sub-optimal way, a way to show a new
time to leverage the purchase technology or as an ex-
or development of new cuse to bring a new
technology technology into the or-
ganization
Short Term Solution project meets short project is focused on project team has been
term need without short-term solution to a explicitly directed to ig-
serious compromise to problem, with little un- nore the long term out-
long term outlook derstanding of what is look and focus on com-
needed in the long term pleting the short term
deliverable
Organization Stability little or no change in some management management or orga-
management or struc- change or reorganiza- nization structure is
ture expected tion expected continually or rapidly
changing
Organization Roles and individuals throughout individuals understand many in the organiza-
Responsibilities the organization under- their own roles and re- tion are unsure or un-
stand their own roles sponsibilities, but are aware of who is respon-
and responsibilities and unsure who is respon- sible for many of the ac-
those of others sible for work outside tivities of the organiza-
their immediate group tion
Policies and Standards development policies development policies no policies or stan-
and standards are and standards are in dards, or they are ill-
defined and carefully place, but are weak or defined and unused
followed not carefully followed
Management Support strongly committed to some commitment, not little or no support
success of project total
Executive Involvement visible and strong sup- occasional support, no visible support; no
port provides help on issues help on unresolved is-
when asked sues
Continued on next page
16
Table 2.1 – Continued from previous page
Risk Factors Low Risk Cues Medium Risk Cues High Risk Cues
Project Objectives verifiable project ob- some project objectives, no established project
jectives, reasonable re- measures may be ques- objectives or objectives
quirements tionable are not measurable
User Involvement users highly involved users play minor roles, minimal or no user in-
with project team, pro- moderate impact on volvement; little user
vide significant input system input
User Experience users highly expe- users have experience users have no previous
rienced in similar with similar projects experience with similar
projects; have specific and have needs in mind projects; unsure of how
ideas of how needs can needs can be met
be met
User Acceptance users accept concepts users accept most of users do not accept any
and details of system; concepts and details of concepts or design de-
process is in place for system; process in place tails of system
user approvals for user approvals
User Training Needs user training needs user training needs con- requirements not iden-
considered; training sidered; no training yet tified or not addressed
in progress or plan in or training plan is in de-
place velopment
User Justification user justification com- user justification pro- no satisfactory justifica-
plete, accurate, sound vided, complete with tion for system
some questions about
applicability
Project Size small, non-complex, or medium, moderate large, highly complex,
easily decomposed complexity, decompos- or not decomposable
able
Reusable Components components available components available, components identified,
and compatible with but need some revision need serious modifica-
approach tion for use
Supplied Components components available components work under components known to
and directly usable most circumstances fail in certain cases,
likely to be late, or in-
compatible with parts
of approach
Budget Size sufficient budget allo- questionable budget al- doubtful budget is suffi-
cated located cient
Budget Constraints funds allocated without some questions about allocation in doubt or
constraints availability of funds subject to change with-
out notice
Cost Controls well established, in system in place, weak in system lacking or
place areas nonexistent
Delivery Commitment stable commitment some uncertain com- unstable, fluctuating
dates mitments commitments
Continued on next page
17
Table 2.1 – Continued from previous page
Risk Factors Low Risk Cues Medium Risk Cues High Risk Cues
Development Schedule team agrees that sched- team finds one phase team agrees that two or
ule is acceptable and of the plan to have a more phases of schedule
can be met schedule that is too ag- are unlikely to be met
gressive
Requirements Stability little or no change ex- some change expected rapidly changing or no
pected to approved set against approved set agreed-upon baseline
(baseline)
Requirements Com- all completely specified some requirements in- some requirements only
pleteness and Clarity and clearly written complete or unclear in the head of the cus-
tomer
Testability product requirements parts of product hard to most of product hard to
easy to test, plans test, or minimal plan- test, or no test plans be-
underway ning being done ing made
Design Difficulty well defined interfaces; unclear how to design, interfaces not well de-
design well understood or aspects of design yet fined or controlled; sub-
to be decided ject to change
Implementation Diffi- content is reasonable content has elements content has components
culty for this team to imple- somewhat difficult for this team will find very
ment this team to implement difficult to implement
System Dependencies clearly defined depen- some elements of the no clear plan or sched-
dencies of the project system are well under- ule for how the whole
and other parts of sys- stood and planned; oth- system will come to-
tem ers are not yet compre- gether
hended
Response or other Per- readily fits boundaries operates occasionally at operates continuously
formance Factors needed; analysis has boundaries at boundary levels
been done
Customer Service Im- requires little change to requires minor changes requires major changes
pact customer service to customer service to customer service ap-
proach or offerings
Data Migration Re- little or no data to mi- much data to migrate, much data to migrate;
quired grate but good descriptions several types of data or
available of structure no good descriptions of
and use what is where
Pilot Approach pilot site (or team) pilot needs to be done only available pilot sites
available and interested with several sites (who are uncooperative or in
in participating are willing) or with one crisis mode already
who needs much help
Alternatives Analysis analysis of alternatives analysis of alternatives analysis not completed,
complete, all consid- complete, some as- not all alternatives con-
ered, assumptions veri- sumptions questionable sidered, or assumptions
fiable or alternatives not fully faulty
considered
Continued on next page
18
Table 2.1 – Continued from previous page
Risk Factors Low Risk Cues Medium Risk Cues High Risk Cues
Commitment Process changes to com- changes to commit- changes to com-
mitments in scope, ments are communi- mitments are made
content, schedule are cated to all involved without review or
reviewed and approved involvement of the
by all involved team
Quality Assurance Ap- QA system established, procedures established, no QA process or estab-
proach followed, effective but not well followed or lished procedures
effective
Development Docu- correct and available some deficiencies, but nonexistent
mentation available
Use of Defined Develop- development process process established, but no formal process used
ment Process in place, established, not followed or is inef-
effective, followed by fective
team
Early Identification of peer reviews are incor- peer reviews are used team expects to find all
Defects porated throughout sporadically defects with testing
Defect Tracking defect tracking defined, defect tracking process no process in place to
consistent, effective defined, but inconsis- track defects
tently used
Change Control for formal change control change control process no change control pro-
Work Products process in place, fol- in place, not followed or cess used
lowed, effective is ineffective
Physical Facilities little or no modification some modifications major modifications
needed needed; some existent needed, or facilities non
existent
Tools Availability in place, documented, available, validated, unvalidated, pro-
validated some development prietary or major
needed (or minimal development needed;
documentation) no documentation
Vendor Support complete support at adequate support little or no support,
reasonable price and in at contracted price, high cost, and/or poor
needed time frame reasonable response response time
time
Contract Fit contract with customer contract has some open contract has burden-
has good terms, com- issues which could in- some document require-
munication with team terrupt team work ef- ments or causes extra
is good forts work to comply
Disaster Recovery all areas following se- some security measures no security measures in
curity guidelines; data in place; backups done; place; backup lacking;
backed up; disaster re- disaster recovery con- disaster recovery not
covery system in place; sidered, but procedures considered
procedures followed lacking or not followed
Continued on next page
19
Table 2.1 – Continued from previous page
Risk Factors Low Risk Cues Medium Risk Cues High Risk Cues
PM Approach product and process planning and monitor- weak or non existent
planning and monitor- ing need enhancement planning and monitor-
ing in place ing
PM Experience PM very experienced PM has moderate expe- PM has no experience
with similar projects rience or has experience with this type of project
with different types of or is new to project
projects management
PM Authority has line management or is able to influence has little authority from
official authority that those elsewhere in the location in the organi-
enables project leader- organization, based on zation structure and lit-
ship effectiveness personal relationships tle personal power to in-
fluence decision-making
and resources
Support of the PM complete support by support by most of no visible support;
team and of manage- team, with some reser- manager in name only
ment vations
Team Member Avail- in place, little turnover available, some high turnover, not
ability expected; few inter- turnover expected; available; team spends
rupts for fire fighting some fire fighting most of time fighting
fires
Mix of Team Skills good mix of disciplines some disciplines inade- some disciplines not
quately represented represented at all
Team Communication clearly communicates team communicates rarely communicates
goals and status be- some of the information clearly within team or
tween the team and some of the time to others who need to
rest of organization be informed
Application Experience extensive experience in some experience with little or no experience
team with projects like similar projects with similar projects
this
Expertise with Applica- good background with some experience with no expertise in domain
tion Area (Domain) application domain domain in team or able in team, no availability
within development to call on experts as of experts
team needed
Experience with high experience average experience low experience
Project Tools
Experience with high experience average experience low experience
Project Process
Training of Team training plan in place, training for some areas no training plan or
training ongoing not available or training training not readily
planned for future available
Team Spirit and Atti- strongly committed to willing to do what it little or no commitment
tude success of project; coop- takes to get the job to the project; not a co-
erative done hesive team
Continued on next page
20
Table 2.1 – Continued from previous page
Risk Factors Low Risk Cues Medium Risk Cues High Risk Cues
Team Productivity all milestones met, de- milestones met, some productivity low, mile-
liverables on time, pro- delays in deliverables, stones not met, delays
ductivity high productivity acceptable in deliverables
Technology Match to technology planned for some of the planned selected technology is a
Project project is good match technology is not well- poor match to the prob-
to customers and prob- suited to the problem or lem or customer
lem customer
Technology Experience good level of experience some experience with no experience with the
of Project Team with technology the technology technology
Availability of Technol- technology experts experts available else- will need to acquire
ogy Expertise readily available where in organization help from outside the
organization
Maturity of Technology technology has been in technology is well un- technology is leading
use in the industry for derstood in the indus- edge, if not ”bleeding
quite some time try edge” in nature
Design Complexity easily maintained certain aspects difficult extremely difficult to
to maintain maintain
Support Personnel in place, experienced, missing some areas of significant discipline or
sufficient in number expertise expertise missing
Vendor Support complete support at adequate support little or no support,
reasonable price and in at contracted price, high cost, and/or poor
needed time frame reasonable response response time
time
21
Chapter 3
Probability theory
3.1.1 Event
In order to define probability, we need to work with events. Let as an example A be the event that
there is an operator error in a control room. This is written:
A = {operator error}
An event may occur, or not. We do not know the outcome in advance prior to the experiment or
a situation in the “real life”. We also use the word event to denote a set of distinct events. For
example the event that we get an even number when tossing a dice.
3.1.2 Probability
When events are defined, the probability that the event occurs is of interest. Probability is denoted
by Pr(·), i.e.
The sample space defines all possible events. As an example let A = {It is Sunday}, B = {It is
Monday}, .. , G = {It is Saturday}. The sample space is then given by S = {A, B, C, D, E, F, G}.
So-called Venn diagrams are useful when we want to analyse a subset of the sample space S. A
rectangle represents the entire sample space, and closed curves such as a circle are used to represent
22
A
subsets of the sample space as illustrated in Figure 3.1. In the following we will illustrate frequently
used combinations of events:
Union. We write A ∪ B to denote the union of A and B, i.e., the occurrence of A or B or (A and
B). Let A be the event that tossing a die results in a “six”, and B be the event that we get an odd
number of eyes. We then have A ∪ B = {1, 3, 5, 6}. S AB
Intersection. We write A ∩ B to denote the intersection of A and B, i.e. the occurrence of both
A and B. As an example, let A be the event that a project is not completed in due time, and let
B be the event that the budget limits are exceeded. A ∩ B then represent the situation that the
project is not completed in due time and the budget limits are exceeded. S AB
Disjoint events. A and B are said to be disjoint if they can not occur simultaneously, i.e. A ∩ B
= Ø = the empty set. Let A be the event that tossing a die results in a “six”, and B be the event
that we get an odd number of eyes. A and B are disjoint since they cannot occur simultaneously,
and we have A ∩ B = Ø. SA B
Complementary events. The complement of an event A is all events in the sample space S
except for A. The complement of an event is denoted by AC . Let A be the event that tossing a die A
results in an odd number of eyes. AC is then the event that we get an even number of eyes. S AC
A B
0 Pr(A) Pr(B) 1
23
2. Pr(S) = 1
3. If A1 , A2 , A3 ,... is a sequence of disjoint events we shall then have:
Pr(A1 ∪ A2 ∪ . . .) = Pr(A1 ) + Pr(A2 ) + . . .
The axioms are the basis for establishing calculation rules when dealing with probabilities, but
they do not help us in establishing numerical values for the basic probabilities Pr(A1 ), Pr(A2 ),
etc. Historically two lines of thoughts have been established, the classical (frequentiest) and the
Bayesian approach. In the classical thinking we introduce the concept of a random experiment,
where Pr(Ai ) is the relative frequency with which the event Ai occurs. The probability could then
be interpreted as a property of the experiment, or a property of the world. By letting nature reveal
itself by doing experiments, we could in principle establish all probabilities that are of interest.
Within the Bayesian framework probabilities are interpreted as subjective believe about whether
Ai will occur or not. Probabilities is then not a property of the world, but rather a measure of the
knowledge and understanding we have about a phenomenon.
Before we set up the basic rules for probability theory that we will need, we introduce the
concepts of conditional probability and independent events.
Conditional probability. Pr(A|B) denotes the conditional probability that A will occur given
that B has occurred.
Independent events. A and B are said to be independent if information about whether B has
occurred does not influence the probability that A will occur, i.e., Pr(A|B) = Pr(A).
Basic rules for probability. The following calculation rules for probability apply:
Pr(A ∪ B) = Pr(A) + Pr(B) − Pr(A ∩ B) (3.1)
Pr(A ∩ B) = Pr(A) · Pr(B) if A and B are independent (3.2)
Pr(A ) = Pr(A does not occur) = 1 − Pr(A)
C
(3.3)
Pr(A ∩ B)
Pr(A|B) = (3.4)
Pr(B)
Example 3.1
Let the two events A and B be defined by A = {It is Sunday} and B = {It is between 6 and 8 pm).
First we note that A and B are independent but not disjoint. We will find Pr(A ∩ B), Pr(A ∪ B)
and Pr(A|B)
1 2 1
Pr(A ∩ B) = Pr(A) · Pr(B) =
· =
7 24 84
1 2 1 9
Pr(A ∪ B) = Pr(A) + Pr(B) − Pr(A ∩ B) = + − =
7 24 84 42
Pr(A ∩ B) 1/84 1
Pr(A|B) = = =
Pr(B) 2/24 7
♢
24
to assess the unconditional probability. Now, we say that A1 , A2 , . . ., Ar is a division of the sample
space if the union of all Ai ’s covers the entire sample space, i.e. A1 ∪ A2 ∪ … ∪ Ar = S and the Ai ’s
are pair wise disjoint, i.e. Ai ∩ Aj = Ø for i ̸= j. An example is shown in Figure 3.3.
A2
A4
A1
A3
S
Let A1 , A2 , . . ., Ar represent a division of the sample space S, and let B be an arbitrary event in S.
The law of total probability now states:
∑
r
Pr(B) = Pr(Ai ) · Pr(B|Ai ) (3.5)
i=1
Example 3.2
A special component type is ordered from two suppliers A1 and A2 . Experience has shown that
components from supplier A1 has a defect probability of 1%, whereas components from supplier
A2 has a defect probability of 2%. In average 70% of the components are provided by supplier A1 .
Assume that all components are put on a common stock, and we are not able to trace the supplier
for a component in the stock. A component is now fetched from the stock, and we will calculate
the defect probability, Pr(B):
∑
r
Pr(B) = Pr(Ai ) · Pr(B|Ai ) = Pr(A1 ) · Pr(B|A1 ) + Pr(A2 ) · Pr(B|A2 ) =
i=1
0.7 · 0.01 + 0.3 · 0.02 = 1.3%
Pr(B|Aj ) · Pr(Aj )
Pr(Aj |B) = (3.6)
∑
r
Pr(Ai ) · Pr(B|Ai )
i=1
25
Example 3.3
We have
Pr(B|A1 ) · Pr(A1 ) 0.01 · 0.7
Pr(A1 |B) = = = 0.54
∑
r
0.013
Pr(Ai ) · Pr(B|Ai )
i=1
Thus, the probability of A1 is reduced from 0.7 to 0.54 when we know that the component is defect.
The reason for this is that components from supplier A1 are the best ones, and hence when we
know that the component was defect, it is less likely that it was from supplier A1 . ♢
To be more precise, a stochastic variable X is a real valued function that assigns a quantitative
measure to each event ei in the sample space S. Often the underlying events, ei are of little
interest. We are only interested in the stochastic variable X measured by some means. Examples
of stochastic variables are given below:
We use subscript X to emphasise the relation to the cumulative distribution function of the quantity
X. The argument (lowercase x) states which values the stochastic variable X could take, or is of
our interest. From the expression we observe that FX (x) states the probability that the random
quantity X is less or equal than (the numeric value of) x. A typical distribution function is shown
26
in Figure 3.4. Note that the distribution function is strictly increasing, and 0 ≤ FX (x) ≤ 1. From
FX (x) we can obtain the probability that X will be within a specified interval, [a,b):
FX (x)
1
x
Figure 3.4: Cumulative distribution function, FX (x)
Example 3.4
Assume that the probability distribution function of X is given by FX (x) = 1 − e−(0.01x) , and we
2
will find the probability that X is in the interval (100,200]. From Equation (3.8) we have:
Probability density function. For a continuous stochastic variable, the probability density
function is given by
d
fX (x) = FX (x) (3.9)
dx
The probability density function expresses how likely the various x-values are. Note that for
fX (x)
x
Figure 3.5: Probability density function, fX (x)
continuous random variables the probability that X will take a specific value vanishes. However,
27
the probability that X will fall into a small interval around a specific value is positive. For each
x-value given in Figure 3.5 fX (x) could be interpreted as the probability that X will fall within a
small interval around x divided by the length of this interval. Especially we have:
∫x
FX (x) = fX (u)du (3.10)
−∞
and
∫b
Pr(a < X ≤ b) = fX (x)dx (3.11)
a
fX (x)
x
a b
Figure 3.6: The shadded area equals Pr(a < X ≤ b)
Random quantities that take discrete values are said to be discretely distributed. For such
quantities we introduce the point probability for X in the point xj :
The expectation can be interpreted as the long time run average of X, if an infinite amount of
observations are available.
Median. The median of a distribution is the value m0 of the stochastic variable X such that
Pr(X ≤ m0 ) ≥ 1/2 and Pr(X ≥ m0 ) ≥ 1/2. In other words, the probability at or below m0 is at
least 1/2, and the probability at or above m0 is at least 1/2.
Mode. The mode of a distribution is the value M of the stochastic variable X such that the
probability density function, or point probability at M is higher or equal than for any other value
of the stochastic variable. We sometimes used the term ‘most likely value’ rather than mode.
28
Variance. The variance of a random quantity expresses the variation in the value X will take in
the long run. We denote the variance of X by:
∫∞
[x − E(X)]2 · fX (x) dx if X is continuous
Var(X) = −∞
∑ (3.14)
[(xj − E(X)]2 · p(xj ) if X is discrete
j
The standard deviation defines an interval which observations are likely to fall into, i.e., if 100
observations are available, we expect that approximate1 67 of these observations fall in the interval
[E(X) − SD(X), E(X) + SD(X)].
1
Precision. The precision, P , is the reciprocate of the variance, i.e. P = Var(X) .
♢
It follows easily that
29
3.2.1 The normal distribution
X is said to be normally distributed if the probability density function of X is given by:
2
1 1 − (x−µ)
fX (x) = √ e 2σ2 (3.20)
2π σ
where µ and σ are parameters that characterise the distribution. The mean and variance are given
by:
E(X) = µ
Var(X) = σ 2 (3.21)
The distribution function for X could not be written on closed from. Numerical methods are
required to find FX (x). It is convenient to introduce a standardised normal distribution for this
purpose. We say that U is standard normally distributed if it’s probability density function is given
by:
1 u2
fU (u) = ϕ(u) = √ e− 2 (3.22)
2π
We then have
∫u ∫u
1 t2
FU (u) = Φ(u) = ϕ(t)dt = √ e− 2 dt (3.23)
2π
−∞ −∞
and we observe that the distribution function of U does not contain any parameters. We therefore
only need one look-up table or function representing Φ(u). A look-up table is given in Table 3.1.
To calculate probabilities in the non-standardised normal distribution we use the following result:
Result 3.3 Let X be normally distributed with parameters µ and σ. We then have:
∫ a ( ) ( )
a−µ a−µ
xf (x)dx = µΦ − σϕ (3.25)
−∞ σ σ
where Φ() and ϕ() are the CDF and PDF for the standard normal distribution respectively. ♢
∫a ∫ (a−µ)/σ
To prove Equation (3.25) first introduce u = (x − µ)/σ yielding −∞ xf (x)dx = −∞ (σu −
µ)ϕ(u)du. The µϕ(u) part of the integral is directly found by the Φ() function whereas for the
√ ∫ (a−µ)2 /2σ2 −z
σuϕ(u) part introduce z = −u2 /2 yielding −σ/ 2π −∞ e dz. The result then follows.
30
Example 3.5 Calculation in the normal distribution
Let X be normally distributed with parameters µ = 5 and σ = 3. We will find Pr(3 < X ≤ 6). We
have:
3−µ X −µ 6−µ 3−5 6−5
Pr(3 < X ≤ 6) = Pr( < ≤ ) = Pr( <U ≤ )
( ) ( ) σ σ σ 3 3
1 −2
=Φ −Φ = Φ(0.33) − (1 − Φ(0.67)) = 0.629 − 1 + 0.749 = 0.378
3 3
Problem 3.1 Consider the example in Example 3.5, and carry out the calculation by means of
the pRisk.xlsm program. ♢
Problem 3.2 Let X be the height of men in a population, and assume X is normally distributed
with parameters µ = 181 and σ = 4. How large percentage of the population is more than 190 cm?
♢
E(X) = 1/λ
Var(X) = 1/λ2 (3.28)
Note that for the exponential distribution, X will always be greater than 0. The parameter λ
is often denoted the intensity in the distribution
Example 3.6
We will obtain the probability that X is greater than it’s expected value. We then have:
FX (x) = 1 − e−(λx)
α
(3.30)
31
and the mean and variance are given by:
[ )
1 1
E(X) = Γ +1
λ α
( ( ) ( )]
1 2 1
Var(X) = 2 Γ +1 −Γ 2
+1 (3.31)
λ α α
where Γ(·) is the gamma function. Note that in the Weibull distribution X will also always be
positive.
For non-integer values of α numerical methods are required to obtain the cumulative distribution
function. The mean and variance are given by:
α
E(X) =
λ
α
Var(X) = 2 (3.34)
λ
If we know the expectation E and the variance V in the gamma distribution, we may obtain the
parameters α and λ by: λ = E/V , and α = λ · E. The gamma distribution is often used as a prior
distribution in a Bayesian approach.
For integer values of α the gamma distribution and in particular the Erlang distribution may
be seen as a distribution for a sum of exponentially distributed stochastic variables:
32
3.2.6 The lognormal distribution
X is said to be lognormal distributed if the probability density function of X is given by:
1 1 1 − 1 2 (log x−ν )2
fX (x) = √ e 2τ (3.37)
2π τ x
We write X ∼ LN(v,τ ). The mean and variance of X is given by
1 2
E(X) = eν+ 2 τ
2 2
Var(X) = e2ν (e2 τ − eτ ) (3.38)
i) n trials are performed, and in each trial we record whether A occurs or not.
When i)-iii) is satisfied, we say that we have binomial trials. Now let X be the number of times
event A occurs in such a binomial trial. X is then a stochastic variable with a binomial distribution.
This is written X ∼ Bin(n, p).
The probability function is given by
( )
n px
Pr(X = x) = (1 − p )n−x for x = 0, 1, 2, .., n (3.39)
x
The cumulative distribution function Pr(X ≤ x) is given in statistical tables. For the binomial
distribution, expectation and variance are given by:
E(X) = np
Var(X) = np(1 − p) (3.40)
33
For the poison distribution, expectation and variance are given by:
E(X) = λ
Var(X) = λ (3.42)
It can be proved that the Poisson distribution is appropriate if the following situation applies:
Consider the occurrence of a certain event (e.g. a component failure) in an interval (a, b), and
assume the following:
1. A could occur anywhere in (a,b), and the probability that A occurs in (t, t + ∆t) is approxi-
mately equal to λ∆t, and is independent of t (∆t should be small).
2. The probability that A occurs several times in (t, t + ∆t) is approximately 0 for small values
of ∆t.
3. Let I1 og I2 be disjoint intervals in (a, b). The event A occurs within I1 is independent of if
the event A occurs in I2 .
When the criteria above are fulfilled we say we have a Poisson point process with intensity λ. The
number of occurrences (X) of A in (a, b) is then Poisson distributed with parameter λ(b − a):
Result 3.6 In a Poisson point process with parameter λ the times between the occurrence of the
event A are exponentially distributed with parameter λ.
E(T ) = µ (3.45)
3
Var(T ) = µ /λ (3.46)
3
We use the symbol T rather than the more general symbol X here since this modell is so explicitly linked to the
time.
34
3.2.10 The triangular distribution
The triangular distribution has a probability density function that comprises a triangle. The lover
left corner points out the lowest value (L), the upper right corner points out the highest value
(H). Finally, the x-value of the third corner points out the most probable value, or mode (M ).The
probability density function for the triangular distribution is given by:
{ 2(x−L)
(M −L)(H−L) if L ≤ x ≤ M
fX (x) = 2(H−x) (3.47)
(H−M )(H−L) if M ≤ x ≤ H
Problem 3.4 Consider Problem 3.3 and assume that a special building method could reduce H
from 350 to 300, leaving L and M unchanged. This will cost 2,000 Euro extra. Do a cost benefit
analysis of this option. ♢
(x − L)α1 −1 (H − x)α2 −1
fX (x) = (3.51)
B(α1 , α2 )(H − L)α1 +α2 −1
where B(·, ·) is the beta function. The cumulative distribution function is given by:
35
Bz (α1 , α2 )
FX (x) = (3.52)
B(α1 , α2 )
where Bz (·, ·) is the incomplete beta function. The mean and variance are given by:
L + 4M + H
E(X) =
6
(E(X) − L)(H − E(X))
Var(X) = (3.53)
7
Problem 3.6 Consider a situation where the unconditional distribution of the duration of a project
groundwork activity is PERT distributed with parameters L = 0.5, M = 1.5 and H = 3.5 days.
By a detailed analysis into the uncertainty of the situation we recognize that frozen soil is a major
factor to the long duration. Let B represent the event that it is frozen soil. We now make the
following assessment: Given frozen soil, the duration of the activity, T B, is PERT distributed with
parameters L = 2, M = 2.5 and H = 3.5, and if the soil is not frozen the duration of the activity,
T B C , is PERT distributed with parameters L = 0.5, M = 1 and H = 2.5. Find p = Pr(B) such
that the expectation in the conditional situation is the same as in the unconditional situation. Hint:
You may use that E(T ) = E(T B) Pr(B) + E(T B C ) Pr(B C ), see Equation ( 3.18). ♢
Problem 3.7 Make a sketch of the unconditional probability distribution function in the situation
in Problem 3.6 when the consideration of frozen soil is taken into account. ♢
Problem 3.8 Find the unconditional variance of the duration in Problem 3.6. Hint: You may use
equation (3.19). ♢
Problem 3.9 Consider again the situation in Problem 3.6, i.e. we let in the first place T ∼
PERT(L = 0.5, M = 1.5, H = 3). Also Let B represent frozen soil and Pr(B) = 0.2. We now
introduce three factors, fB , fB C and fV that relate the conditional situation to the original situation.
The parameters relevant in the conditional situation are {LB , MB , HB } and {LB C , MB C , HB C } in
the situation where B occurs, and B does not occur respectively. We now let MB = fB · M ,
LB = MB − fV · (M − L), HB = MB + fV · (H − M ), MB C = fB C · M , LB C = MB C − fV · (M − L),
and HB C = MB C + fV · (H − M ). Let fB = 1.5 and fV = 0.5. Find by an iterative procedure the
value of fB C such that the expectation of T is equal to the original expectation. Next find fV by
a similar iterative procedure such that the variance of T is equal to the original variance. ♢
36
would use expert judgement to assess the parameters, see e.g., [11] for further discussion on expert
judgement. In this presentation we will very often assume that the uncertainty in a quantity,
e.g. the duration of an activity could be described by a so-called triple estimate {L, M, H}. We
will then as a general rule assume that the corresponding parametric distribution is the PERT
distribution. We will further assume that the L value is the absolute minimum, and that the H
value is the absolute maximum the quantity could take. It is, however, important to realise that in
other presentation the L and H values are treated as lower and upper quantiles in the distribution,
and often a 90% interval is assumed. This is even the situation for the PERT distribution which is
defined for a finite domain. So if we for a given triple estimate should establish the expected value,
and the standard deviation we should be careful regarding the interpretation of the triple estimate.
(∑ n ) ∑n
Var(X1 + X2 + . . . + Xn ) = Var Xi = Var(Xi ) (3.55)
i=1 i=1
(∑ n ) √
∑n
SD Xi = [SD(Xi )]2 (3.56)
i=1 i=1
Note that Equations (3.55) and (3.56) are only valid if the x-es are stochastically independent. If
there is dependency between the x-es we need to include a covariance term, e.g., if we only have
two variables X1 and X2 we have:
37
Result 3.10 Central limit theorem
Let X1 , X2 ,…,Xn be a sequence of identical independent distributed stochastic variables with ex-
pected value µ and standard deviation σ. As n approaches infinity, the average value of the x-es
√
will asymptotically have a normal distribution with expected value µ and standard deviation σ/ n.
Similarly, the sum of the x-es will asymptotically have a normal distribution with expected value
√
nµ and standard deviation σ n. ♢
Several generalizations for finite variance exist which do not require identical distribution but
incorporate some conditions which guarantee that none of the variables exert a much larger influence
than the others. Two such conditions are the Lindeberg condition and the Lyapunov condition.
Now, as n approaches ∑infinity, the sum of the ∑
x-es will asymptotically have a normal distribution
n
with expected value i=1 E(Xi ) and variance ni=1 Var(Xi ).
Problem 3.10 Consider a project consisting of n activities that follow each other in time. Let
each activity have a PERT distribution with parameters L = 3, M = 5 and H = 10. Use the Monte
Carlo simulation procedure in the pRisk.xlsm program to find the cumulative distribution function
for the total duration of the project. Compare the result with using the Central Limit Theorem for
various values of n. How large should n be in order to give a reasonable approximation by using
the normal distribution? ♢
The results for the variance and standard deviation is more complicated, and we only present the
results for n=2.
Var(X1 X2 ) = Var(X1 )Var(X2 ) + Var(X1 ) [E(X2 )]2 + Var(X2 ) [E(X1 )]2 (3.59)
√
SD(X1 X2 ) = Var(X1 )Var(X2 ) + Var(X1 )[E(X2 )]2 + Var(X2 )[E(X1 )]2 (3.60)
Problem 3.11 Show that Equation (3.59) is correct by using the fact that Var(X) = E(X 2 ) −
[E(X)]2 . ♢
Problem 3.12 Use the program pRisk.xlsm to simulate the mean and standard deviation of the
product X1 X2 if both X1 and X2 are independent and normally distributed with expected value
10 and standard deviation 2. Compare the result with the exact result. ♢
38
In this situation we could easily obtain the distribution of the maximum of two stochastic variables,
but it is not so easy to obtain the expectation and variance. However, since the probability density
function, fY (x) is the derivative of FY (x) we find:
∫∞ ∫∞
E(Y ) = x · fY (x) dx = x · [fX1 (x)FX2 (x) + fX2 (x)FX1 (x)] dx (3.62)
−∞ −∞
∫∞
Var(Y ) = [x − E(Y )]2 · [fX1 (x)FX2 (x) + fX2 (x)FX1 (x)] dx (3.63)
−∞
Problem 3.13 Find the expectation and standard deviation of Y = max(X1 , X2 ) if X1 and X2 are
independent and normally distributed with µ1 = E(X1 ) = 10, µ2 = E(X2 ) = 7, σ1 = SD(X1 ) = 2,
and σ2 = SD(X2 ) = 3. Hint: You might use the routine for numerical integration implemented in
the pRisk.xlsm program. ♢
Problem 3.14 Consider the problem above, but now find the result by using the Monte Carlo
simulation procedure in the pRisk.xlsm program. ♢
Problem 3.15 Consider the problem above, but now find the result by using the EMax and VarMax
functions in the pRisk.xlsm program. ♢
39
Table 3.1: The Cumulative Standard Normal Distribution
∫z
1 u2
Φ(z) = Pr(Z ≤ z) = √ e− 2 du
2π
−∞
z .00 .01 .02 .03 .04 .05 .06 .07 .08 .09
0.0 .500 .504 .508 .512 .516 .520 .524 .528 .532 .536
0.1 .540 .544 .548 .552 .556 .560 .564 .567 .571 .575
0.2 .579 .583 .587 .591 .595 .599 .603 .606 .610 .614
0.3 .618 .622 .626 .629 .633 .637 .641 .644 .648 .652
0.4 .655 .659 .663 .666 .670 .674 .677 .681 .684 .688
0.5 .691 .695 .698 .702 .705 .709 .712 .716 .719 .722
0.6 .726 .729 .732 .732 .739 .742 .745 .749 .752 .755
0.7 .758 .761 .764 .767 .770 .773 .776 .779 .782 .785
0.8 .788 .791 .794 .797 .800 .802 .805 .808 .811 .813
0.9 .816 .819 .821 .824 .826 .829 .831 .834 .836 .839
1.0 .841 .844 .846 .849 .851 .853 .855 .858 .860 .862
1.1 .864 .867 .869 .871 .873 .875 .877 .879 .881 .883
1.2 .885 .887 .889 .891 .893 .894 .896 .898 .900 .901
1.3 .903 .905 .907 .908 .910 .911 .913 .915 .916 .918
1.4 .919 .921 .922 .924 .925 .926 .928 .929 .931 .932
1.5 .933 .934 .936 .937 .938 .939 .941 .942 .943 .944
1.6 .945 .946 .947 .948 .949 .951 .952 .953 .954 .954
1.7 .955 .956 .957 .958 .959 .960 .961 .962 .962 .963
1.8 .964 .965 .966 .966 .967 .968 .969 .969 .970 .971
1.9 .971 .972 .973 .973 .974 .974 .975 .976 .976 .977
2.0 .977 .978 .978 .979 .979 .980 .980 .981 .981 .982
2.1 .982 .983 .983 .983 .984 .984 .985 .985 .985 .986
2.2 .986 .986 .987 .987 .987 .988 .988 .988 .989 .989
2.3 .989 .990 .990 .990 .990 .991 .991 .991 .991 .992
2.4 .992 .992 .992 .992 .993 .993 .993 .993 .993 .994
2.5 .994 .994 .994 .994 .994 .995 .995 .995 .995 .995
2.6 .995 .995 .996 .996 .996 .996 .996 .996 .996 .996
2.7 .997 .997 .997 .997 .997 .997 .997 .997 .997 .997
2.8 .997 .998 .998 .998 .998 .998 .998 .998 .998 .998
2.9 .998 .998 .998 .998 .998 .998 .999 .999 .999 .999
3.0 .999 .999 .999 .999 .999 .999 .999 .999 .999 .999
Φ(-z) = 1 - Φ(z)
40
Chapter 4
Schedule
In order to analyse the duration of a project, or a project activity we use flow network models. In
the literature there are two different ways to represent a project in a network model. Activity on
node (AON) networks uses the nodes to represent the project activities whereas the arcs are used
to represent the sequence and dependencies between activities. Activity on arc (AOA) networks
uses the arcs to represent the project activities whereas the nodes are used to represent events
in the project, i.e., starting and finalizing activities. In this presentation we will use the AON
representation of a project.
Visually, a flow network model is similar to a bar chart, or a Gantt diagram. However, we
usually indicate dependencies between activities with arrows, and the y and x axes are usually not
labelled. The symbols used in a flow network used in this presentation are shown in Figure 4.1.
An example flow network diagram is shown in Figure 4.2.
Activity
Uncertain activity
Milestone
S Start point
There exist several methods for analysing flow networks. All these models requires that the
flow network is described completely in terms of dependencies between the activities. Further the
duration of the activities should be described by probability distribution functions with numeric
values for the parameters. When analysing such flow network we differentiate between:
• Analytical methods.
41
D
S A B E F H F
C G
Figure 4.2: Example flow network from [3]
Generally we let T denote the duration of the project we are analysing, or a part of the project,
e.g. a work package. If the project comprises n activities, we often denote these activities Ai , and
the duration of activity Ai is denoted Ti . Sometimes in this presentation we also use the more
simplified notation where each activity is described by a letter, e.g. A, B etc. The main purpose
of the schedule analysis is to establish the cumulative distribution function for the entire project
duration. We might also want to establish the cumulative distribution function for parts of the
project, milestones etc. Another important measure of interest is the probability that an activity
will delay the project, i.e. the criticality index. The methods we will investigate are:
The example diagram shown in Figure 4.2 will be used to demonstrate the various methods. This
example is adapted from [3]. The parameters to describe the duration of each activity are given
in Table 4.2. Fundamental for all methods is to understand the term ’path’. A path in a flow
network is a set of activities from the the starting point to the end point in the network, where
each activity in the set follows another activity in the set except the first activity that follows the
starting point. This means that all activities in a path have to be executed in order to complete
the project. Usually there are several paths in a flow network. Formally, we also include uncertain
activities in a path, even if they might not be necessary to execute.
42
4.1 Critical Path Method (CPM)
The idea of the CPM method is to find all paths in the flow network. Next, we assume that the
duration of all activities are deterministic, and typically equal to the most likely duration (M ).
The duration of each path is given as the sum of duration of all activities in the path. The path
with the longest duration is denoted a critical path, and the duration of the project is found by the
duration of the critical path (or all critical paths in case of several critical paths). In Figure 4.2 we
have the following paths: P1 = {A,B,D,F,H}, P2 = {A,B,E,F,H} and P3 = {A,C,G,H}. Inserting
the duration of each activity, we get the following durations TP1 = 5 + 6 + 7 + 3 + 7 = 28, TP2
= 5 + 6 + 7 + 3 + 7 = 28 and TP3 = 5 + 12 + 5 + 7 = 29 for P1 , P2 and P3 respectively. Since
P3 has the longest duration, P3 is a critical path, and the project duration is found to be 29. A
disadvantage of the CPM method is that it cannot handle the uncertainty in the duration of each
activity, i.e. it is a deterministic approach.
Problem 4.1 Consider the situation above with activities B and C in parallel following activity
A. Further let the expectation and variance of the activity durations be given by: µA = 10, µB = 7,
43
µC = 8, σA2 = 22 , σ 2 = 32 and σ 2 = 22 . Find the expectation and duration of the project by first
B C
treating the two paths {A, B} and {A, C} as independent. Next, carry out an exact calculation
and compare the result with the first result. ♢
To structure the analysis we need some definitions. We define a meeting point where two or more
arrows join before or into an activity or the endpoint. For example in Figure 4.2 the activities D
and E join into a meeting point just before activity F . A branching point is a point where one
activity is followed by two or more activities in parallel, i.e. one branch splits into two or more
branches. For example in Figure 4.2 the activities B and C follow in parallel after activity A, and
the branching point is just right to activity A. We also need some numerical routines for solving
the integrals in Equations (3.62) and (3.63). Assume that we have access to the following routines
EMax=EMax(µ1 , σ12 , µ2 , σ22 ) and VarMax=VarMax(µ1 , σ12 , µ2 , σ22 ) for solving Equations (3.62) and (3.63)
respectively. Here µ1 , σ12 , µ2 and σ22 are expectations and variances for the two variables we are
taking the maximum of. We will only consider the situation where EMax and VarMax are implemented
under the assumption of independent and normally distributed variables. See pRisk.xlsm for such
an implementation.
Problem 4.2 Show that EMax(µ1 , σ12 , µ2 , σ22 ) = ∆µ+ EMax(µ1 −∆µ, σ12 , µ2 −∆µ, σ22 ) and VarMax(µ1 , σ12 , µ2 , σ22 ) =
VarMax(µ1 − ∆µ, σ12 , µ2 − ∆µ, σ22 ) ♢
The procedure for successive schedule planning with respect to describing the project duration is
now as follows1 :
1. For each activity i, establish the expectation, µi and variance σi2 for the duration of activity
i.
2. Identify all meeting points, i.e. where one or more branches join into one arrow.
3. Repeat and follow all activities from left to right in the flow network. This process is iterative
since each activity to the left of the current activity has to be processed before we can proceed.
4. For each activity i establish the expected start (EiS ), and expected finalisation (EiF ). The
expected start is equal to the expected finalisation of the preceding activity (or meeting point
in case of branches are joining just before activity i). The expected finalisation is given by
the expected start plus the expected duration of activity i, i.e. EiF = EiS + µi . Note that this
step cannot be executed if one or more of the activities to the left have not been processed.
5. For each activity i establish the accumulative variance, ViF . Here ViF is the accumulative
variance of the activity (or meeting point) preceding activity i plus σi2 , i.e. ViF = VkF + σi2
where k is the activity preceding activity i.
6. If there is a meeting point in the network just before the entry into an activity, we have to
process this meeting point. Note that this means that two or more branches join together
and the succeeding activity cannot start before all the branches, or paths up to this point,
have been finalised (completed). Technically, we now introduce a virtual node at the meeting
point, representing the finalisation of the two (or more) branches going into the meeting point.
The virtual nodes are enumerated V1 , V2 , . . .. If three or more branches join into one meeting
point, we first process two branches into one virtual node, then this virtual node represent
one branch which is then processed together with the third branch into another virtual node
etc.
1
The presentation is slightly different from the original presentation by Lichtenberg (1990).
44
7. Let Vk be the virtual node we are processing, and assume that it is activities i and j that
are joining into Vk . If one of the activities (or virtual nodes) immediate to the left of Vk
has not been processed, we have to go to the left in the network until we meet processed
activities or nodes. The expectation and variance for the finalisation of activity i are now
given by EiF and ViF respectively. Similar we have EjF and VjF for termination of activity
j. If the two paths up to activity i and j were disjoint, we could easily find the expectation
and variance of the finalisation of the virtual node Vk by Equations (3.62) and (3.63), or
numerically by EMax and VarMax. Typically, the two branches that join together after activity
i and j did split up into two branches from one single branch prior to a branching point. Let
l be the activity at which the branches did split up before joining again at the virtual node
k. When finding the expectation and variance up to the virtual node k we then first find
the expectation and variance up to the branching point l, and then add the expectation and
variance of the maximum of the two branches from the branching point l to the virtual node
k. The accumulated variance along the path from branching point l to the end of activity i
is found by Vi∆F = ViF − VlF . We get similar results for the other branch, i.e. the one with
activity j preceding the virtual node k. The expectation and variance for the finalisation of
the virtual node k is now given by EiF = EMax( EiF , ViF − VlF , EjF , VjF − VlF ) and ViF = VlF +
VarMax( EiF , ViF − VlF , EjF , VjF − VlF ).
8. If there are more branches not processed into the meeting point, repeat until all branches are
processed by creating new virtual nodes.
Example 4.1
We want to demonstrate the calculation process by the flow network shown in Figure 4.2. We
further assume that we have a spread sheet program available. The result of the calculations are
shown in Table 4.2. In addition to the activity row, and the three rows for the low, most likely and
high row, we add four rows for µi , σi2 , EiF and ViF respectively. For each row corresponding to
normal activities we calculate µi = (L+4m+H)/6 and σi2 = (µi −L)(H −µi )/7. Then we calculate
the expected finalisation of each activity, EiF as the expected finalisation of the previous activity (or
virtual node) plus the expected duration of activity i, µi . Similarly the variance of the finalisation
of activity i, ViF is the variance of the finilisation of the previous activity (or virtual node) plus the
variance of activity i, σi2 . For activity A the expected finalisation and variance of the finalisation is
equal to the expectation and variance of the duration of activity A since it is the first activity. We
note that after activity A we have a branching point that meets again after activities F and G. For
Activity B we see that EB F = E F +µ = 5.17+6.17 = 11.3, and V F = V F +σ 2 = 1.73+0.88 = 2.61.
A B B A B
We note that after activity A we have a branching point that meets again after activities F and
G.We proceed similarly with activities C, D and E. We now proceed to the virtual node V1 . It is
convenient to insert a new row in the spread sheet program just before activity F . In order to find
the expectation and variance for this node we take advantage of the functions EMax and VarMax.
The arguments to these functions are the expectation of the finalisation of each of the preceding
activities, and the accumulated variance through the branches from the branching point which
in this case is after activity B. EVF1 = EMax(18.5,3.49-2.61,18.5,4.35-2.61)=19.14. In order to
obtain the variance we use the VarMax function but we have to remember to add the accumulate
variance up to finalisation of activity B, VVF1 = VarMax(18.5,3.49-2.61,18.5,4.35-2.61)+2.61=3.5.
We complete the sheet for the remaining activities, including the virtual node V2 . The expectation
and variance of the duration of the entire project is now given by EH F and V F respectively.
H ♢
45
D
S A B E V1 F V2 H F
C G
Figure 4.3: Example flow network with virutal nodes, adapted from [3]
Note that we in the SSP-method have used the PERT distribution as a basis. The method could
be used for any distribution for the activities, the essential point is to assess the expectation and
variance of each activity duration.
Problem 4.3 Consider Example 4.1 and carry out the calculations by your self in a spread sheet
program. ♢
Problem 4.4 Find the cumulative distribution function for the entire project duration (T ) based
on the calculation in Problem 4.3, and especially find Pr(T > 35). ♢
Table 4.2: Data for the successive schedule planning demonstration in Example 4.1
46
stochastic variables we need. By inserting these stochastic variables into the deterministic model
(e.g. an MS Excel model) we now get one realisation of the system, or more specific the project
duration. Let t1 be the numeric value when this process is done the first time. Now, we might repeat
the process by generating another set of random quantities and insert these into the deterministic
model to yield another value, say t2 . By repeating this process we could think of the generated
values t1 , t2 . . . as realisations of the project, and use the values to obtain statistical properties such
as the mean, the standard deviation, the cumulative distribution function etc.
We will now illustrate how this process could be carried out with the pRisk.xlsm program.
Example 4.2
It will be convenient to establish one row in MS Excel for each activity. The first column (A)
could contain the activity number, the second, third and forth (B, C and D) could then contain the
parameters in the PERT distribution, similar to Table 4.2. Now we introduce three new columns
(E, F and G) to contain the duration, start and finalisation of each activity respectively. We start
to enter the duration of each activity. Assume that activity A is described in row 2 in the Excel
sheet. In cell E2 we now enter the following expression for the duration:
=RndPert(Rand(),B2,C2,D2)
Here the RndPert() function is a pRisk.xlsm specific function, whereas the Rand() function is a
standard Excel function. The procedure is repeated for all activities, and we simply copy the
formula in cell E2 into the cells E3, E4 etc. We will now proceed to the start and finalisation of
each activity. It will be convenient to give the cells containing the start and finalisation names in
Excel. For activity A we give the following names D_A, S_A and F_A for the duration, start and
finalisation respectively. Similarly we give the names S_B, D_B and F_B for the start, duration and
finalisation of activity B respectively, and so on for the remaining activities. By giving name to
the activities, it is easy to access them in formulas in other cells. We now use the convention cell
name = expression where the cell name is the name of the cell we want to assign an expression. By
inspecting the network in Figure 4.2 we easily verify the following statements for the start of the
various activities:
S_A = 0
S_B = F_A
S_C = F_A
S_D = F_B
S_E = F_B
S_F = Max(F_D, F_E)
S_G = F_C
S_H = Max(F_F, F_G)
The finalisation of the activities is given as the start point plus th duration, e.g., for activity A we
enter:
F_A = S_A + D_A
and similar for the other activities.
We have now specified the model and are prepared to simulate several runs. First we note that
each time we press the F9 key Excel updates the model by generating new random numbers since
we used the RAND() function in the cells containing the duration of each activity. Next we switch
to the RunSimul sheet and press the Run button. ♢
Problem 4.5 Consider the example in Figure 4.2. We will now consider an alternative execution
47
method for the last part of the project. Rather than executing activity H as one activity, it is
possible to split this activity into two parallel activities H and I. Each of these activities could be
described by the PERT distribution with parameters L = 3, M = 5 and H = 8. Set up the flow
network for this situation, and use the pRisk.xlsm program to find the expectation and standard
deviation of the project duration by Monte Carlo Simulation. ♢
where fT (t) is the probability density function for the project duration.
In principal we have to perform the integration in Equation 4.1 to find the expected total penalty
for default in a project. In most cases we also need to carry out numerical integration. However, if
we have a Monte Carlo simulation model for the project, we might utilise that for a given project
duration T the total penalty for default is max(0, (T − D)PD), and in e.g. pRisk.xlsm we could
specify in the “Cell to analyse”:
=max(0,PD*( T_End- D_Start )
where D_Start is the name of the cell where we have specified when penalty for default is ini-
tiated, and T_End is the name of the cell where the total project duration could be found.
Problem 4.6 Consider the example in Figure 4.2. Assume that D = 34, and PD = 1, 000 Euro.
Find the total expected penalty for default in this project. ♢
∫a
Problem 4.7 For the normal distribution we have from Equation (3.25) that −∞ xf (x)dx =
( ) ( )
µΦ a−µ
σ − σϕ a−µ
σ . Use this to verify that if the duration, T of a project is normally distributed
with mean µ and standard deviation σ then we have:
∫ ∞
PDTot = (t − D)PDfT (t) dt (4.2)
( [ ])
D
( )
D−µ D−µ
= PD(µ − D) 1 − Φ + PDσϕ (4.3)
σ σ
where Φ() and ϕ() are the cumulative distribution function and probability density function for the
standard normal distribution respectively.
48
4.6 Event uncertainty in the schedule model
When we describe the uncertainty about the duration of an activity by parameters such as L, M
and H these parameters account for all factors and conditions that influence the duration. In some
situations we might want to describe and model some important factors explicit. For example if
the event W denote extremely bad weather conditions, we might describe two set of parameters
{L, M, H}; one if W occurs, and one if W does not occur. That is, the duration of the activity is
described by the parameters {LW , MW , HW } if W occurs, and {LW C , MW C , HW C } if W does not
occur. Next we describe the probability that W occurs by a probability statement, pW = Pr(W ).
We might now include the uncertainty about the weather (event uncertainty) explicit into the
schedule model. The easiest way to include such event uncertainty is to use the Monte Carlo
simulation approach. We have to do the following:
• Describe the event we want to condition on, e.g. W
• The duration for each activity influenced by W is now entered as one expression if W occurs,
and another if W does not occur.
Now, assume that it is activity A which is influenced by the event W . The following information
could then be specified to pRisk.xlsm:
p_W = 0.1
Event_W = IF(Rand() < p_W,1,0)
D_A = IF(Event_W,RndPert(Rand(),5,8,12) ,RndPert(Rand(),2,5,8))
where the probability of bad weather conditions were set to 0.1, and we defined a cell with cell name
Event_W. Finally we have used the following set of parameter values: {LW , MW , HW } = {5, 8, 12}
and {LW C , MW C , HW C } = {2, 5, 8}.
Problem 4.8 Consider the example in Figure 4.2. Let W be the event {Bad weather}. Assume
that this event influences primarily the activities B and C. Let Pr(W ) = 0.1. The parameters
describing the duration of activity B and C is now similar to the situation in Problem 3.9. The
transformation factors now read fW , fW C and fV . Let fW = 1.5 and find fW C and fV similar to
the procedure in Problem 3.9. Follow this procedure both for activity B and C. Now update the
Monte Carlo simulation model for the example in Figure 4.2 when the event W is introduced in
the model, and compare the simulation results with the original results. ♢
49
• If new activities were necessary to add to the project, these are added to the schedule model.
• If some activities were cancelled these are removed from the schedule model.
• If the status of events and other risk factors that were included in the schedule model is
known, we replace the probabilistic statements about these with deterministic statements.
• Other parameters (typically L, M and H) are revised in light of the knowledge available at
this moment, e.g. related to resources available.
Problem 4.9 Consider the example in Figure 4.2 and the penalty for default structure as defined in
Problem 4.6. assume that we now are in the project phase, and activity A has just been completed.
Due to special circumstances the duration of activity A was tA = 10 which is even higher than the
most pessimistic assessment. First calculate the expected project duration, and total penalty for
default. Next we will consider alternative production methods to increase the speed in the project.
A major sub activity in activity H is to produce an element on the construction site. It is, however,
possible to have this element prefabricated in advanced. The extra cost of such a prefabrication
is 1,500 Euro. The gain of such a prefabrication is seen in the new distribution of the duration
of activity H, i.e. we now judge TH ∼ PERT(3, 5, 7). Update the pRisk.xlsm model and find the
expected total cost with and without prefabrication. ♢
50
Problem 4.11 Use pRisk.xlsm to perform the calculations in Example 4.4. ♢
51
Chapter 5
5.1 Introduction
In this section we will give a general introduction to the field of decision theory where uncertainties
are involved. Examples are related to project risk management. As a motivation consider the
following situation where we have to choose between two or more alternatives:
• Choice between double track and single track for a new railway line.
• Choice of tunnel trace now, or perform more investigation into the ground.
52
1. Decision under certainty. In this situation all the outcomes are known, and we will know for
sure what the outcome will be for the different decision alternatives.
2. Decision under risk. In this situation all the possible outcomes are known, but we do not
know which outcome will be the result of our decision. We are able to state probabilities for
the various outcomes.
3. Decision under uncertainty. In this situation all the possible outcomes are known, and we
are unsure about the probabilities for the various outcomes.
4. Decision under ignorance. In this situation we do not know all the possible outcomes, and
we are also unsure about the probabilities of those outcomes we know about.
In this presentation we will only consider decisions under risk and uncertainties. We also note that
many authors claim that it is not meaningful to state uncertainty about the probabilities, it is the
outcome which is uncertain, not the probabilities. When it comes to the final outcome, i.e., related
to the attribute vector Y we agree that there is no uncertainty in the probability distribution of
Y, i.e., the probability distribution contains all the uncertainty about Y. We will therefore not
differentiate between the situation of decisions under risk, and decisions under uncertainty. Most
frequently we will use the term ‘decision under uncertainty’.
As indicated above we will seek the decision d that gives the best “value” of the attribute vector
Y, for example the lowest cost and the shortest execution time of a project. There are, however,
some difficulties in this approach:
• We will not be neutral to the risk. Very often we are willing to make a decision that do
not give the maximum expected revenue, but rather choose an option with a lower expected
revenue but with a lower probability of big losses. We are what is called risk averse.
• The attribute vector Y comprises several dimensions, and it is not straight forward how to
weight these dimensions. For example how should we treat a project with low cost, but a
higher risk of accidents during project execution?
In order to treat such decision situations we introduced the concept of utility theory, and utility
functions. We will only briefly mention the major aspects, and refer to [9] for further discussions.
First we will treat situations where we make only one decision, and the end consequences are
assumed to take effect immediately after our decision. In more complex situations the effects will
come on a later stage, and we could make several decision in a sequence with time delays between
each decision. In such decision problems we often use decision trees to help assisting the decision
process.
53
5.2.1 Discrete end consequences vs attribute vector
Generally we use Y to describe the values of the end consequences resulting from a decision. In
some situations we want to simplify the representation by a set of few end consequences, EC. We
will let the end consequences, ECj , be disjoint. Further we will let pj = Pr(ECj occurs) be the
probability that we get end consequence ECj . It is not always straight forward to determine what is
the most convenient, either to work with the full attribute vector Y, or the set of end consequences
EC1 , EC2 , . . .. This will depend on the level of precision in the risk analysis, the skill of the risk
analyst, or the decision maker etc.
In order to see the difference, consider the occupational safety dimension during project ex-
ecution. If we work with end consequences it could be natural to introduce the following end
consequences:
1. EC0 = No injury
5. EC4 = 1 fatality
In order to describe the expected result we also specify the corresponding probabilities, p0 , p1 , . . . p6 .
These probabilities will be dependent on the decisions we make. If we want to use a full attribute
vector we could use Y = [Y1 , Y2 , Y3 , Y4 ], where Y1 = number of minor injuries, Y2 = number of
major injuries, Y3 = number of fatalities, and Y4 = is the number of gross accidents, i.e., accidents
with five or more fatalities. In this latter situation we specify the expected outcome in terms of
the joint probability distribution function of Y. We then often introduce parameters that depend
on the decision d we make.
Utility function
The utility function expresses the preferences of the decision maker regarding various attribute
vectors or end consequences. A prerequisite for establishing a utility function is that the decision
maker is able to express preferences between different values of the attribute vector. For example in
a one dimensional situation where we set Y = NPV (net present value) this will be rather obvious
in the first place, it is reasonable that all decision makers will prefer a higher value to a lower value.
Now let y1 and y2 denote two arbitrary values Y may take. The following relations are of interest
between y1 and y2 :
The utility function is now a function that assigns a one-dimensional utility value to each value
of the attribute vector or quantity, u = u(y). For the utility function we require:
y1 ∼ y2 ⇔ u(y1 ) = u(y2 )
y1 ≻ y2 ⇔ u(y1 ) > u(y2 )
y1 ≽ y2 ⇔ u(y1 ) ≥ u(y2 )
y1 ≺ y2 ⇔ u(y1 ) < u(y2 )
y1 ≼ y2 ⇔ u(y1 ) ≤ u(y2 )
54
Relation Explanation
y1 ∼ y2 y1 and y2 is considered equal
y1 ≻ y2 y1 is preferred over y2
y1 ≽ y2 y1 is as least as preferable as y2
y1 ≺ y2 y2 is preferred over y1
y1 ≼ y2 y2 is as least as preferable as y1
There exists, however, an infinite number of utility functions that satisfy the above criteria and we
therefore want to fix the utility function for some values. In order to be useful, the utility function
should also express how much we prefer for example, y1 over y2 . Further we also want the utility
function to reflect the fact that there will be uncertainty regarding the future value of the attribute
Y . We still consider the one dimensional situation where Y = NPV (net present value). Y will be
a stochastic variable in the decision point. Now, assume that we could choose between a decision
A that for sure gives the net present value Y = y0 and the decision B that gives the net present
value Y = y1 with probability α and the net present value Y = y2 with probability 1 − α. Further
assume that y1 ≺ y0 ≺ y2 . For a given set of values of y0 , y1 and y2 there will exist a value of
α which makes the decision maker indifferent between the two decisions A and B. This will be
reflected in the utility function which must satisfy:
Equation (5.1) could now in principle be used to establish the utility function. In this process we
might restrict our selves to let the utility function take values between 0 and 1, or 0 and 100.
2. A baseline hour rate of 10 Euro, and an additional value of 50 Euro which will be paid if the
project reaches the targets that have been set up.
We have been studying the progression in the project so far, and assess the probability that the
extra 50 Euro to be paid is 40%. Simple calculations shows that the second alternative gives the
highest expected hour rate (30 Euro vs 20 Euro). However, we are in a cash position which makes
it very difficult for us if we only receive 10 Euro per hour. After some considerations we have found
out that the two alternatives are equal, i.e., neither of them are prefered over the other.
We will now utilise Equation (5.1) to set up our utility function. Three points could be assessed,
u(1 000), u(2 000), and u(6 000). Since we arbitrary may choose the end points, we let u(1 000) = 0
and u(6 000) = 100. We now have (α = 0.6):
The utility function is shown in Figure 5.1 where we have fitted the function u(y) = 55.702 ln(y) −
384.25 and the diamonds represent the assessed values. ♢
Problem 5.1 Consider Example 5.1. What probability would you required for being paid the
additional 50 Euro per hour in order to treat the two alternatives as equal. ♢
55
u(y)
100
y
1000 2000 3000 4000 5000 6000
Figure 5.1: Utility function for Example 5.1
Problem 5.2 Consider Example 5.1 again, but assume that you were going to work 500 hours.
Make a sketch of your utility function in this situation. ♢
For private economies we are usually risk averse. Risk aversion means a concave utility function as
shown in Figure 5.1. Also smaller enterprises will often be risk averse reflecting that rather than
optimising expected revenue, decisions are taken to minimise the probability of big losses which
could lead to bankruptcy. Larger enterprises will often have an almost linear utility function (in
monetary values) because their economical strength is good, and there is no real possibility for
bankruptcy.
• What is the benefit, or utility of saving one (statistical) life vs saving 10 statistical lives?
• What is the benefit, or utility of saving one (statistical) life vs the possibility to earn an extra
million Euro?
In the first situation we deal with the question to rank the consequences within the same main
dimension (safety), whereas we in the second situation need to compare benefits or disadvantages
across dimensions. The discussion below will be very short, and we refer to e.g., [10] for further
discussion on this topic.
The first issue we will discuss is the concept ‘value of prevented fatality’ (VPF). The idea behind
this concept is that in any industrial activity, transportation services etc. there will always be a
risk of accidents, and hence possibilities of severe injuries and fatalities. As a decision maker we
have to face this fact. However, we will make effort to reduce this risk, and we are willing to
spend money to achieve such a reduction. The VPF value states then how much we are willing to
spend in order to prevent one statistical fatality. We use the term ‘statistical’ fatality to emphasise
that this willingness to pay is not related to specific persons, but arbitrary persons where it is not
meaningful or possible to identify single persons. In some presentation also the term ‘value of life’
(VOL) has been used. We feel that this term is not appropriate because the term indicates that
the life it self has a value which could be measured in monetary units. This is not our perspective.
The value of life it self could not be measured. What we could assess figures to, is what we are
willing to pay in order to reduce risk, or the probability of fatalities. Hence, the term VPF make
more sense in our understanding.
56
If we accept that the term VPF make sense, then the next question will be how to assess the value
of VPF. Different approaches exist. One approach is to look into economical considerations from
the society point of view. For example we could calculate the reduction in GNP (Gross National
Product) caused by a fatality. Such calculations have been carried out, and in e.g., Norway this
indicate a value of 3 million Euro for VPF. Another approach has been to ask single persons about
their willingness to pay for risk reduction (see “The change in risk of death” [16]). For example
for buyers of cars, we could ask what they are willing to pay for a given safety system or measure,
for example an improved airbag system. Let assume that the amount one is willing to pay is ∆W ,
and that the assessed risk reduction during the service life of the car is ∆P . It would then be
natural to set VPF = ∆W/∆P . In Norway no such systematic surveys have been conducted, but
more arbitrary surveys at NTNU among ordinary students and continuation students a value of 2.5
million Euro has also been found for VPF. We will emphasise some challenges of such a willingness
to pay approach:
• Different persons have different preferences. For example young people tend to be less willing
to pay for risk reduction compared to older persons with family obligations.
• Individuals are not consistent in their preference statements.
• In real surveys to establish ∆W and ∆P we face the problem that other dimensions than
being killed are involved, e.g., the risk of minor and major injuries. Further one does not only
consider the life of one self, but also the life of family members etc when making decisions
about safety.
• It is not obvious that “what I am willing to pay” is what I want the society to pay for risk
reduction in general, or what I expect my employer to pay for my risk reduction.
It is a tendency to set a lower value for VPF when it comes to the area of public responsibility
compared to industrial activity. For example in the petroleum industry we see VPF values in
the order 10 to 15 million Euro. It is also a tendency to set a higher VPF for multiple fatality
accidents compared to single fatality accidents. This could be interpreted as an aversion against
gross accidents. This aversion should not be confused with risk aversions which would be an aversion
against a high number of fatalities in general, and not the number of fatalities in single accidents.
An another perspective in this discussion is how we should treat injuries in such a framework. One
common approach here is to introduce the concept of ‘equivalent fatality’. For example we could
be willing to pay five times more to prevent a fatality than a severe injury, which corresponds to
an equivalent fatality of 0.2.
In a utility function approach we could now in case of a VPF value 2.5 million Euro let the
utility of one fatality be equivalent to -2.5. If we now extend the situation to include multiple
fatality accidents, and minor and major injuries, we could set up a more general utility function:
where y1 is the number of minor injuries, y2 is the number of major injuries, y3 is the number
of fatalities in accidents with less than five fatalities, and y4 is the number of fatalities in gross
accidents (five or more fatalities in one accident). It is important to emphasise that the utility
function offered in Equation (5.2) is a function that could be used as a start in a discussion about
value trade-offs and preferences, and should not be considered as the “correct utility function”. Also
note that Equation (5.2) includes an aversion against gross accidents, but there is no risk aversion
in terms of a concave utility function in the attributes. If we also want to include attribute y7 as
the profit in a project measured in million Euro we could extend the utility function:
57
u(y1 , y2 , y3 , y4 ) = −0.03y1 − 0.5y2 − 2.5y3 − 7y4 + y7 − ae−by7 (5.3)
where a and b are constant. Reasonable values of these constants are a = 0.08 and b = 0.7.
The utility function in Equation (5.3) is an additive utility function. Very often we use additive
utility functions for simplicity. However, arguments could indicate that a situation with one extra
fatality is “worse” if there is a situation with a gross accident than without such a gross accident.
Such discussions will not be pursued any further, and we refer to [9].
Result 5.1 ∫∞
The optimal decision d is the decision that maximises expected utility, E(u(Y)) = −∞ u(y)fY (y)dy
♢
The basic steps in obtaining the optimal decision is then:
1. Establish an explicit expression for the utility function, u = u(y1 , y2 , . . .) which corresponds
to the preferences and value trade-offs of the decision maker.
2. Establish the probability distribution function for the attribute vector Y = [Y1 , Y2 , . . .] for
each decision alternative, or for each value of a decision variable (d).
3. Calculate the expected utility to each decision alternative by integrating the utility function
over the probability distribution of the attribute vector.
4. Find the decision alternative that gives the maximum expected utility.
Problem 5.3 In this problem you shall first make an attempt to construct the utility function
u(y) for a given decision maker. In the problem there is only one dimension, and the attribute
y is measured in thousand Euro by the procedures we have established in the previous sections.
Assume that u(−100) = 0, and u(400) = 1.
a) Why do we have the freedom to assess two points on the utility function, and why is is suitable
to use these two values.
Now, assume that the decision maker makes the following considerations regarding the outcome of
a project:
• An uncertain project which gives -100 with probability 0.50 and +400 with probability 0.50
is considered as equal attractive as receiving the fixed amount +150.
• An uncertain project which gives -100 with probability 0.50 and +150 with probability 0.50
is considered as equal attractive as receiving the fixed amount +100.
58
• An uncertain project which gives +150 with probability 0.50 and +400 with probability 0.50
is considered as equal attractive as receiving the fixed amount +225.
b) Draw the points on the utility function which you could calculate based on the above infor-
mation, and make a sketch of the utility function in the interval -100 to +400.
c) What does the graph say about the decision makers attitudes to risk?
d) Use the graph to choose the optimum project among the following projects:
A) A project returning -100 with probability 0.2, +150 with probability 0.2 and +350 with
probability 0.6.
B) A project returning 0 with probability 0.4 and +400 with probability 0.6.
e) Which of these two projects would the decision maker choose if he adopts the principle of
maximum expectation. ♢
Problem 5.4 In a tunnel project one could choose between bursting or drilling. Bursting is con-
sidered to be the cheapest alternative, but the risk of personal injuries or fatalities is considered
higher. Assume the utility function given in Equation 5.2 on page 57. Let fi = E(Yi ), i = 1, . . . , 4
be the expected number of minor injuries, serious injures etc. and assume the following numbers:
How much cheaper need bursting be compared to drilling if these two methods should be equally
valued with respect to utility? ♢
Problem 5.5 A company has established the following utility function: u(y) = y − ae−by , where
a = 1/5 and b = 1/2, to be applied for prioritization of projects. y is given in million NOKs and
represents the profit in the projects.
a. Make a sketch of the utility function, and discuss the risk attitude of the company.
b. The company considers to invest in a project which (i) either gives a profit of 10 (million
NOKs), or (ii) a loss of 10 (million NOKs). What probability of success is required to invest
in the project?
c. The company is going to choose between two projects, A and B. By analysis the following
has been derived: YA ∼ N (6, 42 ) and YB ∼ N (6.1, 52 ), where Y is the profit, and ∼ N (µ, σ 2 )
indicates normally distributed quantities. Calculate the expected utility for each project to
decide which project is best. Discuss the result. Hint: Show that if Y is normally distributed
with expected value µ and standard deviation σ, then the expected utility is found by use of
moment generating function to be: E(u(Y )) = µ − ae−bµ+1/2b σ provided the utility function
2 2
59
5.2.3 Examples with one decision node
In this section we will investigate examples where only one decision is going to be made.
Alternative a1 Alternative a2
↓ p U V P U V
Amount
2,000 0.1 3.9 200 0.3 11.7 600
5,000 0.8 72.1 4,000 0.4 36.1 2,000
8,000 0.1 11.6 800 0.3 34.9 2,400
Sum→ 1.0 87.7 5,000 1.0 82.7 5,000
where each term in the sum is calculated in the column for U in Table 5.1. In the V column we
have similarly calculated the expected monetary value. In the last row the sum is shown, and we
observe that the expected utility for a1 and a2 is 87.7 and 82.7 respectively, and hence alternative
a1 has the largest expected utility. When expectations are considered, the two alternatives are
equivalent.
Problem 5.6 Consider Example 5.2, but now assume that the probability distribution for the
payments are ∼ PERT(4000,5000,6000) and ∼ PERT(2000,6000,8000) respectively. Hint: you
might use the program pRisk.xlsm. ♢
60
p(TP)
1
.8
.6
.4
.2
TP
30 35 40 45 50
Figure 5.2: Probability p of getting the project as a function of tender price, TP
LP = Lowest tender prince among our competitors. We assume a lump sum contract.
p = The probability that we get the contract, p = p(TP) = E(Y1 ) = Pr(LP > TP) .
Y3 = PC = project cost.
In order to find p = (TP) we could utilise the pRisk.xlsm program, and the function CDFPert. The
syntax to enter in an EXCEL cell where we store the result is:
= 1 - CDFPert(TP,30,35,50)
where TP is a cell reference or a numeric value for the tender price. Figure 5.2 shows the probability
of getting the contract as a function of the tender price. Given that we get the contract, the expected
utility equals TP-E(PC). The expectation in the PERT distribution is given as (L + 4M + H)/6,
i.e., (10 + 4 · 30 + 80)/6 = 35 million Euro, and the expected utility equals:
By using the result for p = p(TP) from Figure 5.2 we could easily find the expected utility as shown
in Figure 5.3. The optimum tender price is found to be 39 million Euro. ♢
u(y) = y − a · e−by
61
E(u)
1
0
-1
-2
-3
-4
-5 TP
30 35 40 45 50
Figure 5.3: Expected utility as a function of the tender price, TP
E(u)
1
0
-1
-2
TP
35 3637 38 39 40 41
Figure 5.4: Expected utility as a function of the tender price, TP
where a and b er parameters. We set a = 0.2, and b = 0.2, and we let money be measured in million
Euro. With these parameters the utility of 10 million Euro is then 9.97 whereas the utility of a loss
of 20 million Euro is - 30.9 and the utility of -30 is -110. In this example we will also add one if
we get the contract. This extra utility unit could represent the value of competence improvement.
The utility function is then:
It will not be easy to maximise expected utility as a function of the tender price, TP. We therefore
use pRisk.xlsm to carry out a Monte Carlo simulation, and the result is shown in Figure 5.4. We
see that the optimum value is slightly increased from 39 million to almost 40 million. The reason
for this is the concave utility function where we want to reduce the probability of the large losses.
♢
Problem 5.7 Make a sketch of the utility function in Example 5.4. Discuss the utility when y
approaches minus infinity and plus infinity. ♢
Problem 5.8 Use pRisk.xlsm to perform the calculations for the example in Example 5.4. Discuss
the influence of the parameter b and discuss the effect of letting b approach zero. ♢
62
PD = 10,000 Euro = Penalty for default, i.e the amount to pay each day the project is delayed
(Y1 >CD).
BO = 5,000 Euro = Bonus, i.e the amount BO is paid extra for each day the project is completed
before CD.
We will use a linear utility function, and we let the utility of one Euro be equal to one. The
following utility function the applies:
Also here we utilise the pRisk.xlsm program to calculate expected utility. The following statements
are entered into the Excel sheet:
CD=30
L_0=25
M_0=30
H_0=60
PD=10000
BO=5000
Y_2= 10000
LY_2=L_0*(0.5+0.5*exp(-Y_2/50000))
MY_2=M_0*(0.5+0.5*exp(-Y_2/50000))
HY_2=H_0*(0.5+0.5*exp(-Y_2/50000))
Y_1=RndPERT(RAND(),LY_2, MY_2, HY_2)
We obtain the following values for expected utility:
The maximum expected utility is obtained by spending between 30 and 40 thousand Euros in order
to reduce the risk of delays in the project.
Problem 5.9 Use the pRisk.xlsm program to perform the simulation as indicated above. Check
the sensitivity in the results as a function of the number of simulation runs. Then find a more
exact result for the optimum value of the extra effort. ♢
63
5.2.4 Decision trees
The use of decision trees is a fruitful approach when we are going to systemise a decision process
where the decisions are made at different point of times. The main reason for postponing a decision
is to follow the development of e.g., a project, and hence make the most appropriate decision when
more information is available. The drawback is that postponing a decision could yield more costly
solutions. Another drawback could be that it is no time to implement necessary measures in due
time if we wait to take action.
Starting point
Decision node
Chance node
Consecutive costs
Example 5.5
Construction Ltd. is the main contractor for a road tunnel project. During the work more water
penetration than expected is discovered. Physically there are three alternatives to choose among:i)
bursting an outlet drain which is very costly but a satisfactorily solution, ii) build a pumping
station to pump away the water which is a cheaper solution but may not be adequate if there is
very much water, and iii) carry out seal work which is even cheaper, but adequate only in case
of very little water. The amount of water is uncertain at the time being. Below we discuss the
decision process:
The first decision is now (DN1 ), and at this decision node we have the following options:
64
Repeat for all end terminals
(*) Move to the node to the left, and bring with the EMV-value in the current node
IF this is a chance node THEN
Calculate EMVi,j = pi,j · EMV
IF EMV has been calculated for all branches into this node THEN
Calculate EMVi = Σj EMVi,j
GoTo (*)
ELSE
GoTo next terminal node
ENDIF
ELSEIF this is a decision node THEN
IF EMV has been calculated for all branches into this node THEN
Let EMVi = Min j ( EMVi,j )
Optimum decision in DNi is the branch with minimum EMVi,j
GoTo (*)
ELSE
GoTo next terminal node
ENDIF
ELSEIF this is a consecutive node THEN
Add EMV of the consecutive node to EMV
GoTo (*)
ELSEIF this is the start node THEN
We are done
ENDIF
B: Wait half a year until more information about the amount water is available
If we postpone the decision (B) we would have more information about the amount of water in half
a year and a better decision could be made. Two outcomes are foreseen in half a year (CN1 ):
C: It is obviously so much water that bursting the outlet drain is necessary
D: There is still uncertainty regarding the amount of water, and we have a new option in decision
node DN2 :
If the pumping station is build at this time (E) this could result in the following outcomes (CN2 ):
G: The pumping station was sufficient
H: The pumping station was not sufficient, and an outlet drain have to be bursted
If we wanted to postpone the decision (F) there are tree possible outcomes (CN3 ):
I: Bursting the outlet drain is required
65
Drain outlet
A 50 million
D DN2 H Fiasco
90 million
I Drain outlet
70 million
Table 5.2 shows the associate costs (the letter in parentheses corresponds to the alternative above).
P(C|CN1 ) = 30%
P(D|CN1 ) = 70%
P(G|CN2 ) = 90%
P(H|CN2 ) = 10%
P(I|CN3 ) = 10%
P(J|CN3 ) = 40%
P(K|CN3 ) = 50%
Note that in the example we have not used the symbol for consecutive costs. For the calculation
we use the algorithm indicated in Figure 5.6. We start with the upper right terminal node, and
“collect” the EMV = 50 mill. into the decision node to the left, e.g., DN1 . In this decision node we
observe that not all branches (from the right) into node DN1 have been processed, and we therefore
need to go back to a new terminal node. We go back to the next upper terminal node and collect
EMV = 60 mill. which is multiplied with the branch probability (30%) such that we get EMV =
0.3 · 60 mill. = 18 mill. into chance node CN1 . Here, the second branch into the chance node has
not been processed and we again have to go back to the next non-processed terminal node. Here we
66
50 Drain outlet
50 million
7 Drain outlet
10% 70 million
40% Seal work
23.5 CN3 4
10 million
50% Pump station
12.5
25 million
Figure 5.8: Decision tree for tunnel project with calculations
collected EMV = 20 mil which is multiplied with 90% gives EMV = 18 mill. into chacne node CN2 .
Similarily we get EMV = 90 mill. · 10% = 9 mill. for the second branch into chance noe CN2 . We
may now complete the processing of chance node CN2 by adding the EMV values entering the node
from the right, yilding an EMV of 27 mill. This number now goes into decision node DN2 . Now the
remaining end nodes are processed, and we get the EMV to collect from CN3 equal to 23.5 which
again will be the second EMV into decision node DN2 . In decision node DN2 we shall choose the
branch having the lowest EMV value, i.e., branch F with an EMV of 23.5. In decision node DN2 it is
most beneficial to postpone the decision for another half year. We complete the tree similarly, and
find that in decision node DN1 the optimal decision is to postpone any physical activity. We remain
then with an EMV equal to 34.45. The number from these calculations are shown in Figure 5.8.
Also note that we have not taken the discounting aspects into account, something that also would
have been an argument for postponing the decision. ♢
Problem 5.10 An oil company has the rights for a given oil field, and have the options between:
• Drill a well (D)
67
In this problem you should establish a decision tree for the situation. The decision node following
the start node should be:
Probabilities in relation to a seismic investigation (e.g., based on experience figures from geol-
ogists):
No structure: 0.60
Open structure: 0.30
Closed structure: 0.10
The probabilities for finding a profitable oil pool, given the result of the seismic investigation:
No structure: 0.10
Open structure: 0.25
Closed structure: 0.70
The probability for finding oil when no seismic investigation has been performed: 0.20. Find the
optimal decision in each decision node, and formulate the conclusions from your analysis. ♢
68
Chapter 6
6.1 Introduction
In this chapter we will give a short introduction to life cycle cost (LCC) modelling and anaylysis
in connection with project management. The term LCC is defined in IEC 60300: “LCC is the
cumulative cost of a product over its life cycle”. The LCC concept was first introduced in the
US Army and the idea was to establish the cost of development, production and use (operation
and maintenance) of military equipment. In the original use the revenues was not included in the
modelling. However, in order to get a complete picture we will usually also include the possible profit
of a new system or product. Hence the term ‘Life Cycle Profit’ has been introduced. Kawacuchi
and Rausand [8] suggest a process for LCC analysis comprising the following steps:
1. Problem definition
3. System modelling
4. Data collection
6. Evaluation
7. Reporting
In this presentation we will focus on the cost modelling aspects, i.e. mainly step 3 in the above
procedure.
69
(now)
X1 X2 X3 X4 X5 XT −1 XT
time
0 1 2 3 4 5 T −1 T
Figure 6.1: Visualisation of the cach flow, Xt
The net present value of an amount Xt that occurs at the end of year t is:
where r is the discount rate. Similarly, we find the net present value of a cash flow X0 , X1 . . . . , XT :
∑
T
NPV = Xt (1 + r)−t (6.2)
t=0
where X0 represents in or outgoing cash now, and T is the number of years to consider.
Sometimes we want to establish the net present value of a constant yearly (nominal) amount
∑n i.e.i the samen amount each year. By utilising the formula for the sum of a geometric series,
XA ,
i=1 q = q(1 − q )/(1 − q) we obtain:
[ ]
1 − (1 + r)−T
NPV = XA (6.3)
r
XA (1 + r)−l
NPV = (6.6)
1 − (1 + r)−k
70
• The monetary value increases due to general conditions, such as inflation.
• The monetary value increases due to increased operating costs, e.g. physical deterioration
and hence more maintenance is required.
Increased operating costs due to deterioration could usually be reset by a renewal of the system we
are considering, whereas external conditions like inflation is not affected by e.g. a system renewal.
In the modelling we will assume a fixed inflation rate, even if we in a more advanced model also
could let the inflation rate vary. This inflation rate will be denoted v, and we could use Equation 6.4
to calculate the net present value of a amount that changes due to inflation. When we want to
model increased operating cost due to deterioration, we need to introduce a local age parameter.
We will let a denote the age of the system, or the age of the system since the last system renewal.
When we consider degradation, we introduce the degradation rate d where we assume that the
yearly increase due to deterioration equals (1 + d). This corresponds to an exponential growth
which very often is found realistic if we have degradation mechanisms that drive the costs. Now,
let c0 be the yearly cost of operation, maintenance etc now (i.e. at time t = 0). We then have the
yearly cost in year t (occurring at the end of year t):
ct = c0 (1 + d)t (6.7)
In order to obtain the degradation rate d we usually need data about the costs as a function of
time. A very simple approach if we know that c(t) has increased by a growth factor (GF) during
a period of T years. We then have:
d = eln(GF)/T − 1 (6.8)
• Outsourcing of truck-maintenance.
• Point wise refill of ballast in order to postpone the need for a full renewal (ballast cleaning).
• Grinding of rails.
71
There are several aspects to consider when conducting an LCC analysis, for example:
• Visualise the cost picture, enabling the possibility to work actively with eliminating the main
cost drivers, or the effect of these.
• Use the LCC model as a decision support when making decision about the profitability of
projects or measures, and when to conduct or implement these.
• Use the LCC model as a basis for contractual follow-up, e.g. LCC contracts.
Example 6.1 We will consider a railway system where the quality of the ballast has deteriorated
during the last years, and in order to compensate for this it is proposed to do a point wise replace-
ment of the ballast on the line. The age of the ballast is 35 years, and without this point wise refill
of ballast it is expected that a full renewal (ballast cleaning) is necessary within five years. If we
conduct the project we could postpone the ballast cleaning with another five year. The length of
the line we are considering is 10 km. The quantities to include in the LCC model is as follows:
RC = 2.5 million Euro = Renewal cost = 250 Euro per meter for ballast cleaning.
IC = 400,000 Euro= Improvement cost, e.g. cost of point wise ballast refill.
a = ballast age, i.e. effective age relative to the implemented measures. Without point wise
refill of ballast a = 35 years, and with point wise refill of ballast a = 30 year. For a track
that has just being renewed a = 0.
c0 = 25,000 Euro = yearly cost of maintenance and operation of the track, for a new track, i.e.
just being renewed.
c40 = 250,000 Euro = yearly cost of maintenance and operation of the track, for a track that has
reached it’s service life, e.g. 40 years.
d = eln(250000/25000)/40 − 1 = 0.05925
r = 6% = interest rent.
We start by calculating the various LCC-terms (in million Euros) if the improvement project (point
wise refill of ballast) is not executed. The total renewal cost if found by Equation (6.6):
RC(1 + r)−5
LCCRC = = 2.069
1 − (1 − r)−40
The variable cost the next five years (up to the next renewal) is found from Equation (6.4)
( )5
1 − 1+r
1+d
35
LCCVC,1 = c0 (1 + d) = 0.883
r−d
72
After the renewal in five year the variable costs will be reset to v0 , and then start increasing again.
The net present value in one cycle is:
( )40
1 − 1+d
1+r
LCCVC,0 = c0 (1 + d) = 0.986
r−d
The amount LCCVC,0 will then be repeated every 40 year, and the first time will be in five years:
LCCVC,0 (1 + r)−5
LCCVC,∞ = = 0.816
1 − (1 + r)−40
Finally we have the total contribution from variable costs:
If we execute the improvement project, the calculations are similar. We start with the total renewal
cost (first renewal after 10 years):
RC(1 + r)−10
LCCRC = = 1.546
1 − (1 + r)−40
The variable cost the next ten years (up to the next renewal) noting that the effective age after the
improvement project is a = 30:
( )10
1 − 1+d
1+r 30
LCCVC,1 = c0 (1 + d) = 1.322
r−d
After the renewal in ten year the variable costs will be reset to v0 , and then start increasing again.
The net present value in one cycle, LCCVC,0 , is the same as without the improvement project, but
the first cycle will start in ten years:
LCCVC,0 (1 + r)−10
LCCVC,∞ = = 0.610
1 − (1 + r)−40
Finally we have the total contribution from variable costs:
LCCIC = 0.4
Summing up all LCC contributions we find that implementing the improvement project gives a
total LCC of 3.878 million versus not implementing the project gives a total cost of 3.768. Thus
the improvement project is not profitable. ♢
Problem 6.1 Consider Example 6.1. Find the value of the discount r that makes the two alter-
natives equal from a LCC point of view. Why is this value of r higher than the initial one?
Problem 6.2 Consider Example 6.1 and investigate if ballast cleaning after 40 years is optimal.
If not, find the optimal period for ballast cleaning, i.e., the optimal renewal period.
73
Chapter 7
Parameter estimation
7.1 Introduction
In this chapter we will briefly describe principles for parameter estimation. A parameter in this
context is a quantity in the risk analysis for which we assign numerical values. There are two
principles for establishing numerical values (parameter estimates):
If we have access to relevant data we will usually use these data to estimate the parameters.
Often we have little relevant data, and we then have to rely on expert judgements. In some
situations we combine historical data with expert judgements by use of Bayesian methods.
• Assume that we know the probability density function of the observations for which we have
data. Let this distribution be denoted f (x; θ).
The MLE principle now tells us to estimate θ by the value which is most likely given the observed
data. To define “likelihood” we use the probability density function. The simultaneous probability
density for X1 , X2 , …Xn is given by:
∏
n
f (x1 ; θ)f (x2 ; θ) . . . f (xn ; θ) = f (xi ; θ) (7.1)
i=1
74
This density express how likely a given combination of the x-values are, given the value of θ.
However, in our situation the x-values are given, whereas θ is unknown. We therefore interchange
the arguments, and consider the expression as a function of θ:
∏
n
L(θ; x1 , x1 . . . xn ) = f (xi ; θ) (7.2)
i=1
where L(θ; x1 , x1 . . . xn ) in equation (7.2) denotes the likelihood function. The MLE principle will
now be formulated as to choose the θ-value that maximizes the likelihood function. To denote the
MLE estimator we write a “hat” over θ, θ̂. Generally, θ will be a function of the observations:
θ̂ = θ̂(X1 , X2 , . . . Xn ) (7.3)
When we insert numerical values for the x’s we denote the result as the parameter estimate.
Note that the parameter is denoted λ, whereas we generally use θ. Further we denote the observa-
tions with t because we here have failure times. The probability density function in the exponential
distribution is given by f (t) = λe−λt . A common “trick” when maximising the likelihood function
is to take the logarithm. Because the logarithm (ln) function is monotonically increasing, ln L will
also be maximised for the same value as for which L is maximised. We could then find:
∑n
l(λ; t1 , t2 , . . . , tn ) = ln L(λ; t1 , t2 , . . . , tn ) = n ln λ − λti
i=1
By taking the derivative wrt λ and set this expression to zero, we easily obtain:
/∑n
λ̂ = n ti
i=1
Problem 7.1 Find the MLE for µ and σ in the normal distribution. ♢
75
∑
From (7.4) we easily obtain M = (6E(X) − L − H)/4. Now let x̄ = i xi denote the sample mean
(first order moment), and we may estimate M by:
M̂ = (6x̄ − L̂ − Ĥ)/4 (7.6)
The challenge now is to find L̂ ∑
and Ĥ. By rearranging equation (7.5) and inserting x̄ for E(X) and
2
i (xi − x̄) for Var(X) we have:
1
the sample variance S 2 = n−1
Problem 7.2 Assume that we have observed the following durations for a typical activity in
projects: 9.3, 10.5, 9.4, 9.0, 9.4, 8.6, 10.1, 10.7, 12.0, 10.6, 9.8, 13.1, 12.0, 8.6 and 10.9. Estimate
the parameters if you assume that the durations are PERT distributed. Hint: Use the Average()
and STDEV() functions in MS Excel to find x̄ and S. ♢
Equation (7.11) is the starting point for estimating the parameters in so-called regression models.
The most simple formula is given by:
E(Yi ) = β0 + β1 xi (7.12)
In this model x is denoted the independent variable, whereas Y is denoted the dependent variable
because it depends on the independent variable, x.
76
Problem∑7.3 Prove that the estimators for β0 and β1 in equation (7.12) is given by:
(x −x̄)y
β̂1 = ∑i (xi −x̄)2i and β̂0 = ȳ − β̂1 x̄.
i i
♢
The model in equation (7.12) could be extended to cover more independent variables. These
are denoted regression variables , or explanatory variables. To extend the model we introduce an
extra index for each x. We write xij , where index i denotes the i′ th data point, whereas index j
denotes the j’th explanatory variable. The model then reads:
To obtain the LS estimators in this situation, we introduce matrix notation. Let y = [y1 ,y2 , …,
yn ]T be a column vector containing the dependent variables, and let X be the design matrix given
by:
1 x11 ...... x1r
1 x21 x1r
X= :
(7.14)
xij
1 xi1 ....... xnr
It may be shown that the LS estimator for β = [β0 , β1 , β2 , · · · , βr ]T is given as the solution of the
following matrix equation:
XT y = XT Xβ (7.15)
If the design matrix has full rank, XT X will be non-singular, and the solution is given by:
If one has access to a tool for matrix calculus, we easily obtain the LS estimates. We could also
use commercial available statistical programs, or the “analysis” module of MS Excel.
77
From MS Excel we obtain the following parameters: β̂0 = 4.211, β̂1 = 0.0167, β̂2 = 2.196, and β̂3 =
0.0011
♢
The residuals εˆi may now be used as input in the method of moments to estimate PERT parameters.
Below se summarize the procedure to get L, M and H values for an activity, cost element etc. in
a specific project.
1. Estimate regression parameters from data from similar projects: β̂ = (XT X)−1 XT y
2. Calculate the predicted values: yˆi = βˆ0 + βˆ1 xi,1 + βˆ2 xi,2 + · · · + βˆr xi,r
4. Use the estimates εˆi as basis for estimation of L, M and H, i.e., find L̂, M̂ and Ĥ by the
method of moments
5. For the new activity, cost element etc., find the corresponding x-vector, and denote it x =
[x1 , x2 , . . . , xr ]
6. Find the prediction of the new observation by y0 = βˆ0 + βˆ1 x1 + βˆ2 x2 + · · · + βˆr xr
Problem 7.5 Calculate the residuals in equation (7.19) with the data in Example 7.2 and estimate
the parameters by assuming the residuals are PERT distributed. Find the probability that the
duration for observation number 5 is shorter than the observed value of 6.1. From the observations
it seems that duration number 3 is rather long. Conclude on this by applying the regression model.
♢
78
7.5 Bayesian meothds
In Bayesian estimation procedures we utilise prior information about the relevant parameters. The
procedure could briefly be described as follows:
2. Structure reliability data information into a likelihood function, L(θ; x), see equation (7.2).
3. Calculate the posterior uncertainty distribution of the para,eter parameter vector, π(θ|x).
The posterior distribution is found by π(θ|x) ∝ L(θ; x)π(θ), and the proportionality constant
is found by requiring the posterior to integrate to one.
4. The Bayes estimate for the parameter vector is given by the posterior mean, which in principle
could be found by integration.
To establish the likelihood function, we look at the data. In this example we assume that we
have observed identical units for a total time in service, t, equal to 525 600 hours (e.g. 60 detector
years). In this period we have observed n = 1 failure. If we assume exponentially distributed failure
times, we know that the number of failures in a period of length t, N (t), is Poisson distributed with
parameter λ · t. The probability of observing n failures is thus given by:
and we recognize the posterior distribution as a gamma distribution with new parameters α′ = α+n,
and ξ ′ = ξ + t. The Bayes estimate is given by the mean in this distribution:
α+n 5.44 + 1
λ̂ = = = 0.78 · 10−6
ξ+t 7.78 · 106 + 525600
1
This could be based on statements from experts, see Øien et.al (1998), or by analysis of similar components
(empirical Bayesian analysis).
2
π(λ) ∝ λα−1 e−ξλ for the gamma distribution.
79
We note that the maximum likelihood estimator gives a much higher failure rate estimate
(1.9 · 10−6 ), but the “weighing procedure” favours the prior distribution in our example. Generally
we could interpret α and ξ here as “number of failures” and “time in service” respectively for the
“prior information”. ♢
80
Chapter 8
Expert judgements
Note that this chapter is widely based on: Øien, K. & Hokstad, P.1998 ”Handbook for performing
Expert Judgement”. SINTEF report STF38 A98419 (ISBN 82-14-00449-7). Whereas Øien &
Hokstad have a general approach, we will here focus on estimation where the objective is to assess
two or more parameters describing the uncertainty distribution of durations or costs.
8.1 Introduction
When statistical data do not exist or are not available, the alternative is to obtain such information
from experts/resource persons. To handle this in the best possible way, this should be carried out as
a structured and systematic process, both during planning, elicitation and calculation, as proposed
in this handbook. Such a structured collection of information is what we call “expert judgements”.
Thus, expert judgements are carried out to provide necessary input data for our analyses.
ESA [17] gives the following definition of expert judgement data:
“Expert Judgement data are estimates of unknown values about a system made by specialists
who have system-related knowledge.”
8.1.1 Purpose
The purpose of this chapter is primarily to give a simple and complete “recipe” of how expert
judgements may be carried out. Note that the referred handbook by Øyen and Hokstad gives a
significant more detailed picture than this chapter. At the end of the chapter combining expert
judgement findings with statistical data is also discussed.
8.1.2 Extent
Figure 8.1 shows tasks that are covered for performing expert judgements. This chapter covers all
steps of all 3 phases.
8.1.3 Use
It has been an overall objective to make this chapter flexible, that is, also to be applicable for
expert judgements carried out with little resources, and thereby in a simple way. In addition, there
are guidelines of how to perform more comprehensive expert judgements.
Therefore, it is not the intention that all the procedure items are to be carried out every time.
Which items are to be carried out depend on which resources are available, the problem to solve,
81
Problem
description
Evaluation of
available resources
Choosing
experts
Evaluation of
experts
Choosing method
of calculation
Developing a
PHASE I: THE PREPARATION PHASE
questionnaire
Information
to the experts
Accomplishment
of elicitation
Establishing
3-point estimates
Establishing a basis
for giving weights
Final
discussion
Possible adjustments
PHASE II: THE ELICITATION PHASE
of weights
Calibration
of experts
Weighting
of experts
Calculation of
combined estimates
Documentation
Communication/
Presentation
Supplementary
PHASE III: THE CALCULATION PHASE
work
Figure 8.1: The steps in an
82 expert judgement process
and the choices made during the work/process, (among others: choosing method of calculation).
All items should therefore be looked into, in order to evaluate whether they are relevant. The short
version in Section 3 gives a suggestion for what to include for very simple expert judgements.
This chapter represents an alternative to an unstructured and undocumented “engineering”
judgements.
Scenario analyses
Herman Kahn developed this method as a kind of system analyses method, where hypothetical
sequences of events are constructed for the purpose of focusing on cause processes and decision
points. This gives the answers to two kinds of questions:
1. Precisely how might some hypothetical situation come about, step by step?
2. What alternatives exist, for each actor, at each step, for preventing, diverting, or facilitating
the process?
This type of scenario analysis must not be confused with accident scenarios, and are only used to
examine the main trends, which are extrapolated into the future. This is done without evaluating
or using the probabilities of the scenarios to occur.
When the trends are extrapolated into the future, any theoretical or experience based knowledge
that may influence the extrapolation is considered.
1. An “observation team” (the analysts) defines the purposes and choose the respondents (ex-
perts). Normally, the respondents are anonymous to each other, and the answers are anony-
mous.
2. The observation team prepares a temporary questionnaire, which is sent to the respondents
for comments.
5. The observation team analyses the answers and calculates median values and the interquartile
range (the 25% and the 75% estimates).
83
6. The results are returned to the respondents, who are asked whether they want to adjust their
answers. Those who still are outside the interquartile range, are asked to give arguments for
their prediction.
7. The revised predictions are processed in the same way as the first responses, and arguments
for “outliers” are summarised. This information is sent back to the respondents, and the
whole process is iterated (3 - 4 times).
8. The median values on the final round are taken as the best predictions. Generally, the spread
in the last round is smaller than in the first round, which is taken to indicate a degree of
consensus.
This method was very popular in the 1960’s and 1970’s, but some later evaluation studies which have
seen carefully and critically on the Delphi method, have concluded that it violates methodological
rules for common experimental science. Comparisons with other methods have also shown that the
Delphi method was the poorest.
One major criticism is directed to the “reward” given to the expert for changing his estimates
towards the median value through group interaction. It has been shown that this does not increase
the relative frequency of correct estimates. The Delphi method seems to have lost its popularity.
84
Information from per-
SYSTEM sonnel who have knowl-
DEFINITION edge/experience about the
system
Figure 8.2: Expert judgements of input data compared to other type of information
85
modelling is of great importance (and may easily be disregarded or handed over to the analysts
only). This type of information/evaluation is neither what we usually view as an expert judgement.
Input data to our models may either be statistical data (objective /”hard” data) or subjective
(“soft”) data provided by personnel who have the necessary knowledge and experience to provide
this information. Such subjective data estimates is what we here view as expert judgements. Also,
we have restricted this to quantitative data. However, in general also qualitative estimations, as
for instance ranking, may be viewed as expert judgements.
Final results and conclusions are evaluated by a steering committee/client against the project
targets. Such an “expert judgement” of the results is also important, but is not what we usually
mean by the term expert judgements.
Evaluations/quality assurance performed by the person(-s) responsible of the quality assurance
are not illustrated in the figure, but should be performed in all phases of the project. Neither such
an evaluation is what we usually mean by expert judgements.
To summarise, we may say that the type of evaluations that are viewed as expert judgements
in this case is:
Documentation
By using a formal expert judgement, all the steps in the process are well documented. Normally,
by using engineering judgements, very little is documented. A more structured method, as expert
judgements, also promotes documentation of the results. Further, the information that is collected
often is less precisely defined by engineering judgements. Neither, important assumptions that
make a basis for the judgements are not necessarily stated.
86
Table 8.1: Advantages and disadvantages of strictly formalised expert judgements vs. informal
engineering judgements.
Factors Formal expert judge- Engineering judge-
ment ment
Structure Systematic and struc- Unsystematic and
tured method/process. unstructured process.
“Discussion across the
table”.
Specification of Well specified. Only in- Imprecise. Assump-
information formation given as an- tions are not specified.
swers of well defined
questions.
Documentation All steps of the pro- Poor or no documenta-
cedure are well docu- tion.
mented.
Extent of col- Limited. Only that ob- Wide. May cover many
lected informa- tained through prede- aspects of the subject,
tion fined questions. also through follow-up
questions.
Evaluation of ex- “Objective” rules for The confidence in a spe-
perts evaluation and possible cific expert is judged
weighting of the ex- subjectively by the an-
perts. alyst.
Simplicity Extensive and expen- Very simple. Performed
sive. without preparations.
87
answers to the questions, thus making demands to the questionnaire being worked out. There is
little room for follow-up questions and improvisation, something that may impair the completeness
of the information gathered about the subject. In engineering judgements, one may more easily
gain a general insight in addition to precise quantification of a given subject.
Simplicity
With its lack of structure and preparation, engineering judgements are often too simple to become
credible. Formal expert judgements have a tendency to become quite complex and extensive.
Thus, it is a problem that too extensive expert judgements hardly make a relevant alternative to
engineering judgements, due to high costs.
Thus, a conclusion is that nowadays’ engineering judgements have a big problem in respect
to documentation and objectivity. As for how complete the information gets, the picture is more
mixed. The advantage of having a well structured list of specified questions is acknowledged.
However, this can lead to a non-flexible elicitation, and is not necessarily the best in all situations.
A person that satisfies the first or the two first criteria, but not the last, is evaluated by Shanteau
to be a novice and not an expert. In his evaluation, such a person will not be able to train to
satisfy item iii). Thus, he will stay a novice, even though he sees himself as an expert, and also if
the people around him do so.
To us, this is a too rigid judgement of what is required to characterise a person as an expert.
We do not ask for “experts” meeting some fixed criteria, but rather the persons having as much
knowledge/experience on the subject as possible. Thus, we have quite a pragmatic view to what
we mean by the notion “expert”.
88
8.2.5 How should the expert judgement be carried out?
The method of this chapter is based on some basic requirements some of which were also discussed
in Section 8.2.3, were we performed an evaluation of expert judgements vs. engineering judgements.
The basic requirements are:
1. Documentation
2. Objectivity
4. Completeness
5. Simplicity
The requirements 1-4 are necessary to obtain scientific credibility. Requirement number five about
simplicity is included to secure a practical method that leads to a widespread use.
Unbiasedness
Unbiasedness may be defined as the degree of “accuracy” of the assessments, and describes to what
extent the assessments show a systematic deviation from the true value. Systematic deviations may
be caused by psychological factors, or by that the experts’ experience somehow is not representative
for the system in question (e.g. having different operational conditions than the expert is familiar
with).
A definition of biasedness may be:
Biasedness = The degree of systematic deviation from the true value
A common measure of the experts’ biasedness is:
Bias = mean of the estimated values - true value
89
Calibration
Calibration may have two meanings. Calibrate (as a verb) means to correct the estimates of
an expert that shows systematic deviations (giving several estimates), that is provide unbiased
estimates. This corrected (calibrated) estimate will then be the input to the analysis.
Calibration (as an adjective) is used by e.g., Cooke [18] to describe the accuracy of one specific
estimate (independent of systematic or random errors). Cooke defines calibration as:
Calibration = to what extent the estimated probability agree with the observed relative frequency
Thus, calibration (as an adjective) is a characteristic of one estimate (the distance between the
expert’s best estimate and the true value).Unlike unbiasedness (which is applied having more than
one estimate), poor calibration is not necessarily due to biasedness (it might be unsystematic). In
this memo we use the term ‘calibration’ as a verb.
Informativeness
Even though the experts do not show any systematic deviations from the correct values, this
does not mean that he always gives the correct values. He may even have random/unsystematic
deviations, which result in spread or variation of the estimates round the true values. Here we use
the notion informativeness to describe the experts’ unsystematic (unpredictable/random) deviation
from the correct values. This is also referred to as the degree of precision.
Informativeness = the degree of unsystematic variation/spread of the estimates around the true
values. (High informativeness = low variation).
An expert that has great variations in his estimates will, even though he sometimes strikes the
target, be perceived to have little informativeness (is not to “be trusted”), and should be given
small weight when his estimates are weighted with the other experts’ estimates.
One way to measure the informativeness is to use the sum of squares of the deviations between
the experts’ estimates and the true values.
NOTE! Some authors (e.g., [18]) use informativeness about the experts’ own evaluation of how
certain he is of his estimate, that is, the confidence interval he puts around his best estimate. This
we will denote “subjective informativeness”, see below.
Subjective informativeness
Subjective informativeness = the informativeness (degree of precision) of the estimate, as assessed
by the expert himself
This subjective informativeness is given by the confidence interval he puts around his best
estimate. For example in case of PERT distributed random quantities in the model, we ask first
the expert to assess the values of L, M , and H. Then for each of these, we ask the expert to assess
the informativeness regarding these values. For example for the parameter M , the expert gives
90
MEV (expected value for M ), MLV (low value for M ), and MHV (high value for M ). MHV – MLV
is an expression of how confident the expert is in his own estimate.
Note that it has been shown not to be a good correlation between an expert’s calibration
(“accuracy”) and his subjective informativeness (self-confidence). Thus, there is no guarantee
whatsoever that an expert who gives narrow confidence intervals gives better estimates (more
accurate) than one with wide confidence intervals.
Dependence
Dependence is often related to biasedness. If the expert gives biased estimates and this biasedness
shows special patterns, we talk about dependencies. For instance, the biasedness may increase
when the correct values increase. This represents a positive dependence (correlation) between the
correct values and the bias. (Dependence may be seen as a special type of biasedness.)
Resolution
Resolution = the ability to separate the probability of a specific event from the average probability
for the total set of events
An example of this is the ability to estimate the most likely duration of the given project with
it’s characteristics compared to a general project.
Consistency
Consistency = the assessment is independent of the method/approach that is used and also of when
the evaluation is carried out
Consistency means that the judgements are reproducible, that is, we get the same result inde-
pendent of time and method. For each expert this means that he is consistent if e.g. he provides
the same result/estimate, even though the question is phrased in different ways, and that he gives
the same answer if the assessment is repeated at a later stage of the elicitation. If the expert is
inconsistent, he should be given a lower weight than the other experts, when the overall estimate
is calculated.
Coherence
Coherence = conformity with the laws of the theory of probability
This may also be referred to as “logical consistency”.
For example, according to probability theory, Pr(A ∪ B) = Pr(A) + Pr(B) − Pr(A ∩ B), and
Pr(AC ) = 1 − Pr(A). If the expert does not provide answers that conform with these rules, he is
said to be incoherent. (If the experts are not coherent, the analyst should avoid using questions
that include probabilities.)
91
Reproducibility or Inter-expert reproducibility
Reproducibility = the extent to which different experts provide similar results to a given question
(when using the same method)
Lack of reproducibility should indicate that at least some of the experts do not give satisfying
estimates.
Reproducibility may be measured through the spread.
3. State the reason for the need for experts judgements in relation to the existing problem
9. Choose the experts with most knowledge/experience about the problem of interest
11. Avoid having too many experts with a great interest in the project (interest of a specific
result)
92
18. Formulate clear and simple questions
19. Explain to the experts the difference between uncertainty (variability) in the quantity to
assess (e.g., activity duration), and the uncertainty in the parameter estimates
5. Emphasise the importance of the experts being sincere and honest when giving the estimates
6. Call on the experts to not be affected/dominated by the other experts or by the ana-
lyst/process leader
7. Explain to the experts some of the most important phenomenon which causes biased/incorrect
estimates
9. The analyst/process leader must remain neutral and not actively take part in the evaluations
so that the estimates are affected
10. The analyst/process leader must be present during the elicitation to show interest, control,
and to answer questions
11. Estimate the duration of the questioning. (By individual estimation, max. 1 hour before
having a break).
12. The analyst/process leader must see to that not one (or a few) expert dominates or affects
the others
14. For each parameter of interest, be as specific as possible to describe the cost element, the
activity etc for which parameters are required
15. Ask the experts to first indicate their lowest and highest estimate. Thereafter, their best
estimate
93
16. If the variable of interest is PERT distributed, it is recommended to ask the experts to
estimate the most likely value (M ), the P10 and the P90 percentiles1 .
17. To assess the subjective informativeness ask the expert to give low and high values for each
of the three estimates in a typical case (base case)
19. Equal weights of the experts should be used when there is no strong argument for weighting
21. Finally, ask the experts if there are any questions that are not clear, so that these may be
clarified before ending the elicitation
22. In a group processes (after ind. proc.) the experts should affirm for that the result is complete,
before finishing the process.
5. All assumptions (made by both analyst and experts) should be gathered in a separate form
6. The reason why each expert is chosen, plus his strengths and knowledge of the current subject,
must be documented.
8. Any form for manipulation/weighting of the estimates must not be carried out without in-
cluding argumentation/documentation
10. The experts’ names should be stated, but each single estimate does not need to be allocated
to a specific expert, if anonymity is considered to encourage objectivity
11. Present the results to the client; how results are achieved, and who the experts are.
12. Graphical presentation of the results may be advantageous, especially to show the uncertainty.
13. The results, including the handling of each single estimate are presented to the experts (may
be in writing)
1
The Px percentile is defined by Pr(X ≤ P x) = x%. For example Pr(X ≤ P 10) = 10%
94
14. Evaluate whether the results seem reasonable and accurate.
15. Evaluate whether there is need for further work (more data)
18. Make a “debriefing” of the expert judgement process (how did it work?)
19. Correct the items in the procedure for carrying out the expert judgement, when weaknesses
are experienced.
Table 8.2: Estimates and correct values from 3 activities, one expert
Activity xi (true, known value) Yi (estimate)
A 50 70
B 60 90
C 70 140
∑
All Yi in this case are larger than xi , that is, Z = 3, but: 1/n i (Yi /xi ) = 1/3 (7/5 + 9/6 +
14/7) = 1.63 <(6 - n) = 6 - 3 = 3. That is, here it is chosen not to calibrate. ♢
95
8.4.2 Method for performing calibration
In the following we describe the principle for calibration if the criterion for calibration is satisfied.
We assume that there is a linear relation between the true values (xi ’s) and the estimates (Yi ’s):
Note that here we write x as a function of Y and not the other way as we usually do. To estimate
β0 and β1 we apply standard LS methods, i.e.:
∑
(Yi − Ȳ )xi
β̂1 = ∑i ( )2 (8.2)
i Yi − Ȳ
For a given value of Y , say y, we now estimate the parameter of interest by:
Note that if Y is less than Yi,min = mini Yi it is recommended to use a regression line through the
origin, i.e.,
( )
β̂0
x̂ = + β̂1 y (8.5)
xi,min
Example 8.2 With the data given above we find (although calibration was not recommended):
Assume now that we ask the expert to estimate a new value for duration of a new activity D (M =
Most likely value). He assigns the value y = 100. The estimate is now given by:
Equal weighting
Give the experts a weight 1/m, where m = the number of experts. Equal weighting of the experts
is used when there is no basis for performing differentiation of weighting (e.g. control questions,
mutual weighting, knowledge profile.)
96
Based on control questions
Give the experts a weight related to the estimated random error (variance) obtained from evaluation
of the control questions. This expresses the experts’ “true informativeness”. Note that we now write
the relation between x and Y in the normal way:
where the objective is to estimate the variance of the error term. To find the variance of the error
term we estimate the regression line, let SS be the square sum of the residuals for expert k:
∑
SS k = (yi − αˆ0 − α̂1 xi )2 (8.8)
i
i.e., the differences between the observation (yi ’s) and the estimated line. An estimate for the
variance is given by:
where n is the number of control questions given to expert k. If we use e.g., MS Excel to estimate
the parameters we directly find S as the “Standard Error”. If we have m experts the weight of
expert k is then
S −2
wk = ∑ k −2 (8.10)
j Sj
97
Expert no. (who is evauated)
1 2 3 ... m
...
2 p2,1
3 p3,1
...
...
m pm,1
and
Example 8.3 Assume that P̂10 = 6, M̂ = 10 and P̂90 = 14. By applying the pRisk.xls program
we may now generate one cell containing CDFPert(P̂10 , L̂ , M̂ , Ĥ), and one cell containing
CDFPert(P̂90 , L̂ , M̂ , Ĥ). Then we set a third cell equal to the sum of these two cells. By using
the solver requiring the sum to be one, and for example the cell with CDFPert(P̂10 , L̂ , M̂ , Ĥ)
to be equal to 0.1 in the “Constraints” editor, we find L̂ = 2.11 and Ĥ = 17.89. Observe that the
CDFPert(P̂90 , L̂ , M̂ , Ĥ) cell evaluates to 90%. ♢
98
8.4.6 Standard weighting model - Experts and data
In some situations we both have data and experts judgements. In principle we may use Bayesian
updating strategies to find the final estimates to apply. In the following we propose a slightly simpler
approach where we weight the expert judgement results with the estimates found by statistical
analysis of data.
Weight of experts
Equation (8.12) is used to find the combined estimate of the parameter of interest. One way to
weight this value with the data is to calculate the “variance” of the estimate in equation (8.12).
An unbiased estimator for the variance based on the variation of the expert statements is:
1 ∑
m
2
SVE = ∑m wj (x̂j − x̂)2 (8.15)
1− 2
j=1 wj j=1
If there are few experts this measure is not very reliable. A better approach may then to use the
“Subjective informativeness” assessed by the experts. Now, assume that each expert, k, has stated
an estimate, x̂k , but also low (x̂k,L ) and high (x̂k,H ) values for the estimate. Assume that the
low and high values corresponds to the P10 and P90 percentiles. If we assume that the underlying
uncertainty distribution of the expert is PERT distributed, we may apply the approach shown in
the previous example to find the L, M and H values of the uncertainty distribution, and hence the
variance by standard formulas. A more pragmatic approach would be to claim that the standard
deviation is proportional to the distance between the low and high value, and then try to assess
the proportional constant. Following the example we then assess the “self evaluated standard
deviation” by :
If Ŝk−2 is used as basis for the weighting of expert k, and the Ŝk−2 ’s are considered as true variances,
it can be shown that the variance of the weighted means equals the reciprocal of the sum of these
variances, and hence a reasonable estimate for the variance of the weighted estimate based on all
experts is found by:
2 1
SSE =∑ −2
(8.17)
j=1:m Ŝj
If there are few experts it is recommended to apply equation (8.17). If there are a medium number
of experts, say 3 to 5 one may calculate the variance by both equation (8.15) and equation (8.17)
and use the maximum of these two values. Generally we denote the estimate of the variance of the
expert judgement estimator by SE2 .
Weight of data
In some cases it is easy to find the variance of the estimator used when statistical data is available.
For example the standard deviation of the estimator for the mean value in the normal distribution
is found by the sample standard deviation divided by the square root of the number of observations.
In more complicated situations we may use the principle of bootstrapping to find the variance of the
estimator. The procedure is as follows: Let z1 , z2 , ... , zn be observations from the distribution for
which we are seeking estimates. Apply an estimation procedure to estimate the parameter vector.
Denote the result by θ̂. Now repeat, k = 1, 2,…
99
1. Generate n pseudo random numbers from the actual probability distribution with parameter
vector θ̂.
2. Apply the estimator again to find a new estimator for the pseudo generated numbers, and let
the result be denoted θ̂k for the k th iteration.
The sample variance of the j th element of the θ̂k -sequence may now be used as an estimate for the
variance of the estimator based on data. Let SD 2 denote the variance based on the data.
Combining it
If we let x̂E and x̂D denote the estimates from combined experts and data respectively, and further
SE2 and SD 2 the corresponding estimated for variances, we find the final weighted estimate of experts
SE−2 x̂E + SD
−2
x̂D
x̂ = −2 −2 (8.18)
SE + SD
All experts are asked to give the most likely value (M ) for a new activity D, with corresponding
uncertainties in the estimate shown in Table 8.4:
Table 8.4: New values and self assessed uncertainty by the experts
Activity Y (M in PERT) Low Y (P10 ) High Y (P90 )
Expert 1 100 75 150
Expert 2 80 60 95
Expert 3 60 55 70
Observation from previous project shows the following costs: 53.2, 60.4, 52.4, 66.8, 56.2, 72.2,
71.7 and 73.2.
Find a combined estimate from the experts, and an estimate from the data. Then combine
these two estimates into a final estimate.
We decide to calibrate the expert independent of the “calibration test”. Regression analysis
from MS Excel is shown in Table 8.5.
Predictions of a new value for a new activity, D, by the experts are shown in Table 8.5. Low
and high values, specified by the expert are given in Table 8.6.
100
Parameter estimate Expert 1 Expert 2 Expert 3
β̂0 = x̄∑− β̂1 Ȳ 33.08 45.00 36.92
(Yi −Ȳ )xi
β̂1 = ∑
i
2 0.27 0.25 0.35
i (Yi −Ȳ )
Table 8.5: New values predicted by the experts with calibrated values
Quantity Expert 1 Expert 2 Expert 3
Y = Prediction by expert 100 80 60
x̂ = β̂0 + β̂1 y 60.00 65.00 57.69
Table 8.6: Low and high values specified by the expert and derived results
Quantity Expert 1 Expert 2 Expert 3
Low Y by expert =P10 75 60 55
High Y by expert =P90 150 95 70
Calibrated low, (P10 = x̂k,L ) 53.27 60.00 55.96
Calibrated high, (P90 = 73.46 68.75 61.15
x̂k,H )
Ŝk = 0.37(x̂k,H − x̂k,L ) 7.471 3.238 1.921
−2
Ŝk 0.018 0.095 0.271
Ŝ −2
wk = ∑m k −2 0.047 0.248 0.705
j=1 Ŝj
101
Weighted calibrated prediction based on self evaluated weights:
∑
m
x̂ = wj x̂j = 59.61
j=1
2 1
SSE =∑ −2
= 2.60
j=1:m Ŝj
Problem 8.1 Assume that we have three experts. They are all calibrated with three control
questions shown in Table 8.3 above. All experts are asked to give a high value (H) for a new
activity D, with corresponding uncertainties in the estimate as shown in Table 8.7. Observation
from previous project shows the following costs: 53.2, 60.4, 52.4, 66.8, 56.2, 72.2, 71.7 and 73.2 (all
in thousands).
b. In the remaining of this you should calibrate the experts independent of the result from (a).
Calibrate the experts.
c. Find an estimate of the variance of each expert (i.e., around the regression line).
d. Use the variance of each expert as a basis for weights to each expert. Then find a weighted
estimate for the high value.
e. Estimate the parameters in the distribution by assuming a PERT distribution based on the
data given.
g. Combine the estimates from data and experts to find a final estimate for the high value (H).
Table 8.7: New values and self assessed uncertainty by the experts for the high value in the
distribution
Activity Y (H in PERT) Low Y (P10 ) High Y (P90 )
Expert 1 140 105 210
Expert 2 96 72 114
Expert 3 72 66 84
102
Chapter 9
9.1 Introduction
Turnarounds are scheduled events wherein an entire process unit of an industrial plant such as an
oil & gas production platform is taken shut-down for an extended period for modification and/or
renewal. Sometimes the term ‘revision stop’ is used to denote a turnaround.
Turnarounds are expensive - both in terms of lost production while the process unit is offline
and in terms of direct costs for the labour, tools, heavy equipment and materials used to execute
the project. Turnarounds have unique project management challenges due to the large number of
activities to execute, logistics for labour and spare parts, and safety issues due to many activities
in parallel.
In this chapter we mainly discuss how maintenance activities either could be conducted individ-
ually, or if they should be in cooperated in a turnaround to save cost since the production already
is down.
103
inspection. The main rationale for preventive maintenance is that it is cheaper to prevent a failure
from occurring by a preventive maintenance task than it will be if a failure occurs. A range of
optimization models have been derived in order to determine the appropriate level of preventive
maintenance. The reason why maintenance is addressed within project risk management is that
turnarounds essential is a huge package of preventive maintenance activities organized as a project.
where CPM is the cost of a preventive maintenance action (to prevent failures), CCM is the cost
of a corrective maintenance (CM) action (given that a failure did occur), λE (τ ) is the effective
failure rate, i.e., the expected number of failures per time unit when the component is preventively
maintained every τ time unit, CEP is the expected production losses upon a component failure,
and finally CES is the expected safety cost upon a component failure, including material damages
and environmental losses.
In the following we let CU = CCM + CEP + CES denote the expected unplanned cost upon a
failure. The effective failure rate depends on the life time distribution of the component. The
Weibull distribution is a widely used distribution for aging components. In the case of Weibull
distributed life times we may find approximation formulas for the effective failure rate. If we know
the mean time to failure , MTTF (without maintenance), and the ageing parameter (α) of the
lifetime distribution of the component, the effective failure rate may be approximated by:
( )
Γ(1 + 1/α) α α−1
λE (τ ) = τ (9.2)
MTTF
where Γ(·) is the gamma function. The approximation is good when the maintenance interval is
small compared to the MTTF. If the maintenance interval is approaching the MTTF value, the
approximation in equation (9.2) is not very accurate, and we might use the following improved
approximation:
( )
Γ(1 + 1/α) α α−1 [ ]
λE (τ ) = τ 1 − 0.1α(τ /MTTF)2 + (0.09α − 0.2)τ /MTTF (9.3)
MTTF
In MS Excel the gamma function could be found by Gamma(x). In the following we will always
assume that the approximation in equation (9.2) is sufficient for our purpose. By setting the
derivative of C(τ ) in equation (9.1) equal to zero, we find the optimal interval to be:
( )1/α
MTTF CPM
τ∗ = (9.4)
Γ(1 + 1/α) CU (α − 1)
1 fX (x)
The hazard rate is given by hX (x) = 1−F X (x)
, where fT (t) is the probability density function of the lifetime
T of the unit, and FT (t) is the cumulative distribution function. The hazard rate is also to be understood as the
conditional probability of a failure in (t, t + ∆t) given that the unit has survived up to time t. In some textbooks
the hazard rate is also dented the failure rate function.
104
9.2.3 Single activity - Consideration for inclusion in turnaround - Static con-
sideration
A turnaround is usually conducted every year, every two years, or every three years. There is an
aim to conduct the turnaround in a period where working conditions are good (summer period),
and where also production losses are as low as possible which historically also has been in the
summer period since energy prices usually reduces when the demand goes down. It is therefore a
common practice to conduct the turnarounds in the summer. The interval between turnarounds is
denoted τTA .
The formula for the average cost C(τ ) in equation (9.1) needs to be modified if maintenance is
considered. Let CTA be the preventive maintenance cost if the maintenance of the component is
included as part of the turnaround. Usually we will have that CTA <CPM . We will now investigate
whether the activity should be included in the turnaround. The criterion for inclusion in the
turnaround is that the average cost is reduced. An example is given to demonstrate the steps. The
relevant parameters are given in Table 9.1 where the time unit is months, and cost unit is thousand
NOKs.
Table 9.1: Parameters for the decision regarding inclusion in turnaround, τ ∗ >τT A
Parameter Value
α 2.5
MTTF 48
CPM 5
CTA 4
CU 50
τTA 12
We first apply equation (9.4) and get τ ∗ ≈ 18.3 months. The total average cost in the situation
where the PM activity is not included in the turnaround is thus ≈ 0.455 per month by inserting
τ = τ ∗ in equation (9.1). If the PM is included in the turnaround, the average cost is found by
inserting τ = τTA and CPM = CTA in equation (9.1). This gives a monthly cost of ≈ 0.43 <0.455.
It is thus in average cheaper to include the PM into the turnaround. We observe that τTA <τ ∗
<2τTA . We might therefore also consider to insert τ = 2τTA and CPM = CTA in equation (9.1).
This gives a monthly cost of ≈ 0.44 which is still better than the situation without including the
PM in the turnaround, but less favourable comparing by including the PM in every turnaround.
The optimal strategy is thus found to be to include the PM activity in every turnaround with the
given assumptions.
Problem 9.1 We will consider a preventive maintenance action (PM) for inclusion in a turnaround.
For detailed description of the situation, see the introductory example above and use the data in
Table 9.1. Find the optimum interval for the maintenance interval if the PM is conducted outside
the turnaround by using the Excel Solver. ♢
Problem 9.2 Since the maintenance interval is rather close the MTTF the suggested approxima-
tion formula for the effective failure rate is not very accurate. Find the optimum interval if the
improved approximation in equation 9.3 is used. ♢
Problem 9.3 In the example above it was shown that the long run average cost will be lower both
in the case we include the PM in every turnaround, and in the case where we include it in every
105
second turnaround. Since the individual optimal interval is close to 18 months, it seems reasonable
to include the PM in every third turnaround. This means that average PM cost will be reduced.
Find the total cost for this strategy. ♢
If we choose to skip the PM activity for the coming turnaround, and wait till the next we get
the following total cost:
where λE (τTA + x) · CU · (τTA + x) is the expected “unplanned” cost from the last PM until the
next turnaround, and λE (x) · CU · x is the expected “unplanned” cost from the last PM up to now
(i.e., what already has been “paid”). With the example data we get:
and
Hence, the PM activity should not be included in the coming turnaround but in the turnaround
coming next year. Note that we might consider to execute the PM in between the two coming
turnarounds, but the example in Section 9.2.3 indicates that the gain by synchronizing with a
turnaround is always the best alternative.
106
In the modelling we still assume Weibull distributed lifetimes. For the Weibull distribution2 we
have:
α
fT (t) = αβ(βt)α−1 e−(βt) (9.7)
α
R (t) = Pr (T > t) = 1 − FT (t) = e−(βt) (9.8)
Γ(1/α + 1)
β= (9.9)
MTTF
In the modelling we now we still assume that the cost of an unplanned failure is CU . But in addition
to the immediate cost of a failure we assume that for the remaining time until the next opportunity
for maintenance there will be a cost of CW per time unit until the component could be repaired.
If we include the PM in the coming turnaround the total cost in the next period (i.e., from now
until the PM next year) is:
∫ τTA
CFirst TA = CTA + [1 − R(τTA )] CU + fT (t) (τTA − t) CW dt (9.10)
t=0
If we choose to skip the PM activity for the coming turnaround, and wait till the next we get the
following total cost in between the two turnarounds:
∫ τTA
CSecond TA = [1 − R(τTA + x)/R(x)] CU + fT (t + x) (τTA − t) CW dt/R(x) (9.11)
t=0
We proceed by the same data as in the previous examples but we now also let CW = 10. We
then get:
∫ τTA
CFirst TA = CTA + [1 − R(τTA )] CU + fT (t) (τTA − t) CW dt ≈ 5.93
t=0
and
∫ τTA
CSecond TA = [1 − R(τTA + x)/R(x)] CU + fT (t + x) (τTA − t) CW dt/R(x)
t=0
≈ 7.06
Note that we here have assumed that even if we include the PM activity in the first turnaround,
it will also be included in the second turnaround. In a more general setting, it might be that if
we include now, we might skip the activity in the next turnaround. The analysis then becomes
more complicated because to assess the costs that follow one needs to take into account different
strategies after the second turnaround.
2
Note that various parametrization exist for the Weibull distribution, and also note that the the symbol for the
scale parameter often is λ rather than β as used here.
107
9.3 Single component - Impact on turnaround duration
Up to now we have considered that the cost of including the PM activity in the turnaround could be
described by a single number, CTA . Since the inclusion of any activity in a turnaround will influence
the duration and complexity of the turnaround, it might be an oversimplification to consider a fixed
cost incurred by the PM activity we are considering. At the end, we might still be able to calculate
an expected cost for CTA , but we will now investigate the situation in some detail.
Assume that there are several activities that are to be included in the turnaround, and that the
PM activity we are considering is the last activity to include. For the time being, we assume that
the duration of the turnaround is fixed, say DTA . This is to say, that the planned duration is DTA .
The actual duration is a random quantity which we denote TTA . Now, assume that we have an
oversight over the remaining activities to include. The question is whether we should include the
PM activity or not. Let ATA,0 denote the set of activities and their relations already considered
to include in the turnaround. Further let ATA,1 be the activities to consider if we include the PM
activity we are considering.
where fTI (t) and fTO (t) are the probability density functions for the project duration with and
without the PM activity included respectively (I = Included, O = Outside the turnaround). DTA is
the due date for the turnaround, where there will be no production loss if project duration T is
less than DTA , and the production loss is T -DTA if T is greater than DTA . The situation is now
identical to the situation in Section 9.2.3 if we set:
Thus the total cost of the PM activity execution within the turnaround is the fixed cost plus
the random cost, E(CTA,R ), caused by the increased risk of exceeding the due date. Note that a
first approximation is presented here. In reality the proses of select activities for inclusion is more
challenging than considering one and one activity individually. Some kind of dynamic programming
to select the most appropriate PM activities out of a huge list is required.
108
if the turnaround for some reasons is delayed. Assume there is a critical milestone upfront of the
execution of the PM activity. If a critical time is passed, one need to consider to skip the PM, and
rather execute it at its optional time (τ ∗ ). Let Let pD denote the probability that the turnaround
is delayed and one needs to consider to skip the PM activity as part of the turnaround. If the PM
activity is skipped, it is assumed to be conducted at its optimal time, i.e., at τ ∗ -τTA time units
after the turnaround. The decision diagram is shown in Figure 9.1.
PM in TA
& problems
Keep inside TA
CN
The PM is planned 1
inside the TA, (τTA ) PM in TA without
No problems any problems
109
when the coming turnaround is executed. The duration of the turnaround is so short that we may
ignore this duration in comparison to τTA . Further we assume that the previous execution of the
PM took place at time t0 − τTA , i.e., at the previous turnaround. The following cost equations may
now be found for each end consequence, cf Figure 9.1:
PM outside TA
PM in TA & problems
∫ ∞
C3 = CTA,E + CTA,P + CU λE (τTA )τTA + CPL fTI (t)(t − DTA )dt (9.16)
t=DTA
With these cost figures the optimal decision is found by processing the decision tree in Figure 9.1.
• Currently there exist tasks that from a safety point of view need to be conducted at intervals
of length τTA,0 , where τTA,0 is the . In order to extend the turnaround interval it is therefore
necessary to improve the reliability of these components, or propose more extensively condi-
tion monitoring in the period between turnarounds. In the following we will only consider
the possibilities to increase component reliability.
• Currently there exist tasks where the “optimum” interval is close to τTA,0 , and for these
components we will also consider to increase the reliability by conducting upgrade projects
prolonging the MTTF.
110
The PM is planned for the second turnaround PM in second
TA
PM in TA
& problems
Keep inside TA
CN1
The PM is included in
the coming turnaround PM in TA without
No problems any problems
• τTA,0 = 2 years in the example, and the question to be raised is whether it pays off to extend
the interval, i.e., τTA = 3.
111
Table 9.2: Reliability and cost data treating safety issues only
i ni CUG,i CPM,i CTA,i MTTFi αι CU,i i ∈ {SC}
1 4 8 4 1 Yes
2 3 2 1 0.5 Yes
3 4 3 2 1 Yes
4 4 2 5 4 20 No
5 8 2 6 3 30 No
6 6 3 6 2.5 25 No
where we have not taken discounting into account. By calculating C(τTA ) for τTA = 2 and
τTA = 3 in equation (9.18) we may calculate the yearly gain or loss by changing the interval of the
turnaround. If there is a gain, we also need to consider the (yearly) cost of the upgrading project:
∑
CUG = ni CUG,i /T (9.19)
i∈{SC}
where T is the number of years to consider, e.g., T = 10 if the installation will be disposed after
10 years. CU G should not exceed the gain by increasing τTA from 2 to 3 years.
Problem 9.4 Consider the example data above, and calculate the yearly cost for τTA = 2 and
τTA = 3 in order to determine if the turnaround interval could be increased. ♢
It is a simplification to ignore the discounting of the cost elements in this situation. We will now
discuss how to discount the variable cost related to failures. λE (τTA ) represents the yearly expected
number of failures if the time unit is year as we assume in this example. It is then reasonable to
calculate the expected number of failures each year j, j = 1,.., τTA and discount the cost to present
values. We assume that all costs are incurred at the end of the year, although it could also been
argued that it is more reasonable to assume that a failure occur in the middle of a year. We now
introduce ΛE,i as a measure of “discounted number of failures” in a period between two turnarounds
where the turnaround interval is τTA :
∑
τTA
ΛE,i (τTA , r) = (jλE,i (j) − (j − 1)λE,i (j − 1)) (1 + r)−j (9.20)
j=1
where r is the interest rate. Further we let n(T ,τTA ) be the number of turnarounds for an installation
that is disposed in year T (no turnaround the last year) when the turnaround interval is τTA .
Table 9.3 shows n(T, τTA ) as a function of τTA and T .
112
We always assume that the turnaround is executed at time t = 0, then after time τTA , 2τTA
and so on. We still assume that the upgrade of all safety critical components takes place. The
turnaround related cost up to time of disposal is given by:
C A (τTA ) =
∑n(T,τTA ) ( ∑ ∑ )
−(j−1)τTA
(1 + r) CTA,B + ni CTA,i + ni CU,i ΛE,i (τTA , r) +
j=1 i∈{SC} i∈{NC}
∑
∆CFLP + ni CUG,i (9.21)
i∈{SC}
where ∆CFLP is a correction term to account for failures the last period, i.e., the difference between
ΛE,i (τTA ,r) calculated after the last turnaround, and the actual discounted number of failures
multiplied with the failure cost and summed over all activities that are not safety critical. There
might be a difference because either we skip a turnaround the last year and hence there is an extra
failure cost for that year not accounted for, or because we accumulate failure costs in the calculation
formula for the period after T . If T is large, i.e., in the order of magnitude 10 years, we may ignore
∆CFLP due to discounting effects.
The accumulated turnaround related cost for the remaining life of the installation may now be
compared by applying equation (9.21) for τTA = 2 and τTA = 3 respectively, where we for τTA = 2
do not include upgrade costs, CUG,i .
If upgrade cost, CUG,i , is high for some components we may also consider to exclude the corre-
sponding activities as part of the turnaround. This means to remove the corresponding terms for
ni CTA,i and ni CUG,i in equation (refeq:turnaroundLCC1), and similarly add the cost of preventive
∑n(T,τTA,0 )
maintenance: j=1 (1 + r)−(j−1)τTA,0 ni CPM,i to the total cost for τTA equal the extended
turnaround interval (i.e., τTA = 3) and τ TA,0 is the original turnaround interval.
Problem 9.5 Consider the example data above, and calculate the total accumulated cost for the
remaining period up to year T for τTA = 2 and τTA = 3 in order to determine if the turnaround
interval could be increased from a life cycle cost perspective. ♢
Note that the cost terms CUG,i might contain some implicit interrelations. Since the upgrade is
part of the next turnaround, it is reasonable that adding more and more upgrade projects totally
increases the cost more than the individual activities requires due to the increase of complexity,
and impact on the shutdown duration. This is not pursued further in this presentation.
113
If there are no interrelations between upgrading cost of the production related activities we may
perform individual analysis for each of the activity types. If there are interrelations the situation
becomes much more complicated. A dynamic programming approach might be required, or at least
we need to consider the activities pair wise.
In the following we assume that we may treat each activity individually.
Keep the activity in the turnaround, upgrade in order to prolong the MTTF
We apply equation (9.21) but need to add the cost of upgrading, i.e.,ni CUG,i . Since MTTF is
prologned the term ΛE,i (τTA , r) will also be reduced.
n(T,τi∗ )
∑ ∗
(1 + r)−(j−1)τi (ni CPM,i + ni CU,i ΛE,i (τi∗ , r)) (9.22)
j=1
Note that τi∗ is not an integer so we may need to rewrite the n() and Λ() functions.
Problem 9.6 Assume that we increaseτTA to 3 years. For each i ∈ {NC}find out whether it pays
off to pay the upgrading cost. ♢
• Activity: F - Install new Flare tip system, may start when all the other activities are com-
pleted.
114
Table 9.4: Activities to include and durations (normal conditions)
ID Activity L M H
0 Shutdown 1 2 3
A Flare gas meters 1 2 3
B Change flare tip 1 2 3
C LP Flare drum 4 5 6
D NF HP Flare drum 3 4 5
E HP Flare drum 3 4 5
F Install new Flare tip system 5 6 8
In Table 9.5 we have listed some threats related to activity: B - Change flare tip. We will consider
only those activities where a schedule impact is explicitly listed. The turnaround is scheduled for
14 days. It cost 10 million NOKs each day the turnaround is delayed. Costs in Table 9.5 are given
in million NOKs.
Problem 9.7 We will consider the impact of including the PM into the turnaround. So far we
have included the following activities (PERT distributed durations) as given in Table 9.6
First all A-activities are conducted in serial, then all the B-activities also follows in serial, but
in parallel with all the C-activities. The C-activities also follows each other in serial. The duration
of the shutdown is 14 days. After 14 days there is a penalty of 50 (thousand NOKs) per day.
Assume the fixed cost of the PM activity (i.e., the cost contribution not related to the duration)
is 2 (thousand NOKs). Find the total expected cost of including the PM activity after activity C3
(there is some expected slack in the “C-branch” of the model). Assume the duration of the PM is
∼ PERT(1.5,2,4)-distributed. ♢
Problem 9.8 Conduct a cost benefit analysis of the measure where one wants a more powerful
helicopter taking the risk elements of Table 9.5 into account. ♢
115
Table 9.5: Threats related to activity: B - Change flare tip
Threat Probability Schedule impact Comment/Measure Cost
element
Technical 0.05 1 Technician on type (TLM) 0.05
problem be available for repair.
with heli-
copter or
equipment
Failure in 0.01 Accepted and tested strap
clearance of arrangement for the flare tip,
lifting specified by OM.
Wind force 0.08 1 per windy day Given wind force over 30 1
over 30 knots knots (critical limit) the day
of execution, the probabil-
ity of the wind problem re-
peats the subsequent days is
50% for each day. Proposed
measure: More powerful he-
licopter that may operate in
stronger wind. To model the
impact, we assume that all
the probabilities are reduced
by 30%
Failure in ra- 0.02 Stop all operations until
dio commu- the problem is solved/re-
nication established.
Loosing 0.01 Insist take off and delivering
object fly area. Cancel the take-off and
around rotor postpone operation until the
wind area is safe. This has to be a
continuous process.
Failure with 0.03 Use of certificated and
lifting strap approved lifting equipment.
Prohibited zone for flying
with hanging load has to be
defined.
Failure with 0.02 All equipment are to be func-
lifting hook tion test before operation.
Prohibited zone for flying
with hanging load has to be
defined.
Unexpected 0.05 Use of certificated and ap-
rotation on proved straps and lifting
load equipment. In case the pilot
can’t stop rotation by chang-
ing of the speed, the load has
to be return back to helideck
and defined again.
Failure with 0.01 Personnel shall not stay un-
116
motor or der the loading zone where
loss of motor the load is lowering.
craft
Table 9.6: Activity durations
Activity L M H
A1 0.5 1 2
A2 0.2 1.5 3
A3 2 2.5 4
B1 1 2 4
B2 1 3 5
B3 1 2 3
C1 0.5 1 2
C2 1 3 5
C3 0.2 0.5 4
117
Chapter 10
Portfolio management
10.1 Introduction
A project portfolio is a group of projects to be carried out under the sponsorship of a particular
organization. The portfolio may be seen from various stakeholders. For example it might be a
set of oil fields to be developed for a major oil company. Further, it might be a set of installation
project for a sub-sea equipment manufacturer. A third stakeholder will be a financial institute
(banks etc.) supporting the project financially in a limited time period. In this chapter we will
discuss the following:
• Probability that actual cost is larger than a factor, f , times the budget asked for, for example
f =1.5, and f = 2
118
Table 10.1: Project key numbers
Project Name NPV Pr(Cost overrun) Return ratio
A 10 0.3 3
B 5 0.2 2
C 8 0.1 2
D 4 0.45 5
E 7 0.3 4
F 4 0.2 3
Bubble diagrams are convenient to visualise tree of the dimensions at a time. To cover more
dimensions several bubble diagrams are required. We will consider the data shown in Table 10.1:
The net present value (NPV) represents the total discounted cost and income in the project,
either for a period of years, or an infinite time horizon if relevant. The corresponding bubble
diagram is shown in Figure 10.1.
Pr(Cost overrun)
0.6
0.5
D
0.4
0.3 E A
0.2 F B
0.1 C
0
2 4 6 8 10 12
Net present value
Figure 10.1: Example of bubble diagram where the size of the bubbles represents the return ratio
Si = Σj wj · xij (10.1)
We return to the example presented in Table 10.1, where the following weights has been introduced:
w1 = 4 = Weight of (expected) NPV
w2 = -2 = Weight of probability of overrun (e.g., more than a factor of 1.5)
w3 = 3 = Weight of return ratio
The calculated costs are shown in Table 10.2. The most promising projects according to the
applied weights are A, E and C respectively. Projects might be chosen according to the calculated
119
scores, where budget constraints might limit the number of projects. Assume that the budget is 8
units.
Starting from the top of the list sorted on the score gives projects A, D, E and F where we
always fulfil the budget constraints. We then spend 7.2 cost units, and the total score of the selected
projects is 142.5. In general the problem is a dynamic programming problem. In order to set the
weights often the Analytic Hierarchy Process (AHP) is applied, see e.g., [31].
Note that the score of each project is a weighted sum of various criteria. It is, however, not
obvious that such a weighted score make sense in selecting a range of project. If there are budget
constraints as indicated above, the best strategy would obvious be to chose project with a low cost
and a reasonable high score. An alternative approach would be to derive a “utility” based score.
Ideally we could find a utility function and then search for a portfolio of projects that maximizes
expected utility under the budget constraints. By introducing a utility function the probability of
cost overrun could be discarded from the utility function since the uncertainty related to overrun
is reflected in the concave utility function. The return ratio would also be of less importance.
Obvious it seems reasonable to consider the return ratio, but it might be argued that how much we
get back of the money we invest is not the most appropriate measure, the most interesting measure
is the total profit at the end of the day. A pragmatic approach is now to use the NPV value as the
baseline for the score of a project, but introduce a penalty for projects with high probability of cost
overrun, and projects with high cost, i.e., those projects that lock capital. A proposed measure is:
where the weights 0.3 and 0.2 represent how negative overruns and lack of capital are considered
respectively.
Table 10.3 is prepared for using the Excel solver to optimize which projects to include. In the
column “Included” a binary variable is introduced, where 1 means that the project is included, and
0 means that the project is not included. In the Excel solver all the “Included” variables are defined
as binary variables in the “Constraints” (Underlagt begrensningene) in Figure 10.2. Further the
sum of the project cost (cell J8) is limited to 8 units. The target cell is here set to cell I8 for the
alternative score, S∗ . The optimization in this particular situation gives the same result as for the
original score. The reason for this is that project C locks much capital even though the score is
high (S∗ = 4.8 which is the second highest) .
120
Table 10.3: Project summary, alternative score S∗
ProjectNPV Pr(Cost Return Project Score IncludedScore Score* Pr.Cost
Name over- ratio cost |Incl |Incl |Incl
run)
A 10 0.3 3 3.3 48.4 1 48.40 6.33 3.33
B 5 0.2 2 2.5 25.6 0 0.00 0.00 0.00
C 8 0.1 2 4.0 37.8 0 0.00 0.00 0.00
D 4 0.45 5 0.8 30.1 1 30.10 2.64 0.80
E 7 0.3 4 1.8 39.4 1 39.40 4.55 1.75
F 4 0.2 3 1.3 24.6 1 24.60 2.53 1.33
Σ= 142.50 16.06 7.22
121
portfolio analysis we distinguish between random variation and systematic variation. Random
variation represents variation within one project independent of the other projects. Systematic
variation represents variation due to one or more factor possible affecting two or more projects. In
portfolio analysis it is the systematic variation which usually is of concern. The random variation
will be wiped out due to the “law of large numbers”. We will therefore in the following discuss
systematic variation. We now propose a 8 step method for selecting projects taking systematic
variation into account.
2. Choose the most promising projects based on a preliminary score (Si ), but include more
projects than the total budget constraints
4. For each project the vulnerability factors are considered explicitly. Further for each combi-
nation of the vulnerability factors recalculate the expected values of the key numbers. Let
{V}be the set of all combination of the vulnerability factors.
5. Present the corresponding projects key figures by bubble diagrams for the corresponding
values of {V}in order to visualise the impact of the vulnerabilities.
6. Recalculate the expected values of the weighted sum for various subsets of projects using
the law of total probability, where pv is the probability for element v in {V}. This means
that we for each v in {V}first calculate the sum of scores, Σi Si |v, for projects included by
conditioning on the value v, then we sum these weighted scores over all v in {V}in a new
weighted sum using the pv -values.
9. Highlight the vulnerability factors as an important input to the overall portfolio management
(e.g., risk registers for each project, and global enterprise risk registers).
Table 10.4 shows the situation where only one vulnerability factor is considered. Two values are
considered, © = Good, § = Bad.
122
10.4.1 Dynamic programming - Portfolio selection
The situation described inTable 10.4 is of a dynamic programming type. The knapsack problem is
one of the standard example used to illustrate the knapsack problem:
• Determine the number of each item to include in a collection so that the total weight is less
than or equal to a given limit and the total value is as large as possible
An introduction to dynamic programming is given in Section 11.3. Note that for dynamic pro-
gramming we need to do the programming, whereas if we use mixed integer programming we use
Excel without any “programming” (coding).
• Define the objective function to be the sum of score for included projects. The objective
function is the function to maximize (Set Objective in Excel.
• Each project is described by a binary variable. If we arrange the projects as a table in Excel,
we introduce a column named: Included.
• The set of binary variables are the decision variables for maximizing the objective function.
In excel we name the range of all “included” cells, i.e., we give them the name Included
• In Excel, we use the Subject to the Constraints: to ensure our binary variables to be
binary, i.e., we use the bin option in the constraints editor.
• We also need to add the constraints regarding maximum cost, maximum vulnerability etc in
the Subject to the Constraints:-window.
Note that we typically define a cell in Excel to hold the objective function. This cell could then
be SUMPRODUCT(Scores,Included) where Scores is the named range of the scores of all projects,
and Included is the named range of our binary variables.
Problem 10.1 Discuss steps 1-8. Maximise expected NPV under the constraints that the total
spending given the bad value (§) of the vulnerability factor should not exceed 12. You may either
use mixed integer programming or dynamic programming. ♢
123
10.5 Scheduling of projects within a portfolio
Given a selected portfolio of projects we need to schedule them in time. In some situations the
scheduling might be unproblematic, but in other situations we need to take resource constraints
into account. We will in the following only consider one critical resource (e.g., labour force of a
critical discipline). We assume that we have done a preliminary assessment of resources required in
each project in the various project phases. We will not give a complete approach to the scheduling
problem, but just illustrate some aspect of the scheduling problem.
Table 10.5 shows the example data we will use. The variables are discussed in the following:
• DDi = Due date project i
• AR =Available resources per time unit. Here we consider AR = 100 = fixed value for the
entire planning horizon.
We assume that all projects run through three phases, “ramp up”, “full speed” and “ramp
down”. From Table 10.5 we see that progress per time unit is less for ramp up and ramp down
compared to full speed. But we also see that the demand for resources is lower in the ramp up and
ramp down phases.
124
• xFS,i = Starting full speed project i
• xi = End of project i
Figure 10.3 illustrates the resources needed in the various phases, and we also indicate the progress
µ per unit time. We will simplify and make some assumptions:
Resources
• We can not run all 3 projects on full speed simultaneously due to constraint limitations
• The cost of using resources equals the resources used, i.e., measuring on the same scale as
penalties. In general we need some scaling factors.
• xRU,i ≤ xFS,i ≤ xRD,i ≤ x,i , i.e., the chronological order of the phases, cf. Figure 10.3
Z=MAX(X_1-DD_1,0)*PD_1
+MAX(X_2-DD_2,0)*PD_2
+MAX(X_3-DD_3,0)*PD_3
+(XFS_1-XRU_1)*RRUD_1+(X_1-XRD_1)*RRUD_1+(XRD_1-XFS_1)*RFS_1
+(XFS_2-XRU_2)*RRUD_2+(X_2-XRD_2)*RRUD_2+(XRD_2-XFS_2)*RFS_2
+(XFS_3-XRU_3)*RRUD_3+(X_3-XRD_3)*RRUD_3+(XRD_3-XFS_3)*RFS_3
125
The contribution from project 1 is:
• MAX(X_1-DD_1,0)*PD_1 = number of days exceeding the due date multiplied with penalty
per day
There are different types of constraints. First, all projects need to be completed, i.e., progress
should be 100. For project 1 we define the following cell:
progress\_1=muRUD\_1*(XFS\_1-XRU\_1)+muFS\_1*(XRD\_1-XFS\_1)+muRUD\_1*(X\_1-XRD\_1)
and similarly for project 2 and 3. Then in the constraint editor we requires this cell to be greater
or equal to 100.
Next we need to ensure the chronological order of the projects, again for project 1 we need cells
defined by:
=XRU_1-XFS_1
=XFS_1-XRD_1
=XRD_1-X_1
In the constraints window these cells should then be less or equal to 0. We repeat for project 2 and
3.
To ensure that project 1 is the first one to start, we will require the following cell to be less or
equal to 0:
=XRU_1-MIN(XRU_2,XRU_3)
Finally, we need to ensure that not all three projects run on full speed simultaneously, this means
that we either need to ramp down project 1 or 2 before we give “full speed” for project 3, i.e.,
MIN(XRD_1,XRD_2)<=XFS_3}
or we ned to ramp down project 1 or 3 before project 2 is given “full speed”, i.e.,
MIN(XRD_1,XRD_3)<=XFS_2}
One of these inequalities need to be true. To formulate this constraint into a cell to be non-positive,
we use:
=IF(OR(MIN(XRD_1,XRD_2)<=XFS_3,MIN(XRD_1,XRD_3)<=XFS_2),0,1)
126
Table 10.6: Optimal scheduling
Project RampUp FullSpeed RampDown EndOfProject
1 0 0 125 125
2 0 0 100 140
3 0 100 162.5 162.5
127
Chapter 11
• Linear programming
• Dynamic programming
• Non-linear programming
• Stochastic programming
The learning objective of this chapter is to understand these type of programming and obtain skills
to apply the programming techniques. This means that after the end of this chapter the student
shall be able to formulate problems and solve problems by appropriate tools such as MS Excel.
The deeper understanding of the mathematical ideas behind these methods are left for courses in
operations research.
This chapter was originally developed for the course TPK4161 - Supply chain analytics, and we
have simplified to meet the requirements for TPK5115.
2. The objective function is given as a linear function of the decision variables. The linear
assumption implies that only the first powers of the decision variables are included, and no
cross terms are allowed.
128
11.2.1 Motivating example
As an introduction example consider the shipment of goods from Trondheim to Oslo. A truck has
a capacity to take 100 units. There are four products types that could be transported. The profit
per unit depends on the the product type and are given by c1 = 10, c2 = 5, c3 = 8 and c4 = 4 for
product type 1, 2, . . . , 4. Let xi denote the number of units to take of product type i. The objective
function is then given by:
Z(x1 , x2 , x3 , x4 ) = 10x1 + 5x2 + 8x3 + 4x4 = c1 x1 + c2 x2 + c3 x3 + c4 x4 (11.1)
There are some constraints in addition to the truck capacity. The customer in Oslo has storage
restrictions implying that 2x1 + 3x2 cannot exceed 50. Further there is a maximum of 50 units
available in Trondheim of product type 3. The constrains can thus be formulated by:
x1 + x2 + x3 + x4 ≤ 100 = b1
2x1 + 3x2 ≤ 50 = b2 (11.2)
x3 ≤ 50 = b3
In this problem it is quite obvious that the number of units cannot be negative, hence x1 ≥ 0, x2 ≥
0, . . . , x4 ≥ 0.
The LinearProgrammingIntroExample.xlsx file provides a solution to this transportation
problem. The optimal solution is given by x1 = 25, x2 = 0, x3 = 50 and x4 = 25 giving a
final profit equal to Z = 750.
Subject to: Ax = b
x≥0
b≥0
129
A linear programming problem can be solved by the SIMPLEX method which we will come back to.
The motivating example was not exactly on standard form because there were some inequalities in
the constraints, i.e., · · · ≤ bi rather than · · · = bi . In other cases there are other types of restrictions
that do not completely match the standard format of the problem. The following steps may be
used to transform a linear programming problem into standard form:
• If the problem is to minimize Z convert the problem to a problem that maximizes −Z.
• If there is a less or equal inequality in one or more rows of the constraints, i.e., ai1 x1 + ai2 x2 +
. . . ain xn ≤ bi , convert it into an equality constraint by adding a nonnegative slack variable
si yielding the resulting constraint: ai1 x1 + ai2 x2 + . . . ain xn + si = bi , where si ≥ 0.
• If there is a greater or equal inequality in one or more rows of the constraints, i.e., ai1 x1 +
ai2 x2 + . . . ain xn ≥ bi , convert it into an equality constraint by subtracting a nonnegative
surplus variable si yielding the resulting constraint: ai1 x1 + ai2 x2 + . . . ain xn − si = bi , where
si ≥ 0.
• If some of the bi ’s are negative, multiply that constraint equation by -1.
• If one of the decision variables, say xj is unrestricted in sign, replace it by xaj − xbj everywhere
in the problem formulation and add to the constraints: xaj ≥ 0 and xbj ≥ 0.
Note that si , xaj and xbj will be renamed to maintain a set of n decision variables, xj , j = 1 . . . n.
130
3. Define the decision variables, i.e., those variables we could change in order to find the best
solution. (By Changing Variable Cells:)
4. Define the constraints, i.e., specify the equations that define the constraints. Note that MS
Excel will accept inequalities in the constraints, hence we do not need to add slack variables.
(Subject to the Constraints:)
5. Define the solving method. For linear models we always choose the Simplex LP method. For
nonlinear models the Simplex LP method does not work. (Select a solving method:)
Figure 11.1 shows the specification screen for the MS Excel Solver. Note the following:
• The cell containing the objective function has been named Z. Further, to calculate Z we used
the =SUMPRODUCT() function. This function accepts two arguments, the first argument is the
x-values, and the second argument is the objective function coefficients.
• The cells containing the x-variables has been given the name xValues.
• To simplify the specification of the constraints it is convenient to move the ∑ constant term to
the left-hand side of the inequality sign, i.e., we can create cells containing ai,j xj −bi . These
cells are then used in the specification of constraints. Also here the =SUMPRODUCT() function is
useful for calculating the sum of products. The three cells containing the constraint equations
have been given the name Constraints. When the ≤ comparison is specified it applies to all
the constraint equations.
131
possible to use overtime to produce up to 200 extra units per week in week 2 and 3 but then at
an additional item cost of $5. Excess production can be stored at a cost of $3 per item (from one
month to the other). Table 11.1 shows the decision variables to use and other important quantities
for the problem.
Minimize: Z = 5x1 + 5x2 + 10x3 + 10x4 + 10x5 + 15x6 + 3x7 + 3x8 + 3x9 + 3x10
132
Figure 11.2: MS Excel solution of Example 11.1
and bound algorithm is presented below but technical details regarding the implementation is left
out and the reader is referred to a operations research text book for further reading. A mixed
integer programming problem (MIP) is generally given on the following form:
Maximize: Z = cx (11.5)
Subject to: Ax = b
x≥0
b≥0
xj is an integer forj ∈ I
133
LP-1 Z1 = 9
x2 ≤ 1 x2 ≥ 2
x1 ≤ 4 x1 ≥ 5
LP-2 has an optimal solution of Z2 = 8 and LP-3 has an optimal solution of Z2 = 8.5. Assume
that LP-3 has a feasible solution, i.e., that integer restrictions are fulfilled.
We proceed with LP-3 and split into two new problems, say LP-4 and LP-5 and so on. Here
the integer variable x1 is used to put new restrictions on the LP-problems. Occasionally the new
restrictions put on one the integer variables will give integer value for all the integer variables in
the optimal value in the solution, and we are closer to the solution. If this is the case we have
found a solution for that particular LP-problem. It is no guarantee that it is optimal, but it is at
least a lower bound for the optimal solution. This also means all LP-problems with a Z-value lower
than the best feasible solution up to now can be ignored for further investigation. In the example
in Figure 11.3 LP-4 has a solution less favourable compared to the best feasible solution in LP-2,
and since LP-5 gave no feasible solution the best solution found is the solution found in LP-2.
There is no guarantee that an optimal solution is found, but we are able to squeeze the solution
into limits which might be good enough. When one LP-problem is branched into two new problems
it is not evident which one to proceed with. Some principles are discussed in the literature, and
again we refer to an operation research textbook for further reading.
In MS Excel we may specify that some decision variables are integer or binary variables. This
is done when specifying the constraint comparison symbol where we can use ”int” or ”bin”.
• Sn = input state, i.e., the system state when the system enters state n
• S̃n = output state, i.e., the system state when the system leaves state n
• rn (Sn , dn ) = return function, typically a cost or profit function that depends on the system
state and the decision made
Figure 11.4 shows the main elements of an N -stage multistage system. The stage numbering is
typically in an opposite direction compared to the flow of information.
Usually we solve the problem computationally by starting at stage 1 and proceed from right
to left to stage N . Such an approach is denoted backward analysis or backward recursion. In a
134
rN rN −1 rn r1
dN dN −1 dn d1
Figure 11.4: An N -stage multistage system
backward analysis we always have that Sn−1 = S̃n . The idea is that for the rightmost stage it is
rather easy to find the optimal solution for a given input, i.e., for a given S1 we easily find the
best decision d1 . When proceeding left to stage 2 we then know all the stage 1 optimal decisions
and corresponding return functions r1 for each of the output values S̃2 . Then adding r1 and r2 for
possible d2 decisions makes it possible to find optimal decisions at stage 2 for a given input S2 . We
then proceed till stage N is reached.
The change in system state in a given stage is governed by some transition function, i.e.,:
The return function rn (Sn , dn ) specifies what is the cost or profit related to what happens in stage
n. The accumulated total return calculated over n stages, i.e., from right to left or numerically
from 1 to n is denoted fn (Sn , dn ). Further fn∗ (Sn ) is the accumulated optimal returns up to stage
n.
Mathematically it is a challenge to be explicit on the formulation of fn (Sn , dn ) and fn∗ (Sn ).
Typically these functions are formulated recursively.
135
where opt is the optimization with respect to decision dn at stage n. Now, assume that S_N is the
initial state of the system, the optimal strategy is found by calling f(N, S_N).
Note that f() needs to be implemented in a programming language that allows recursive calls
because this function calls itself. Further note that the all the functions need to be carefully
designed to take care off all boundary conditions, for example to stop when state 0 is invoked!
Further, since the function calls are recursive, the number of calls may explode in magnitude.
The store purchases pairs of shoes from a manufacturer at a purchasing cost of 4 $ per pair
(1976 prices). The supplier only sells in lots of 10, 20, 30, 40 or 50 pair. A discount is achieved for
large orders as given in Table 11.3.
For each order there is a fixed cost of 10 $ to account for transportation, insurance, packaging,
and so on. The shop has limited capacity and can not carry an inventory level higher than 40 at
the end of each month. There is a holding cost of 0.3 $ based on the end-of-month inventory level.
The season starts with an empty inventory, and should also end with an empty inventory. Table
11.4 gives summarized information.
The cost structure favours large orders, but the downside is a holding cost at the end of each
month. Also large order could save the fixed cost if we could escape from ordering one month. It
is assumed that the pair of shoes arrives the first day of the month.
The backward analysis means that the system runs through 6 stages, where October corresponds
to n = 6, November to n = 5 and so on to March where n = 1. The decision variables, dn is the
number of pairs to order for each month. The state variable Sn is the number of pairs in the
136
Table 11.4: Additional parameters
Parameter Value Description
cP 4 Purchasing cost per unit
cF 10 Fixed cost per order
cH 0.2 Holding cost per unit at end of month
SL 40 Storage limitation at end of month
S6 0 Inventory level at the beginning of the season
S0 = S̃1 0 Inventory level at the end of the season
beginning of the month when the order from the manufacturer has arrived. From Table 11.2 we
can read the demand, Dn each month, again remembering that October corresponds to n = 6 and
so on.
The transition function must relate the state variable at stage n to that of stage n − 1:
Sn−1 = Sn + dn − Dn (11.7)
The return function comprises the cost of an order and holding cost at the end of the month:
where ϕ(dn ) is the cost of an order comprising the fixed cost and the discounted variable cost per
pair of shoes. It is assumed that ϕ(0) = 0. Now, assume that the transition function in Equation
(11.7) and the return function in Equation (11.8) are implemented by the functions g(n,S,d) and
r(n,S,d) respectively. Then we need to write a function for the accumulated return. Below is
given the main structure of the computer code we need. The programming language is VBA, i.e.,
the built in Visual Basic for Microsoft applications.
137
Table 11.5: Intermediate results for n = 1
S1 d∗1 f1∗
0 20 86
10 10 48
20 0 0
The recursive formulation causes the f(n,S) function to call it self recursively starting from
n = N = 6. At the lowest level, i.e., n = 1 there are only 3 feasible values of S1 to consider, i.e., 0,
10 and 20. It is easy to verify that the optimal decisions are given in Table 11.5
These results for n = 1 are needed several times when the function is called from stage n = 2.
Therefore it would be a good idea to save the optimal values. This applies also for all other sates. In
fact, for this example we need to calculate fn () more than 1 100 times, which reduces to around 30
calculations by clever saving of intermediate results. In computing, memoization is an optimization
technique used primarily to speed up computer programs by storing the results of expensive function
calls and returning the cached result when the same inputs occur again. Memoization here means
to save calculated values for f ∗ for (S, n) combinations.
The intermediate results for the other (Sn , n) combinations, i.e., n > 1 is not given in this
presentation.
Note that since the function f(n,S) only returns the accumulated cost, but we have not kept
track of the optimal decision at the various levels. A technical solutions to this challenge is left for
the reader as an exercise.
Table 11.6 summarizes the final results for the DP analysis.
Subject to: Ax = b
x≥0 (11.10)
b≥0
138
The notation used is that x is a decision vector which is controlled by the decision maker, c is
the profit/cost vector, A is the coefficient matrix and b is the requirement vector. Up to now we
have assumed that c, A and b are all known at the time of the decision to be made, that is we
have a deterministic optimization problem at hand. In real life, we often have to make decisions
when some of these quantities are not known. Typically the requirement vector could be uncertain,
and we need to treat this vector as a stochastic variable. The requirement vector often represents
resource constraints which is only or partly known at a later stage in the decision process.
Stochastic programming is a technique to solve such challenges. In this presentation we limit
the presentation of stochastic programming only to deal with linear programming situations.
To motivate for the formalism to be introduced, consider a fish processing facility where we have
to plan tomorrow’s production before we actually know the catch, i.e., the amount and quality of
fish available for production. Fish of high quality can be used for production of fresh fillet. Frozen
fish can be produced from both high and low quality fish. It is assumed that the selling price of
fresh fillet is higher than of frozen fish, but production of fresh fillet requires more workload. Today
we have to decided upon how many workers, x to allocate for the production. x is often referred
to as the here an now decision. The production profile, y can be determined at a later stage,
i.e., when we know the catch. The objective function is now split into two objectives function,
Z1 and Z2 . The maximization (minimization) of Z1 = cx subject to some Ax = b is denoted
the first stage optimization. In the example this corresponds to finding the number of workers to
allocate for tomorrows production. This is obvious not straight forward since we need to take into
account the catch during the night. Now, assume that we can describe tomorrows situation by
some stochastic vector U. Given that we know the value of this vector, say U = u there is a second
stage optimization problem defined by:
Since the decision to make tomorrow is influenced by the decision we make today through the
constraints B(u)x we are facing the classical two stage linear stochastic programming problem
were the decision the two days are interconnected. In the first stage we need to take into account
what will be the result of the second stage decision but this again depends on the unknown vector
U. The idea now is to add the expected value of the best solution tomorrow to the today’s objective
function:
Subject to: Ax = b
x≥0 (11.14)
b≥0
where the expectation is taken with respect to the stochastic vector U and Q(x, u) is the solution
to:
139
Subject to: B(u)x + C(u)y = d(u)
y≥0 (11.16)
d(u) ≥ 0
In the second stage problem the profit/cost vector q(u) and the requirement vector d(u) both depend
on the value, u, the stochastic vector U will take. Further the coefficient matrix in B(u)x + C(u)y
is built up by two terms: B(u) relates to the first stage decision vector x and C(u) relates to the
second stage decision vector y. Both matrices depend on the value the stochastic vector U will
take.
In some presentations the second stage problem is considered as a recourse action where the
term C(u) compensates for a possible inconsistency of the system B(u)x = d(u) and q(u)y is the
cost of this recourse action.
In the example −cx represents the cost of allocating x workers for tomorrows production. q(u)y
is the profit of the tomorrows production. B(u)x + C(u)y = d(u) gives the constraints taking the
number of workers and catch into account.
11.4.2 Discretization
To solve the two-stage stochastic problem numerically it is common to perform a discretization
where the stochastic vector U can only take a finite number of possible realizations. These re-
alizations are often denoted scenarios, say u1 , u2 . . . , uk , with corresponding probabilities pi =
Pr(U = ui ). The expectation in the second-stage problem is then given by:
∑
k
EU [Q(x, U)] = pi Q(x, uk ) (11.17)
i=1
It is now possible to write the stochastic problem on a so-called extensive form, that is a determin-
istic equivalent problem:
∑
k
Maximize: Z1,2 = cx + pi q(ui )yi (11.18)
i=1
Subject to: Ax = b
B(ui )x + C(ui )yi = d(ui ), i = 1, 2, . . . , k
x≥0
(11.19)
b≥0
yi ≥ 0
d(ui ) ≥ 0, i = 1, 2, . . . , k
For the scenarios i = 1, 2, . . . , k we have to specify B(ui ), C(ui ) and d(ui ) one by one based on the
understanding of the problem at hand.
140
Table 11.7: Supply scenarios
Scenario, i pi High quality Low quality
1 0.15 16 000 12 000
2 0.5 8 000 6 000
3 0.35 6 000 3 000
For fish of high quality the most beneficial product is fresh fillet despite the higher workload
required. This means that if we knew the supply we will always allocate work-force to process
all high quality fish to fresh fillet, and the low quality to frozen fillet. However, at the decision
point for the work-force we do not know the supply. If we allocate sufficient work-force for the best
scenario, i.e., 250 man-hours (some 30 persons), and actually get less raw material, we have to pay
for man-hours we are not utilizing.
The following decision variables are introduced:
In the objective function the profit coefficient corresponding to x1 is −cMHR , the profit coefficients
corresponding to y1,i are pi ρFi pFi and the profit coefficients corresponding to y2,i are pi ρFz pFz .
In the constraint matrix for the deterministic equivalent problem we have for each scenario to
make sure that y1,i not exceeds supply of high quality fish, and that y1,i + y2,i not exceeds total
supply of fish. Further for each scenario we have to make sure that y1,i /µFi + y2,i /µFz ≤ x1 .
Solving the deterministic equivalent problem gives x1 ≈ 134. If scenario 1 occurs, 3 739 kg of
high quality fish is used for fresh fillet, and 24 260 kg (= remaining high quality fish + low quality
fish ) is used to produce frozen fish fillet. The Excel model used is shown in Figure 11.5.
141
Figure 11.5: Specification of the stochastic programming example in Excel
that to solve the two stage problem the decision maker anticipate the value of U by u = E(U),
and optimize:
Subject to: Ax = b
B(u)x + C(u)y = d(u)
x≥0
(11.21)
b≥0
y≥0
d(u) ≥ 0
The first stage solution is denoted xEV . The first stage solution is optimal if the stochastic variable
U takes the value u. However, since the problem is stochastic in nature, other values could also be
the case. Therefore, the decision maker needs to modify the second stage decision when the value
of the stochastic variable U is revealed. The expected value of the expected value solution is now
the value of the following optimization problem:
∑
k
Maximize: ZEEV = cxEV + pi q(ui )yi (11.22)
i=1
142
11.4.4 Expected value of perfect information
In decision theory, the expected value of perfect information (EVPI) is the price that one would
be willing to pay in order to gain access to perfect information. In our context this means the
difference between the expected value of the stochastic solution and expected value of the solution
if we knew u at the time of the first stage decision. For the latter case we often say that this
correspond to a wait and see situation were we imagine that we can wait and observe u before we
need to decide upon x.
The optimization problem for the wait and see situation given that we observe u = ui is given
by:
∑
k
EVPI = pi Zi∗ − Z1,2
∗
(11.27)
i=1
There are several reasons why we would like to calculate the EVPI. For example if we know EVPI
we can decide upon how much effort we should pay to reduce uncertainty. For example in Example
11.2 we could imagine that we are able to establish a better prediction model for the catch. Such a
model could take weather forecast, type of catch tools and so on into account. A better prediction
model will not reduce variability in the catch, but will reduce the epistemic uncertainty, and hence
we are able to take better decisions.
1 ∑
N
Maximize: Z1,2 = cx + q(ui )yi (11.28)
N
i=1
143
subject to the same constraints as in Section 11.4.2.
Another approach is to investigate element by element in the stochastic vector U. That is, for
each element Ui we first calculate the VSS where we use 3 values, say low, medium and high for
the discretization. If SSV is relatively small, we may be tempted to ignore uncertainty of element
Ui .
However, if SSV is large the uncertainty of element Ui should be taken into account. The next
question is then if a discretization using only 3 values is sufficient. If Ui is a continuous variable we
would expect to get a better result if we approximate Ui with a 5-point discrete variable compared
to a 3-point discrete variable. Again, we calculate the expected value of the stochastic solution
(SSV). The situation is now more challenging. We have to compare the 5-point solution with the 3-
point solution. If we assume that the 5-point discretization of Ui is “perfect”, we can then calculate
the expected cost of assuming a 3-point discretization for the first stage decision, and then face
the 5-point discretization for the second stage decision. The procedure for calculating the expected
profit for the 3-point discretization is then:
1. Formulate the problem as a two stage optimization problem on extended form where a 3-point
discretization is used
2. Find the solution for both x and yi , i = 1, 2, 3. Denote the solution for x by x3p
3. Switch to the 5-point discretization
4. For each of the 5 scenarios, solve the second stage problem assuming that first stage solution
is bounded to x3p
5. For each of the 5 scenarios, calculate the expected profit
6. Calculate an average weighted profit by using corresponding pi ’s for each scenario
The calculated weighted average is then subtracted from the expected value where a 5-point dis-
cretization is used in the ordinary manner.
If the difference is significant, we could compare the 5-point discretization with the situation
where a 7-point discretization is assumed to be “perfect”. We proceed in this manner until the
value of a higher point discretization is relatively low. Here we need to be pragmatic, so that the
total number of combinations taking all elements of U into account is manageable.
144
Table 11.9: Standardized distances from mid point
k = # of scenarios dk,1 dk,2 dk,3 dk,4 dk,5
3 1.22474
5 0.87889 1.31436
7 0.16787 0.49042 1.79758
9 0.21902 0.60872 0.67431 1.90442
11 0.26459 0.43883 0.70498 0.89051 1.98681
145
Bibliography
[1] Fischhoff, B., S. Lichtenstein, P. Slovic, S.L. Derby, and R.L. Keeney. Acceptable Risk. Cam-
bridge University Press, New York, 1981.
[4] Klakegg, O.J.: Trinnvis-prosessen, Institutt for bygg- og anleggsteknikk NTH, 1993.
[5] Chapman, C. & Ward, S., 1997. Project Risk Management; Processes, Techniques and Insights.
John Wiley & Sons, England.
[6] Hokstad, P. Life Cycle Cost Analysis in Railway Systems. SINTEF Report STF38 A98424.
ISBN 82-14-00450-0. 1988.
[7] IEC 60300. International Standard, IEC 60300-3-3, Dependability management - Part 3: Ap-
plication Guide - Section 3: Life Cycle Costing
[8] Kawauchi, Y. and Rausand, M. Life Cycle Cost (LCC) analysis in oil and chemical process
industries. NTNU report . 1999.
[9] Keeney, R. L. and H. Raiffa. Decisions with Multiple Objectives: Preference and Value Trade-
offs. New York: Wiley. 1976.
[10] Vatn, J. 1998. A discussion of the acceptable risk problem. Reliability Engineering and System
Safety, 61(1-2):11-19, 1998.
[11] Øien, K, P.R. Hokstad, and R. Rosness. 1998. Handbook for performing Expert Judgement.
SINTEF report STF38 A98419. Could be ordered from http://www.sintef.no.
[14] Kaplan, S (1997): The Words of Risk Analysis. Risk Analysis 17(4), 407-417.
[15] Klinke, A. and Renn, O. (2001): Precautionary principle and discursive strategies: classifying
and managing risk. Journal of Risk Research 4 (2), 159-173.
[16] Needleman, K. Methods of valuing life. In Technological Risk. University of Waterloo, 1982
146
[17] ESA. 1991 European Space Agency - ESA, Expert Judgment, Requirements and Methods,
PSS-01-405, Issue 1 Draft 1-Nov. 1991, Noordwijk, the Netherlands.
[18] Cooke, R., Experts in Uncertainty - Opinion and Subjective Probability in Science, Oxford
University Press, 1991.
[19] Gossens, L.H.J., Cooke, R.M., van Steen, J.F.J., Expert Opinions in Safety Studies, Vol. 1:
Final Report, TUDelft, 1989.
[20] van Steen, J.F.J., Oortman Gerlings, P.D., Expert Opinions in Safety Studies, Vol. 2: Litera-
ture Survey Report, TUDelft, 1989.
[21] Hokstad, P. and Øien, K., Expert Judgment - For Assessing Input Data to a Small/medium
Size Reliability Study, Presentation at the Growth Point Centre Group Meeting, Trondheim
1994-08-25, STF75 S94018.
[22] Hokstad, P., Reinertsen, R. and Øien, K., Recommendations on the use of expert judgment
in safety and reliability engineering studies. Recommendations on the use of expert judgment
in safety and reliability engineering studies. Two offshore case studies Reliabiliity Engineering
and System Safety 61 (1998) 65-76. case studies. To appear in Reliability Engineering and
System Safety, 1997.
[23] Klakegg, O. J., The stepwise process (In Norwegian), Institutt for bygg- og anleggsteknikk,
NTH, 1993.
[24] Hauge, S. and Tor Onshus, T (2004). Reliability Data for Safety Instrumented Systems – PDS
Data Handbook, 2010 Edition. SINTEF report SINTEF A13502. Trondheim 1994.
[25] Austeng, K. og Hugsted, R., Stepwise calculation (In Norwegian), Institutt for bygg- og an-
leggsteknikk, NTH, 1995.
[26] The PS 2000 project team (Hokstad, P., Klakegg, O.J., Rosness, R. and Øien, K.)
[27] Kahneman, D., Slovic, P. and Tversky, A., Judgment under Uncertainty. Heuristics and Biases.
Cambridge, Cambridge University Press, 1982.
[28] Svenson, O., On expert judgments in safety analysis in the process industries. Reliability
Engineering and Systems Safety, 25, 219-256, 1989.
[29] Arentz, C., Bakken, J., Kilde, H.S., Klakegg, O.J., Krogh, J., Competence as a steering
parameter - The basis for development (in Norwegian), NTH Rapport NTH 95010, PS2000,
1995.
[30] Norman Dalkey, Olaf Helmer (1963) An Experimental Application of the Delphi Method to
the use of experts. Management Science, 9(3), Apr 1963, pp 458-467
[31] Saaty, T.L. 1980 The Analytic Hierarchy Process: Planning, Priority Setting, Resource Allo-
cation, ISBN 0-07-054371-2, McGraw-Hill
[32] Archer, N. and Chasemzadeh, F. (2004) Project Portfolio Selection and Management. In The
Wiley Guide to Managing Projects Morris, P and Pinto, J.K. (Editors). Willey. ISBN: 978-0-
471-23302-2
147
[33] Høyland, K. and Wallace, SW. (2001) Generating Scenario Trees for Multistage Decision Prob-
lems. Management Science, 47(2), pp 295 - 307
[34] Bjerke, MM. and Bykvist, HR. (2018) Managing Supply Uncertainty in the Operational Pro-
duction Planning in a Whitefish Value Chain - A Stochastic Programming Approach. Master
thesis. NTNU - Norwegian University of Science and Technology.
[35] Birge, JR. and Louveaux, F. (2011). Introduction to Stochastic Programming. Springer Inter-
national Publishing, second edition.
[36] Phillips, DT., Ravindran, A. and Solberg, J. (1976) Operations research: Principles and prac-
tice. John Wiley&Sons. New York.
148
Index
α-percentiles, 29 Dependence, 91
Dependency, 5
Activity on arc, 41 Stochastic, 6
Activity on node, 41 Design matrix, 77
Ageing parameter, 104 Deterministic equivalent problem, 140
Arctic maintenance, 110 Discount rate, 70
Decision tree, 111 Disjoint events, 23
Single activity, 106 Distribution
Maximum values, 38
Bayes theorem, 25
Product, 38
Bayesian methods, 79
Sums, 37
Biasedness, 89
Double expectation, 29
Overestimation, 90
Dynamic programming, 134
Underestimation, 90
Binary variables, 132 Effective failure rate, 104
Binomial distribution, 33 EMax function, 44
Bootstrapping, 99 Engineering judgement, 86
Branch and bound algorithm, 132 Erlang distribution, 32
Bubble diagram, 119 Event, 22
Event uncertainty, 49
Calibration, 90
Expectation, 28
Method for, 96
Project expectation, 9
Central limit theorem, 38
Expert judgement, 37, 81
Checklists, 10
Data, 81
Coherence, 91
Explanaotry variable, 77
Commitment, 9
Exponential distribution, 31
Complementary event, 23
Extensive form, 140
Completeness, 86
Conditional probability, 24 Failure rate function, see Hazard rate
Confidence level, 12
Consistency, 91 Gamma distribution, 32
Cost modelling, 11
Critical path method, 43 Hazard, 10
Cumulative distribution function, 26 Hazard and Operability Study, 10
Hazard rate, 104
Decision trees, 64
Decision variable, 128 Independent events, 24
Degradation rate, 71 Informativeness, 90
Delphi method, 83 Subjective, 90
149
Input state, 134 Conditional, 24
Intersection, 23 Density function, 27
Inverse-Gauss distribution, 34 Total, 24
Inverted gamma distribution, 32 Program Evaluation and Review Technique
(PERT), 43
Life cycle cost, 69 Project portfolio, 118
Life cycle profit, 69 Project selection, 119
Likelihood function, 75 Scheudling, 124
Linear programming, 128 Visualization, 118
Solving by MS Excel, 130 Vulnerability, 120
standard form, 129
Lognormal distribution, 33 Random variable, see Stochastic variable
LP, see Linear programming Regression variable, 77
LS principle, 76 Reproducibility, 92
Residual, 78
Maximum likelihood principle, 36 Resolution, 91
Mean time to failure, see MTTF Return function, 134
Median, 28 Risk, 6
Memoization, 138 Acceptance criterion, 6
Method of moments, 75 Identification, 9
Mixed integer programming, 132 Management, 8
MLE, 74 Picture, 6
Mode, 28 Risk and opportunity reg, 12
Monte Carlo simulation, 46
MTTF, 104 Scenario, 10
Scenario analyse, 83
Net present value, 70 Schedule, 6, 41
Normal distribution, 30 Model, 11
SIMPLEX method, 130
Objective function, 128 Slack variable, 130
Observable quantity, 5 Stakeholder, 118
Output state, 134 Standard deviation, 29
Overconfidence, 91 Stochastic programming, 138
discretization, 140
Parameter, 6 expected value of perfect information,
Estimate, 6 143
Estimation, 74 scenario building, 143
Penalty for default, 48 value of the stochastic solution, 141
PERT Stochastic variable, 6, 26
Distribution, 35 Successive schedule planning, 43
Method, 43 Surplus variable, 130
Poisson distribution, 33
Posterior distribution, 79 Target, 9
Precision, 29 Task analysis, 10
Preliminary hazard analysis, 10 Threat, 10
Preventive maintenance, 103 Total probability, 24
Priror distribution, 79 Triangular distribution, 35
pRisk, 5 Triple estimate, 37
Probability, 6, 22 Turnaround, 103
150
Changing the frequency, 110 Utility
Exclude activities, 114 Maximising expected, 58
Two stage stochastic programming Utility function, 54
formulation, 139
Variance, 29
VarMax function, 44
Unbiasedness, 89
Venn diagram, 22
Uncertainty
Aleatory, 5, 7 Weibull distribution, 31
Cost and schedule model, 11 Used in maintenance models, 104
Decision under, 52 Weight of data, 99
Epistemic, 5 Weighting of experts, 96
Event, 49 Based on control questions, 97
Underconfidence, 91 Based on knowledge profile, 97
Undesired event, 10 Based on mutual evaluation, 97
Union, 23 Equal weighting, 96
151