Modeling Extreme Events From Computer Simulations
Modeling Extreme Events From Computer Simulations
Part II
Emmanuel Vazquez
Outline of Part II
1 Structural reliability
Limit-state functions
Reliability indexes
Limit-state function defined on the factor space
Affine limit-state functions
First-order reliability method
Reliability of systems
References
x gR,S (R, S)
◮ The success of a design (or serviceability) is seen in the verification of the inequality
gR,S (R, S) = R − S ≥ 0
◮ The quantity gR,S (R, S) = R − S is called the (safety) margin
◮ gR,S is called the limit-state function.
11111111
00000000 ◮ The yield point of a material is defined as the stress
00000000
11111111
00000000
11111111
00000000
11111111
at which a material begins to deform plastically
(prior to the yield point the material will deform
elastically and will return to its original shape when
the applied stress is removed)
◮ Stress: σ = F /A (Pa)
2
cross section A (m ) ◮ Yield limit: σy (Pa)
◮ Design rule: F /A ≤ σy
◮ Set
0
1
0
1
x = (F , A, σy ) ∈ R3
11
00 0
1
1111
0000
00
11 0
1 F ·L (m)
00
11 0
1 ∆L = A·E r (x) = Aσy
0
1
00 0000
11 1111
0
1
0
1
0
1 s(x) = F
◮ Define gR,S : (R, S) 7→ R − S. The design must
load F (N) verify
gR,S (r (x), s(x)) = Aσy − F ≥ 0
◮ The operating scenario is the availability of a resistance R greater than the stress S,
such that
gR,S (R, S) = R − S ≥ 0
◮ The failure scenario is
gR,S (R, S) = R − S < 0
and the failure region is defined as the region of R2+
Probability of failure
◮ In practice, the load applied on a system is unknown, and the design parameters are
subjected to dispersions
→ the parameter vector x is uncertain, and can be modeled by a random vector
X ∼ PX
◮ Then, R = r (X ) and S = s(X ) are also random variables
◮ Let fR,S be the joint pdf of (R, S) (wrt to the Lebesgue measure)
◮ The probability of failure of the system corresponds to
Z
α = P{R − S < 0} = fR,S (r , s)dr ds
r −s<0
◮ Reliability is defined as 1 − α
Probability of failure
◮ Note that the point of view of structural reliability is not different from that
presented in Part I
◮ In practice, fR and fS cannot be determined directly by the user → only the
distribution PX of the vector of uncertain factors can
◮ Thus, it is generally easier to express a limit-state function gX in the factor space:
α = P{gX (X ) < 0} , X ∼ PX
or
α = PX {x ∈ X; gX (x) < 0} = PX {gX < 0}
(The probability of failure is the volume of excursion of g above zero.)
◮ Assume X ∼ N(0, Id ) ∈ Rd
◮ Assume moreover that gX is affine: ∀x ∈ Rd
fX = cst
x⋆
gX = 0
fX = cst
x⋆
gX = 0
◮ A failure corresponds to the event {U > β} = {(X , x ⋆ ) > β 2 }, which has probability
◮ The sensitivity of the probability of failure with respect to changes in the factors is
an important information for the design of a system (makes it possible to understand
which factor are most important to control)
◮ We have
a0
∇gX (x) = a = − η
β
1 ⋆
with η = β
x
◮ Thus, if the i th component η[i] of the unit vector η is large, the margin will vary
rapidly as we move along the i th direction
➟ η[i] accounts for the importance of the ith factor
Extensions
fX = cst
x⋆
gX = 0
fX = cst
x⋆
gX = 0
subject to gX (x) = 0
is a constrained optimization problem, for which many algorithms has been proposed
in the literature:
◮ Penalty methods
◮ Augmented Lagrangian
◮ Projected gradient
◮ BFGS
◮ SQP
◮ ...
(they all are local optimization algorithms)
◮ In general, finding a good approximation of x ⋆ can be done with a moderate number
of evaluations of gX
◮ However: gX generally need to be convex and differentiable
➟ If there exist (even small) numerical instabilities when computing gX , as may be
the case in models based on the numerical solution of some partial differential
equations, then the optimization of gX using standard techniques can fail direly
◮ For some applications, finding x ⋆ can be done with only a few evaluations of gX →
interesting when gX is expensive to evaluate
◮ However, having found x ⋆ does not tell if the approximation α ≈ Φ(−β) is good or
not
◮ The probability of failure can be overestimated or underestimated
◮ Consider, for instance the following domain of failure:
Γ = {x ∈ Rd ; kxk > β0 }
2 2
◮ Let V = X[1] + · · · X[d] , so that V ∼ χ2 (d). The failure event is {V > β02 }.
◮ Hence, we have α = 1 − FV (β02 ) and α
b = 1 − Φ(β0 )
◮ b ≈ 2.9 · 10−7 but α ≈ 0.2!
Example: suppose d = 20 and β0 = 5, we obtain α
◮ The FORM approximation should be used only when prior knowledge about the
shape of ∂Γ is available
fX = cst
x⋆
β → Second-Order Reliability Method (SORM)
gX = 0
◮ Such an approximation requires the approximation of x ⋆ and the estimation of the curvature
of the limit-state at x ⋆
◮ A quadratic approximation is more expensive to obtain, but there is, in general, no
guarantee that the approximation of the probability of failure will be better than the
first-order approximation
◮ In fact, we can have
bFORM > α
α>α bSORM
◮ Another possibility proposed in the literature: compute several design points, that is,
several local minimizers of kxk subject to the constraint gX (x) = 0
β1
Γ = {x ∈ Rd ; kxk > β0 }
◮ Then, Q
b = P{X ∈ b
α Γ} = 1 − i (Φ(β0 ) − Φ(−β0 )) = 1 − (1 − 2Φ(−β0 ))d ≈ 2dΦ(−β0 )
b
=⇒ for d >> 1, α >> α
4 4
3 3
2 2
1 1
0 0
T
−1 −1
−2 −2
−3 −3
−4 −4
−4 −2 0 2 4 −4 −2 0 2 4
◮ When the components of X are independent but non Gaussian, the idea is to search
for one-to-one transformations Ti : Rd → Rd such that for each i = 1, . . . , d,
U[i] = Ti (X[i] ) is a standard Gaussian random vector
◮ Assume that the cdf Fi of X[i] is continuous and strictly increasing, then
◮ This case will be examined separately: in fact it raises two difficult issues:
◮ how to specify the distribution of a random vector in the case of non-independent
components?
◮ what is the influence of the input distribution on the probability of failure?
Reliability of systems
◮ Until now, we have implicitly considered the case of the failure of a unique
component
◮ In a real system, a failure can happen due to the failure of just one of its (possibly
many) components
◮ The designer can also choose to have redundancy on critical components; in this
case the failure of the system happens when all redundant components fail
◮ To deal with these issues, the domain of structural reliability generally introduces the
notions of parallel and series systems
◮ In structural reliability, a system is called a parallel system if the failure of all events
is necessary for the failure of the whole system
◮ A parallel system is a principle a redundancy
◮ Using the notations above, the domain of failure for a parallel system corresponds to
\
Γ= Γi
i
◮ Again, the failure event for the whole system can be characterized by a single
limit-state function
(1) (I )
gX : x 7→ gX (x) ∨ · · · ∨ gX (x)
◮ Conclusion: parallel systems can also be dealt with using the framework we have
exposed previously
Some references
◮ A very difficult issue: to choose a probability distribution for modeling the uncertain
factors of a system
◮ The process of choosing a distribution for the factors is called elicitation1 in the
literature of decision analysis
◮ Elicitation is particularly difficult when doing risk analysis about new and untried
technologies, for which little data are available
◮ Very often, risk analysis relies on expert judgment
◮ Elicitation of subjective probability distributions is often subject to a number of
serious biases, such as overconfidence in the ability to quantify uncertainty
1
from Latin elicere: draw forth, bring out
E. Vazquez Extreme events modeling II Summer School CEA-EDF-INRIA, 2011 32 / 43
Elicitation of subjective probability distributions Definition
◮ X ∼ N(µ, σ 2 )
0.5
2
0.45 ◮ Y ∼ logN(m, s 2 ), with s 2 = log( σ + 1)
µ2
and m = log µ − s 2 /2.
0.4
0.35
0.15
◮
0.1
0.05
Assume µ = 3 and σ 2 = 1. Then,
0 P(X > 7) = 3.1 · 10−5 and
P(Y > 7) = 2.8 · 10−3
0 1 2 3 4 5 6 7 8
◮ For risk analysis, “knowing” only the mean and the standard deviation of a random variable
is a very poor information
◮ In fact, in many applications, the occurrence of a failure is likely to be related to an extreme
event in the factor space
→ it is probably more important to characterize the tail behavior of the factors than the
central behavior (EVT can help)
Copulas
◮ Consider a random vector X = (X1 , . . . , Xd ) ∈ Rd . The dependence between the
component random variables X1 , . . . , Xd is completely described by the joint cdf
F (x1 , . . . , xd ) = P{X1 ≤ x1 , . . . , Xd ≤ xd }
◮ For simplicity, assume that the components Xi , i = 1, . . . , d, have continuous,
strictly increasing, marginal cdfs Fi
◮ The concept of copula: separate F into a part that describes the dependence
structure and parts which describe the marginal behavior only
◮ Transform X component-wise to obtain standard-uniform marginal
distributions U([0, 1])
T: dom F → [0, 1]d
(x1 , . . . , xd ) 7 → (F1 (x1 ), . . . , Fd (xd ))
◮ The joint cdf of U = T (X ) is called the copula of the random vector X
◮ It follows that for x ∈ dom F
F (x1 , . . . , xd ) = C (F1 (x1 ), . . . , Fd (xd ))
and for u ∈ [0, 1]d
C (u1 , . . . , ud ) = F (F1−1 (u1 ), . . . , Fd−1 (ud ))
E. Vazquez Extreme events modeling II Summer School CEA-EDF-INRIA, 2011 35 / 43
Copulas
◮ Some copulas/dependence structures:
◮ The independent copula: Cind (u) = u1 u2 · · · ud
◮ The Gaussian copula
Cg,R (u1 , . . . , ud ) =
Z Φ−1 (u1 ) Z Φ−1 (ud ) 1
···(2π)−d/2 (det R)−1/2 exp − u T R −1 u du1 · · · dud
−∞ −∞ 2
h β i
◮ The bivariate Gumbel copula CGu,β (u, v ) = exp − (− log u)1/β + (− log v )1/β
Nataf transformation
T2 : Rd → [0, 1]d
v = (v1 , . . . , vd ) 7 → (Φ(v1 ), . . . , Φ(vd ))
T3 : [0, 1]d → Rd
w = (w1 , . . . , wd ) 7 → (F1−1 (w1 ), . . . , Fd−1 (wd ))
Dependence measures
Dependence measures
Tail dependence
◮ In the context of risk analysis, it might be relevant to study the dependence between
extreme values
◮ Let X = (X1 , X2 ) be a vector of continuous random variables with marginals F1
and F2 . The coefficient of upper tail dependence of (X1 , X2 ) is
Some references
◮ Embrechts P., McNeil A., Strautmann D. (1999), Correlation and dependence in risk
management: properties and pitfalls
◮ Embrechts P., Lindskog F., McNeil A. (2001), Modelling Dependence with Copulas
and Applications to Risk Management
◮ Lebrun R. and Dutfoy A. (2009), A generalization of the Nataf transformation to
distributions with elliptical copula