Modeling and Quantification of Physical Systems Uncertainties in a Probabilistic Framework 2017_Américo_Cunha
Modeling and Quantification of Physical Systems Uncertainties in a Probabilistic Framework 2017_Américo_Cunha
net/publication/316519530
CITATION READS
1 90
1 author:
Americo Cunha Jr
Rio de Janeiro State University
102 PUBLICATIONS 135 CITATIONS
SEE PROFILE
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Americo Cunha Jr on 22 March 2018.
Americo Cunha Jr
1 An Introductory Overview on UQ
1
2 Americo Cunha Jr
Decision making: Some risk decisions, which negative result can cause catas-
trophic failure or huge financial costs, need to be well analysed before a final
opinion by the responsible party. The possible variabilities that generate un-
certain scenarios need to be taken into account in the analysis. The evaluation
of these uncertain scenarios has the task of assisting the responsible party to
minimize the chances of a wrong decision. Briefly, and in this context, UQ is
essential to provide the necessary certification for a risk decision.
Model validation: Experimental data are widely used to check the accuracy
of a computational model which is used to emulate a real system. Although
this procedure is already being used by scientists and engineers for many
decades, there is still no universally accepted criteria to ensure the model
quality. However, it is known that any robust criteria of model validation
must take into account the simulation and experiment uncertainties.
system response
1
−1
−2
simulation
experiment
−3
0 1 2 3 4 5 6
system input
(a)
3
simulation
experiment
2
confidence band
system response
−1
−2
−3
0 1 2 3 4 5 6
system input
(b)
Fig. 1 (a) Comparison between simulation and experimental data, without an envolepe
of reliability for the simulation, and (b) including this envelope.
tal value (defined by the point and its error bar) is contained within the
reliability envelope around the simulation.
OPTIMAL VIOLATION
Fig. 2 Illustration of a possible notion of validated model.
4 Americo Cunha Jr
This chapter is organised into six sections. Besides this introduction, there
is a presentation of some fundamental concepts of UQ in section 2; a brief re-
view on probability theory basics in section 3; an exposure of the fundamental
aspects of probabilistic modeling of uncertainties, through a simplistic exam-
ple, in section 4; the presentation of the uncertainty propagation problem in
section 5; and the final remarks in section 6.
It is noteworthy that many of the ideas that are presented in this
manuscript are very influenced by courses taught by the author’s doctoral
supervisor, Prof. Christian Soize [13], [14] and [15]. Lectures of Prof. Gi-
anluca Iaccarino, Prof. Alireza Doostan, and collaborators were also very
inspiring [16], [17] and [18].
designed
system
model input
real system + computational
real input real response model parameters model model response
(real parameters)
(data uncertainty)
(uncertain system)
Fig. 3 Schematic representation of the relationship between the designed system, the real
system and the computational model [19].
Modeling and Quantification of Physical Systems ... 5
Numerical errors: The response obtained with the computational model is, in
fact, an approximation to the model equation’s true solution. Inaccuracies,
intrinsic to the discretization process, are introduced in this step giving rise
to numerical errors [19]. Other source of errors are: (i) the finite precision
arithmetic that is used to perform the calculations, and (ii) possible bugs in
the computer code implementation of the computational model.
errors uncertainties
w w
w
6= w
mathematical physical
nature nature
Today verification and validation, also called V&V, are two concepts of fun-
damental importance for any carefully done work in UQ. Early works advo-
cating in favor of these ideas, and showing their importance, date back to
the late 1990s and early 2000s [20], [21], [22], [23]. The impact on the nu-
merical simulation community was not immediate, but has been continuously
growing over the years, conquering a prominent space in the last ten years,
especially after the publication of Oberkampf and Roy’s book [24].
Verification:
Are we solving the equation right?
Validation:
Are we solving the right equation?
solution model
verification validation
w w
w
6= w
mathematics physics
g
y0
m ÿ(t) + m g = 0, (1)
ẏ(0) = v0 ,
y(0) = y0 ,
˙ := d /dt.
where the upper dot is an abbreviation for a time derivative, i.e.,
This toy model is obtained from Newton’s 2nd law of motion and considers
the weight as the only force acting on the skydiver body.
Imagine that we have developed a computer code to integrate this IVP
using a standard 4-th order Runge-Kutta method [25]. The model response
obtained with this computer code is shown in Figure 7.
solution verification
2500
2000
heigth (m)
1500
1000
500
simulation
0
0 5 10 15 20
time
1
y(t) = − g t2 + v0 t + y0 . (2)
2
In Figure 8(a) we can see the comparison between toy model response
(solid blue curve —) and the reference solution (dashed red curve - - -). We
note that both curves are in excellent agreement, but if we look at Figure 8(b),
which shows the difference between numerical and analytical solutions, it is
evident the effectiveness of the numerical method and the robustness of its
implementation become ever clearer.
Here the verification was made taking as reference the real solution of the
model equation. In the most frequent case, the model equations solution is
not known. In such a situation, the verification task can be performed, for
instance, using the method of manufactured solutions [26], [27], [24], [28].
Modeling and Quantification of Physical Systems ... 9
solution verification
2500
2000
heigth (m)
1500
1000
500
simulation
analytical
0
0 5 10 15 20
time
(a)
0.8
absolute error
0.6
0.4
0.2
0
0 5 10 15 20
time
(b)
Fig. 8 (a) Solution verification: comparison between toy model response and reference
solution; (b) absolute error of Runge-Kutta method approximation.
Now let’s turn our attention to model validation, and compare simulation
results with experimental data, such as shown in Figure 9(a). We note that
the simulation is completely in disagreement with experimental observations.
In other words, the model does not provide an adequate representation of the
real system behaviour.
The toy model above take into account the gravitational force which at-
tracts the skydiver toward the ground, but neglects air resistance effects. This
is the major reason for the observed discrepancy, the model deficiency (model
uncertainty). If the air drag force effects are included, the improved model
below is obtained
10 Americo Cunha Jr
model validation
2500
2000
heigth (m)
1500
1000
500
toy model
experiment
0
0 5 10 15 20
time
(a)
model validation
2500
2000
heigth (m)
1500
1000
Fig. 9 (a) Model validation: comparison between experimental data and the toy model,
(b) comparison between experimental data, the toy model, and the improved model.
1 2
m ÿ(t) + m g − ρ A CD ẏ(t) = 0, (3)
2
ẏ(0) = v0 ,
y(0) = y0 ,
where ρ is the air mass density, A is the cross-sectional area of the falling
body, and CD is the (dimensionless) drag coefficient.
With this new model, a better agreement between simulation and experi-
ment is expected. In Figure 9(b) the reader can see the comparison between
experimental data and the responses of both models, where we note that the
improved model provides more plausible results.
An important message, implicit in this example, is that epistemic uncer-
tainties can be reduced by increasing the actual knowledge about the phe-
nomenon/system of interest [22], [24].
Modeling and Quantification of Physical Systems ... 11
This section presents a brief review of probability basic concepts. Such expo-
sition is elementary, being insufficient for a solid understanding of the theory.
Our objective is only to equip the reader with basic probabilistic vocabu-
lary necessary to understand UQ scientific literature. For deeper studies on
probability theory, we recommend the references [38], [39], [40] and [41].
Sample space: The set which contains all possible outcomes (events) for a
certain random experiment is called sample space, being represented by Ω. An
elementary event in Ω is denoted by ω. Sample spaces may contain a number
of events that is finite, denumerable (countable infinite) or non-denumerable
(non-countable infinity). The following three examples, respectively, illustrate
the three situations:
Example 1 (finite sample space). Rolling a given cube-shaped fare die, where
the faces are numbered from 1 through 6, we have Ω = {1, 2, 3, 4, 5, 6}.
PX (x) = P {X ≤ x} . (5)
PX has the following properties:
• 0 ≤ PX (x) ≤ 1 (it is a probability);
• PX is non-decreasing, and right-continuous;
• lim PX (x) = 0, and lim PX (x) = 1;
x→−∞ x→+∞
so that
Z x
PX (x) = dPX (ξ), (6)
ξ=−∞
and
Z
dPX (x) = 1. (7)
R
Z {X}
mX = E
(11)
= x pX (x) dx,
R
n o
2
σX2 = E (X − mX )
Z (12)
= (x − mX )2 pX (x) dx,
R
and
q
σX = σX2 , (13)
which are the mean value, variance, and standard deviation of X, respectively.
Note further that
n o
σX2 = E X2 − m2X . (14)
The ratio between standard deviation and mean value is called coefficient
of variation of X
σX
δX = , mX 6= 0. (15)
mX
These scalar values are indicators of the random variable behaviour. Specif-
ically, the mean value mX is a central tendency indicator, while variance σX2
and standard deviation σX are measures of dispersion around the mean. The
difference in these dispersion measures is that σX has the same unit as mX
while σX2 is measured in mX unit squared. Once it is dimensionless, the coef-
ficient of variation is a standardized measure of dispersion.
For our purposes, it is also convenient to define the entropy of pX
n o
S (pX ) = −E ln pX (X) , (16)
The inequality expressed in (18) implies that E {X} < +∞ (mX is also finite).
Consequently, with the aid of Eq.(14), we see that a second-order random
variable X has finite variance, i.e., σX2 < +∞.
This class of random variables is very relevant for stochastic modeling,
once, for physical considerations, typical random parameters in physical sys-
tems have finite variance.
and
Z Z
dPX Y (x, y) = 1. (21)
R2
PX Y is also known as joint cumulative distribution function.
If the partial derivative ∂ 2 PX Y /∂x ∂y exists, for any x and y, then it is called
joint probability density function of X and Y, being denoted by
∂ 2 PX Y
pX Y (x, y) = (x, y). (22)
∂x ∂y
Hence, we can write dPX Y (x, y) = pX Y (x, y) dy dx, so that
Modeling and Quantification of Physical Systems ... 17
Z x Z y
PX Y (x, y) = pX Y (ξ, η) dη dξ. (23)
ξ=−∞ η=−∞
Consider the pair of random events {X ≤ x} and {Y ≤ y}, where the prob-
ability of occurrence of the second one is non-zero , i.e., P {Y ≤ y} > 0.
The conditional probability of event {X ≤ x}, given the occurrence of event
{Y ≤ y}, is defined as
n o P {X ≤ x} ∩ {Y ≤ y}
P {X ≤ x} {Y ≤ y} = . (25)
P {Y ≤ y}
This section discusses the use of the parametric probabilistic approach to de-
scribe uncertaintis in physical systems. Our goal is to provide the reader with
some key ideas behind this approach and call attention to the fundamental
issues that must be taken into account. The exhibition is based on [13] and
[15] and use a simplistic example to discuss the theory.
Consider the mechanical system which consists of a spring fixed on the left
side of a wall and being pulled by a constant force on the right side (Fig-
ure 10). The spring stiffness is k, the force is represented by f , and the spring
displacement is denoted by u. A mechanical-mathematical model to describe
this system behaviour is given by
k u = f, (31)
from where we get the system response
u = k −1 f. (32)
k
f
We are interested in studying the case where the above mechanical system is
subject to uncertainties on the stiffness parameter k. To describe the random
behaviour of the mechanical system, we employ the parametric probabilistic
approach.
Let us use the probability space (Ω, Σ, P), where the stiffness k is modeled
as the random variable K : Ω → R. Therefore, due to result of the relationship
imposed by Eq.(32), the displacement u is also uncertain, being modeled as a
random variable U : Ω → R, which respects the equilibrium condition given
by the following stochastic equation
K U = f. (33)
It is reasonable to assume that the deterministic model is minimally rep-
resentative, and corresponds to the mean of K, i.e., mK = k. Additionally, for
physical reasons, K must have a finiten variance.
o Thus, K is assumed to be a
2
second-order random variable, i.e., E K < +∞.
whose support is the entire real line, i.e., Supp pK = (−∞, +∞).
The attentive reader may question, at this point, that from the physical
point of view, make no sense use a Gaussian distribution to model a stiffness
parameter, since K is always positive. This is true and makes the arbitrary
choice of a Gaussian distribution inappropriate. However, this is not the only
reason against this choice.
For physical considerations, it is necessary that the model
n oresponse U be a
second-order (finite variance) random variable, i.e., . E U2 < +∞. Is this
possible when we arbitrate the probability distribution as Gaussian? No way!
Just do a simple calculation
n o n o
E U2 = E K−2 f 2
Z
= k −2 f 2 pK (k) dk
R
(35)
( )
Z +∞ 2
1 (k − mK )
= k −2 f 2 p exp − dk
k=−∞ 2π σK
2 2 σK2
= +∞.
−2δ −2
( )
1 δK K δK−2 −1 k/mK
pK (k) = 1(0,+∞) (k) k/mK exp − 2 , (38)
mK Γ δ −2 δK
K
√
where 0 ≤ δK = σK /mK < 1/ 2 is a dispersion parameter, and Γ denotes
the gamma function
Z +∞
Γ (α) = tα−1 e−t dt. (39)
t=0
Bm = (m − 1) h, m h , m ∈ Z, (44)
being h the bin width. Then we count the number of samples in each of the
bins Bm , denoting this number by νm . After that, we normalize the counter
(dividing by N h) to obtain the normalized relative frequency νm / (N h). Fi-
nally, for each bin Bm , we plot a vertical bar with height νm / (N h) [44],
[43].
In analytical terms (see [44] and [43]) we can write this as PDF estimator
as
+∞
1 X
p N (x) =
b νm 1Bm (x), (45)
N h m=−∞
2
sample value
−1
−2
−3
−4
20 40 60 80 100
x
1
exact CDF
empirical CDF
0.8
probability
0.6
0.4
0.2
0
−3 −2 −1 0 1 2 3
x
(a)
0.4
exact PDF
0.35 histogram
0.3
probability density
0.25
0.2
0.15
0.1
0.05
0
−3 −2 −1 0 1 2 3
x
(b)
Fig. 12 (a) Estimators for the probability distribution of X: the empirical CDF, and (b)
a histogram.
When very little or no experimental data is available, to the best of the au-
thor’s knowledge, the most conservative approach uses the Maximum Entropy
Principle (MEP) [15], [45], [46], [48], with parametric statistical estimation,
to construct the random parameter distribution. If no experimental data
is available, this approach takes into account only theoretical information
which can be inferred from the model physics and its mathematical structure
to specify the desired distribution.
The MEP can be stated as follows: Among all the (infinite) probability dis-
tributions, consistent with the known information about a random parameter,
the most unbiased is the one which corresponds to the maximum of entropy
PDF.
Using it to specify the distribution of a random variable X presupposes
finding the unique PDF which maximizes the entropy (objective function)
Modeling and Quantification of Physical Systems ... 25
Z
S (pX ) = − pX (x) ln pX (x) dx, (47)
R
where gk are known real functions, with g0 (x) = 1, and µk are known real
values, being µ0 = 1. The restriction associated with k = 0 corresponds to
the normalization condition of pX , while the other constraints, typically, but
not exclusively, represent statistical moments of X.
To solve this problem, the method of Lagrange multipliers is employed, and
introduces other (N +1) unknown real parameters λk (Lagrange multipliers).
We can show that if this optimization problem has a solution, it actually
corresponds to a maximum and is unique, being written as
XN
pX (x) = 1K (x) exp (−λ0 ) exp − λk gk (x), (49)
k=1
where K = Supp pX here denotes the support of pX , and 1K (x) is the indicator
function of K.
The Lagrange multipliers, which depend on µk and K, are identified with
the aid of the restriction defined in Eq.(48) using techniques of parametric
statistics.
In this section we exemplify the use of the MEP to consistently specify the
probability distribution of a random variable X.
Suppose that Supp pX = [a, b] is the only information we know about X. In
this case, a consistent (unbiased) probability distribution for X is obtained
solving the following optimization problem:
Maximize
Z
S (pX ) = − pX (x) ln pX (x) dx
ZRb
=− pX (x) ln pX (x) dx,
x=a
subjected to the constraint
26 Americo Cunha Jr
Z
1= pX (x) dx
ZRb
= pX (x) dx.
x=a
To solve this optimization problem, first we define the Lagrangian
!
Z b
Z b
L (pX , λ0 ) = − pX (x) ln pX (x) dx − (λ0 − 1) pX (x) dx − 1 ,
x=a x=a
(50)
where λ0 − 1 is the associated Lagrange multiplier. It is worth mentioning
that λ0 depends on the known information about X, i.e. λ0 = λ0 (a, b).
Then we impose the necessary conditions for an extreme
∂L ∂L
(pX , λ0 ) = 0, and (pX , λ0 ) = 0, (51)
∂pX ∂λ0
whence we conclude that
Z
pX (x) = 1[a,b] (x) e−λ0 , and pX (x) dx = 1. (52)
R
The first equation in Eq.(52) provides the PDF of X in terms of the La-
grange multiplier λ0 , while the second equation corresponds to the known
information about this random variable (the normalization condition).
In order to represent pX in terms of the known information (a and b),
we need to find the dependence of λ0 with respect to these parameters. To
this end, let’s go to replace the expression of pX into the second equation of
Eq.(52), so that
Z
1
1[a,b] (x) e−λ0 dx = 1 =⇒ e−λ0 (b − a) = 1, =⇒ e−λ0 = , (53)
R b−a
Once one or more of the model parameters are described as random objects,
the system response itself becomes random. To understand how the variabili-
ties are transformed by the model, and influence in the response distribution,
is a key issue in UQ, known as uncertainty propagation problem. This problem
can only be attacked after the construction of a consistent stochastic model.
Very succinctly, we understand the uncertainty propagation problem as to
determine the probability distribution of model response once we know the
distribution of model input/parameters. A schematic representation of this
problem is can be seen in Figure 13.
Uncertainty Propagation
computational
model
input output
PDF PDF
[a, b] {X} o
En = mX ∈ [a, b]
E X2 = m2X + σX
2
pX (x) = 1[a,b] (x) exp −λ0 − x λ1 − x2 λ2
λ0 = λ0 (a, b, mX , σX )
λ1 = λ1 (a, b, mX , σX )
λ2 = λ2 (a, b, mX , σX )
[0, 1] E ln (X) = p, |p| < +∞
Γ (a + b) a−1
(1 − x)b−1
E ln (1 − X) = q, |q| < +∞ pX (x) = 1[0,1] (x) x
Γ (a) Γ (b)
2 2
a = mX /δX 1/mX − δX −1
b = a 1/mX − 1
(beta with shape parameters a and b)
1 x
(0, +∞) E {X} = mX > 0 pX (x) = 1(0,+∞) (x)exp −
mX mX
(exponential with mean mX )
stochastic deterministic
model model
When the use of MC method is unfeasible, the state of art strategy is based on
the so-called stochastic Galerkin method. This spectral approach was orig-
inally proposed by [65] and [66], and became very popular in the last 15
years, especially after work of [67]. It uses a Polynomial Chaos Expansion
(PCE) to represent the stochastic model response combined with a Galerkin
projection to transform the original stochastic equations into a system of
deterministic equations. The resulting unknowns are the coefficients of the
linear combination underlying to the PCE.
Once PCE theory is quite rich and extensive, we do not have space in
this manuscript to cover it in enough detail, but to the reader interested in
digging deeper on this subject is encouraged to see the references [2], [3], [68],
[69], [70] and [8].
6 Concluding Remarks
Acknowledgements
tion for the Improvement of Higher Education Personne) and FAPERJ (Re-
search Support Foundation of the State of Rio de Janeiro).
References
48. J. N. Kapur, Maximum Entropy Models in Science and Engineering. New Age, 2009.
49. F.E. Udwadia, Response of uncertain dynamic systems. I. Applied Mathematics and
Computation, 22 (1987), 115–150.
50. F.E. Udwadia, Response of uncertain dynamic systems. II. Applied Mathematics and
Computation, 22 (1987), 151–187.
51. F.E. Udwadia, Some results on maximum entropy distributions for parameters known
to lie in finite intervals. SIAM Review, 31 (1989), 103–109.
52. K. Sobezyk and J. Trçbicki, Maximum entropy principle in stochastic dynamics. Prob-
abilistic Engineering Mechanics, 5 (1990), 102–110.
53. K. Sobezyk and J. Trȩbicki, Maximum entropy principle and nonlinear stochastic
oscillators. Physica A: Statistical Mechanics and its Applications, 193 (1993), 448–
468.
54. J. Trȩbicki and K. Sobezyk, Maximum entropy principle and non-stationary distribu-
tions of stochastic systems. Probabilistic Engineering Mechanics, 11 (1996), 169–178.
55. A. Cunha Jr, R. Nasser, R. Sampaio, H. Lopes, and K. Breitman, Uncertainty quantifi-
cation through Monte Carlo method in a cloud computing setting. Computer Physics
Communications, 185 (2014), 1355-1363.
56. N. Metropolis and S. Ulam, The Monte Carlo Method. Journal of the American
Statistical Association, 44 (1949), 335–341.
57. C. Lemieux, Monte Carlo and Quasi-Monte Carlo Sampling. Springer, 2009.
58. D. P. Kroese, T. Taimre, and Z. I. Botev, Handbook of Monte Carlo Methods. Wiley,
2011.
59. J. S. Liu, Monte Carlo Strategies in Scientific Computing. Springer, 2001.
60. G. Fishman, Monte Carlo: Concepts, Algorithms, and Applications. Springer, cor-
rected edition, 2003.
61. R. Y. Rubinstein and D. P. Kroese, Simulation and the Monte Carlo Method. Wiley,
2nd edition, 2007.
62. S. Asmussen and P. W. Glynn, Stochastic Simulation: Algorithms and Analysis.
Springer, 2007.
63. R. W. Shonkwiler and F. Mendivil, Explorations in Monte Carlo Methods. Springer,
2009.
64. C. P. Robert and G. Casella, Monte Carlo Statistical Methods. Springer, 2010.
65. R. Ghanem and P. D. Spanos, Polynomial chaos in stochastic finite elements. Journal
of Applied Mechanics, 57 (1990) 197–202.
66. R. Ghanem and P. D. Spanos, Stochastic Finite Elements: A Spectral Approach.
Dover Publications, 2nd edition, 2003.
67. D. Xiu and G. E. Karniadakis, The Wiener-Askey Polynomial Chaos for stochastic
differential equations. SIAM Journal on Scientific Computing, 24 (2002) 619–644.
68. P. Vos, Time-dependent polynomial chaos. Master Thesis, Delft University of Tech-
nology, Delft, 2006.
69. P. Constantine, A Primer on Stochastic Galerkin Methods. Lecture Notes, 2007.
70. A. O’Hagan, Polynomial Chaos: A tutorial and critique from a statistician’s perspec-
tive. (submitted to publication), 2013.