SF 2940 Forms
SF 2940 Forms
pages)
These pages (+ Appendix 2 of Gut)
are permitted as assistance at the exam.
6 september 2011
• Bivariate probability
• Transforms
• Stochastic processes
• Gaussian processes
• Poisson process
• Convergence
1 Probability
1.1 Two inequalities
• A ⊆ B ⇒ P (A) ≤ P (B)
1
1.2 Change of variable in a probability density
Let X = (X1 , X2 , . . . , Xm )T have the probability density fX (x1 , x2 , . . . , xm ).
Define a new random vector Y = (Y1 , Y2 , . . . , Ym )T by
Yi = gi (X1 , . . . , Xm ) , i = 1, 2, . . . , m,
where gi are continuously differentiable and (g1 , g2 , . . . , gm ) is invertible (in
a domain) with
Xi = hi (Y1 , . . . , Ym ) , i = 1, 2, . . . , m,
where hi are continuously differentiable. Then the density of Y is (in the
domain of invertibility)
fY (y1 , . . . , ym ) = fX (h1 (y1 , y2 , . . . , ym ) , . . . , hm (y1 , y2 , . . . , ym )) | J |,
where J is the Jacobian determinant
∂x ∂x1 ∂x1
∂y1
1
∂y2
... ∂ym
∂x2 ∂x2 ∂x2
...
∂y ∂y2 ∂ym
J = . 1 .
. .. .
. . . . . ..
∂xm ∂xm
. . . ∂x m
∂y1 ∂y2 ∂ym
2
where
• fX,Y (x, y) ≥ 0,
R +∞ R +∞
• −∞ −∞ fX,Y (x, y)dxdy = 1
Marginal distribution:
R +∞
• fX (x) = −∞ fX,Y (x, y)dy,
R +∞
• fY (y) = −∞ fX,Y (x, y)dx.
Distribution function
Z x
FX (x) = P (X ≤ x) = fX (u)du.
−∞
• X | Y = y,
fX,Y (x, y)
fX|Y =y (x) := ,
fY (y)
if fY (y) > 0.
• Y |X=x
fX,Y (x, y)
fY |X=x (y) := ,
fX (x)
if fX (x) > 0.
Bayes’ formula
fX|Y =y (x) is a a posteriori density for X and fX (x) is a priori density for
X.
3
2.1.2 Independence
X and Y are independent iff
fX,Y (x, y) = fX (x) · fY (y) for every (x, y).
and
Φ(−x) = 1 − Φ(x).
4
2.2 Mean and variance
The expectations or means E(X), E(Y ) are defined (if they exist) by
Z +∞
E(X) = xfX (x)dx,
−∞
Z +∞
E(Y ) = yfY (y)dy,
−∞
respectively. We have
5
2.5 Covariance
We have
n n n−1 n
a2i Var(Xi ) + 2
X X X X
Var( ai Xi ) = ai aj Cov(Xi , Xj ),
i=1 i=1 i=1 j=i+1
n
X Xm Xn Xm
Cov( ai Xi , bj Xj ) = ai bj Cov(Xi , Xj ).
i=1 j=1 i=1 j=1
Cov(X, Y )
ρ := ρX,Y := q .
Var(X) · Var(Y )
6
4 Conditional Expectation w.r.t to a Sigma-
Field
a and b are real numbers, E [| Y |] < ∞, E [| Z |] < ∞, E [| X |] < ∞ and
H, G F are sigma fields, G ⊂ F, H ⊂ F.
1. Linearity:
E [aX + bY | G] = aE [X | G] + bE [Y | G]
2. Double expectation :
E [E [Y | G]] = E [Y ]
5. Tower Property : If H ⊂ G,
E [E [Y | G] | H] = E [Y | H]
6. Positivity: If Y ≥ 0,
E [Y | G] ≥ 0.
5 Covariance matrix
5.1 Definition
Covariance matrix
h i
CX := E (X − µX ) (X − µX )T
7
• Covariance matrix is nonnegative definite, i.e., for all x 6= 0 we have
xT CX x ≥ 0
Hence
det CX ≥ 0.
Remark 6.1 These are measurable maps X(ω), ω ∈ Ω, from a basic pro-
bability space (Ω, F, P ) (= outcomes, a sigma field of of subsets of Ω and
probability measure P on F), to X .
X and Y are two discrete state spaces, whose generic elements are called
values or instantiations and denoted by xi and yj , respectively.
X = {x1 , · · · , xL }, Y = {y1 , · · · , yJ }.
8
6.1 Joint Probability Distributions
A two dimensional joint (simultaneous) probability distribution is a proba-
bility defined on X × Y
p(xi , yj ) := P (X = xi , Y = yj ). (6.1)
Hence 0 ≤ p(xi , yj ) and Li=1
P PL
j=1 p(xi , yj ) = 1.
Marginal distribution for X:
J
X
p(xi ) = p(xi , yj ). (6.2)
j=1
9
Next X
PX (A) := p(xi ) (6.6)
xi ∈A
6.4 Independence
X and Y are independent random variables if and only if
for all pairs (xi , yj ) in X ×Y. In other words all events {X = xi } and {Y = yj }
are to be independent. We say that X1 , X2 , . . . , Xn are independent random
variables if and only if the joint distribution
pX1 ,X2 ,...,Xn (xi1 , xi2 . . . , xin ) = P (X1 = xi1 , X2 = xi2 , . . . , Xn = xin ) (6.11)
equals
pX1 ,X2 ,...,Xn (xi1 , xi2 . . . , xin ) = pX1 (xi1 ) · pX2 (xi2 ) · · · pXn (xin ) (6.12)
for every xi1 , xi2 . . . , xin ∈ X n . We are here assuming for simplicity that
X1 , X2 , . . . , Xn take values in the same alphabet.
10
6.5 A Chain Rule
Let Z be a (discrete) random variable that assumes values in Z = {zk }K
k=1 .
If p(zk ) > 0,
p(xi , yj , zk )
p(xi , yj | zk ) = .
p(zk )
Then we obtain as an identity
p(xi , yj , zk ) p(yj , zk )
p(xi , yj | zk ) = ·
p(yj , zk ) p(zk )
and again by definition of conditional probability the right hand side is equal
to
p(xi | yj , zk ) · p(yj | zk ).
In other words,
7 Miscellaneous
7.1 A Marginalization Formula
Let Y be discrete and X be continuous, and let their joint distribution be
Z x
P (Y = k, X ≤ x) = P (Y = k | X = u) fX (u)du.
−∞
Then Z ∞ ∂
P (Y = k) = P (Y = k, X ≤ u) du
−∞ ∂u
Z ∞
= P (Y = k | X = x) fX (x)dx.
−∞
11
7.2 Factorial Moments
X is an integer-valued discrete R.V.,
def
µ[r] = E [X(X − 1) · · · (X − r + 1)] =
X
= (x(x − 1) · · · (x − r + 1)) fX (x).
x:integer
8 Transforms
8.1 Probability Generating Function
8.1.1 Definition
Let X have values k = 0, 1, 2, . . . ,.
∞
X
tk fX (k)
X
gX (t) = E t =
k=0
= E [X]
12
•
dr
µ[r] = E [X(X − 1) · · · (X − r + 1)] = gX (1)
dtr
!2
d2 d d
Var[X] = 2 gX (1) + gX (1) − gX (1)
dt dt dt
•
gZ (t) = E tZ =
E tX+Y = E tX · E tY = gX (t) · gY (t).
• Y ∈ Bin(n, p)
gY (t) = (1 − p + pt)n
• Z ∈ Po (λ)
gZ (t) = eλ·(t−1)
13
8.2 Moment Generating Functions
8.2.1 Definition
Moment generating function, for some h > 0,
def
h i
ψX (t) = E etX , |t| < h.
( P
etxi f (x ) X discrete
ψX (t) = R ∞xi tx X i
−∞ e fX (x)dx X continuous
• Y ∈ Exp (a)
1
ψY (s) = , at < 1.
1 − at
14
8.2.4 Characteristic function
Characteristic function h i
ϕX (t) = E eitX .
exists for all t.
15
9.1 Four product rule
(X1 , X2 , X3 , X4 )T ∈ N (0, C).Then
E [X1 X2 X3 X4 ] = E [X1 X2 ]·E [X3 X4 ]+E [X1 X3 ]·E [X2 X4 ]+E [X1 X4 ]·E [X2 X3 ]
Cov(X, Y )
ρ= q .
Var(X) · Var(Y )
If ! !
X Z1
=µ+B ,
Y Z2
then !
X
∈ N µ, C
Y
with !
σ12 ρσ1 σ2
C= .
ρσ1 σ2 σ22
16
10 Stochastic Processes
A stochastic process X = {X(t) | t ∈ T }. The mean function µX (t) of the
process is
def
µX (t) = E(X(t))
and the autocorrelation function is
def
RX (t, s) = E (X(t) · X(s)) .
and we have
CovX (t, s)(t, s) = RX (t, s) − µX (t)µX (s).
A stochastic process X = {X(t) | t ∈ T =] − ∞, ∞[} is called (weakly)
stationary if
where a = t0 < t1 < . . . < tn−1R < tn = b and maxi |ti − ti−1 | → 0 as n → ∞.
The mean square integral ab X(t)dt of {X(t)|t ∈ T } exists over [a, b] ⊆ T
if and only if the double integral
Z b Z b
E [X(t)X(u)] dtdu
a a
17
and "Z #
b Z b Z b
V ar X(t)dt = CovX (t, u)dtdu. (10.6)
a a a
1) W (0) = 0.
2) The sample paths t 7→ W (t) are almost surely continuous.
3) {W (t) | t ≥ 0} has stationary and independent increments.
4) W (t) − W (s) ∈ N (0, t − s) for t > s.
The meanR square integral with respect to the Wiener process or the Wiener
integral ab f (t)dW (t) is the mean square limit
n Z b
X 2
f (ti−1 ) (W (ti ) − W (ti−1 )) → f (t)dW (t), (10.7)
i=1 a
18
• "Z #
b
E f (t)dW (t) = 0. (10.8)
a
• "Z #
b Z b
Var f (t)dW (t) = f 2 (t)dt (10.9)
a a
• !
Z b Z b
2
f (t)dW (t) ∈ N 0, f (t)dt . (10.10)
a a
Rb Rb
• If a f 2 (t)dt < ∞ and a g 2 (t)dt < ∞,
"Z #
b Z b Z b
E f (t)dW (t) g(t)dW (t) = f (t)g(t)dt. (10.11)
a a a
• Z t
Y (t) = h(s)dW (s).
0
Z min(t,s)
E [Y (t) · Y (s)] = h2 (u)du. (10.12)
0
11 Poisson process
N (t) = number of occurrences of some event in in (0, t].
Definition 11.1 {N (t) | t ≥ 0} is a Poisson process with parameter λ > 0,
if
1) N (0) = 0.
2) The increments N (tk ) − N (tk−1 ) are independent stochastic variables
1 ≤ k ≤ n, 0 ≤ t0 ≤ t1 ≤ t2 ≤ . . . ≤ tn−1 ≤ tn and all n.
3) N (t) − N (s) ∈ Po(λ(t − s)), 0 ≤ s < t.
19
12 Convergence
12.1 Definitions
We say that
P
Xn → X, as n → ∞
if for all > 0
P (| Xn − X |> ) →0, as n → ∞
We say that
q
Xn → X
if
E |Xn − X|2 →0, as n → ∞
We say that
d
Xn → X, as n → ∞
if
FXn (x) → FX (x), as n → ∞
for all x, where FX (x) is continuous.
20
12.3 Law of Large Numbers
X1 , X2 , . . . are independent, identically distributed (i.i.d.) random variables
with finite expectation µ. We set
Sn = X1 + X2 + . . . + Xn , n ≥ 1.
Then
Sn P
→ µ, as n → ∞.
n
Lemma 12.1 Let {Ak }k≥1 be arbitrary events. If ∞ n=1 P (An ) < ∞, then it
P
holds that P (E) = P (An i.o) = 0, ie., with probability one finitely many of
An occur.
Lemma 12.2 Let {Ak }k≥1 be independent events. If ∞ n=1 P (An ) = ∞, then
P
it holds that P (E) = P (An i.o) = 1, ie., with probability one infinitely many
of An occur.
21
13 Series Expansions and Integrals
13.1 Exponential Function
•
∞
x
X xk
e = − ∞ < x < ∞.
k=0 k!
• n
cn
cn → c ⇒ 1 + → ec .
n
22
13.5 A formula (with a probabilistic proof )
k−1
Z ∞ 1 k k−1 −λx (λt)j
e−λt
X
λ x e dx =
t Γ(k) j=0 j!
23