Đồ_án_CSXS (1)
Đồ_án_CSXS (1)
VINH UNIVERSITY
Group 13
Nghe An - 2024
MEMBER OF GROUP 13
1 Preliminaries 4
1.1 Definition . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 4
1.2 Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 6
2 Convergence concepts 7
2.1 Convergence in mean . . . . . . . . . . . . . . . . . . . . . . . . 7
2.1.1 Definition and examples . . . . . . . . . . . . . . . . . . 7
2.1.2 Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . 9
2.2 Convergence in probability . . . . . . . . . . . . . . . . . . . . . 10
2.2.1 Definition and examples . . . . . . . . . . . . . . . . . . 10
2.2.2 Theorem . . . . . . . . . . . . . . . . . . . . . . . . . . . 11
2.2.3 The Cauchy criterion and the continuous mapping the-
orem . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 13
2.3 Convergence in distribution . . . . . . . . . . . . . . . . . . . . 17
1
Conclusion 25
References 26
2
Preface
Limit theorems is a vital field within probability theory and continue to be a
contemporary focus of research in the current stage. We recognize the impor-
tance of the convergence concepts theme in the exploration of limit theorems.
Convergence in mean, convergence in probability, and convergence in distri-
bution carry significant meaning in the study of the Law of Large Numbers
and the Central Limit Theorem. Therefore, the purpose of this article is to
provide readers with the basic knowledge needed to investigate limit theorems
by discussing important kinds of convergence such as convergence in mean,
convergence in probability, and convergence in distribution. The content of
this paper is divided into 3 chapters.
The first chapter introduces some definitions and theorems as preparation
knowledge for chapter 2.
Chapter 2 presents the definitions, examples and theorems of convergence in
mean, convergence in probability and convergence in distribution.
In chapter 3, we present the pairwise relationships among three types con-
vergence, include convergence in mean implies convergence in probability,
convergence in probability implies convergence in distribution, and examples
show that the converse does not hold in general.
3
Chapter 1
Preliminaries
1.1 Definition
Definition 1.1.1. The triple (Ω, F, P) is called a probability space. Inside, Ω
is a non-empty set, called the sample space, F is an σ - field of events, and P
is a function define in F , satisfy Kolmogorov axiomatic system:
Definition 1.1.2. Let Ω = R and A = {(a, b) : −∞ < a < b < +∞}, the
collection of open intervals. Then σ(A) is called the Borel σ -field, denoted by
B(R). Elements A ∈ B(R) are called Borel sets. The pair (R, B(R)) is called
Borel space.
is at most countable.
a.s.
Denote: Xn −−→ X as n → ∞.
5
1.2 Theorem
Theorem 1.2.1. (Jensen’s inequality) Let X be an integrable random vari-
able. Let g : R → R be a convex function. Then g(E[x]) ≤ E[g(x)].
1q
Theorem 1.2.2. (Lyapunov’s inequality) For all 0 < q < p then E[X]q ≤
p1
E[X]p .
a.s.
(i) Xn −−→ X .
a.s.
Theorem 1.2.6. (Continuous mapping theorem). Let Xn −−→ X and f : R →
R be a continuous function. Then
a.s.
f (Xn ) −−→ f (X).
6
Chapter 2
Convergence concepts
2.1 Convergence in mean
2.1.1 Definition and examples
Definition 2.1.1. Let p > 0. A sequence of random variables {Xn , n ≥ 1} is
said to converge in mean of order p to a random variable X as n → ∞ if
Then we write
Lp
Xn −→ X as n → ∞.
Lp
Xn −→ X.
Proof. We have
1 1 n−1
E|Xn = 1| = (1 − 1) 1 − 2 + (n − 1) 2 = 2 → 0 as n → ∞.
n n n
1 L
Therefore, Xn −→ 1.
On the other hand,
1
1
n − 1 2
E|Xn = 1|2 = (1 − 1)2 1 − 2 + (n − 1)2 2 = → 1 as n → ∞.
n n n
7
2 L
Hence, Xn −→ 1.
|
Example 2.1.3. Suppose that for every n ≥ 1, Xn is a discrete random
variable having a following values table
Xn −n2 0 n2
1 1 1
P 1−
2n5 n5 2n5
3 L
Then Xn −→ 0 as n → ∞.
Proof. We have
E|Xn |3 = n → 0 as n → ∞.
3 L
Then Xn −→ 0 as n → ∞.
Proof. We have,
1 1
r r
E|Xn | = 0 · 1 − α + nr · α
n n
→ 0 for r <α
r−α
=n =1 for r = α, as n → ∞
→ +∞ for r > α
8
2.1.2 Theorem
Lp Lp
Theorem 2.1.7. Assume that Xn −→ X and Yn −→ Y . Then the following
statements hold:
Lp
(i) Xn ± Yn −→ X ± Y .
Lp
(ii) aXn −→ aX , where a is constant.
Lp
Proof. (i) Since Xn −→ X . this implies that lim E(|Xn − X|p ) = 0.
n→∞
Lp
Since Yn −→ Y . this implies that lim E(|Yn − Y |p ) = 0.
n→∞
Consider the sum Xn + Y n, applying the triangle inequality, we have
Then
Then
lim E(|Xn + Yn − X − Y |p ) = 0.
n→∞
Lp
In other words, Xn + Yn −→ X + Y .
Lp
Similarly, we will prove Xn − Yn −→ X − Y .
Lp
(ii) Consider a sequence aXn , we want to show that aXn −→ aX .
Indeed, we have
|aXn − aX|p = |a||Xn − X|p .
9
2.2 Convergence in probability
In this section we will introduce convergence in probability. This notion will
play a preliminary for the weak law of large numbers which we will study
later.
In brief, a sequence of random variables {Xn , n ≥ 1} is said to converge to a
random variable X in probability if for all ε > 0, probability of event (|Xn −
X| > ε) is small when n large enough.
We then write
P
Xn −
→ X as n → ∞ or lim Xn = X in probability.
n→∞
Xn −n2 0 n2
1 1 1
P 1−
2n5 n5 2n5
P
Then Xn −
→ 0 as n → ∞.
Then
1
P(|Xn − 0| > ε) ≤ P(|Xn | =
̸ 0) = → 0.
2n5
P
Therefore, Xn −
→ 0 as n → ∞.
10
2.2.2 Theorem
Theorem 2.2.3. (The uniqueness of the limit). If a sequence of random
variables {Xn , n ≥ 1} converges to both X and Y in probability, then X ≡ Y
a.s., i.e.,
P (X = Y ) = 1.
P P
Since Xn −
→ X and Xn −
→ Y,
ε ε
lim P(|Xn − X| > ) = 0 and lim P(|Xn − Y | > ) = 0. (2.3)
n→∞ 2 n→∞ 2
1
P(|X − Y | > ) = 0. (2.5)
k
This implies
∞
[ ∞
1 X 1
P |X − Y | > ≤ P |X − Y | > = 0. (2.6)
k k
k=1 k=1
a.s.
Proof. Assume that Xn −−→ X . Then for any ε > 0, we have that
0 = lim P sup |Xk − X| > ε
n→∞ k≥n
≥ lim P(|Xn − X| > ε) ≥ 0.
n→∞
P
Theorem 2.2.5. If Xn −
→ X , then there exists a subsequence {Xnk } of {Xn }
c a.s.
such that Xnk →
− X and therefore Xnk −−→ X .
P
Proof. Assume that Xn −
→ X . We choose a subsequence {Xnk } such that
Let ε > 0. With k large enough, we have 2−k < ε and so P(|Xnk − X| > ε) ≤
P(|Xnk − X| > 2−k ) ≤ 2−k . This implies
P
nk P(|Xnk − X| > ε) < ∞,
c
i.e., Xnk →
− X as nk → ∞. By proposition about complete convergence implies
a.s.
almost sure convergence, we conclude that Xnk −−→ X .
The following corollary gives a necessary and sufficient condition for conver-
gence in probability. We will use this criterion many times when we prove a
sequence of random variables converging in probability.
P
Corollary 2.2.6. Xn −
→ X if and only if for every subsequence of {Xn } con-
tain another subsequence which converges to X a.s.
thereby proving lim P(|Xm − Xn | > ε) = 0 for all ε > 0. Conversely, assume
m,n→∞
that {Xn , n ≥ 1} is a Cauchy sequence in probability. For k ≥ 1, let nk be
such that
and
This implies
∞
1
X
P |Xnk+1 − Xnk | > k < ∞.
2
k=1
or
1
P lim inf |Xnk+1 − Xnk | ≤ k = 1. (2.9)
2
1
For ω ∈ lim inf |Xnk+1 − Xnk | ≤ k , it is easy to verify that {Xnk (ω), k ≥ 1}
2
is a Cauchy sequence in R so it converges in R. Thus (2.9) implies that
13
{Xnk , k ≥ 1} converges a.s., to a random variable X. Let ε > 0. We need to
prove that
a.s.
Since Xnk −−→ X ,
P(|Xn − X| > 2ε) ≤ P(|Xn − Xnk | > ε) + P(|Xnk − X| > ε). (2.12)
P
Theorem 2.2.8. (Continuous mapping theorem) Assume that Xn −
→ X and
f : R → R is a mapping function. Then
P
f (Xn ) −
→ f (X).
P
Proof. Assume that {f (Xnk )} is a subsequence of {f (Xn )}. Since {Xnk } −
→ X,
we have from Theorem 2.2.5. that there exists a subsequence {Xnkl } of {Xnk }
a.s.
such that Xnkl −−→ X . By applying continuous mapping theorem for a.s.
a.s.
convergence, we have f (Xnkl ) −−→ f (X). Corollary 2.2.6. then ensures that
P
f (Xn ) −
→ f (X).
P P
Theorem 2.2.9. Assume that Xn −
→ X as n → ∞ and Yn −
→ Y . Then the
following statements hold.
P
(i) Xn ± Yn −
→X ±Y.
P
(ii) Xn Yn −
→ XY .
P
Proof. (i) Since Xn −
→ X for all ε > 0, we have
P
Since Yn −
→ Y for all ε > 0, we have
|Xn + Yn − (X + Y )| > ε
⇔ |(Xn − X) + (Yn − Y )| > ε.
P P
Since n → ∞, Xn −
→ and Yn −
→Y,
P
Xn + Yn −
→X +Y.
15
P
(ii) Since Xn −
→ X for all ε > 0, we have
P
Since Yn −
→ Y for all ε > 0, we have
P P
Since n → ∞, Xn −
→ and Yn −
→Y,
lim P(|Xn (Yn − Y )| > 2ε ) = 0 and lim P(|Y (Xn − X)| > 2ε ) = 0.
n→∞ n→∞
P
Xn Yn −
→ XY .
16
Theorem 2.2.10. If {Xn , n ≥ 1} is a monotone sequence (increasing or de-
a.s. P
creasing), then Xn −−→ X if and only if Xn −
→ X.
Then we write:
d
Xn → X as n → ∞.
Condition lim Fn (x) = F (x) for all x ∈ C(F ) instead of for all x ∈ R is
n→∞
quite strange for the first encounter.
X ≡ 0, Xn ≡ n1 , n ≥ 1.
Then
d
Xn → X.
It is clear that
0 if x < 0
F (x) =
1 if x ≥ 0,
and
0 if x < 1/n
Fn (x) =
1 if x ≥ 1/n.
We see that x = 0 is the unique discontinuous point of F and Fn (x) only
converge to F (x) if x ̸= 0.
17
Example 2.3.3. Let Z ∼ N (0, 1), and let {Xn , n ≥ 1} and {Yn , n ≥ 1} be two
sequences of random variables such that Xn = Z, Yn = −Z for all n ≥ 1. Since
−Z ∼ N (0, 1),
d d
Xn → Z and Yn → Z.
d d d
Xn + Yn →
− Z + Z, Xn − Yn →
− Z − Z, Xn YN →
− Z × Z.
|
|
Example 2.3.4. This example concerns Poisson approximation of the bino-
mial distribution. For the sake of illustration we assume, for simplicity, that
λ
p = pn = λ/n. Suppose that Xn ∈ Bin n, . Then
n
d
Xn → P (λ) as n → ∞.
Example 2.3.5. Toss a symmetric coin, set X = 1 for heads and X = 0 for
tails, and let X2n = X and X2n−1 = 1 − X , n ≥ 1. Since X, X1 , X2 ..., all have
the same distribution, it follows, in particular, that
d
Xn → X as n → ∞.
18
Chapter 3
Relations between
convergence concepts
Let X and X1 , X2 , ... be random variables. The following implications hold
as n → ∞:
Lp P d
Xn −→ X =⇒ Xn −
→ X =⇒ Xn → X.
then
P
Xn −
→ X.
E(|Xn − X|p )
P(|Xn − X|p ≥ εp ) ≤ .
εp
19
Lp
Since Xn −→ X , that is, lim E(|Xn − X|p ) = 0.
n→∞
So,
E(|Xn − X|p )
lim = 0.
n→∞ εp
Hence
1
0 ≤ P(|Xn − 0| > ε) = P(|Xn | > ε) ≤ P(|Xn | > 0) = 1 − P(|Xn | = 0) = → 0 as n → ∞.
n
Then
lim P(|Xn − 0| > ε) = 0.
n→∞
P
Hence, Xn −
→ 0.
On the other hand, since
1
E|Xn − 0|2 = E|Xn |2 = |a|2 × → 0 as n → ∞.
n
2 L
Therefore, Xn −→ 0.
The following examples will show that the converse does not hold.
20
Proof. For any r ≥ 1, we can write
E(Xn )r = (Xi − 0)r P(Xi = 0) + (Xi − 0)r P(Xi = n2 )
1 1
= 0r 1 − + (n2 )r
n n
= n2r−1 .
Example 3.1.4. Let the probability space be ([0, 1], B([0, 1]), λ) where λ is
Lebesgue measure and set
Xn = 2n 1(0, 1 )
n
Then
1
1
P[|Xn | > ε] = P 0, = → 0.
n n
but
1
E(|Xn |p ) = 2np → ∞.
n
Example 3.1.5. (The continuous mapping theorem does not hold for conver-
gence in mean). Let {Xn , n ≥ 1} be a sequence of random variables satisfying
1 1
P(Xn = 2) = 1 − 3
and P(Xn = n2 ) = 3 , n ≥ 1.
n n
Then
1
E|Xn − 2| = |n2 − 2| × → 0.
n3
However,
1
E|Xn2 − 4| = |n4 − 4| × → ∞.
n3
Example 3.1.6. (Convergence in mean does not imply almost sure conver-
gence.) Consider the function {Xn } defined on ([0, 1], B([0, 1]), λ) where λ is
Lebesgue measure.
X1 = 1[0, 1 ] , X2 = 1[ 1 ,1] ,
2 2
X3 = 1[0, 1 ] , X4 = 1[ 1 , 2 ] ,
3 3 3
X5 = 1[ 2 ,1] , X6 = 1[0, 1 ] ,
3 4
21
and so on. Note that for any p > 0,
1 1
E(|X1 |p ) = , E(|X2 |p ) = ,
2 2
1 1
p
E(|X3 | ) = , · · · , E(|X6 |p ) = .
3 4
So E(|Xn |p ) → 0 and
Lp
Xn −→ 0.
then
d
Xn → X.
Proof. Let ε > 0. Then for all n ≥ 1 and for all x ∈ R, we have
Fn (X) = P(Xn ≤ x)
22
( Because P(A) = P(AΩ) = P(A(B ∪ B c )) = P(AB ∪ AB c ) = P(AB) + P(AB c )
≤ P(AB) + P(B c ) and Xn ≤ x, −ε ≤ X − Xn ≤ ε then P((Xn ≤ x) ∩ (|Xn − X| ≤
ε)) ⊂ P(X ≤ x + ε).)
P
Since Xn → X , this implies that
Similarly, we have
Then
Z −1 1
1 1
P
2 2
−Z −1 1
1 1
P
2 2
23
d P
Then F−Z (x) → FZ (x) as n → ∞, but −Z −
→ Z as n → ∞.
|
Indeed, we see that FZ (x) ≡ F−Z (x). Put Xn = −Z , then FXn (x) ≡ FZ (x) for
1
all x ∈ R. However, since Xn − Z = −2Z and ε = , we have
2
1 1 1
lim P |Xn − Z| > = P |2Z| > = P |Z| > = 1.
n→∞ 2 2 4
P
Then Xn −
→ Z as n → ∞.
|
C(F ) = R \{c} .
= 1 − Fn (c + ε) + Fn (c − ε) − P(Xn = c + ε)
≤ 1 − Fn (c + ε) + Fn (c − ε)
→ 1 − F (c + ε) + F (c − ε) = 1 − 1 + 0 = 0.
The proof of the theorem is completed.
24
Conclusion
In conclusion, the concepts of convergence in mean, convergence in probabil-
ity, and convergence in distribution provide valuable tools for understanding
the limiting behavior of sequences of random variables. Convergence in mean
Lp
(Xn −→ X ) focuses on the convergence of expected values and moments, offer-
ing insights into the behavior of random variables in terms of their averages.
P
Convergence in probability (Xn −
→ X ) addresses the likelihood that random
variables in a sequence get arbitrarily close to their limiting value, emphasiz-
d
ing a probabilistic perspective. Convergence in distribution (Xn →
− X ) deals
with the convergence of cumulative distribution functions, capturing the lim-
iting distributional characteristics. While these concepts are related, they
do not always coalesce, highlighting the nuanced nature of stochastic con-
vergence in various scenarios. Understanding their distinctions is pivotal for
sound applications in probability theory and statistics.
25
References
[1] L. V. Thành, Cơ sở lý thuyết xác suất.
[2] Le Van Thanh, Foundations of probability theory, Vinh University Pub-
lisher.
[3] Allan Gut, Probability: A Graduate Course, Second Edition, Springer,
2013.
[4] Sidney I. Resnick, A Probability Path, 2014.
26