Probability
1.1 Definition
prop If E ∩ F = ∅ then P(E ∪ F ) = P(E ) + P(F )
Bayes’ Theorem
1
Chapter 2
Discrete Distributions
2.1 Random variables
def A random variable is any rule that associates a number with each outcome in a sample
space.
Random variables are customarily denoted by uppercase letters. A particular value of the ran-
dom variable is denoted by a lowercase letter.
def A Bernoulli random variable is a rv whose only possible values are 0 and 1.
def A parameter is a quantity that can be assigned any one of a number of possible values,
with each different value determining a different probability distribution.
The collection of all probability distributions for different values of the parameter is called a
family of probability distributions.
def The cumulative distribution function (cdf) of a discrete random variable X is the probability
that X will take a value less than or equal to x.
X
F (x) = P(X ≤ x) = p(y )
y :y ≤x
2
2.3 Expected values
def The expected value or mean value of a rv X is the average value of X on performing repeated
trials of an experiment.
The expectation of a discrete random variable is the weighted average of all possible outcomes,
where the weights are the probabilities of realizing each given value.
X
E (x) = µX = x · p(x)
x∈D
prop E [X + Y ] = E [X ] + E [Y ]
prop V (cX ) = c 2 V (X )
def The moment generating function for X is denoted by mX (t) and defined as mX (t) = E [e tX ]
3
prop if the moments of a specified order exist, then all the lower order moments automatically
exist.
prop The mgf of the sum of a number of independent random variables is equal to the product
of their respective mgfs.
MX1 +X2 +...+Xn (t) = MX1 (t) · MX2 (t) · . . . · MXn (t)
def A Bernoulli trial is a random experiment or a trial whose outcome can be classified as either
a success or a failure.
def The binomial random variable X denotes the number of successes that occur in n indepen-
dent Bernoulli trials. It takes the parameters n and p, where p is the probability of success,
which remains same for every trial.
If X is a binomial random variable with parameters (n, p), then we write it as X ∼ Bin(n, p).
The pmf of a binomial random distribution having parameters (n, p) is given by:
n p x q n−x , x = 0, 1, . . . , n
b(x; n, p) = x
0, otherwise
where q = 1 − p.
prop E [X ] = np
Proof.
prop V (X ) = np(1 − p)
4
2.6 Geometric random variable
def A geometric random variable X is one which has a geometric distribution with parameter
p, 0 < p < 1.
1
prop E [X ] =
p
q
prop V (X ) =
p2
pe t
prop mX (t) =
1 − qe t
def The Poisson random variable X is one which has a Poisson distribution with parameter k.
e −k k x
p(x; k) = ; for x = 0, 1, 2, . . . and k > 0.
x!
∞ ∞
k
X kx X e −k · k x
e = ; and so =1
x=0
x! x=0
x!
prop E [X ] = k
prop V [X ] = k
t −1)
prop mX (t) = e k(e ∀t ∈ R
prop For any binomial experiment in which n is large (¿50) and p is small, b(x; n, p) ≈ p(x; k),
where k = np.
5
2.7.1 Poisson process
A Poisson process is a counting process with rate λ, such that Xt is the Poisson random variable
(with parameter λt), or in other words the number of events that occur during the interval
[0, t).
def The hypergeometric random variable X is the number of objects with the trait of interest
in the random sample. It takes the parameters (N, n, r ).
r
prop E [X ] = n
N
r N − r N − n
prop V (X ) = n
N N N −1
6
Problem 1
An urn contains 4 white and 4 black balls. We randomly choose 4 balls. If 2 of them are white
and 2 are black, we stop. If not, we replace the balls in the urn and again randomly select 4
balls. This continues until exactly 2 of the 4 chosen are white. What is the probability that we
shall make exactly three selections?
Problem 2
An urn contains N white and M black balls. Balls are randomly selected, one at a time, until
a black one is obtained. If we assume that each ball selected is replaced before the next one
is drawn, what is the probability that (i) exactly n draws are needed (ii) at least k draws are
needed?
1
Problem 3
The probability that a machine produces a defective item is 0.02. Each item is checked as it is
produced. Assuming that these are independent trials, what is the probability that at least 100
items must be checked to find one that is defective?
2
Problem 1
It is known that disks produced by a certain company will be defective with probability 0.01
independently of each other. The company sells the disks in packages of 10 and offers a money-
back guarantee that at most 1 of the 10 disks is defective. (i) What proportion of packages is
returned? (ii) If someone buys three packages, what is the probability that exactly one of them
will be returned?
Problem 2 If the independent random variables X and Y are binomially distributed, respec-
tively with n = 3, p = 13 , and n = 5, p = 13 , write down the probability that X + Y ≥ 1.
Solution:
Let X ∼ Binomial(3, 13 ) and Y ∼ Binomial(5, 31 ). We want:
1
8 32 256
P(X + Y = 0) = × = .
27 243 6561
Finally:
256 6305
P(X + Y ≥ 1) = 1 − = .
6561 6561
2
Problem 1
If electricity power failures occur according to a Poisson distribution with an average of 3 failures
every twenty weeks, calculate the probability that there will not be more than one failure during
a particular week.
Problem 2
If the number of claims handled daily by an insurance company follows Poisson distribution
with mean 5, what is the probability that there will be 4 claims each in exactly 3 of the next 5
days ? Assume that the number of claims on different days is independent.
1
Problem 3
Suppose that the annual number of tropical cyclones that are formed off the U.S. Pacific coast
has a Poisson distribution with mean 15. What is the probability that a given year has at most
5 cyclones?