Unit I Random Variables
Unit I Random Variables
Prerequisites:
If an experiment is repeated under essentially homogeneous and similar conditions one
generally comes across two types of situations:
(i) The result or what is usually known as the 'outcome' is unique or certain.
(ii) The result is not unique but may be one of the several possible outcomes.
The phenomena covered by (i) are known as 'deterministic' or 'predictable' phenomena. By a
deterministic phenomenon the result can be predicted with certainty.
For example:
(a) The velocity ′𝑣′ of a particle after time ′𝑡′ is given by 𝑣 = 𝑢 + 𝑎𝑡 where ‘𝑢’ is the initial
velocity and ‘𝑎’ is the acceleration. This equation uniquely determines ‘𝑣’ if the right-hand
quantities are known.
(b) Ohm’s Law, viz., 𝐶 = 𝐸/𝑅 where 𝐶 is the flow of current, 𝐸 the potential difference between
the two ends of the conductor and 𝑅 the resistance, uniquely determines the value 𝐶 as soon
as 𝐸 and 𝑅 are given.
A deterministic model is defined as a model which stipulates that the conditions under which
an experiment is performed to determine the outcome of the experiment. For a number of
situations, the deterministic model suffices. However, there are phenomena (as covered by
(ii) above) which do not lend themselves to deterministic approach and are known as
'unpredictable' or 'probabilistic' phenomena. For example:
(a) In tossing of a coin one is not sure if a head or tail will be obtained.
(b) If a light tube has lasted for t hours, nothing can be said about its further life. It may fail to
function any moment.
In such cases chance or probability comes into picture which is taken to be a quantitative
measure of uncertainty.
Some basic definitions:
Trial and Event: Consider an experiment which, though repeated under essentially identical
conditions, does not give unique results but may result in any one of the several possible
Fourth Semester 1 Probability theory and Linear Programming (MA241TA)
outcomes. The experiment is known as a trial and the outcomes are known as events or casts.
For example:
(i) Throwing of a die is a trial and getting 1 (𝑜𝑟 2 𝑜𝑟 3, . . . 𝑜𝑟 6) is an event.
(ii) Tossing of a coin is a trial and getting head (𝐻) or tail (𝑇) is an event.
(iii) Drawing two cards from a pack of well-shuffled cards is a trial and getting a king and a
queen are events.
Exhaustive Events: The total number of possible outcomes in any trial is known as
exhaustive events or exhaustive cases. For example:
(i) In tossing of a coin there are two exhaustive cases, viz., head and tail.
(ii) In throwing of a die, there are six, exhaustive cases since anyone of the 6 faces 1, 2, . . . ,6
may come uppermost.
(iii) In drawing two cards from a pack of cards the exhaustive number of cases is 52𝐶2, since
2 cards can be drawn out of 52 cards in 52𝐶2 ways.
Favourable Events or Cases: The number of cases favourable to an event in a trial is the
number of outcomes which entail the happening of the event. For example:
(i) In drawing a card from a pack of cards the number of cases favourable to drawing of an
ace is 4, for drawing a spade 13 and for drawing a red card is 26.
(ii) In throwing of two dice, the number of cases favourable to getting the sum 5 is : (1,4)
(4,1) (2,3) (3,2), i.e., 4.
Mutually exclusive events: Events are said to be mutually exclusive or incompatible if the
happening of any one of them precludes the happening of all the others, i.e., if no two or
more of them can happen simultaneously in the same trial. For example:
(i) In throwing a die all the 6 faces numbered 1 to 6 are mutually exclusive since if any one
of these faces comes, the possibility of others, in the same trial, is ruled out.
(ii) Similarly in tossing a coin the events head and tail are mutually exclusive.
Equally likely events: Outcomes of a trial are set to be equally likely if taking into
consideration all the relevant evidences, there is no reason to expect one in preference to the
others. For example:
(i) In tossing an unbiased or uniform coin, head or tail are likely events.
(ii) In throwing an unbiased die, all the six faces are equally likely to come.
Independent events: Several events are said to be independent if the happening (or non-
happening) of an event is not affected by the supplementary knowledge concerning the
occurrence of any number of the remaining events. For example:
(i) In tossing an unbiased coin the event of getting a head in the first toss is independent of
getting a head in the second, third and subsequent tosses.
Since the number of cases favourable to the 'non-happening' of the event 𝐸 are (𝑛 − 𝑚), the
probability '𝑞' that 𝐸 will not happen is given by
𝑛−𝑚 𝑚
𝑞= = 1 − = 1 − 𝑝 gives 𝑝 + 𝑞 = 1
𝑛 𝑛
Obviously 𝑝 as well as 𝑞 are non-negative and cannot exceed unity, i.e,
0 ≤ 𝑝 ≤ 1, 0 ≤ 𝑞 ≤ 1 .
Statistical or Empirical Probability: If a trial is repeated a number of times under
essentially homogeneous and identical conditions, then the limiting value of the ratio of the
number of times the event happens to the number of trials, as the number of trials become
indefinitely large, is called the probability of happening of the event. (It is assumed that the
limit is finite and unique).
Symbolically, if in 𝑛 trials an event 𝐸 happens 𝑚 times, then the probability ′𝑝′ of the
𝑚
happening of 𝐸 is given by 𝑝 = 𝑃(𝐸) = lim .
𝑛→∞ 𝑛
Axiomatic Probability: Let 𝐴 be any event in the sample space 𝑆, then 𝑃(𝐴) is called the
probability of event 𝐴, if the following axioms are satisfied.
Axiom 1: 𝑃(𝐴) ≥ 0
Axiom 2: 𝑃(𝑆) = 1, 𝑆 being the sure event
Axiom 3: For two mutually exclusive events 𝐴 & 𝐵, 𝑃(𝐴 ∪ 𝐵) = 𝑃(𝐴) + 𝑃(𝐵)
Some important results:
1. The probability of an event always lies between 0 and i.e., 0 ≤ 𝑃(𝐴) ≤ 1.
2. If 𝐴 and 𝐴′ are complementary events to each other defined on a random experiment then
𝑃(𝐴) + 𝑃(𝐴′ ) = 1.
3. Addition Theorem: If 𝐴 and 𝐵 are any two events with respective probabilities 𝑃(𝐴) and
𝑃(𝐵), then the probability of occurrence of at least one of the events is given by
𝑃(𝐴 ∪ 𝐵) = 𝑃(𝐴) + 𝑃(𝐵) − 𝑃(𝐴 ∩ 𝐵).
4. The probability of null event is zero i.e., 𝑃(∅) = 0.
Random variable
A Random variable is a real number 𝑋 connected with the outcome of a random experiment
𝐸. For example, if 𝐸 consists of three tosses of a coin, one can consider the random variable
which is the number of ℎ𝑒𝑎𝑑𝑠 (0, 1, 2 𝑜𝑟 3).
Let 𝑆 denote the sample space of a random experiment. A random variable means it is a rule
which assigns a numerical value to each and every outcome of the experiment. Thus, random
variable is a function 𝑋(𝜔) with domain 𝑆 and range (−∞, ∞) such that for every real
number 𝑎, the event [𝜔: 𝑋(𝜔) ≤ 𝑎] ∈ 𝐵 the field of subsets in 𝑆. It is denoted as 𝑓: 𝑆 → 𝑅.
Note that all the outcomes of the experiment are associated with a unique number. Therefore,
𝑓 is an example of a random variable. Usually, a random variable is denoted by letters such
as 𝑋, 𝑌, 𝑍 etc. The image set of the random variable may be written as 𝑓(𝑆) = {0, 1, 2, 3}.
There are two types of random variables. Namely;
1. Discrete Random Variable (DRV)
2. Continuous Random Variable (CRV).
Discrete Random Variable: A discrete random variable is one which takes only a countable
number of distinct values such as 0, 1, 2, 3, ⋯. Discrete random variables are usually (but not
necessarily) counts. If a random variable takes at most a countable number of values, it is
called a discrete random variable. In other words, a real valued function defined on a
discrete sample space is called a discrete random variable.
Examples of Discrete Random Variable:
This function 𝑝 is called the probability mass function of the random variable 𝑋 and the
set {𝑥𝑖 , 𝑝(𝑥𝑖 )} is called the probability distribution of the random variable 𝑋.
Remarks:
1. The set of values which 𝑋 takes is called the spectrum of the random variable.
2. For discrete random variable, knowledge of the probability mass function enables us to
compute probabilities of arbitrary events. In fact, if 𝐸 is a set of real numbers, (𝑋 ∈ 𝐸) =
∑𝑥∈𝐸∩𝑆 𝑝(𝑥), where 𝑆 is the sample space.
Cumulative Distribution Function: For the random variable 𝑋 = {𝑥1 , 𝑥2 , 𝑥3 , … }. The
cumulative distribution function is given by 𝐹(𝑥) = 𝑃(𝑋 ≤ 𝑥) = ∑𝑥𝑖 ≤𝑥 𝑝 (𝑥𝑖 ).
Mean/Expected Value, Variance, and Standard Deviation of DRV:
The mean or expected value of a DRV 𝑋 is defined as
𝐸(𝑋) = 𝜇 = ∑ 𝑥𝑖 𝑃(𝑋 = 𝑥𝑖 ) = ∑ 𝑝(𝑥𝑖 )𝑥𝑖 .
The variance of a DRV 𝑋 is defined as
𝑉𝑎𝑟 (𝑋) = 𝜎 2 = ∑ 𝑃(𝑋 = 𝑥𝑖 ) (𝑥𝑖 − 𝜇)2 = ∑ 𝑝𝑖 (𝑥𝑖 − 𝜇)2 = ∑ 𝑝𝑖 (𝑥𝑖2 − 𝜇 2 ) .
The standard deviation of DRV 𝑋 is defined as
𝑆𝐷 (𝑋) = 𝜎 = √𝜎 2 = √𝑉𝑎𝑟 (𝑋) .
Continuous Random Variable: A continuous random variable is not defined at specific
values. Instead, it is defined over an interval of values. Thus, a random variable 𝑋 is said to
be continuous if it can take all possible values between certain limits. In other words, a
random variable is said to be continuous when its different values cannot be put in 1-1
correspondence with a set of positive integers. Here, the probability of observing any single
Probability Density Function: The probability density function (p.d.f) of a random variable
𝑋 usually denoted by 𝑓𝑥 (𝑥) or simply by 𝑓(𝑥) has the following obvious properties:
i) 𝑓(𝑥) ≥ 0, −∞ < 𝑥 < ∞
∞
ii) ∫−∞ 𝑓(𝑥)𝑑𝑥 = 1
iii) The probability 𝑃(𝐸) given by 𝑃(𝐸) = ∫ 𝑓(𝑥)𝑑𝑥 is well defined for any event 𝐸.
If 𝑓(𝑥) is the p.d.f of 𝑥, then the probability that 𝑥 belongs to 𝐴, where 𝐴 is some interval
(𝑎, 𝑏) is given by the integral of 𝑓(𝑥) over that interval.
𝑏
i.e., 𝑃(𝑋 ∈ 𝐴) = ∫𝑎 𝑓(𝑥)𝑑𝑥
Cumulative Distribution Function: Cumulative Distribution function of a continuous
𝑥
random variable is defined as 𝐹(𝑥) = ∫−∞ 𝑓(𝑡)𝑑𝑡 for − ∞ < 𝑥 < ∞ .
Mean/Expectation, Variance and Standard deviation of CRV:
∞
The mean or expected value of a CRV 𝑋 is defined as 𝜇 = 𝐸(𝑋) = ∫−∞ 𝑥 𝑓(𝑥)𝑑𝑥
∞
The variance of a CRV 𝑋 is defined as 𝑉𝑎𝑟(𝑋) = 𝜎 2 = ∫−∞ 𝑥 2 𝑓(𝑥)𝑑𝑥 − 𝜇 2
Find (𝑖)𝑘, (𝑖𝑖) 𝐹(4), (𝑖𝑖𝑖)𝑃(𝑋 ≥ 5), (𝑖𝑣)𝑃(2 ≤ 𝑋 < 5), (𝑣) 𝐸(𝑋) 𝑎𝑛𝑑 (𝑣𝑖)𝑉𝑎𝑟(𝑋).
(32)(17
0) 3
𝑓(2) = 𝑃(𝑋 = 2) = = .
(20
2) 190
4. If a car agency sells 50% of its inventory of a certain foreign car equipped with side
airbags, find a formula for the probability distribution of the number of cars with side airbags
among the next 4 cars sold by the agency.
Solution: Since the probability of selling an automobile with side airbags is 0.5, the 24 = 16
events in the sample space are equally likely to occur. Therefore, the denominator for all
probabilities, and also for our function, is 16. To obtain the number of ways of selling 3 cars
with side airbags, it is required to consider the number of ways of partitioning 4 outcomes
into two cells, with 3 cars with side airbags assigned to one cell and the model without side
airbags assigned to the other. This can be done in (43) = 4 ways. In general, the event of
selling x models with side airbags and 4 − 𝑥 models without side airbags can occur in (𝑥4)
ways, where 𝑥 can be 0, 1, 2, 3, or 4. Thus, the probability distribution 𝑓(𝑥) = 𝑃(𝑋 = 𝑥) is
1
𝑓(𝑥) = (16 ) (𝑥4) 𝑓𝑜𝑟 𝑥 = 0, 1, 2, 3, 4.
(iii) Determine the number 𝑏 such that 𝑃(𝑋 < 𝑏) = 𝑃(𝑋 > 𝑏).
⇒ 3𝑏 2 − 2𝑏 3 = [1 − 3𝑏 2 + 2𝑏 3 ]
⇒ 4𝑏 3 − 6𝑏 2 + 1 = 0
⇒ (2𝑏 − 1)(2𝑏 2 − 2𝑏 − 1) = 0
1
From this 𝑏 = 2 is the only real value lying between 0 and 1 and satisfying the given
condition.
8. Suppose that the error in the reaction temperature, in 𝐶 0 , for a controlled laboratory
experiment is a continuous random variable 𝑋 having the probability density function
𝑥2
𝑓(𝑥) = { 3 , −1 < 𝑥 < 2
0 , elsewhere
(i) Verify that 𝑓(𝑥) is a probability density function.
(ii) Find 𝑃(0 < 𝑋 ≤ 1).
Solution:
∞ 2 𝑥2
(i) ∫−∞ 𝑓(𝑥)𝑑𝑥 = ∫−1 3
𝑑𝑥 = 1. Hence the given function is a p.d.f.
1 𝑥2 1
(ii) 𝑃(0 < 𝑋 ≤ 1) = ∫0 3
𝑑𝑥 = 9 .
9. The length of time (in minutes) that a certain lady speaks on telephone is found to be a
−𝑥
10. Suppose 𝑋 is a continuous random variable with the following probability density
function
𝑓(𝑥) = 3𝑥 2 for 0 < 𝑥 < 1 . Find the mean and variance of 𝑋.
∞
Solution: Mean = 𝜇 = ∫−∞ 𝑥𝑓(𝑥)𝑑𝑥
0 1 ∞
= ∫−∞ 𝑥𝑓(𝑥)𝑑𝑥 + ∫0 𝑥𝑓(𝑥)𝑑𝑥 + ∫1 𝑥𝑓(𝑥)𝑑𝑥
1 1 3
= 0 + ∫0 𝑥 ∗ 3𝑥 2 𝑑𝑥 + 0 = ∫0 3𝑥 3 𝑑𝑥 = .
4
∞
Variance = 𝜎 2 = ∫−∞ 𝑥 2 𝑓(𝑥)𝑑𝑥 − 𝜇 2
1
= ∫0 𝑥 2 𝑓(𝑥)𝑑𝑥 − 𝜇 2
1 3 2
= ∫0 𝑥 2 ∗ 3𝑥 2 𝑑𝑥 − (4)
1 3 2 3
= ∫0 3𝑥 4 𝑑𝑥 − (4) = .
80
11. In a certain city, the daily consumption of electric power (in million 𝑘𝑊/ℎ𝑟) is a
continuous random variable having the probability density function
1 −𝑥/3
𝑓𝑋 (𝑥) = {9 𝑥𝑒 , 𝑥≥0
0, 𝑥<0
If the cities power plant has a daily capacity of 12 million 𝑘𝑊/ℎ𝑟. What is the probability
that this power supply will be insufficient on any given day?
1
𝑥𝑒 −𝑥/3 , 𝑥≥0
Solution: Given pdf, 𝑓𝑋 (𝑥) = {9
0, 𝑥<0
The power supply will be insufficient on any given day if power consumption is greater than
12 million 𝑘𝑊/ℎ𝑟.
𝑥 𝑥 ∞
− −
∞ ∞1 1 𝑥𝑒 3 𝑒 3 45
i.e., 𝑃(𝑋 > 12) = ∫12 𝑓𝑋 (𝑥) 𝑑𝑥 = ∫12 9 𝑥𝑒 −𝑥/3 𝑑𝑥 = 9[ 1 − 1 2
] = 9
𝑒 −4 .
− (− )
3 3 12
1, 0 ≤ 𝑥 ≤ 1
12. The density function of 𝑋 is given by 𝑓𝑋 (𝑥) = { .
0, 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒
Fourth Semester 11 Probability theory and Linear Programming
(MA241TA)
Let 𝑌 = 𝑒 𝑋 . (i) Find pdf of 𝑌, (ii) 𝐸(𝑌).
Solution:
Cumulative distribution function of 𝑌 is
log 𝑦
𝑋
𝐹𝑌 (𝑦) = 𝑃(𝑌 ≤ 𝑦) = 𝑃(𝑒 ≤ 𝑦) = 𝑃(𝑋 ≤ log 𝑦) = ∫ 𝑓𝑋 (𝑥) 𝑑𝑥
−∞
0 log 𝑦
= ∫ 𝑓𝑋 (𝑥) 𝑑𝑥 + ∫ 𝑓𝑋 (𝑥) 𝑑𝑥 , 𝑤ℎ𝑒𝑛 0 ≤ 𝑥 ≤ 1, 1 ≤ 𝑦 ≤ 𝑒 ⇒ 0 ≤ log 𝑦 ≤ 1
−∞ 0
log 𝑦
⇒ 𝐹𝑌 (𝑦) = ∫ 1 𝑑𝑥 = log 𝑦
0
By differentiating, we get
𝑑 1
𝑓𝑌 (𝑦) = 𝑑𝑦 𝐹𝑌 (𝑦) = 𝑦 , 1 ≤ 𝑦 ≤ 𝑒.
𝑒 1
Hence 𝐸[𝑌] = ∫1 𝑦𝑓𝑌 (𝑦)𝑑𝑦 = 𝑒 − 1 OR 𝐸[𝑌] = ∫0 𝑒 𝑥 𝑑𝑥 = 𝑒 − 1.
13. Given the probability density function 𝑓𝑋 (𝑥), find the cumulative distribution function.
0, 𝑥 < −2
0.25, −2 ≤ 𝑥 < 1
𝑓𝑋 (𝑥) = {
0.5, 1 ≤ 𝑥 < 1.5
0, 1.5 ≤ 𝑥
Solution:
𝑥
(i) if 𝑥 < −2, 𝐹𝑋 (𝑥) = ∫−∞ 𝑒 𝑥 𝑑𝑥 = 0
𝑥 𝑥
(ii) if −2 ≤ 𝑥 ≤ 1, 𝐹𝑋 (𝑥) = ∫−∞ 𝑓𝑋 (𝑥)𝑑𝑥 = ∫−2 0.25 𝑑𝑥 = 0.25(𝑥 + 2)
𝑥 −2 1 𝑥
(iii) if 1 ≤ 𝑥 ≤ 1.5, 𝐹𝑋 (𝑥) = ∫−∞ 𝑒 𝑥 𝑑𝑥 = ∫−∞ 0 𝑑𝑥 + ∫−2 0.25 𝑑𝑥 + ∫1 0.5 𝑑𝑥
= 0 + 0.25(1 + 2) + 0.5(𝑥 − 1) = 0.5𝑥 + 0.25
𝑥 1 1.5 𝑥
(iv) if 1.5 ≤ 𝑥, 𝐹𝑋 (𝑥) = ∫−∞ 𝑒 𝑥 𝑑𝑥 = ∫−2 0.25 𝑑𝑥 + ∫1 0.5 𝑑𝑥 + ∫1.5 0 𝑑𝑥
= 0.25(1 + 2) + 0.5(1.5 − 1) + 0 = 1
Thus the cumulative distribution function is given by
0, 𝑥 < −2
0.25𝑥 + 0.5, −2 ≤ 𝑥 < 1
𝐹𝑋 (𝑥) = {
0.5𝑥 + 0.25, 1 ≤ 𝑥 < 1.5
1, 1.5 ≤ 𝑥
14. Determine the probability density function for the following cumulative distribution
function.
0, 𝑥<0
0.2𝑥, 0≤𝑥<4
𝐹𝑋 (𝑥) = {
0.04𝑥 + 0.64, 4≤𝑥<9
1, 9≤𝑥
Fourth Semester 12 Probability theory and Linear Programming
(MA241TA)
∞ 𝑑
Solution: If 𝑓(𝑥) is a p.d.f. then ∫−∞ 𝑓(𝑥) 𝑑𝑥 ⇒ 𝑓(𝑥) = 𝑑𝑥 𝐹(𝑥).
𝑃(𝑋 𝑏𝑒𝑖𝑛𝑔 𝑎𝑛 𝑜𝑑𝑑 𝑛𝑢𝑚𝑏𝑒𝑟) by first establishing that 𝑃(𝑥) is a probability function.
3. The probability mass function of a random variable 𝑋 is zero except the points 𝑥 = 0,1,2. At
these points it has the values 𝑝(0) = 3𝑐 3 , 𝑝(1) = 4𝑐 − 10𝑐 2 and
𝑝(2) = 5𝑐 − 1 for some 𝑐 > 0.
a) Determine the value of 𝑐.
b) Compute the probabilities 𝑃(𝑋 < 2)𝑎𝑛𝑑 𝑃(1 < 𝑋 ≤ 2).
c) Find the largest 𝑥 such that (𝑥) < 1/2 .
d) Find the smallest 𝑥 such that 𝐹(𝑥) ≥ 1/3.
1
4. If 𝑋 is a random variable with 𝑃(𝑋 = 𝑥) = 2𝑥 , where 𝑥 = 1,2,3, … ∞.
1
7. Find the mean and variance of the probability density function 𝑓(𝑥) = 2 𝑒 −|𝑥|
Answers: 1) 0.1392 2) 3/5 3) 1/3, 1/3, 2/3, 1, 1 4) 1, 1/3, 1/7 5) 1/18, 1, 13/27
6) 1/2, 1/2 7) Mean = 0 and Variance = 2 8) Unit area and 0, 1.
JOINT PROBABILITY
Two or more random variables:
So far, only single random variables were considered. If one chooses a person at random and
measures his or her height and weight, each measurement is a random variable – but taller
people also tend to be heavier than shorter people, so the outcomes will be related. In order to
deal with such probabilities, joint probability distribution of two random variables are studied
in detail.
Joint Probability distribution for discrete random variables
Joint Probability Mass Function:
Let 𝑋 and 𝑌 be random variables defined on the same sample space 𝑆 with respective range
spaces 𝑅𝑋 = {𝑥1 , 𝑥2 , … , 𝑥𝑛 } and 𝑅𝑌 = {𝑦1 , 𝑦2 , … , 𝑦𝑚 }. The joint distribution or joint
probability function of 𝑋 and 𝑌 is the function ℎ on the product space 𝑅𝑋 × 𝑅𝑌 defined by
ℎ(𝑥𝑖 , 𝑦𝑗 ) ≡ 𝑃(𝑋 = 𝑥𝑖 , 𝑌 = 𝑦𝑗 ) ≡ 𝑃({𝑠 ∈ 𝑆 ∶ 𝑋(𝑠) = 𝑥𝑖 , 𝑌(𝑠) = 𝑦𝑗 })
The function ℎ has the following properties:
(i) ℎ(𝑥𝑖 , 𝑦𝑗 ) ≥ 0
(ii) ∑𝑖 ∑𝑗 ℎ(𝑥𝑖 , 𝑦𝑗 ) = 1
Thus, ℎ defines a probability space on the product space 𝑅𝑋 × 𝑅𝑌 .
𝑌
𝑦1 𝑦1 … 𝑦𝑗 … 𝑦𝑚 ∑ 𝑦𝑖
𝑋 𝑖
The functions 𝑓 and 𝑔 on the right side and the bottom side, respectively, of the joint
distribution table are defined by
𝑓(𝑥𝑖 ) = ∑𝑗 ℎ(𝑥𝑖 , 𝑦𝑗 ) and 𝑔(𝑦𝑗 ) = ∑𝑗 ℎ(𝑥𝑖 , 𝑦𝑗 ).
That is, 𝑓(𝑥𝑖 ) is the sum of the entries in the 𝑖 𝑡ℎ row and 𝑔(𝑦𝑗 ) is the sum of the entries in the
𝑗 𝑡ℎ column. They are called the marginal distributions of 𝑋 and 𝑌, respectively.
𝑐𝑜𝑣(𝑋, 𝑌) = ∑ 𝑥𝑖 𝑦𝑗 ℎ(𝑥𝑖 , 𝑦𝑗 ) − 𝜇𝑋 𝜇𝑌
𝑖,𝑗
where 𝐴 and 𝐵 are now the events defined by 𝑋 = 𝑥 and 𝑌 = 𝑦, respectively, then
𝑃(𝑋=𝑥,𝑌=𝑦) ℎ(𝑥,𝑦)
𝑃(𝑌 = 𝑦|𝑋 = 𝑥) = = , provided 𝑓(𝑥) > 0,
𝑃(𝑋=𝑥) 𝑓(𝑥)
Problem 1. A coin is tossed three times. Let 𝑋 be equal to 0 or 1 according as a head or a tail
occurs on the first toss. Let 𝑌 be equal to the total number of heads which occurs. Determine
(i) the marginal distributions of 𝑋 and 𝑌, and (ii) the joint distribution of 𝑋 and 𝑌, (iii)
expected values of 𝑋, 𝑌, 𝑋 + 𝑌 and 𝑋𝑌, (iv) 𝜎𝑋 and 𝜎𝑌 , (v) 𝐶𝑜𝑣(𝑋, 𝑌) and 𝜌(𝑋, 𝑌).
Solution:
Here the sample space is given by
𝑆 = {𝐻𝐻𝐻, 𝐻𝐻𝑇, 𝐻𝑇𝐻, 𝐻𝑇𝑇, 𝑇𝐻𝐻, 𝑇𝐻𝑇, 𝑇𝑇𝐻, 𝑇𝑇𝑇}
(i) The distribution of the random variable 𝑋 is given by the following table
variable 𝑌)
which is the marginal distribution of the random variable 𝑌.
(ii) The joint distribution of the random variables 𝑋and 𝑌 is given by the following table
𝑌 0 1 2 3
𝑋 (zero Heads) (one Head) (two Head) (three
Head)
0
1 2 1
(First toss 0
8 8 8
Head
1
1 2 1
(First toss 0
8 8 8
Tail)
4 4 4
(iii) E[X] = μX = ∑ xi P(xi ) = x1 P(x1 ) + x2 P(x2 ) = 0 × + 1 × =
8 8 8
E[Y] = μY = ∑ yj P(yj ) = y1 P(y1 ) + y2 P(y2 ) + y3 P(y3 )
1 3 3 1 12
=0× +1× +2× +3× =
8 8 8 8 8
E[X + Y] = ∑ ∑ Pij (xi + yj )
= P11 (x1 y1 ) + P12 (x1 y2 ) + P13 (x1 y3 ) + P14 (x1 y4 ) + P21 (x2 y1 ) + P22 (x2 y2 )
+ P23 (x2 y3 ) + P24 (x2 y4 )
1 2 2 1
= 0(0 × 0) + (0 × 1) + (0 × 2) + (1 × 1) + (1 × 2) + 0(1 × 3) = 2.
8 8 8 8
(iv) σ2X = E[X 2 ] − μ2X = ∑ xi2 P(xi ) − [E(X)]2 = x12 P(x1 ) + x22 P(x2 )
4 4 4 2 1
= 02 × 8 + 12 × 8 − (8) = 4
= ∑ yi2 P(yi ) − [E(Y)]2 = y12 P(y1 ) + y22 P(y2 ) + y32 P(y3 ) + +y42 P(y4 )
1
2 2
3 2
3 2 1 2 2 3
= 0 × + 1 × + 2 × +3 × − ) =
(
8 8 8 8 8 4
1 1 3 1
(v) Cov(X, Y) = E(XY) − μX μY = − × = −
2 2 2 4
Cov(X,Y) −1/4 1
ρ(X, Y) = = (1/2)( =− .
σX σY √3/2) √3
Problem 2: The joint distribution of two random variables 𝑋 and 𝑌 is given by the following
table:
Y
2 3 4
X
1 0.06 0.15 0.09
2 0.14 0.35 0.21
Determine the individual distributions of X and Y. Also, verify that X and Y are stochastically
independent.
Solution:
X takes values 1, 2 and Y takes the values 2, 3, 4. Also, h11 = 0.06, h12 = 0.15
h13 = 0.09, h21 = 0.14, h22 = 0.35, h23 = 0.21
Therefore, f1 = h11 + h12 + h13 = 0.3, f2 = h21 + h22 + h23 = 0.7,
g1 = h11 + h21 = 0.2, g 2 = h12 + h22 = 0.5, g 3 = h13 + h23 = 0.3.
The distribution of X is given by
xi 1 2
𝑓𝑖 0.3 0.7
Fourth Semester 18 Probability theory and Linear Programming
(MA241TA)
The distribution of 𝑌 is given by
yj 2 3 4
gj 0.2 0.5 0.3
f1 g1 = 0.06 = h11 , f1 g 2 = 0.15 = h12, f1 g 3 = 0.09 = h13,
f2 g1 = 0.14 = h21 , f2 g 2 = 0.35 = h22, f2 g 3 = 0.21 = h23,
Thus, fi g j = hij for all values of i and j so, X and Y are stochastically independent.
Problem 3.The joint distribution of two random variables 𝑋 and 𝑌 is given by the following
table:
Y
0 1
X
0 0.1 0.2
1 0.4 0.2
2 0.1 0
(a) Find P(X + Y > 1)
(b) Determine the individual (marginal) probability distributions of X and Y and verify that X
and 𝑌 are not independent.
(c) Find 𝑃(𝑋 = 2|𝑌 = 0).
(d) Find the conditional distribution of 𝑋 given 𝑌 = 1.
Solution:Note that 𝑋 takes the values 0, 1, 2 and 𝑌 takes the values 0, 1
ℎ11 = 0.1, ℎ12 = 0.2, ℎ21 = 0.4, ℎ22 = 0.2, ℎ31 = 0.1, ℎ32 = 0,
(a) The event 𝑋 + 𝑌 > 1 occurs only when the pair (𝑋, 𝑌) takes the values (1,1), (2,0) and (2,1).
The probability that this event occurs is therefore
P(𝑋 + 𝑌 > 1) = ℎ22 + ℎ31 + ℎ32 = 0.2 + 0.1 + 0 = 0.3.
(b) 𝑓1 = ℎ11 + ℎ12 = 0.1 + 0.2 = 0.3.
𝑓2 = ℎ21 + ℎ22 = 0.4 + 0.2 = 0.6.
Problem 4. The joint distribution of two random variables 𝑋 and 𝑌 is given by 𝑝𝑖𝑗 =
𝑘(𝑖 + 𝑗), 𝑖 = 1, 2, 3, 4; 𝑗 = 1, 2, 3. Find (i) 𝑘 and (ii) the marginal distributions of 𝑋 and 𝑌.
Show that 𝑋 and 𝑌 are not independent.
Solution: For the given 𝑝𝑖𝑗 ,
4 3 4 3
∑ ∑ ℎ𝑖𝑗= ∑ ∑ ℎ = 𝑘 ∑ ∑(𝑖 + 𝑗 )
𝑖 𝑗 𝑖=1 𝑗=1 𝑖=1 𝑗=1
4 4
2 𝑗 +5
𝑔𝑗 = ∑𝑖 ℎ𝑖𝑗 = ∑4𝑗=1 ℎ𝑖𝑗 =𝑘 ∑3𝑗=1(𝑖 + 𝑗 ) =
27
Problem 6. The joint probability function for two discrete random variables 𝑋 and 𝑌 is given
𝑐(2𝑥 + 𝑦), 𝑖𝑓 𝑥, 𝑦 ∈ ℤ 𝑎𝑛𝑑 0 ≤ 𝑥 ≤ 2, 0 ≤ 𝑦 ≤ 3,
by 𝑝(𝑥, 𝑦) = {
0, 𝑜𝑡ℎ𝑒𝑟𝑤𝑖𝑠𝑒
Find (i) the value of the constant 𝑐, (ii) 𝑃(𝑋 = 2, 𝑌 = 1), (iii) 𝑃(𝑋 ≥ 1, 𝑌 ≤ 2),
(iv) 𝑃(𝑋 + 𝑌 ≤ 1), (v) 𝑃(𝑋 + 𝑌 > 1).
Solution: (i) If 𝑝(𝑥, 𝑦) is joint p.m.f., then ∑𝑥 ∑𝑦 𝑝(𝑥, 𝑦) = 1
Y
0 1 2 3
X
0 0 C 2C 3C
1 2C 3C 4C 5C
2 4C 5C 6C 7C
(i) ∑𝑥 ∑𝑦 𝑝(𝑥, 𝑦) = 0 + 𝐶 + 2𝐶 + 3𝐶 + 2𝐶 + 3𝐶 + 4𝐶 + 5𝐶 + 4𝐶 + 5𝐶 + 6𝐶 +
1
7𝐶 = 1 ⇒ 42𝐶 = 1 ⇒ 𝐶 = 42.
1
(ii) 𝑃(𝑋 = 2, 𝑌 = 1) = 5 × 42
4
(iii) 𝑃(𝑋 ≥ 1, 𝑌 ≤ 2) = (2𝐶 + 3𝐶 + 4𝐶 + 4𝐶 + 5𝐶 + 6𝐶) = 7
1
(iv) 𝑃(𝑋 + 𝑌 ≤ 1) = 0 + 𝐶 + 2𝐶 + 3𝐶 =
7
1 6
(v) 𝑃(𝑋 + 𝑌 > 1) = 1 − 𝑃(𝑋 + 𝑌 ≤ 1) = 1 − 7 = 7
Problem 7. Two ballpoint pens are selected at random from a box that contains 3 blue pens,
2 red pens, and 3 green pens. If 𝑋 is the number of blue pens selected and 𝑌 is the number of
red pens selected, find the conditional distribution of 𝑋, given that 𝑌 = 1, and use it to
determine 𝑃(𝑋 = 0|𝑌 = 1).
Solution: The possible pair of values (𝑥, 𝑦) are {(0, 0), (0, 1), (0, 2), (1, 0), (1, 1), (2, 0)}
Problems to practice:
1) The joint probability distribution of two random variables X and Y is given by the
following table.
Y 5
-2 -1 4
X
1 0.1 0.2 0 0.3
2 0.2 0.1 0.1 0
(a) Find the marginal distribution of 𝑋 and 𝑌, and evaluate 𝑐𝑜𝑣(𝑋, 𝑌).
(b) Also determine whether 𝜇𝑋 and 𝜇𝑌 .
(c) Find 𝑃(𝑌 = −1|𝑋 = 1) and 𝑃(𝑋 = 2|𝑌 = 4)
2) Two textbooks are selected at random from a shelf containing three statistics texts,
two mathematics texts and three engineering texts. Denoting the number of books
selected in each subject by S, M and E respectively, find (a) the joint distribution of S
and M, (b) the marginal distributions of S, M and E, and (c) Find the correlation of the
random variables S and M.
3) Consider an experiment that consists of 2 throws of a fair die. Let 𝑋 be the number of
4s and 𝑌 be the number of 5s obtained in the two throws. Find the joint probability
distribution of 𝑋 and 𝑌. Also evaluate 𝑃(2𝑋 + 𝑌 < 3).
where ℎ(𝑥, 𝑦) is the joint density function of 𝑋 and 𝑌, ℎ1 (𝑥) is the marginal density
function of 𝑋.
Fourth Semester 25 Probability theory and Linear Programming
(MA241TA)
𝑃(𝑎 < 𝑋 < 𝑏, 𝑐 < 𝑌 < 𝑑)
𝑃(𝑐 < 𝑌 < 𝑑|𝑎 < 𝑥 < 𝑏) =
𝑃(𝑎 < 𝑥 < 𝑏)
2 −𝑦 1 2 22 1
= 𝑒 ∫ (𝑥 + 1) 𝑑𝑥 = 𝑒 −𝑦 { − }
3 0 3 2 2
= 𝑒 −𝑦 , 𝑦 > 0.
Therefore, ℎ1 (𝑥)ℎ2 (𝑦) = ℎ(𝑥, 𝑦) and hence 𝑥 and 𝑦 are stochastically independent.
Problem 2: The life time 𝑥 and brightness 𝑦 of a light bulb are modeled as continuous
random variables with joint density function
Fourth Semester 26 Probability theory and Linear Programming
(MA241TA)
ℎ(𝑥, 𝑦) = 𝛼𝛽𝑒 −(𝛼 𝑥+ 𝛽𝑦) , 0 < 𝑥 < ∞, 0 < 𝑦 < ∞.
Where 𝛼 and 𝛽 are appropriate constants. Find (i) the marginal density functions of 𝑥 and 𝑦,
and (ii) the compound cumulative distributive function.
Solution: For the given distribution, the marginal density function of 𝑥 is
∞ ∞
ℎ1 (𝑥) = ∫ ℎ(𝑥, 𝑦) 𝑑𝑦 = ℎ1 (𝑥) = ∫ 𝛼𝛽𝑒 −(𝛼 𝑥+ 𝛽𝑦) 𝑑𝑦
−∞ 0
∞
= 𝛼𝛽 𝑒 −𝛼𝑥 ∫0 𝑒 −𝛽𝑦 𝑑𝑦 = 𝛼 𝑒 −𝛼𝑥 , 0 < 𝑥 < ∞
the marginal density function of 𝑦 is
∞
ℎ2 (𝑦) = ∫ ℎ(𝑥, 𝑦) 𝑑𝑥 = 𝛽 𝑒 −𝛽𝑦 , 0 < 𝑦 < ∞.
−∞
1 1
= 𝛼𝛽 { (1 − 𝑒 −𝛼 𝑢 )} { (1 − 𝑒 −𝛽 𝑣 )}
𝛼 𝛽
= (1 − 𝑒 −𝛼 𝑢 )(1 − 𝑒 −𝛽 𝑣 ), 0 < 𝑢 < ∞, 0 < 𝑣 < ∞.
Problem 3: The joint probability density function of two random variables 𝑥 and 𝑦 is given
2, 0 < 𝑥 < 𝑦 < 1
by h(x, y) = {
0, 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
(i) Find the covariance between 𝑥 and 𝑦.
Solution: The marginal density function of 𝑥 is
1
∞
ℎ1 (𝑥) = ∫ ℎ(𝑥, 𝑦) 𝑑𝑦 = {∫𝑥 2 𝑑𝑦 = 2(1 − 𝑥), 0 < 𝑥 < 1
−∞
0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
The marginal density function of 𝑦 is
𝑦
∞
∫ 2 𝑑𝑥 = 2𝑦, 0<𝑦<1
ℎ2 (𝑦) = ∫ ℎ(𝑥, 𝑦) 𝑑𝑥 = { 0
−∞
0 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
∞ 1
E[x] = ∫ 𝑥 ℎ1 (𝑥)𝑑𝑥 = ∫ 𝑥{2 (1 − 𝑥)} 𝑑𝑥
−∞ 0
1
1 1 1
= 2 ∫ (𝑥 − 𝑥 2 ) 𝑑𝑥 = 2 ( − ) = ,
0 2 3 3
Therefore,
1 1 2 1
𝐶𝑜𝑣 (𝑥, 𝑦) = 𝐸[𝑥𝑦] − 𝐸[𝑥]𝐸[𝑦] = − . = .
4 3 3 36
𝑒 −(𝑥+𝑦) , 𝑥 ≥ 0, 𝑦 ≥ 0
Problem 4: Verify that f (x, y) = { is a density function of a joint
0, 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
probability distribution. Then evaluate the following:
1
(i) 𝑃 ( < 𝑥 < 2, 0 < 𝑦 < 4) (ii) 𝑃(𝑥 < 1) (iii) 𝑃(𝑥 > 𝑦) (iv) 𝑃(𝑥 + 𝑦 ≤ 1),
2
∞ ∞ ∞ ∞ ∞ ∞
𝑓(𝑥, 𝑦) = ∫ ∫ 𝑓(𝑥, 𝑦)𝑑𝑥 𝑑𝑦 = ∫ ∫ 𝑒 −(𝑥+𝑦) 𝑑𝑥 𝑑𝑦 = ∫ 𝑒 −𝑥 𝑑𝑥 ∫ 𝑒 −𝑦 𝑑𝑦
−∞ −∞ −∞ −∞ −∞ 0
= (0 + 1)(0 + 1) = 1.
Therefore, 𝑓(𝑥, 𝑦) is a density function.
1 2 4 2 4
(i) 𝑃 (2 < 𝑥 < 2, 0 < 𝑦 < 4) = ∫1/2 ∫0 𝑓(𝑥, 𝑦) 𝑑𝑦 𝑑𝑥 = ∫1/2 ∫0 𝑒 −(𝑥+𝑦) 𝑑𝑦 𝑑𝑥
2 4
= ∫1/2 𝑒 −𝑥 𝑑𝑥 ∫0 𝑒 −𝑦 𝑑𝑦 = (𝑒 −1/2 − 𝑒 −2 )(1 − 𝑒 −4 ).
∞ ∞ ∞
ℎ1 (𝑥) = ∫ 𝑓(𝑥, 𝑦) 𝑑𝑦 = ∫ 𝑒 −(𝑥+𝑦) 𝑑𝑦 = 𝑒 −𝑥 ∫ 𝑒 −𝑦 𝑑𝑦 = 𝑒 −𝑥
−∞ 0 0
1 1 1
Therefore, 𝑃(𝑥 < 1) = ∫0 ℎ1 (𝑥) 𝑑𝑥 = ∫0 𝑒 −𝑥 𝑑𝑥 = 1 − .
𝑒
∞ 𝑦 ∞ 𝑦
(iii) 𝑃(𝑥 ≤ 𝑦) = ∫0 {∫0 𝑓(𝑥, 𝑦) 𝑑𝑥} 𝑑𝑦 = ∫0 {∫0 𝑒 −(𝑥+𝑦) 𝑑𝑥} 𝑑𝑦
∞ 𝑦 ∞
= ∫ 𝑒 −𝑦 (∫ 𝑒 −𝑥 𝑑𝑥) 𝑑𝑦 = ∫ 𝑒 −𝑦 (1 − 𝑒 −𝑦 ) 𝑑𝑦
0 0 0
∞
1 1
= ∫ (𝑒 −𝑦 − 𝑒 −2𝑦 ) 𝑑𝑦 = 1 − =
0 2 2
1 1
Therefore, 𝑃(𝑥 > 𝑦) = 1 − 𝑃(𝑥 ≤ 𝑦) = 1 − = .
2 2
1 1−𝑥 1 1−𝑥
= ∫ ∫ 𝑓(𝑥, 𝑦) 𝑑𝑦 𝑑𝑥 = ∫ {∫ 𝑒 −(𝑥+𝑦) 𝑑𝑦} 𝑑𝑥
𝑥=0 𝑦= 0 0 0
Fourth Semester 28 Probability theory and Linear Programming
(MA241TA)
1 1−𝑥 1
=∫ 𝑒 −𝑥
{∫ 𝑒 −𝑦
𝑑𝑦} 𝑑𝑥 = ∫ 𝑒 −𝑥 {1 − 𝑒 − (1−𝑥) } 𝑑𝑥
0 0 0
1
2
= ∫ (𝑒 −𝑥 − 𝑒 −1 ) 𝑑𝑥 = 1 − .
0 𝑒
𝑃(0 < 𝑥 < 1|𝑦 = 2)
(v) 𝑃(0 < 𝑥 < 1|𝑦 = 2) = 𝑃(𝑦=2)
[putting 𝑦 = 2]
1
∫0 𝑒 −(𝑥+2) 𝑑𝑥 1
𝑃(0 < 𝑥 < 1|𝑦 = 2) = ∞ =1− = 0.63
∫0 𝑒 −(𝑥+2) 𝑑𝑥 𝑒
𝐶 𝐶
⇒− + = 1 ⇒ 𝐶 = 15
10 6
1 2
(i) 𝑃(𝑋 < 1, 𝑌 < 2) = ∬𝐷 𝑓𝑋,𝑌 (𝑥, 𝑦)𝑑𝐴 = ∫𝑥=0 ∫𝑦=𝑥 15 𝑒 −2𝑥 𝑒 −3𝑦 𝑑𝑦 𝑑𝑥
1
1 2 1
𝑒 −3𝑦 15
= 15 ∫ 𝑒 −2𝑥 [ ] 𝑑𝑥 = ∫ 𝑒 −2𝑥 [𝑒 −6 − 𝑒 −3𝑥 ]𝑑𝑥
−3 𝑦=𝑥 −3
𝑥=0 𝑥=0
1
𝑒 −6 𝑒 −2𝑥 𝑒 −5𝑥
= −5 [ − ] = 0.9879
−2 −5 0
∞
2 ∞ 2 𝑒 −3𝑦
(ii) 𝑃(1 < 𝑋 < 2) = ∫𝑥=1 ∫𝑦=𝑥 15 𝑒 −2𝑥 𝑒 −3𝑦 𝑑𝑦 𝑑𝑥 = 15 ∫𝑥=1 𝑒 −2𝑥 [ ] 𝑑𝑥
−3 𝑦=𝑥
2 2 2
15 5𝑒 −5𝑥
= ∫ 𝑒 −2𝑥 [0 − 𝑒 −3𝑥 ]𝑑𝑥 = 5 ∫ 𝑒 −5𝑥 𝑑𝑥 = [ ] = 0.0067
−3 −5 1
𝑥=1 𝑥=1
𝑦
∞ 𝑦 ∞ 𝑒 −2𝑥
(iii) 𝑃(𝑌 > 3) = ∫𝑦=3 ∫𝑥=0 15 𝑒 −2𝑥 𝑒 −3𝑦 𝑑𝑥 𝑑𝑦 = 15 ∫𝑦=3 𝑒 −3𝑦 [ ] 𝑑𝑦
−2 𝑥=0
∞ ∞
15 −5𝑦 −3𝑦
15 𝑒 −5𝑦 𝑒 −3𝑦
= ∫ [𝑒 − 𝑒 ]𝑑𝑦 = [ − ] = 0.000308
−2 −2 −5 −3 3
𝑦=3
2 2
(iv) 𝑃(𝑋 < 2, 𝑌 < 2) = ∫𝑥=0 ∫𝑦=𝑥 15 𝑒 −2𝑥 𝑒 −3𝑦 𝑑𝑦 𝑑𝑥
2 2 2
−2𝑥
𝑒 −3𝑦 15
= 15 ∫ 𝑒 [ ] 𝑑𝑥 = ∫ 𝑒 −2𝑥 [𝑒 −6 − 𝑒 −3𝑥 ]𝑑𝑥
−3 𝑦=𝑥 −3
𝑥=0 𝑥=0
2
𝑒 −6 𝑒 −2𝑥 𝑒 −5𝑥
= −5 [ − ] = 0.9939
−2 −5 0
(v) 𝐸[𝑋] = ∫𝐷 ∫ 𝑥 𝑓𝑋,𝑌 (𝑥, 𝑦) 𝑑𝐴, where
𝐷 = {(𝑥, 𝑦)| 𝑥 ≤ 𝑦 ≤ ∞, 0 ≤ 𝑥 ≤ ∞}
∞ ∞
as at 𝑥 = 1, 𝑦 𝑣𝑎𝑟𝑖𝑒𝑠 𝑓𝑟𝑜𝑚 1 𝑡𝑜 ∞
∞
𝑒−3𝑦+3 𝑒−3𝑦+3
𝐸[𝑌|𝑋 = 1] = 3 [𝑦 − ] = 4/3
−3 9 1
2 2
(x) 𝑃(𝑌 < 2|𝑋 = 1) = ∫𝑦=1 𝑓𝑌|𝑋 (𝑦|1)𝑑𝑦 = ∫𝑦=1 3𝑒 −3𝑦+3 𝑑𝑦
2
𝑒 −3𝑦+3
= 3[ ] = 1 − 𝑒 −3
−3 1
(xi) Conditional probability distribution of 𝑋 given by 𝑌 = 2
∞ 𝑦
15 −5𝑦 15 −3𝑦
𝑓𝑌 (𝑦) = ∫ 𝑓𝑋,𝑌 (𝑥, 𝑦) 𝑑𝑦 = ∫ 15 𝑒 −2𝑥 𝑒 −3𝑦 𝑑𝑦 = − 𝑒 + 𝑒
2 2
𝑥=−∞ 𝑥=0
15 −6
𝑓𝑌 (2) = (𝑒 − 𝑒 −10 )
2
𝑓𝑋,𝑌 (𝑥, 2) 15𝑒 −2𝑥−6
𝑓𝑋|𝑌 (𝑥|2) = =5 ,0 ≤ 𝑥 ≤ 𝑦
𝑓𝑌 (2) 𝑒 −5 (𝑒 −6 − 𝑒 −10 )
2
Problems to practice:
1
(6
f (x, y) = {8 − 𝑥 − 𝑦), 0 < 𝑥 < 2, 0<𝑦<4
0, 𝑒𝑙𝑠𝑒𝑤ℎ𝑒𝑟𝑒
Evaluate (i) 𝑃(𝑥 < 1, 𝑦 < 3), (ii)𝑃(𝑥 + 𝑦 < 3), (iii) the covariance between 𝑥 and 𝑦
and (iv) 𝑃(𝑥 < 1|𝑦 < 3)
. Video Links:
https://www.youtube.com/watch?v=82Ad1orN-NA
https://www.youtube.com/watch?v=eYthpvmqcf0
https://www.youtube.com/watch?v=L0zWnBrjhng
https://www.youtube.com/watch?v=Om68Hkd7pfw
https://www.youtube.com/watch?v=RYIb1u3C13I