Statistics
Statistics
• Statistics is a body of knowledge that is useful for collecting, organizing, presenting, analyzing,
and interpreting data (collections of any number of related observations) and numerical facts.
• Descriptive statistics deals with the presentation and organization of data. These types of
statistics summarize numerical information.
• Inferential statistics deals with the use of sample data to infer, or reach, general
conclusions about a much larger population.
Types of Sampling
Probability Sampling (Random Non-Probability Sampling
Sampling) (Non-random Sampling)
⇓
⇓
The decision whether a particular element is The decision whether a particular
included in the sample or not, is governed element is included in the sample
by Chance. or is not governed by Chance.
⇓ ⇓
Each element in the population has equal Each element in the population
probability (>0) of incorporating in the sample may not have an equal chance of
being incorporated in the sample.
Sampling Methods
Probability Sampling
Non-probability Sampling
(i) Sampling Random Sampling.
(i) Conveniences Sampling.
(a) With replacement.
(b) Without replacement. (ii) Purposive Sampling.
(ii) Systematic Sampling. (iii) Judgment Sampling.
(iii) Stratified Sampling. (iv) Cluster Sampling. (iv) Quota Sampling
Data Description
• Systematic gathering of crude information regarding any subject which is relevant for
our decision making problem.
Nature of Data
Qualitative Categorical Do you practice yoga? Yes/ No
Measurement Scale
Scale of Characteristics of Basic Empirical Examples
Measurement Measurement Operations
Nominal Identifies groups which can not Gender, Nationality, place
Order: No
be ordered of work
Distance: No
Unique origin: No
7
Statistical Methods for Multivariate Data
Syllabus
• Probability and Probability Distributions
• Sampling Distributions and Estimating
• Hypothesis Testing
• Chi-square Tests and Factorial Experiments • Index
Number and Forecasting Techniques
UNIT 1-Probability and Probability Distributions
Content
• Definition of Probability- Unconditional Probability statement
• Conditional Probability statement and its implications
• Bayes’ Theorem and its applications
• Joint Probabilities and its implications
• Mathematical Expectations
• Theoretical Probability Distributions – Binomial, Poisson and Normal (Their
Characteristics and Applications in Business).
Introduction
• The concept of ‘Probability’ is as old as civilization itself – originated from an age old malaise,
Gambling used to make bets.
• The probability theory was first applied to Gambling and later to other socio-economic
problems.
• Lately, quantitative analysis has become the backbone of statistical application in business
decision making and research.
• Statistics and especially the theory of probability play a vital role in making decisions in
situations, when there is a lack of certainty.
Terminology
Experiment🡺 An act which can be repeated under some given condition.
Random 🡺 Depending on chance. Random Experiment is an experiment whose result depends on
chance.
Properties:
• They can be repeated physically or conceptually.
• Outcome set, ‘S’ is specified in advance. • Repetition may not yield some
result
Outcome / sample space (S)
Random experiment => Result => Outcome Examples
(1) Tossing a coin=>Result: H or T =>Sample space = {H, T}
(2) Throwing a die=>Result: 1, 2,3,4,5 or 6 =>Sample space = {1, 2, 3, 4, 5, 6}
Trial: To conduct an experiment once is trial.
Events: Possible outcomes or combination of outcomes is termed as events.
Example: Trial - Tossing a coin;
Event – Occurrence of either H or T
Events
1) Elementary – which can not be decomposed into further simpler events.
Example: Tossing a coin sample space= {H, T}
2) Composite – which can be further decomposed into further simpler events
Example: Throwing a die sample space = {odd no., even no.} decomposed further = {1, 2, 3, 4, 5, 6}
• Equally likely events: Two or more events are said to be equally likely if none of them
expected to be in better performance than the other.
• Mutually exclusive events: Two or more events are said to be mutually exclusive if they
cannot occur simultaneously.
• Mutually Exhaustive Events: Two or more events are said to be exhaustive if at least one
of them must occur.
Set Theory
S: Universal set / sample space/ outcome set φ: Null Set
Let Ai ⊆S, ∀ i= 1, 2
And A1 = {x| xϵA1, A1 ⊆ S}
A2 = {x| x ϵ A2, A2 ⊆ S}
Then, A1 ∪ A2 = {x| x ϵ A1and / or x ϵ A2} A1 ∩ A2 = {x| x ϵ A1and x ϵ A2}
Ᾱ1 = {x∉ A1, A1 ⊆ S}
• Result 1 A1 ∪ Ᾱ1 = S, Mutually exhaustive.
• Result 2 A1 ∩Ᾱ1 =φ, mutually exclusive.
• Result 3 [combining (1) & (2)] A1and Ᾱ1 are disjoint partition of the sample space.
• Disjoint union of sample space (forA1,A2) Mutually exclusive A1∩A2 = φ Mutually exhaustive
A1 ∪ A2 = S
• Result 4 For every A1,A2 ⊆ S. A1∩A2 ⊆A1 and A1∩A2 ⊆A2. A1, A2 ⊆ (A1 ∪ A2)
7/21/2024 27
Probability of an event
Let A⊆S be any event.
Subjective Probability
🡺 Probability is based on the experience and judgment of the person making the estimate.
🡺 This may differ from person to person, depending on one’s perception of the situation and
the past experience.
Sometimes, subjective probability is also termed as the set theoretic approach to probability.
Rules of Probability
Rule 1 Let A1,A2 ⊆ S. If A1and A2are mutually exclusive events then A1∩A2 =φ
⇒ P (A1∪A2) = P (A1) + P (A2).
Rule 2 Let A1,A2 ⊆ S. For any A1,A2 ⊆ S. P (A1∪A2) = P (A1) + P (A2) – P (A1∩A2)
Conditional Probability
To discuss conditional probability, it would be better to start from the concept of
unconditional probabilities.
Under unconditional probabilities, when even an event A is considered, already an
outcome set S of which A ⊆ S is considered.
Hence, it would be more appropriate to call the probability of A as the probability of A
given S. This may be written as P (A) = P (A/S).
Conditional Probability …………………………………
Conditional Probability
Independence of Events:
Two events A1and A2are said to be independent iff P (A1∩A2) = P (A1) ⋅ P (A2)
• Result 1: If A1and A2are independent events then Ᾱ1and Ᾱ2 are independent.
• Result 2: If A1and A2are independent events then Ᾱ1and A2 are independent.
• Result 3: If A1and A2are independent events then A1and Ᾱ2 are independent.
Beyes’ Rule
Lemma If A1,A2,………. An denoted a disjoint partition of the outcome set and if P (Aj) ≠ 0, ∀ j = 1, 2,……..n.
Probability Distribution
• A probability distribution is essentially an extension of the theory of probability.
• To discuss the concept of probability distribution, we should start with distinguishing
frequency distribution with probability distribution.
• Probability distribution is essentially an extension of the relative frequency distribution
approach.
Probability Distribution
Consider the case of sample space associated with tossing a coin twice.
The sample space S = {HH, HT, TH, TT}.
Let X: be any variable which denotes the number of head appear.
No. of Heads (x) Frequency (f) Relative frequency / Probability
0 1 ¼
1 2 1/2
2 1 ¼
Total N=4 100 % = 1
Probability Distribution
X 0 1 2 Total
Here, the variable ‘X’ has a special characteristics i.e. corresponding to each and
every value of X, ∃ an associated probability.
⇒X is known as the random variable, as the occurrence of any value depends on chance.
Definition: A real Random Variable (x) is a variable defined on the outcome set of a
random experiment such that the probability statement P {X≤C} is
defined for every real number C
Notation
• P {a ≤ x ≤ b} ⇒ Probability that x lies between ‘a’ and ‘b’ (both ‘a’ and ‘b’ inclusive)
• P {a ≤ x <b} ⇒ Probability that x lies between ‘a’ and ‘b’ (only ‘a’ inclusive)
• P {x ≤ c} ⇒ Probability that ‘x’ takes value less than or equal to c
• P {x ≥ d} ⇒ Probability that x takes the value greater than or equal to ‘d’
Mathematical Expectation
Define. If X is a random variable with probability function P (x) and φ (x) is a function
of X which is again a random variable then the mathematical expectation of P (X) is defined
as: E {φ (x)} = Σφ (x) ⋅ P (x) if x is discrete
-∝<x<+∝
Case 1. φ (x) = X, E (X) = Σ X⋅ P (x) = μ
-∝<x<+∝
2 2 2
Case 2. φ (x) =X , E (X ) = Σ X ⋅ P (x).
-∝<x<+∝
Define σ2 = var (x) = E {x-E (x)}2 = E (x2) – {E(x)}2 So, Standard deviation = σ
Standardized variable z ={ x – E (x)} / s.d(X)
⇒ E(z) = 0;Var(z) = 1
Discrete Probability Distribution
• Binomial Probability Distribution
• Poisson Probability Distribution
Note: Poisson Probability Distribution is a limiting case of Binomial
Probability Distribution
Binomial Probability Distribution
The binomial distribution describes discrete data resulting from a random experiment known
as Bernoulli process. Binomial distribution is a probability distribution expressing the
probability of one set of dichotomous alternatives i.e. success or failure.
The Binomial probability model is appropriate in the following experimental situations:
suppose that a random experiment is such that.
(i) Any trial results in a success or a failure;
(ii) There are n repeated trials which are independent;
(iii) The probability of success in any trial; 0 <p< 1 => Probability of failure = (1-p)
Binomial Probability Distribution
Let, : denote the exact number of success in n trials.
Then X ~ Bin (n, P)
The corresponding probability function is:
1/2024 53
1 06 08
2 30 28
3 52 56
4 67 70
5 56 56
6 32 28
7 10 08
8 01 01
Step2: The Probabilities of various values of the random variables (x) to be determined by
using the p.m.f.
Step3 Multiply each and every term in step 2 by ‘n’ to get the expected frequencies.
Fitting of a Poisson distribution……………
0 214 212.62
1 92 93.34
2 20 20.49
3 03 3.00
4 01 0.33
N = 330
𝑥−μ
Z= σ
~ 𝑁 (0, 1)
Step 1: If x ~ N (μ, σ2) convert it into
Z = x - μ/ σ~ N (0, 1)
Step 2: The sum of area under the Normal curve is 1,
Moreover f (z) = f (-z).
Step 3: Looking from the table identify the required probability.
SUBRATA SUTRADHAR