Lesson 13
Lesson 13
Why we need
Statistics
Denny Vell M. Devaras, RPm
Chapter Objectives
Type 2 Error: We can fail to reject the null hypothesis even though it is really
false
• The probability of making a Type 2 error is affected by the amount of
overlap between the populations being sampled.
Applying Statistical Inference: An
Example
Experimental Errors
• The probability of making a Type 2 error is represented by the greek letter
β (beta)
• But say we knew that the odds of making a Type 2 error were equal to
exactly .75 in our experiment. Then the odds of making a correct decision
would be equal to 1 – .75, or .25. If we are likely to be wrong 3 out of 4
times (.75), then we should be right 1 out of 4 times (.25)
• The odds of correctly rejecting the null hypothesis when it is false are
always equal to 1 – β
• This 1-β is also referred to as the power of the statistical test
Experimental Errors
How can we reduce the value of β?
• Even though we cannot measure the precise value of β, we can reduce it
by increasing our sample size
• we can also reduce β by reducing the variability in our sample data
• We can also reduce β by using more powerful statistical tests, called
parametric tests
• We can also reduce β if we accept a less extreme significance level.
Experimental Errors
• The probability of making a Type 1 error is represented by the Greek letter
α (alpha)
• There is some chance (α) that we will reject the null hypothesis when we
should have conditionally retained it
• There is a 1 – α chance that we will fail to reject the null hypothesis when
that is the correct decision
The Odds
of Finding
Significance
The importance of variability
Variability – the amount of change or fluctuation we see in something
• As the amount of variability in the distribution goes up, the critical regions
fall farther from the center of the distribution
• When there is more variability, larger differences between means of
samples are required to reject the null hypothesis
One-tailed and Two-tailed Tests
One-tailed Test – A statistical procedure used when a directional prediction
has been made
• The critical region of the distribution of the test statistic is measured in just
one tail of the distribution
Organizing Data
• Statistical work will go more quickly and be more accurate if you begin by
organizing the data and labeling them in a clear and orderly way.
Summarizing Data
• Raw Data – The data we record as we run the experiment
• Summary Data – Descriptive statistics computed from the raw data of an
experiment including the measures of central tendency and variability
Measures of Central Tendency
Measures of Central Tendency – summary statistics that describe what is
typical of a distribution of scores
• Mode – the score that occurs most often
• Median – the score that divides the distribution in half so that the half of
the scores in the distribution fall above the median, half below
• Mean – the arithmetic average (add all the scores and divide by the total
number of scores)
• Skewed distribution – the mean median and mode will be different and
each can lead to different impressions about the data.
Measures of Variability
Range – the difference between the largest and smallest scores in a set of
data
• The range is often a useful measure; it can be computed quickly, and it
gives a straightforward indication of the spread between the high and low
scores of the distribution
• The problem with using the range is that it does not reflect the precise
amount of variability in all the scores
Measures of Variability
Variance – the average squared deviation of scores from their mean
• The variance tells us something about how much scores are spread out,
or dispersed, around the mean of the data.