Qa Ken Black All Chapters Solution PDF
Qa Ken Black All Chapters Solution PDF
Chapter 1
Introduction to Statistics
LEARNING OBJECTIVES
1. Define statistics.
of the meaning of the term. For this reason, several perceptions of the word
statistics are given in the chapter.
CHAPTER OUTLINE
Marketing
Management
Finance
Economics
Nominal Level
Page | 3
Ordinal Level
Interval Level
Ratio Level
KEY TERMS
finance - World bank bond rates, number of failed savings and loans,
measured risk of common stocks, stock dividends, foreign exchange
rate, liquidity rates for a single-family, etc.
1.5 a) ratio
b) ratio
c) ordinal
d) nominal
e) ratio
f) ratio
g) nominal
h) ratio
1.6 a) ordinal
b) ratio
c) nominal
d) ratio
e) interval
f) interval
Page | 9
g) nominal
h) ordinal
1.7 a) The population for this study is the 900 electric contractors who
purchased Rathburn wire.
Chapter 2
Charts and Graphs
LEARNING OBJECTIVES
The overall objective of chapter 2 is for you to master several techniques for
summarizing and depicting data, thereby enabling you to:
CHAPTER OUTLINE
Class Midpoint
Relative Frequency
Cumulative Frequency
Histograms
Frequency Polygons
Ogives
Pie Charts
Pareto Charts
KEY TERMS
Ogive
P a g e | 13
2.1
10 - under 25 9
25 - under 40 13
40 - under 55 11
55 - under 70 9
70 - under 85 8
50
10 - under 18 7
18 - under 26 3
26 - under 34 5
34 - under 42 9
42 - under 50 7
50 - under 58 3
58 - under 66 6
66 - under 74 4
74 - under 82 4
P a g e | 14
82 - under 90 2
2.2 One possible frequency distribution is the one below with 12 classes
and class intervals of 2.
39 - under 41 2
41 - under 43 1
43 - under 45 5
45 - under 47 10
47 - under 49 18
49 - under 51 13
51 - under 53 15
53 - under 55 15
55 - under 57 7
57 - under 59 9
59 - under 61 4
61 – under 63 1
The distribution reveals that only 13 of the 100 boxes of raisins contain
50 ± 1 raisin (49 -under 51). However, 71 of the 100 boxes of raisins
contain between 45 and 55 raisins. It shows that there are five boxes
that have 9 or more extra raisins (59-61 and 61-63) and two boxes that
have 9-11 less raisins (39-41) than the boxes are supposed to contain.
2.3
5 - 10 8 7.5 .0930 14
10 - 15 17 12.5 .1977 31
15 - 20 23 17.5 .2674 54
20 - 25 18 22.5 .2093 72
25 - 30 10 27.5 .1163 82
30 - 35 4 32.5 .0465 86
TOTAL 86 1.0000
2.4
inventory build-up,
2.6 Histogram:
P a g e | 18
P a g e | 19
Frequency Polygon:
2.7 Histogram:
P a g e | 20
P a g e | 21
Frequency Polygon:
2.8 Ogive:
P a g e | 22
P a g e | 23
21 2, 8, 8, 9
22 0, 1, 2, 4, 6, 6, 7, 9, 9
23 0, 0, 4, 5, 8, 8, 9, 9, 9, 9
24 0, 0, 3, 6, 9, 9, 9
25 0, 3, 4, 5, 5, 7, 7, 8, 9
26 0, 1, 1, 2, 3, 3, 5, 6
27 0, 1, 3
Deere .246 89
Illinois Too Works .144 52
Eaton .121 44
Pie Chart:
P a g e | 24
Annual Sales
Deere; 25%
P a g e | 25
Southwest .202 73
Delta .198 71
American .181 65
United .140 50
Northwest .108 39
US Airways .094 34
Continental .078 28
Pie Chart:
Em planem ents
Continental; 8%
Southw est; 20%
US Airw ays; 9%
Delta; 20%
United; 14%
American; 18%
P a g e | 26
Merck .125 45
Wyeth .095 34
Pie Chart:
Wyeth; 9%
Merck; 13%
Johnson & Johnson; 26%
P a g e | 27
P a g e | 28
1 3, 6, 7, 7, 7, 9, 9, 9
2 0, 3, 3, 5, 7, 8, 9, 9
3 2, 3, 4, 5, 7, 8, 8
4 1, 4, 5, 6, 6, 7, 7, 8, 8, 9
5 0, 1, 2, 2, 7, 8, 9
6 0, 1, 4, 5, 6, 7, 9
7 0, 7
8 0
The stem and leaf plot shows that the number of passengers per flight
were
Customer Complaints
800
700
100
600 80
Percent
500
Count
400
60
300 40
200
100
20
0 0
C1
2.15
3500
3000
2500
2000
Industrial Products
1500
1000
500
0
3000 3500 4000 4500 5000 5500 6000 6500
Hum an Food
P a g e | 31
2.16
180
160
140
120
100
Sales
80
60
40
20
0
1 2 3 4 5 6 7 8 9 10 11
Adve rtising
16 - under 23 6
23 - under 30 9
30 - under 37 4
37 - under 44 4
44 - under 51 4
51 - under 58 3
TOTAL 30
P a g e | 32
2.18
50 - under 60 13
60 - under 70 27
70 - under 80 43
80 - under 90 31
90 - under 100 9
TOTAL 123
Histogram:
Frequency Polygon:
P a g e | 34
Ogive:
A 55 .180 65
B 121 .397 143
P a g e | 35
C 83 .272 98
D 46 .151 54
Pie Chart:
D; 15% A; 18%
C; 27%
B; 40%
28 4, 6, 9
29 0, 4, 8
30 1, 6, 8, 9
31 1, 2, 4, 6, 7, 7
32 4, 4, 6
33 5
P a g e | 36
1 673 26.96
2 29 1.16
3 108 4.33
4 379 15.18
5 73 2.92
6 564 22.60
7 12 0.48
8 402 16.11
9 54 2.16
10 202 8.09
2496
Pareto Chart:
P a g e | 37
Problems
2500 100
2000 80
1500 60
Percent
Count
1000 40
500 20
0 0
C5 1 6 8 4 10 3 5 9 2 Other
Count 673 564 402 379 202 108 73 54 29 12
Percent 27.0 22.6 16.1 15.2 8.1 4.3 2.9 2.2 1.2 0.5
Cum % 27.0 49.6 65.7 80.8 88.9 93.3 96.2 98.4 99.5 100.0
2.23
P a g e | 38
16
14
12
10
8
Y
0
4 6 8 10 12 14 16 18
32 - under 37 1
37 - under 42 4
42 - under 47 12
47 - under 52 11
52 - under 57 14
57 - under 62 5
62 - under 67 2
P a g e | 39
67 - under 72 1
TOTAL 50
P a g e | 40
2.25
25 – 30 6 27.5 .1132 14
30 – 35 5 32.5 .0943 19
35 – 40 12 37.5 .2264 31
40 – 45 15 42.5 .2830 46
45 – 50 7 47.5 .1321 53
TOTAL 53 .9999
10 - under 20 2
20 - under 30 3
30 - under 40 9
40 - under 50 7
50 - under 60 12
60 - under 70 9
70 - under 80 6
80 - under 90 2
50
P a g e | 41
Histogram:
P a g e | 42
Frequency Polygon:
The normal distribution appears to peak near the center and diminish
towards the
end intervals.
2.27
Class Cumulative
20 – 25 8 8
25 – 30 6 14
30 – 35 5 19
35 – 40 12 31
40 – 45 15 46
45 – 50 7 53
TOTAL 53
P a g e | 44
Histogram:
Frequency Polygon:
P a g e | 45
P a g e | 46
b. Ogive:
2.28 Cumulative
P a g e | 47
86
P a g e | 48
Histogram:
30
25
20
15
Frequency
10
0
90000 110000 130000 150000 170000 190000
Class Midpoints
Frequency Polygon:
P a g e | 49
30
25
20
15
Frequency
10
0
90000 110000 130000 150000 170000 190000
Class Midpoints
Ogive:
P a g e | 50
100
90
80
70
60
50
Frequencies
40
30
20
10
0
80000 100000 120000 140000 160000 180000 200000
Class Endpoints
$ 0 - under $100 3 3
57
P a g e | 51
P a g e | 52
Histogram:
Frequency Polygon:
P a g e | 53
P a g e | 54
Ogive:
2.30 Cumulative
87
P a g e | 56
Histogram:
Frequency Polygon:
P a g e | 57
Ogive:
P a g e | 58
Pie Chart:
P a g e | 59
Classical; 3% Latin; 3%
Metal ; 5%
R&B; 29%
Soundtrack; 11%
Country; 13%
Alternative; 21%
Rap; 15%
P a g e | 60
2.32
700
600
500
400
M anufactured Goods
300
200
100
0
0 5 10 15 20 25 30 35
Agricultural Products
P a g e | 61
2.33
Transportation
Electrical
Pie Chart:
P a g e | 62
Petro.; 3% Elec.; 1%
Fab. Metals; 4%
Food; 4%
Trans. Equip.; 5%
Chem.; 37%
Plas. & Rubber; 5%
Paper; 11%
2.34
500 100
400 80
Percent
300 60
Count
200 40
100 20
0 0
C1 Fault in plastic Thickness Broken handle Labeling Discoloration
Count 221 117 86 44 32
Percent 44.2 23.4 17.2 8.8 6.4
Cum % 44.2 67.6 84.8 93.6 100.0
44 20, 40, 59
P a g e | 64
45 12
46 53, 54
47 30, 34, 58
49 63
50 48, 49, 90
51 66
53 38, 66, 66
54 31, 78
55 56
56 69
57 37, 50
59 19, 23
22 00, 68
25 24, 55
27 42, 60, 64
28 14, 30
30 02, 10
2.38 Family practice is most prevalent with about 20% with pediatrics next at
slightly
2.39 The fewest number of audits is 12 and the most is 42. More companies
(8)
2.40 There were relatively constant sales from January through August ($4 to
6 million).
P a g e | 66
Chapter 3
Descriptive Statistics
LEARNING OBJECTIVES
It can be emphasized in this chapter that there are at least two major
dimensions along which data can be described. One is the measure of
central tendency with which statisticians attempt to describe the more
central portions of the data. Included here are the mean, median, mode,
percentiles, and quartiles. It is important to establish that the median is a
useful device for reporting some business data, such as income and housing
costs, because it tends to ignore the extremes. On the other hand, the mean
utilizes every number of a data set in its computation. This makes the mean
an attractive tool in statistical analysis.
CHAPTER OUTLINE
Mode
Median
Mean
Percentiles
Quartiles
Range
Interquartile Range
Variance
Standard Deviation
Empirical Rule
Chebyshev’s Theorem
Z Scores
Coefficient of Variation
P a g e | 70
Mean
Mode
Measures of Variability
Skewness
Coefficient of Skewness
Kurtosis
Correlation
KEY TERMS
Correlation Platykurtic
3.1 Mode
2, 2, 3, 3, 4, 4, 4, 4, 5, 6, 7, 8, 8, 8, 9
The mode = 4
P a g e | 72
2, 2, 3, 3, 4, 4, 4, 4, 5, 6, 7, 8, 8, 8, 9
Since there are an odd number of terms, the median is the middle
number.
The median = 4
n 1 151
2 2
th
at the term = = 8th term
3.3 Median
073, 167, 199, 213, 243, 345, 444, 524, 609, 682
Since there are an even number of terms, the median is the average of
the
(2 4 3 3 4 5) 5 8 8
2 2
Median = = 294
P a g e | 73
n 1
2
th
Using the formula, the median is located at the term
10 1 11
2 2
n = 10 therefore = 5.5th term.
3.4 Mean
17.3
31.6
40.0
x
52.8 = x/n = (333.6)/8 = 41.7
38.8
30.1
3.5 Mean
-2
P a g e | 74
5 µ = x/N = -12/12 = -1
-3
x
-6 = x/n = -12/12 = -1
-7
-4
-5
11, 13, 16, 17, 18, 19, 20, 25, 27, 28, 29, 30, 32, 33, 34
35
i (15) 5.25
100
55
i (15) 8.25
100
25
i (15) 3.75
100
Q1 = P25 but
th
15 1
8th term
2
Q2 = Median but: The median is located at the
Q2 = 25
75
i (15) 11.25
100
Q3 = P75 but
80, 94, 97, 105, 107, 112, 116, 116, 118, 119, 120, 127,
128, 138, 138, 139, 142, 143, 144, 145, 150, 162, 171, 172
n = 24
20
i ( 24) 4.8
100
For P20:
47
i (24) 11.28
100
For P47:
83
i (24) 19.92
100
For P83:
Q1 = P25
P a g e | 77
25
i (24) 6
100
For P25:
Q2 = Median
th
24 1
12.5th term
2
The median is located at the:
Q3 = P75
75
i (24) 18
100
For P75:
x 18,245 1216.33
N 15
3.8 Mean = The mean is 1216.33.
th
15 1
2
The median is located at the = 8th term
Median = 1,233
P a g e | 78
Q2 = Median = 1,233
63
i (15) 9.45
100
For P63,
29
i (15) 4.35
100
For P29,
th
12 1
6.5 th
2
3.9 The median is located at the position
75
i (12) 9
100
For Q3 = P75:
20
i (12) 2.4
100
For P20:
60
i (12) 7.2
100
For P60:
80
i (12) 9.6
100
For P80:
93
i (12) 11.16
100
For P93:
x 61 3.588
N 17
3.10 n = 17; Mean = The mean is 3.588
P a g e | 80
th
17 1
2
The median is located at the = 9th term, Median = 4
75
(17) 12.75
100
Q3 = P75: i=
11
(17) 1.87
100
P11: i=
35
(17) 5.95
100
P35: i=
58
(17) 9.86
100
P58: i=
67
(17) 11.39
100
P67: i=
4 0.2857 0.0816
9 4.7143 22.2246
1 3.2857 10.7958
3 1.2857 1.6530
5 0.7143 0.5102
x 30
4.2857
N 7
a.) Range = 9 - 1 = 8
x 14.2857
N 7
b.) M.A.D. = 2.0408
( x ) 2 43.4284
N 7
c.) 2 = = 6.2041
P a g e | 82
( x ) 2
6.2041
N
d.) = = 2.4908
25
(7)
100
Q1 = P25 i = = 1.75
75
(7)
100
Q3 = P75: i = = 5.25
IQR = Q3 - Q1 = 6 - 2 = 4
6 4.2857
2.4908
f.) z = = 0.69
2 4.2857
2.4908
z = = -0.92
P a g e | 83
4 4.2857
2.4908
z = = -0.11
9 4.2857
2.4908
z = = 1.89
1 4.2857
2.4908
z = = -1.32
3 4.2857
2.4908
z = = -0.52
5 4.2857
2.4908
z = = 0.29
xx ( x x) 2
3.12 x
4 0 0
3 1 1
0 4 16
5 1 1
2 2 4
9 5 25
4 0 0
P a g e | 84
5 1 1
xx ( x x ) 2
x = 32 = 14 = 48
x 32
x
n 8
=4
a) Range = 9 - 0 = 9
xx 14
n 8
b) M.A.D. = = 1.75
( x x) 2 48
n 1 7
c) s2 = = 6.8571
( x x) 2
6.857
n 1
d) s = = 2.6186
e) Numbers in order: 0, 2, 3, 4, 4, 5, 5, 9
25
(8)
100
Q1 = P25 i = = 2
75
(8)
100
Q3 = P75 i = = 6
3.13 a.)
x (x-µ) ( x - µ) 2
23 1.833 3.360
19 -2.167 4.696
26 4.833 23.358
24 2.833 8.026
23 1.833 3.360
x 127
N 6
= = 21.167
( x ) 2 126.834
21.139
N 6
= = 4.598 ORIGINAL
FORMULA
P a g e | 86
b.)
x x2
12 144
23 529
19 361
26 676
24 576
23 529
=
(x ) 2 (127) 2
x 2 2815
N 6 2815 2688.17 126.83
21.138
N 6 6 6
3.14 s2 = 433.9267
s = 20.8309
x = 1387
x2 = 87,365
n = 25
x
= 55.48
3.15 2 = 58,631.295
= 242.139
x = 6886
x2 = 3,901,664
n = 16
µ = 430.375
3.16
14, 15, 18, 19, 23, 24, 25, 27, 35, 37, 38, 39, 39, 40, 44,
25
(25)
100
Q1 = P25 i = = 6.25
75
(25)
100
Q3 = P75 i = = 18.75
IQR = Q3 - Q1 = 59 - 25 = 34
1 1 3
1 2
1 .75
2 4 4
3.17 a) .75
1 1
1 2
1 .84
2 .5 6.25
b) .84
1 1
1 2
1 .609
1 .6 2.56
c) .609
P a g e | 89
1 1
1 2
1 .902
3 .2 10.24
d) .902
P a g e | 90
3.18
Set 1:
x 262
1 65.5
N 4
(x ) 2 (262) 2
x 2 17,970
1 N 4
N 4
= 14.2215
Set 2:
x 570
2 142.5
N 4
(x ) 2 (570) 2
x 2 82,070
2 N 4
N 4
= 14.5344
14.2215
(100)
65.5
CV1 = = 21.71%
P a g e | 91
14.5344
(100)
142.5
CV2 = = 10.20%
P a g e | 92
x xx ( x x) 2
3.19
7 1.833 3.361
5 3.833 14.694
10 1.167 1.361
12 3.167 10.028
9 0.167 0.028
8 0.833 0.694
14 5.167 26.694
3 5.833 34.028
11 2.167 4.694
13 4.167 17.361
8 0.833 0.694
6 2.833 8.028
106 32.000
121.665
x 106
x
n 12
= 8.833
xx 32
n 12
a) MAD = = 2.667
( x x) 2 121.665
n 1 11
b) s2 = = 11.06
P a g e | 93
s 2 11.06
c) s = = 3.326
Q1 = P25: i = (.25)(12) = 3
Q3 = P75: i = (.75)(12) = 9
6 8.833
3.326
e.) z = = - 0.85
(3.326)(100)
8.833
f.) CV = = 37.65%
3.20 n = 11 x x-
768 475.64
429 136.64
323 30.64
306 13.64
P a g e | 94
286 6.36
262 30.36
215 77.36
172 120.36
162 130.36
148 144.36
145 147.36
x 1313.08
N 11
b.) MAD = = 119.37
(x ) 2 (3216) 2
x 2
1,267,252
N 11
N 11
c.) 2 = = 29,728.23
29,728.23
d.) = = 172.42
x 172 292.36
172.42
z = = -0.70
172.42
(100) (100)
292.36
g.) CV = = 58.98%
3.21 µ = 125 = 12
3.22 µ = 38 =6
x1 - µ = 50 - 38 = 12
x2 - µ = 26 - 38 = -12
x1 12
6
= 2
x 2 12
6
= -2
1 1 1 3
1 2
1 2 1
k 2 4 4
= .75
P a g e | 97
x1 - µ = 62 - 38 = 24
x2 - µ = 14 - 38 = -24
x1 24
6
= 4
x 2 24
6
= -4
k=4
1 1 1 15
1 2
1 2 1
k 4 16 16
= .9375
1
k2
1- =.89
P a g e | 99
1
k2
.11 =
.11 k2 = 1
1
.11
k2 =
k2 = 9.09
k = 3.015
1
k2
3.23 1- = .80
1
k2
1 - .80 =
P a g e | 100
1
k2
.20 = and .20k2 = 1
k2 = 5 and k = 2.236
3.24 µ = 43. 68% of the values lie µ + 1 . Thus, between the mean, 43,
and one of
1 = 46 - 43 = 3
within 99.7% of the values lie µ + 3. Thus, between the mean, 43,
and one of
3 = 51 - 43 = 8
= 2.67
1
k2
1- = .77
Solving for k:
1
k2
.23 = and therefore, .23k2 = 1
k2 = 4.3478
k = 2.085
2.085 = 4
4
2.085
= = 1.918
3.25 µ = 29 =4
x1 21 29 8
4 4
= -2 Standard Deviations
x 2 37 29 8
8 4
= 2 Standard Deviations
P a g e | 102
Exceed 37 days:
Since 95% fall between 21 and 37 days, 5% fall outside this range.
Since the
Exceed 41 days:
x 41 29 12
4 4
= 3 Standard deviations
The empirical rule states that 99.7% of the values fall within µ ± 3 =
29 ± 3(4) =
29 ± 12. That is, 99.7% of the values will fall between 17 and 41 days.
0.3% will fall outside this range and half of this or .15% will lie
above 41.
P a g e | 103
x 25 29 4
4 4
= -1 Standard Deviation
29 ± 1(4) = 29 ± 4
Therefore, between 25 and 33 days, 68% of the values lie and 32% lie
outside this
3.26 x
97
109
111
118
120
130
132
133
137
137
x
x = 1224 x = 151,486
2
n = 10 = 122.4 s = 13.615
P a g e | 104
Bordeaux: x = 137
137 122.4
13.615
z = = 1.07
Montreal: x = 130
130 122.4
13.615
z = = 0.56
Edmonton: x = 111
111 122.4
13.615
z = = -0.84
Hamilton: x = 97
97 122.4
13.615
z = = -1.87
P a g e | 105
3.27 Mean
Class f M fM
0- 2 39 1 39
2- 4 27 3 81
4- 6 16 5 80
6- 8 15 7 105
8 - 10 10 9 90
10 - 12 8 11 88
12 - 14 6 13 78
f=121 fM=561
fM 561
f 121
= = 4.64
3.28
Class f M fM
f=850 fM=1902
fM 1902
f 850
Mean: = = 2.24
3.29 Class f M fM
20-30 7 25 175
30-40 11 35 385
40-50 18 45 810
50-60 13 55 715
60-70 6 65 390
70-80 4 75 300
Total 59 2775
fM 2775
f 59
= = 47.034
M-µ ( M - µ) 2 f ( M - µ) 2
f ( M ) 2 10,955.93
f 59
2 = = 185.694
185.694
= = 13.627
P a g e | 108
5- 9 20 7 140 980
9 - 13 18 11 198 2,178
13 - 17 8 15 120 1,800
17 - 21 6 19 114 2,166
21 - 25 2 23 46 1,058
(fM ) 2 (618) 2
fM 2 8182
n 54 8182 7071.67
n 1 53 53
s2 = = 20.931
s 2 20.9
s = = 4.575
P a g e | 109
18 - 24 17 21 357
7,497
24 - 30 22 27 594
16,038
30 - 36 26 33 858
28,314
36 - 42 35 39 1,365
53,235
42 - 48 33 45 1,485
66,825
48 - 54 30 51 1,530
78,030
54 - 60 32 57 1,824
103,968
60 - 66 21 63 1,323
83,349
66 - 72 15 69 1,035
71,415
b.) Mode. The Modal Class = 36-42. The mode is the class midpoint =
39
P a g e | 110
(fM ) 2 (10,371) 2
fM
2
508,671
n 231 43,053.5065
n 1 230 230
c.) s2 = = 187.189
187.2
d.) s = = 13.682
P a g e | 111
3.32
Class f M fM fM2
a.) Mean
fM 258.5
f 137
= = 1.887
c.) Variance:
(fM ) 2 ( 258.5) 2
fM 2 682.25
N 137
N 137
2 = = 1.4197
2 1.4197
= = 1.1915
P a g e | 112
P a g e | 113
3.33 f M fM fM2
40-50 1 45 45 2025
50-60 0 55 0 0
70-80 1 75 75 5625
a.) Mean:
fM 760
f 20
= = 38
c.) Variance:
(fM ) 2 (760) 2
fM 2 33,900
N 20
N 20
2 = = 251
2 251
= = 15.843
P a g e | 115
fM 1,970,000
f 49
= = 40,204
The actual mean for the ungrouped data is 37,816. This computed
group
might be less than the class midpoint since the actual mean is less
than the
(fm) 2 (1,970,000) 2
fM
2
1.185 x10
11
N 49
N 49
2 = = 801,999,167
P a g e | 116
= 28,319.59
The actual standard deviation was 29,341. The difference again is due
to
the grouping of the data and the use of class midpoints to represent
the
data. The class midpoints due not accurately reflect the raw data.
median = $33
mode = $21
The stock prices are skewed to the right. While many of the stock
prices
are at the cheaper end, a few extreme prices at the higher end pull the
mean.
P a g e | 117
3.36 mean = 51
median = 54
mode = 59
The distribution is skewed to the left. More people are older but the
most
3( M d ) 3(5.51 3.19)
9.59
3.37 Sk = = 0.726
3.38 n = 25 x = 600
x
= 24 s = 6.6521 Md = 23
3( x M d ) 3(24 23)
s 6.6521
Sk = = 0.451
3.40 n = 18
th
(n 1)th (18 1)th 19
2 2 2
Median: = 9.5th term
Median = 74
Q1 = P25:
25
(18)
100
i = = 4.5
Q1 = 5th term = 66
Q3 = P75:
75
(18)
100
i = = 13.5
Q3 = 14th term = 90
Therefore, IQR = Q3 - Q1 = 90 - 66 = 24
There are no extreme outliers. The only mild outlier is 21. The
x y
xy n
( x ) 2
( y ) 2
x y
2 2
n n
r = =
(80)(69)
624
7
(80)
2
(69) 2 164.571
1,148 815
7 7 ( 233.714)(134.857)
r = = =
164.571
177.533
r = = -0.927
x y
xy n
( x ) 2 ( y ) 2
x y2
2
n n
r = =
(1,087)( 2,032)
507,509
5
(1,087)
2
(2,032) 2
322,345 878, 686
5 5
r = =
65,752.2 65,752.2
(86,031.2)(52,881.2) 67,449.5
r = = = .975
P a g e | 123
47.6 15.1
46.3 15.4
50.6 15.9
52.6 15.6
52.4 16.4
52.7 18.1
x y
xy n
( x ) 2 ( y ) 2
x y2
2
n n
r = =
(302.2)(96.5)
4,870.11
6
(302.2) 2 (96.5) 2
15, 259. 62 1,557 .91
6 6
r = = .6445
x y
xy n
( x ) 2
( y ) 2
x y
2 2
n n
r = =
(6,087)(1,050)
1,130,483
9
(6,087)
2
(1,050) 2
6, 796,149 194 ,526
9 9
r = =
420,333 420,333
(2,679,308)(72,026) 439,294.705
r = = = .957
x y
xy n
( x ) 2
( y ) 2
x y
2 2
n n
r = =
P a g e | 125
(17.09)(15.12)
48.97
8
(17.09)
2
(15.12) 2
58. 7911 41 . 7054
8 8
r = =
16.6699 16.6699
(22.28259)(13.1286) 17.1038
r = = = .975
x y
xy n
( x ) 2 ( y ) 2
x y2
2
n n
r = =
(15.12)(15.86)
41.5934
8
(15.12)
2
(15.86) 2
41. 7054 42 . 0396
8 8
r = =
P a g e | 126
11.618 11.618
(13.1286)(10.59715) 11.795
r = = = .985
x y
xy n
( x ) 2 ( y ) 2
x y2
2
n n
r = =
(17.09)(15.86)
48.5827
8
(17.09) 2 (15.86) 2
58.7911 42.0396
8 8
r =
14.702 14.702
( 22.2826)(10.5972) 15.367
r = = = .957
P a g e | 127
1, 1, 1, 1, 1, 1, 1, 1, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2, 2,
3, 3, 3, 3, 3, 3, 4, 4, 5, 6, 8
x 75
x
n 30
Mean: = 2.5
th
n 1 30 1 31
2 2 2
The median is located at = 15.5th position.
Range = 8 - 1 = 7
25
(30)
100
Q1 = P25: i = = 7.5
P a g e | 129
75
(30)
100
Q3 = P75: i = = 22.5
IQR = Q3 - Q1 = 3 - 1 = 2
P a g e | 130
10
( 40)
100
3.47 P10: i = = 4
80
( 40)
100
P80: i = = 32
25
( 40)
100
Q1 = P25: i = = 10
75
( 40)
100
Q3 = P75: i = = 30
Range = 81 - 19 = 62
P a g e | 132
x 126,904
N 20
3.48 µ = = 6345.2
n 1
2
The median is located at the th value = 21/2 = 10.5th value
P30: i = (.30)(20) = 6
P60: i = (.60)(20) = 12
P90: i = (.90)(20) = 18
Q1 = P25: i = (.25)(20) = 5
Q1 = (4464+4507)/2 = 4485.5
Q3 = P75: i = (.75)(20) = 15
Q3 = (6796+8687)/2 = 7741.5
( x ) 2 (87.95) 2
x 2 1130.9027
N 10
N 10
= = 5.978
P a g e | 134
x
N
3.50 a.) µ= = 26,675/11 = 2425
Median = 1965
c.) Variance:
(x ) 2 (26,675) 2
x 2 86,942,873
N 11
N 11
2 = = 2,023,272.55
Standard Deviation:
2 2,023,272.55
= = 1422.42
d.) Texaco:
x 1532 2425
1422.42
z = = -0.63
P a g e | 135
Exxon Mobil:
x 6300 2425
1422.42
z = = 2.72
e.) Skewness:
3( M d ) 3(2425 1965)
1422.42
Sk = = 0.97
x 32.95
n 14
3.51 a.) Mean: = 2.3536
1.79 2.07
2
Median: = 1.93
Mode: No Mode
1
(14) 3.5
4
Q1: Located at the 4th term. Q1 = 1.68
3
(14) 10.5
4
Q3: Located at the 11th term. Q3 = 2.87
P a g e | 136
xx x x 2
= 11.6972 = 13.8957
xx
11.6972
n 14
MAD = = 0.8355
x x
2
13.8957
n 1 13
s2 = = 1.0689
s 2 1.0689
s= = 1.0339
3( x M d ) 3(2.3536 1.93)
s 1.0339
Sk = = 1.229
There is one mild outlier. The 4.73 recorded for Arizona is outside
the upper
inner fence.
Mineral Production
1 2 3 4 5
Mineral Production
P a g e | 138
3.52 f M fM
2
fM
a.) Mean:
fM 5680
f 170
= = 33.412
Mode: The Modal Class is 30-35. The class midpoint is the mode =
32.5.
b.) Variance:
P a g e | 139
(fM ) 2 (5680) 2
fM 2
199,662.5
n 170
n 1 169
s2 = = 58.483
Standard Deviation:
s 2 58.483
s = = 7.647
P a g e | 140
0 - 20 32 10 320 3,200
20 - 40 16 30 480 14,400
40 - 60 13 50 650 32,500
60 - 80 10 70 700 49,000
Mode: The Modal Class is 0-20. The midpoint of this class is the
mode = 10.
(fM ) 2 (3860) 2
fM 2 253,000
n 90 253,000 165,551.1
n 1 89 89
87,448.9
982.572
89
s =
= 31.346
P a g e | 141
y = 44 y2 = 300
xy = 188 n =7
x y
xy n 188
(36)( 44)
7
( x )
2
( y ) 2 (36)
2
(44)
x y2
2
256 300
n n 7 7
r = =
38.2857 38.2857
(70.85714)( 23.42857) 40.7441
r = = = -.940
P a g e | 142
x 3.45
(100%) (100%)
x 32
3.55 CVx = = 10.78%
y 5.40
(100%) (100%)
y 84
CVY = = 6.43%
3.56 µ = 7.5 Each of the numbers, 1 and 14, are 6.5 units away from
the mean.
3.57 µ = 419, = 27
Each of the points, 359 and 479 is a distance of 60 from the mean,
µ = 419.
400 419
27
c.) Since x = 400, z = = -0.704. This worker is in the lower
half of
3.58 a.) x x2
x 41,100
N 4
= = 10,275
(x ) 2 (41,100) 2
x 2 498,930,000
N 4
N 4
= = 4376.86
b.) x x2
x=41,100 x2=467,550,000
x 41,100
N 4
= = 10,275
(x ) 2 (41,100) 2
x 2 467,550,000
N 4
N 4
= = 3363.31
c.)
P a g e | 145
1 4376.86
(100) (100)
1 10,275
CV1 = = 42.60%
2 3363.31
(100) (100)
2 10,275
CV2 = = 32.73%
Median $31,369
Mode $29,500
Since these three measures are not equal, the distribution is skewed.
The
xy = 314.9091 n =8
x y
xy n
( x ) 2
( y ) 2
x y
2 2
n n
r = =
(36.62)(57.23)
314.9091
8
(36.62)
2
(57.23) 2
217.137 479.3231
8 8
r = =
P a g e | 147
52.938775
(49.50895)(69.91399)
r = = .90
25
(20)
100
3.61 a.) Q1 = P25: i = = 5
75
(20)
100
Q3 = P75: i = = 15
th th
n 1 20 1
2 2
Median: = 10.5th term
Inner Fences:
Outer Fences:
P a g e | 149
b.) and c.) There are no outliers in the lower end. There are two
extreme
190.9). There is one mild outlier at the upper end (New York,
145.9).
1.459 = 32
= 21.93
2.425 = 44
= 18.14
Chapter 4
Probability
LEARNING OBJECTIVES
The main objective of Chapter 4 is to help you understand the basic principles of
probability, specifically enabling you to
CHAPTER OUTLINE
Subjective Probability
Experiment
Event
Elementary Events
Sample Space
Independent Events
Complimentary Events
Probability Matrices
Complement of a Union
Independent Events
KEY TERMS
A Priori Intersection
Experiment Union
4.1 Enumeration of the six parts: D1, D2, D3, A4, A5, A6
D = Defective part
A = Acceptable part
Sample Space:
D1 D2, D2 D3, D3 A5
D1 D3, D2 A4, D3 A6
D1 A4, D2 A5, A4 A5
D1 A5, D2 A6, A4 A6
D1 A6, D3 A4, A5 A6
two is:
9/15 = .60
a) X Z = {1, 2, 3, 4, 5, 7, 8, 9}
b) X Y = {7, 9}
c) X Z = {1, 3, 7}
d) X Y Z = {1, 2, 3, 4, 5, 7, 8, 9}
e) X Y Z= {7}
h) X or Y = X Y = {1, 2, 3, 4, 5, 7, 8, 9}
i) Y and Z = Y Z = {2, 4, 7}
4.3 If A = {2, 6, 12, 24} and the population is the positive even numbers
through 30,
A’ = {4, 8, 10, 14, 16, 18, 20, 22, 26, 28, 30}
4.5 Enumeration of the six parts: D1, D2, A1, A2, A3, A4
D = Defective part
A = Acceptable part
Sample Space:
P a g e | 158
A1 A3 A4, A2 A3 A4
6!
3!3!
6 C3 = = 20
1 defective part.
20!
6!14!
4.7 C6 =
20 = 38,760
4.8 P(A) = .10, P(B) = .12, P(C) = .21, P(A C) = .05 P(B C) = .03
4.9
D E F
A 5 8 12 25
B 10 6 4 20
C 8 2 5 15
23 16 21 60
4.10
E F
who have ridden both or determine if these two events are mutually exclusive.
union.
P a g e | 163
d)
Faculty
References
Y
N
4.15
C D E F
A 5 11 16 8 40
B 2 3 5 7 17
7 14 21 15 57
c) P(D E) = .0000
d) P(A B) = .0000
4.16
D E F
a) P(E B) = .09
b) P(C F) = .06
c) P(E D) = .00
a) (without replacement)
P a g e | 165
6 5 30
50 49 2450
P(D1 D2) = P(D1) P(D2 D1) = = .0122
b) (with replacement)
6 6 36
50 50 2500
P(D1 D2) = P(D1) P(D2) = = .0144
c)
Yes No
Yes .15
I No .85
P a g e | 166
.78 .22
or .11 of the total (.78) people in urban areas. (.11)(.78) = .0858 which
belongs in
the “YES-YES" cell. The answer to b) is found in the Yes for U and no for
I cell.
It can be determined by taking the marginal, .78, less the answer for
a), .0858.
d. P(NU I) is found in the no for U column and the yes for I row (1st row and
2nd column). Take the marginal, .15, minus the yes-yes cell, .0858, to get
.0642.
P a g e | 167
The matrix:
Yes No
The matrix:
P NP
4.21
The matrix:
Yes No
The matrix:
Yes No
4.23
E F G
A 15 12 8 35
B 11 17 19 47
C 21 32 27 80
D 18 13 12 43
65 74 66 205
P a g e | 171
d) P(E G) = .0000
P a g e | 172
4.24
C D
c) P(A B) = .0000
4.25
Calculator
Yes No
Yes 46 3 49
No 11 15 26
Computer
57 18 75
46 49
57 75
?
.8070 .6533
Since this is one example that the conditional does not equal
the marginal in
calculator.
P a g e | 174
1258
P (C S ) 83,384
P(S ) 8010
83,384
c) P(C S) = = .15705
1258
P (C S ) 83,384
P (C ) 10,867
83,384
d) P(S C) = = .11576
P ( NS NC ) 1 P (C S )
P ( NC ) P ( NC )
e) P(NS NC) =
P ( NS C ) P(C ) P (C S )
P (C ) P (C )
f) P(NS C) =
The matrix:
Yes No
Could have been worked as: P(not TD EM) = P(EM) P(not TD EM)
=
(.54)(.80) = .4320
The matrix:
TD
Yes No
No .46
EM
1.00
P a g e | 179
The matrix:
Yes No
4.30 Let R = agreed or strongly agreed that lack of role models was a
barrier
The matrix:
Yes No
D = defective product
P(D Ei)
P(D)=.093
a) P(B) = .45
c)
P(I Ei)
P(I)=.1265
d)
P(K Ei)
P(K)=.8735
P a g e | 186
P(V Ei)
P(V)=.272
P(T Ei)
P(T)=.3720
4.35
Variable 1
D E
A 10 20 30
Variable 2 B 15 5 20
P a g e | 188
C 30 15 45
55 40 95
50/95 = .52632
P(A B) .0 0 0 0
P (B ) 20 / 95
h) P(A B) = = .0000 (A and B are mutually
exclusive)
D E F G 4.36
A 3 9 7 12 31
B 8 4 6 4 22
C 10 5 3 7 25
21 18 16 23 78
P (A B) .0 0 0 0
P (B ) 22 / 78
b) P(A B) = = .0000 (A and B are mutually
exclusive)
4.37
Age(years)
a) P(35-44) = .28
The matrix:
Yes No
The matrix:
Yes No
{(.20)(.83)}/(.84) = .1976
.25 .40
P(N) = .51
S N
Reduce
Yes No
B NB
C
P a g e | 203
Y N
Q N . . .65
39375 25625
. . 1.00
65625 34375
P a g e | 204
I = inappropriate
C = customer
P = payment dispute
S = specialty
R = prescription drugs
c) P(P) = ??
Y N
To solve for P(P), using what is given, since we know that 90%
of P lies
.90P = .36
Y N
Y . .38 .38
0000
M
N .21 .41 .62
4.49 Let F = Flexible Work Let V = Gives time off for Volunteerism
Solve for P(NF NV) = P(NV) – P(F NV) = .695 - .164 = .531
V
P a g e | 209
Y N
P(B Ei)
P(B) = .0623
(1 - .4370)/.65 = .8662
Y N
4.52
P(RB Ei)
P(RB) = .
25066
Chapter 5
Discrete Distributions
LEARNING OBJECTIVES
long-run average that can be appropriately adjusted for various intervals. For
example, if a store is averaging customers in 5 minutes, then it will also be
averaging 2 customers in 10 minutes. On the other hand, x is a one-time
observation and just because x customers arrive in 5 minutes does not mean
that 2x customers will arrive in 10 minutes.
In this text (as in most) because of the number of variables used in its
computation, only exact probabilities are determined for hypergeometric
distribution. This, combined with the fact that there are no hypergeometric
tables given in the text, makes it cumbersome to determine cumulative
probabilities for the hypergeometric distribution. Thus, the hypergeometric
distribution can be presented as a fall-back position to be used only when the
binomial distribution should not be applied because of the non independence
of trials and size of sample.
P a g e | 217
CHAPTER OUTLINE
Probabilities
KEY TERMS
1.836444
= = 1.355155
0 .103
.000 7.573504
0.780071
1 .118 .118 3.069504
0.362201
2.752496
= = 1.6591
1.278064
= = 1.1305
P a g e | 222
.96260
= = .98112
c) n = 10 p = .60 q = .40
+10C10(.60)10(.40)0 =
d) n = 12 p = .45 q = .55
a) n = 20 p = .50
P(x=12) = .120
b) n = 20 p = .30
c) n = 20 p = .70
d) n = 20 p = .90
e) n = 15 p = .40
f) n = 10 p = .60
5.7 a) n = 20
p = .70
q = .30
µ = np = 20(.70) = 14
n p q 20(.70)(.30) 4.2
= = 2.05
b) n = 70 p = .35 q = .65
n p q 70(.35)(.65) 15.925
= = 3.99
µ = np = 100(.50) = 50
n p q 100(.50)(. 50) 25
= = 5
0 .001
P a g e | 227
1 .010
2 .060
3 .185
4 .324
5 .303
6 .118
b) n = 20 p = .50 x Prob
0 .000
1 .000
2 .000
3 .001
4 .005
5 .015
6 .037
7 .074
8 .120
9 .160
P a g e | 228
10 .176
11 .160
12 .120
13 .074
14 .037
15 .015
16 .005
17 .001
18 .000
19 .000
20 .000
P a g e | 229
0 .000
1 .000
2 .001
3 .009
4 .046
5 .147
6 .294
7 .336
8 .168
5.9 a) n = 20 p = .78 x = 14
b) n = 20 p = .75 x = 20
c) n = 20 p = .70 x < 12
5.10 n = 16 p = .40
x Prob
9 .084
10 .039
11 .014
12 .004
13 .001
.142
x Prob
3 .047
4 .101
5 .162
6 .198
.508
n = 13 p = .88
5.11 n = 25 p = .60
a) x > 15
x Prob
15 .161
16 .151
17 .120
18 .080
19 .044
20 .020
21 .007
22 .002
.585
b) x > 20
P(x > 20) = P(x = 21) + P(x = 22) + P(x = 23) + P(x = 24) +
P(x = 25) =
x Prob.
9 .009
8 .003
7 .001
<6 .000
.013
P a g e | 235
x Prob.
11 .067
12 .028
13 .009
14 .002
15 .000
16 .000
.106
5.13 n = 15 p = .20
e)
5.14 n = 18
x Prob
P a g e | 237
8 .081
9 .039
10 .015
11 .005
12 .001
.141
c) n = 18 p = .34
C2(.34)2(.66)16 +
18 C3(.34)3(.66)15 +
18 C4(.34)4(.66)14 =
18
d) n = 18 p = .30 x=0
C0(.30)0(.70)18 = .00163
18
n = 18 p = .34 x=0
C0(.34)0(.66)18 = .00056
18
more likely that none of the CPA financial advisors would fall in this
category.
P a g e | 238
2 .3
2 .3 5 e ( 6 4 . 3 6 3 4 3 ) ( .1 0 0 2 5 9 )
5! 120
5.15 a) P(x=5 = 2.3) = = .0538
3 .9
3 .9 2 e ( 1 5 .2 1 ) ( .0 2 0 2 4 2 )
2! 2
b) P(x=2 = 3.9) = = .1539
4 .1
4 .1 3 e ( 6 8 .9 2 1 ) ( .0 1 6 5 7 3 )
3! 6
= .1904
4 .1
4 .1 2 e ( 1 6 .8 1 ) ( .0 1 6 5 7 3 )
2! 2
= .1393
4 .1
4 .1 1 e ( 4 .1 ) ( . 0 1 6 5 7 3 )
1! 1
= .0679
4 .1
4 .1 0 e ( 1 ) ( .0 1 6 5 7 3 )
0! 1
= .0166
d) P(x=0 = 2.7) =
P a g e | 239
2 .7
2 .7 0 e ( 1 ) ( .0 6 7 2 1 )
0! 1
= .0672
e) P(x=1 = 5.4)=
5 .4
5 .4 1 e ( 5 .4 ) ( .0 0 4 5 1 6 6 )
1! 1
= .0244
4 .4 4 .4 4 .4
4 .4 5 e 4 .4 6 e 4 .4 7 e
5! 6! 7!
+ + =
( 1 6 4 9 .1 6 2 2 ) ( . 0 1 2 2 7 7 3 4 ) ( 7 2 5 6 .3 1 3 9 ) ( .0 1 2 2 7 7 3 4 )
120 720
+ +
( 3 1 , 9 2 7 .7 8 1 ) ( .0 1 2 2 7 7 3 4 )
5040
b) P(x>7 = 2.9):
x Prob
8 .0068
9 .0022
10 .0006
P a g e | 240
11 .0002
12 .0000
.0098
x Prob
3 .1852
4 .1944
5 .1633
6 .1143
7 .0686
8 .0360
9 .0168
.7786
x Prob
0 .0550
1 .1596
2 .2314
3 .2237
4 .1622
5 .0940
P a g e | 241
6 .0455
.9714
x Prob
6 .1594
7 .1298
8 .0925
.3817
6.3
5.17 a) = 6.3 mean = 6.3 Standard deviation = = 2.51
x Prob
0 .0018
1 .0116
2 .0364
3 .0765
4 .1205
5 .1519
6 .1595
7 .1435
8 .1130
9 .0791
10 .0498
P a g e | 242
11 .0285
12 .0150
13 .0073
14 .0033
15 .0014
16 .0005
17 .0002
18 .0001
19 .0000
P a g e | 243
1.3
b) = 1.3 mean = 1.3 standard deviation = = 1.14
x Prob
0 .2725
1 .3542
2 .2303
3 .0998
4 .0324
5 .0084
6 .0018
7 .0003
8 .0001
9 .0000
P a g e | 244
x Prob
0 .0001
1 .0012
2 .0054
3 .0160
4 .0357
5 .0635
6 .0941
7 .1197
8 .1332
9 .1317
10 .1172
11 .0948
12 .0703
13 .0481
14 .0306
15 .0182
16 .0101
17 .0053
18 .0026
19 .0012
20 .0005
21 .0002
P a g e | 245
22 .0001
P a g e | 246
0.6
d) = 0.6 mean = 0.6 standard deviation = = .775
x Prob
0 .5488
1 .3293
2 .0988
3 .0198
4 .0030
5 .0004
6 .0000
P a g e | 247
a) P(x=6 = 2.8)
b) P(x=0 = 2.8) =
x Prob.
5 .0872
6 .0407
7 .0163
8 .0057
9 .0018
10 .0005
11 .0001
.1523
5 .1697
6 .1584
7 .1267
8 .0887
9 .0552
10 .0309
11 .0157
12 .0073
13 .0032
14 .0013
15 .0005
16 .0002
17 .0001
.6579
a) P(x = 0) = .0302
= 7.0 10 minutes
e) P(x = 8 15 minutes)
= 10.5 15 minutes
P a g e | 251
a) P(x=0 = 5.6):
b) P(x=6
= 5.6):
x Prob.
15 .0005
16 .0002
17 .0001
.0008
question the Lambda value as too low for this period. Perhaps the
value of
P a g e | 252
a) P(x=0 = 0.6):
b) P(x=1
= 0.6):
2 .0988
3 .0198
4 .0030
5 .0004
6 .0000
.1220
P(x < 3
= 1.8):
0 .1653
1 .2975
2 .2678
3 .1607
.8913
e) P(x=4 6 years):
P(x=4 = 3.6):
a) P(x=0 = 1.2):
b) P(x=2 2 months):
P(x=2
= 0.6):
0 .1653
1 .2975
.4628
P a g e | 256
The result is likely to happen almost half the time (46.26%). Ship
channel and
a) P(x=0 = 1.2):
4 .0260
5 .0062
6 .0012
7 .0002
8 .0000
.0336
P a g e | 258
Since n > 20 and np < 7, the Poisson approximation to this binomial
problem is
close enough.
7 .0595
8 .0298
9 .0132
10 .0053
11 .0019
12 .0006
13 .0002
14 .0001
.1106
11 .0019
P a g e | 259
12 .0006
13 .0002
14 .0001
.0028
c) P(x = 0) = .1653
5.26 If 99% see a doctor, then 1% do not see a doctor. Thus, p = .01 for
this problem.
P a g e | 261
a) P(x = 5):
C 3 3 C1 (56)(3)
8
11 C 4 330
= .5091
P(x = 1) + P (x = 0) =
c) P(x=0 N = 9, A = 2, n = 3)
C 0 7 C 3 (1)(35)
2
9 C3 84
= .4167
(1)(105)
77520
+ 5C6 (impossible) + 5C7(impossible) = .0014
P a g e | 264
5.28 N = 19 n = 6
a) P(x = 1 private) A = 11
C1 8 C 5 (11)(56)
11
19 C 6 27,132
= .0227
b) P(x = 4 private)
C 4 8 C 2 (330)( 28)
11
19 C 6 27,132
= .3406
c) P(x = 6 private)
C 6 8 C 0 (462)(1)
11
19 C 6 27,132
= .0170
d) P(x = 0 private)
C 0 8 C 6 (1)( 28)
11
19 C 6 27,132
= .0010
P a g e | 265
8 C 0 9 C 4 (1)(126)
17 C 4 2380
a) P(x = 0) = = = .0529
8 C 4 9 C 0 (70)(1)
17 C 4 2380
b) P(x = 4) = = = .0294
9 C 2 8 C 2 (36)( 28)
17 C 4 2380
c) P(x = 2 non computer) = = = .4235
16 C 4 4 C1 (1820)( 4)
20 C 5 15504
a) P(x = 4 white) = = = .4696
4 C 4 16 C1 (1)(16)
20 C 5 15504
b) P(x = 4 red) = = = .0010
P a g e | 266
4 C 5 16 C 0
20 C 5
c) P(x = 5 red) = = .0000 because 4C5 is impossible to
determine
The participant cannot draw 5 red beads if there are only 4 to draw
from.
P a g e | 267
5.31 N = 10 n=4
a) A = 3 x = 2
C 2 7 C 2 (3)( 21)
3
C
10 4 210
P(x = 2) = = .30
b) A = 5 x = 0
C 0 5 C 4 (1)(5)
5
10 C 4 210
P(x = 0) = = .0238
c) A = 5 x = 3
C 3 5 C1 (10)(5)
5
C
10 4 210
P(x = 3) = = .2381
C 3 12 C 0 (4)(1)
4
16 C 3 560
b) P(x = 3) = = .0071
4 C 2 12 C1
16 C 3
c) P(x > 2) = P(x=2) + P(x=3) = + .0071 (from part b.)
=
(6)(12)
560
+ .0071 = .1286 + .0071 = .1357
9 C8(.85)8(.15)1 + 9C9(.85)9(.15)0 =
14 C3(.70)3(.30)11 + C2(.70)2(.30)12 +
14
14 C1(.70)1(.30)13 + C0(.70)0(.30)14 =
14
(364)(.3430)(.00000177) + (91)(.49)(.000000531)=
(14)(.70)(.00000016) + (1)(1)(.000000048) =
x Prob.
0 .028
1 .121
2 .233
3 .267
4 .200
.849
x Prob.
12 .063
13 .022
14 .005
15 .000
.090
P a g e | 272
x Prob.
21 .000
22 .000
23 .000
24 .000
25 .000
.000
P a g e | 273
5.36
1 .2 5
( 1 .2 5 4 ) ( e ) ( 2 .4 4 1 4 ) ( .2 8 6 5 )
4! 24
= .0291
6 .3 7 6 .3 7
( 6 .3 7 ) 1 ( e ) ( 6 .3 7 0 ) ( e ) ( 6 .3 7 ) ( .0 0 1 7 ) ( 1 ) ( .0 0 1 7 )
1! 0! 1 1
2 .4 2 .4 2 .4 2 .4 2 .4
( 2 .4 6 ) ( e ) ( 2 .4 7 ) ( e ) ( 2 .4 8 ) ( e ) ( 2 .4 9 ) ( e ) ( 2 .4 10
)(e )
6! 7! 8! 9! 10!
for values x > 11 the probabilities are each .0000 when rounded off
to 4
decimal places.
P a g e | 274
x Prob.
0 .0369
1 .1217
2 .2008
3 .2209
4 .1823
.7626
x Prob.
3 .1890
4 .0992
5 .0417
6 .0146
7 .0044
8 .0011
9 .0003
10 .0001
11 .0000
.3504
P a g e | 275
x Prob.
3 .1852
4 .1944
5 .1633
.5429
P a g e | 276
C 3 1 C1 (10)(1)
5
6 C4 15
5.38 a) P(x = 3 N = 6, n = 4, A = 5) = = .6667
5 C1 5 C 2 5 C 0 5 C 3 (5)(10) (1)(10)
10 C 3 10 C 3 120 120
P(x = 1) + P(x = 0) = + =
P(x > 10): P(x = 10) + P(x = 11) + . . . + P(x = 25) = .012 + .004
+ .001 = .017
P a g e | 277
P(x = 8) = .180
P(x = 0) = .000
P(x > 12) = P(x = 12) + P(x = 13) + . . . + P(x = 20) = .035 + .015 + .
005 + .001 = .056
x=8
x Prob.
5 .0176
6 .0047
7 .0011
8 .0002
.0236
(increased).
P a g e | 279
P a g e | 280
5.41 N = 32 A = 10 n = 12
10 C 3 22 C 9 (120)( 497,420)
32 C12 225,792,840
a) P(x = 3) = = = .2644
10 C 6 22 C 6 (210)(74,613)
32 C12 225,792,840
b) P(x = 6) = = = .0694
10 C 0 22 C12 (1)(646,646)
32 C12 225,792,840
c) P(x = 0) = = = .0029
d) A = 22
x Prob.
0 .2466
1 .3452
2 .2417
3 .1128
.9463
P a g e | 282
x Prob.
8 .233
9 .121
10 .028
.382
P a g e | 283
15C0(1/3)0(2/3)15 = .0023
c) n = 7 p = .53
Probably the 53% figure is too low for this population since the
probability of
5.45 n = 12
p = .20
a) P(x = 5):
P a g e | 285
b) P(x = 0):
x Prob
7 .0005
8 .0001
.0006
P a g e | 286
x Prob.
0 .001
1 .008
2 .041
3 .124
.174
chance. Other reasons for such a low number of walk-ins might be that
she is
5.48 n = 25 p = .20
x Prob.
11 .004
12 .001
13 .000
.005
c) Since such a result would only occur 0.5% of the time by chance, it
is likely
that the analyst's list was not representative of the entire state of
Idaho or the
x Prob.
3 .0198
4 .0030
5 .0004
.0232
Assume one trip is independent of the other. Let F = flat tire and NF =
no flat tire
5.50 N = 25 n=8
4 C1 21 C 7 (4)(116,280)
25 C 8 1,081,575
= = .4300
C 4 15 C 4 (210(1365)
10
25 C 8 1,081,575
= .2650
C 0 20 C8 (1)(125,970)
5
C
25 8 1,081,575
= .1165
C 3 22 C 5 (1)( 26,334)
3
25 C 8 1,081,575
= .0243
C 6 16 C 0 (28)(1)
8
24 C 6 134,596
a) P(x = 6) = = .0002
P a g e | 290
C 0 16 C 6 (1)(8008)
8
24 C 6 134,596
b) P(x = 0) = = .0595
d) A = 16 East Side
C 3 8 C 3 (560)(56)
16
24 C 6 134,596
P(x = 3) = = .2330
n p q 25(.20)(. 80)
µ = 25(.20) = 5 = = 2
13 .0000
The values for x > 12 are so far away from the expected value that
they are very
unlikely to occur.
If this value (x = 14) actually occurred, one would doubt the validity of
the
x Prob.
6 .0241
7 .0083
P a g e | 292
8 .0025
9 .0007
10 .0002
11 .0000
.0358
P(x = 1) = .3293
P(x = 0) = .5488
.8781
P a g e | 293
8 .0002
9 .0000
.0002
2 .2584
3 .1378
4 .0551
5 .0176
6 .0047
.4736
P a g e | 294
= np = (1,000)(.005) = 5
c) P(x = 0) = .0067
P a g e | 295
5.57 N = 34
a) n = 5 x=3 A = 13
C 3 21 C 2 (286)( 210)
13
34 C 5 278,256
= .2158
b) n = 8 x<2 A=5
5 C 0 29 C8 5 C1 29 C 7 5 C 2 29 C 6
34 C 8 34 C 8 34 C 8
+ + =
P a g e | 296
c) n = 5 x=2 A=3
5.58 N = 14 n=4
C 4 4 C 0 (210((1)
10
14 C 4 1001
= .2098
C 4 10 C 0 (1)(1)
4
14 C 4 1001
= .0010
3.840 e 3.84
0!
P(x = 0) = = .0215
P a g e | 298
b) = 7.68 2,000
15(.36)(.64)
= = 1.86
The most likely values are near the mean, 5.4. Note from the printout
that the
most probable values are at x = 5 and x = 6 which are near the mean.
5.61 This printout contains the probabilities for various values of x from zero
to eleven from a Poisson distribution with = 2.78. Note that the
highest probabilities are at x = 2 and
The mean is np = 22(.64) = 14.08 and the standard deviation is:
P a g e | 300
n p q 22(.64)(. 36)
= = 2.25
5.63 This is the graph of a Poisson Distribution with = 1.784. Note the
high
probabilities for values of x > 8 are near to zero because they are so
far away
Chapter 6
Continuous Distributions
LEARNING OBJECTIVES
This chapter contains a section dealing with the solution of binomial distribution
problems by the normal curve. The correction for continuity is emphasized. In this text,
the correction for continuity is always used whenever a binomial distribution problem is
worked by the normal curve. Since this is often a stumbling block for students to
comprehend, the chapter has included a table (Table 6.4) with rules of thumb as to how to
apply the correction for continuity. It should be emphasized, however, that answers for
this type of problem are still only approximations. For this reason and also in an effort to
link chapters 5 & 6, the student is sometimes asked to work binomial problems both by
methods in this chapter and also by using binomial tables (A.2). This also will allow the
student to observe how good the approximation of the normal curve is to binomial
problems.
The exponential distribution can be taught as a continuous distribution,
which can be used in complement with the Poisson distribution of chapter 5
to solve inter-arrival time problems. The student can see that while the
Poisson distribution is discrete because it describes the probabilities of whole
number possibilities per some interval, the exponential distribution describes
the probabilities associated with times that are continuously distributed.
CHAPTER OUTLINE
KEY TERMS
1 1 1
b a 240 200 40
a) f(x) = = .025
a b 200 240
2 2
b) = = 220
b a 240 200 40
12 12 12
= = 11.547
240 230 10
240 200 40
c) P(x > 230) = = .250
220 205 15
240 200 40
d) P(205 < x < 220) = = .375
225 200 25
240 200 40
e) P(x < 225) = = .625
P a g e | 305
6.2 a = 8 b = 21
1 1 1
b a 21 8 13
a) f(x) = = .0769
a b 8 21 29
2 2 2
b) = = 14.5
b a 21 8 13
12 12 12
= = 3.7528
17 10 7
21 8 13
c) P(10 < x < 17) = = .5385
a b 2.80 3.14
2 2
= = 2.97
b a 3.14 2.80
12 12
= = 0.098
3.10 3.00
3.14 2.80
P(3.00 < x < 3.10) = = 0.2941
1 1
b a 12.03 11.97
Height = = 16.667
12.03 12.01
12.03 11.97
P(x > 12.01) = = .3333
12.01 11.98
12.03 11.97
P(11.98 < x < 12.01) = = .5000
P a g e | 307
b a 3800 400
12 12
= = 981.5
1 3800 400
ba 12
Height = = .000294
b) P (z < 0.73):
x 635 604
56.8
z = = 0.55
x 20 48
12
z = = -2.33
x 150 111
33.8
z = = 1.15
x 100 111
33.8
z = = -0.33
x 250 264
10.9
z = = -1.28
x 255 264
10.9
z = = -0.83
x 35 37
4.35
z = = -0.46
P a g e | 312
x 170 156
11.4
z = = 1.23
6.8 µ = 22 =4
x 17 22
4
z = = -1.25
x 13 22
4
z = = -2.25
x 31 22
4
z = = 2.25
x 25 22
4
z = = 0.75
6.9 µ = 60 = 11.35
x 85 60
11.35
z = = 2.20
x 45 60
11.35
z = = -1.32
x 70 60
11.35
z = = 0.88
x 65 60
11.35
z = = 0.44
x 75 60
11.35
z = = 1.32
x 40 60
11.35
z = = -1.76
x 2000 1332
725
z = = 0.92
x 0 1332
725
z = = -1.84
x 100 1332
725
z = = -1.70
x 700 1332
725
z = = -0.87
x 45,000 30,000
9,000
z = = 1.67
x 15,000 30,000
9,000
z = = -1.67
x 50,000 30,000
9,000
z = = 2.22
x 5,000 30,000
9,000
z = = -2.78
x 20,000 30,000
9,000
z = = -1.11
lower half of the distribution and .9082 - .5000 = .4082 lie between
x and µ.
x
Solving for : z =
7,000 30,000
-1.33 =
= 17,293.23
the upper half of the distribution and .7995 - .5000 = .2995 of the
values lie
x
Solving for µ: z =
33,000
9,000
0.84 =
µ = $25,440
Since 50% of the values are greater than the mean, µ = 200, 10%
or .1000 lie
between x and the mean. From Table A.5, the z value associated
with an area
x
z =
x 200
47
-0.25 =
P a g e | 323
x = 188.25
Since x is only less than 17% of the values, 33% (.5000- .1700) or .
3300 lie
between x and the mean. Table A.5 yields a z value of 0.95 for an
area of
x
z =
x 200
47
0.95 =
x = 244.65
Since 22% of the values lie below x, 28% lie between x and the
mean
x
z =
x 200
47
-0.77 =
x = 163.81
x
z =
x 200
47
0.13 =
x = 206.11
P a g e | 325
6.13 = 625. If 73.89% of the values are greater than 1700, then
23.89% or .2389
lie between 1700 and the mean, µ. The z value associated with .
2389 is -0.64
x
z =
1700
625
-0.64 =
µ = 2100
x
z =
P a g e | 326
x 2258
625
0.48 =
x = 2558
6.14 µ = 22 = ??
Since 72.4% of the values are greater than 18.5, then 22.4% lie
between 18.5 and µ. x = 18.5 is below the mean. From table A.5, z =
- 0.59.
18.5 22
-0.59 =
-0.59 = -3.5
3.5
0.59
= = 5.932
.2100 of the area. The z score associated with this area is -0.55.
Solving for µ:
x
z =
20
4
-0.55 =
= 22.20
6.16 µ = 9.7 Since 22.45% are greater than 11.6, x = 11.6 is in the
upper half of the distribution and .2755 (.5000 - .2245) lie between x
and the mean. Table A.5 yields a z = 0.76 for an area of .2755.
Solving for :
x
z =
11.6 9.7
0.76 =
= 2.5
P a g e | 328
µ = np = 30(.70) = 21
n p q 30(.70)(.30)
= = 2.51
n p q 25(.50)(. 50)
= = 2.5
µ = np = 40(.60) = 24
P a g e | 329
n p q 40(.60)(. 40)
= = 3.10
n p q 16(.45)(.55)
= = 1.99
n p q 8(.50)(. 50)
= = 1.414
µ ± 3 = 4 ± 3(1.414) = 4 ± 4.242
n p q 18(.80)(.20)
= = 1.697
c) n = 12 and p = .30
n p q 12(.30)(. 70)
= = 1.587
n p q 14(.50)(. 50)
= = 1.87
µ ± 3 = 7 ± 3(1.87) = 7 ± 5.61
n p q 25(.40)(. 60)
= = 2.449
µ ± 3 = 10 ± 3(2.449) = 10 ± 7.347
7.5 10
2.449
z = = -1.02
8.5 10
2.449
z = = -0.61
n p q 20(.60)(. 40)
= = 2.19
µ ± 3 = 12 ± 3(2.19) = 12 ± 6.57
x 12.5 12
2.19
z = = 0.23
n p q 15(.50)(.50)
= = 1.9365
x 6.5 7.5
1.9365
z = = -0.52
n p q 10(.70)(.30)
=
µ ± 3 = 7 ± 3(1.449) = 7 ± 4.347
n p q 120(.37)(. 63)
= = 5.29
39.5 44.4
5.29
z = = -0.93
n p q 70(.59)(.41)
µ = n(p) = 70(.59) = 41.3 and = = 4.115
34.5 41.3
4.115
z = = -1.65
µ = 300(.53) = 159
n p q 300(.53)(.47)
= = 8.645
which lies between 0 and 300. It is okay to use the normal distribution
as an approximation on parts a) and b).
175.5 159
8.645
z = = 1.91
n p q 300(.60)(.40)
= 300(.60) = 180 = =
8.485
P a g e | 338
199.5 180
8.485
z = = 2.30
n p q 130(.25)(.75)
= = 4.94
36.5 32.5
4.94
z = = 0.81
19.5 32.5
4.94
z = = -2.63
6.24 n = 95
n p q 95(.52)(. 48)
= = 4.87
43.5 49.4
4.87
z = = -1.21
52.5 49.4
4.87
z = = 0.64
56.5 49.4
4.87
z = = 1.46
P a g e | 342
c) Joint Venture:
p = .70, n = 95
n p q 95(.70)(. 30)
= = 4.47
59.5 66.5
4.47
z = = -1.57
54.5 66.5
4.47
z = = -2.68
62.5 66.5
4.47
z = = -0.89
6.25 a) = 0.1
x0 y
0 .1000
1 .0905
2 .0819
3 .0741
4 .0670
5 .0607
6 .0549
7 .0497
8 .0449
9 .0407
10 .0368
b) = 0.3
x0 y
0 .3000
P a g e | 345
1 .2222
2 .1646
3 .1220
4 .0904
5 .0669
6 .0496
7 .0367
8 .0272
9 .0202
c) = 0.8
x0 y
0 .8000
1 .3595
2 .1615
3 .0726
P a g e | 346
4 .0326
5 .0147
6 .0066
7 .0030
8 .0013
9 .0006
d) = 3.0
x0 y
0 3.0000
1 .1494
2 .0074
3 .0004
P a g e | 347
4 .0000
5 .0000
P a g e | 348
6.26 a) = 3.25
1 1
3.25
µ= = 0.31
1 1
3.25
= = 0.31
b) = 0.7
1 1
.007
µ = = 1.43
1 1
.007
= = 1.43
c) = 1.1
1 1
1.1
µ = = 0.91
1 1
1.1
= = 0.91
P a g e | 349
d) = 6.0
1 1
6
= = 0.17
1 1
6
= = 0.17
P a g e | 350
= .9918
6.28 µ = 23 sec.
1
= = .0435 per second
P a g e | 351
b) = .0435/sec
P(x > 3
= 2.61/min) =
6.29 = 2.44/min.
1 1
2.44
d) Expected time = µ = min. = .41 min = 24.6 sec.
1 1
1.12
a) µ = = .89 hr. = 53.4 min.
1 1
3.39
µ= = 0.295
(0.295)(1,000) = 295
6.32 µ = 20 years
1
20
= = .05/year
1 .9512
2 .9048
3 .8607
foundations will last at least 2 years without major repair and only
9.52% will
6.33 = 2/month
1 1
2
Average number of time between rain = µ = month = 15
days
= µ = 15 days
2
30
Change to days: = = .067/day
6.34 a = 6 b = 14
1 1 1
b a 14 6 8
f(x) = = .125
a b 6 14
2 2
= = 10
P a g e | 357
b a 14 6 8
12 12 12
= = 2.309
14 11 3
14 6 8
P(x > 11) = = .375
12 7 5
14 6 8
P(7 < x < 12) = = .625
x 21 25
4
z = = -1.00
x 77 50
9
z = = 3.00
x 47 50
6
z = = -0.50
x 13 23
4
z = = -2.50
x 29 23
4
z = = 1.50
x 105 90
2.86
z = = 5.24
µ = np = 25(.60) = 15
n p q 25(.60)(. 40)
= = 2.45
µ ± 3 = 15 ± 3(2.45) = 15 ± 7.35
x 11.5 15
2.45
z= = -1.43 From Table A.5, area = .4236
x 12.5 15
2.45
z= = -1.02 From Table A.5, area = .3461
n p q 15(.50)(.50)
= = 1.94
5.5 7.5
1.94
z = = -1.03
µ = np = 10(.50) = 5
n p q 10(.50)(. 50)
= = 1.58
µ ± 3 = 5 ± 3(1.58) = 5 ± 4.74
3 .5 5
1.58
z = = -0.95
µ = np = 15(.40) = 6
P a g e | 363
n p q 15(.40)(.60)
= = 1.90
± 3 = 6 ± 3(1.90) = 6 ± 5.7
7.5 6
1.9
z = = 0.79
let x0 = 3
Let x0 = 2
Let x0 = 1
Let x0 = 3
P a g e | 365
P(1 < x < 3) = P(x > 1) - P(x > 3) = .1920 - .0071 = .1849
Let x0 = 2
6.38 µ = 43.4
Solving for :
x
z =
48 43.4
1.175 =
4.6
1.175
= = 3.915
P a g e | 367
µ = 150(.20) = 30
150(.20)(.80)
= = 4.899
50.5 30
4.899
z = = 4.18
µ = 1/ = 1
a) 1 hour interval
b) 10 to 30 minutes
x0 = .5, x0 = 1.5
x0 = 5/20 = .25
80 90.28
8.53
z = = -1.21
95 90.28
8.53
z = = 0.55
83 90.28
8.53
z = = -0.85
87 90.28
8.53
z = = -0.38
6.42 = 83
x
z =
P a g e | 371
2655
83
1.88 =
= 2498.96 million
P a g e | 372
6.43 a = 18 b = 65
50 25 25
65 18 47
P(25 < x < 50) = = .5319
a b 65 18
2 2
= = 41.5
1 1 1
b a 65 18 47
f(x) = = .0213
1 1
1 .8
a) = = .5556(15 sec.) = 8.33 sec.
x0 = 1 min. = 60/15 = 4
6.45 µ = 951 = 96
x 1000 951
96
z = = 0.51
x 900 951
96
z = = -0.53
x 1100 951
96
z = = 1.55
x 825 951
96
z = = -1.31
x 925 951
96
z = = -0.27
x 700 951
96
z = = -2.61
6.46 n = 60 p = .24
n p q 60(.24)(. 76)
= = 3.308
Since 4.476 to 24.324 lies between 0 and 60, the normal distribution
can be used
x 16.5 14.4
3.308
z = = 0.63
x 22.5 14.4
3.308
z = = 2.45
x 50,000 45,970
4,246
z = = 0.95
x 40,000 45,970
4,246
z = = -1.41
x 35,000 45,970
4,246
z = = -2.58
x 39,000 45,970
4,246
z = = -1.64
x 47,000 45,970
4,246
z = = 0.24
6.48 µ = 9 minutes
= 6.67/hour
Let x0 = 5
6.49 = 88 = 6.4
x 70 88
6.4
z = = -2.81
P a g e | 381
x 80 88
6.4
z = = -1.25
x 100 88
6.4
z = = 1.88
x 90 88
6.4
z = = 0.31
µ = 162
n p q 200(.81)(.19)
= = 5.548
150.5 162
5.548
z = = -2.07
154.5 162
5.548
z = = -1.35
158.5 162
5.548
z = = -0.63
143.5 162
5.548
z = = -3.33
P a g e | 384
n p q 150(.75)(.25)
= = 5.3033
x 104.5 112.5
5.3033
z = = -1.51
109.5 112.5
5.3033
z = = -0.57
120.5 112.5
5.3033
z = = 1.51
95.5 112.5
5.3033
z = = -3.21
ab
2
6.52 = = 2.165
a + b = 2(2.165) = 4.33
b = 4.33 - a
1
ba
Height = = 0.862
1 = 0.862b - 0.862a
1 = 3.73246 - 1.724a
1.724a = 2.73246
6.53 µ = 85,200
The 60% can be split into 30% and 30% because the two x values are
equal distance from the mean.
x
z =
94,800 85,200
.84 =
= 11,428.57
n p q 75(.81)(. 19)
1 = = 3.3974
n p q 75(.44)(. 56)
2 = = 4.2988
µ + 3 = 33 + 3(4.299) = 33 + 12.897
20.103 to 45.897 lies between 0 and 75. It is okay to use the normal
66.5 60.75
3.3974
z = = 1.69
22.5 33
4.2988
z = = -2.44
Let x0 = 1
Let x0 = 0.5
P(x < 0.5 month) = 1 - P(x > 0.5 month) = 1 - .7408 = .2592
6.56 n = 50 p = .80
= np = 50(.80) = 40
n p q 50(.80)(. 20)
= = 2.828
P a g e | 391
34.5 40
2.828
z = = -1.94
41.5 40 47.5 40
2.828 2.828
z = = 0.53 z = = 2.65
z.30 = -.84
x
z =
x 2087
175
-.84 =
x = 1940
z.15 = -0.39
x
z =
x 2087
175
-.39 =
x = 2018.75
P a g e | 394
z.35 = 1.04
x
z =
x 2087
175
1.04 =
x = 2269
Let x0 = 1
Let x0 = 2.5
2,200,000 2,106,774
50,940
z = = 1.83
2,000,000 2,106,774
50,940
z = = -2.10
12
(32 - 11)/ = 6.06. Almost 81% of the time there are less than or
equal to 28
sales associates working. One hundred percent of the time there are
less than or
equal to 34 sales associates working and never more than 34. About
23.8% of
the time there are 16 or fewer sales associates working. There are 21
or fewer
6.62 The weight of the rods is normally distributed with a mean of 227 mg
and a
standard deviation of 2.3 mg. The probability that a rod weighs less
than or
or equal to 227 is .5000 (since 227 is the mean), less than 231 mg is .
9590, and
P a g e | 398
6.63 The lengths of cell phone calls are normally distributed with a mean
of 2.35
less than or equal to 2.60 minutes, almost 82% are less than or equal
to 2.45
minutes, over 32% are less than 2.3 minutes, and almost none are less
than
2 minutes.
Chapter 7
Sampling and Sampling Distributions
LEARNING OBJECTIVES
The two main objectives for Chapter 7 are to give you an appreciation for the
proper application of sampling techniques and an understanding of the sampling
distributions of two statistics, thereby enabling you to:
x p̂
6. Use the sampling distributions of and .
Virtually every analysis discussed in this text deals with sample data.
It is important, therefore, that students are exposed to the ways and means
that samples are gathered. The first portion of chapter 7 deals with
sampling. Reasons for sampling versus taking a census are given. Most of
these reasons are tied to the fact that taking a census costs more than
sampling if the same measurements are being gathered. Students are then
exposed to the idea of random versus nonrandom sampling. Random
sampling appeals to their concepts of fairness and equal opportunity. This
text emphasizes that nonrandom samples are non probability samples and
cannot be used in inferential analysis because levels of confidence and/or
probability cannot be assigned. It should be emphasized throughout the
discussion of sampling techniques that as future business managers (most
students will end up as some sort of supervisor/manager) students should be
aware of where and how data are gathered for studies. This will help to
assure that they will not make poor decisions based on inaccurate and poorly
gathered data.
n
standard error of the mean ( ). As the student sees the central limit
theorem unfold, he/she begins to see that if the sample size is large enough,
sample means can be analyzed using the normal curve regardless of the
shape of the population.
Chapter 7 presents formulas derived from the central limit theorem for
both sample means and sample proportions. Taking the time to introduce
these techniques in this chapter can expedite the presentation of material in
chapters 8 and 9.
CHAPTER OUTLINE
7.1 Sampling
Frame
Systematic Sampling
Nonrandom Sampling
Convenience Sampling
Judgment Sampling
Quota Sampling
Snowball Sampling
P a g e | 402
Sampling Error
Nonsampling Errors
x
7.2 Sampling Distribution of
p̂
7.3 Sampling Distribution of
KEY TERMS
Start between 0 and 20. The human resource department probably has a list
of
company employees which can be used for the frame. Also, there might be a
7.9 a) i. Counties
c) i. States
ii. Counties
7.10 Go to the district attorney's office and observe the apparent activity of
various
attorney's at work. Select some who are very busy and some who seem to
be
less active. Select some men and some women. Select some who appear
to
be older and some who are younger. Select attorneys with different ethnic
backgrounds.
7.12 Suppose 40% of the sample is to be people who presently own a personal
computer and 60% , people who do not. Go to a computer show at the city's
conference center and start interviewing people. Suppose you get enough
people who own personal computers but not enough interviews with those
who do not. Go to a mall and start interviewing people. Screen out personal
computer owners. Interview non personal computer owners until you meet
the 60% quota.
x
a) P( > 52):
x 52 50
10
n 64
z = = 1.6
x
P( > 52) = .5000 - .4452 = .0548
x
b) P( < 51):
P a g e | 408
x 51 50
10
n 64
z = = 0.80
x
P( < 51) = .5000 + .2881 = .7881
x
c) P( < 47):
x 47 50
10
n 64
z = = -2.40
x
P( < 47) = .5000 - .4918 = .0082
x
d) P(48.5 < < 52.4):
P a g e | 409
x 48.5 50
10
n 64
z = = -1.20
x 52.4 50
10
n 64
z = = 1.92
x
P(48.5 < < 52.4) = .3849 + .4726 = .8575
P a g e | 410
x
e) P(50.6 < < 51.3):
x 50.6 50
10
n 64
z = = 0.48
x 51.3 50
1.04
10
n 64
z =
x
P(50.6 < < 51.3) = .3508 - .1844 = .1664
x
a) n = 10, P( > 22):
P a g e | 411
x 22 23.45
3.8
n 10
z = = -1.21
x
P( > 22) = .3869 + .5000 = .8869
x
b) n = 4, P( > 26):
x 26 23.45
3.8
n 4
z = = 1.34
x
P( > 26) = .5000 - .4099 = .0901
x
7.15 n = 36 µ = 278 P( < 280) = .86
x
.3600 of the area lies between = 280 and µ = 278. This probability is
x
n
z =
280 278
36
1.08 =
6
1.08 = 2
12
1.08
= = 11.11
x
7.16 n = 81 = 12 P( > 300) = .18
Solving for µ:
P a g e | 413
x
n
z =
300
12
81
0.92 =
12
9
0.92 = 300 -
1.2267 = 300 -
x
P( < 76.5):
x 76.5 75
N n 6 1000 60
n N 1 60 1000 1
z = = 2.00
P a g e | 414
x
P( < 76.5) = .4772 + .5000 = .9772
b) N = 90 n = 36 µ = 108 = 3.46
x
P(107 < < 107.7):
x
P(107 < < 107.7) = .4871 - .2486 = .2385
x
P( > 36):
x 36 35.6
N n 4.89 250 100
n N 1 100 250 1
z = = 1.05
x
P( > 36) = .5000 - .3531 = .1469
x
P( < 123):
x
P( < 123) = .5000 - .3770 = .1230
P a g e | 416
7.18 µ = 99.9 = 30 n = 38
x
a) P( < 90):
x 90 99.9
30
n 38
z = = -2. 03
x
P( < 90) = .5000 - .4788 = .0212
x
b) P(98 < < 105):
x 105 99.9
30
n 38
z = = 1.05
x 98 99.9
30
n 38
z = = -0.39
P a g e | 417
x
P(98 < < 105) = .3531 + .1517 = .5048
x
c) P( < 112):
x 112 99.9
30
n 38
z = = 2.49
x
P( < 112) = .5000 + .4936 = .9936
x
d) P(93 < < 96):
x 93 99.9
30
n 38
z = = -1.42
x 96 99.9
30
n 38
z = = -0.80
x
P(93 < < 96) = .4222 - .2881 = .1341
X
P( > $185,000):
X 185,000 177,000
N n 8,500 1500 100
n N 1 100 1500 1
z = = 9.74
X
P( > $185,000) = .5000 - .5000 = .0000
P a g e | 419
x x0
7.20 µ = $65.12 = $21.45 n = 45 P( > ) = .2300
x x0
Prob. lies between and µ = .5000 - .2300 = .2700
x0
Solving for :
x0
n
z =
x 0 65.12
21.45
45
0.74 =
x0 x0
2.366 = - 65.12 and = 65.12 + 2.366 = 67.486
x
a) P( > 52):
P a g e | 420
x 52 50.4
11.8
n 42
z = = 0.88
x
P( > 52) = .5000 - .3106 = .1894
x
b) P( < 47.5):
x 47.5 50.4
11.8
n 42
z = = -1.59
x
P( < 47.5) = .5000 - .4441 = .0559
x
c) P( < 40):
x 40 50.4
11.8
n 42
z = = -5.71
P a g e | 421
x
P( < 40) = .5000 - .5000 = .0000
d) 71% of the values are greater than 49. Therefore, 21% are between the
z.21 = -0.55
x
n
z =
49 50.4
42
-0.55 =
= 16.4964
P a g e | 422
7.22 p = .25
p̂
a) n = 110 P( < .21):
pˆ p .21 .25
pq (.25)(. 75)
n 110
z = = -0.97
p̂
P( < .21) = .5000 - .3340 = .1660
p̂
b) n = 33 P( > .24):
pˆ p .24 .25
pq (.25)(. 75)
n 33
z = = -0.13
p̂
P( > .24) = .5000 + .0517 = .5517
p̂
c) n = 59 P(.24 < < .27):
P a g e | 423
pˆ p .24 .25
pq (.25)(. 75)
n 59
z = = -0.18
pˆ P .27 .25
pq (.25)(. 75)
n 59
z = = 0.35
p̂
P(.24 < < .27) = .0714 + .1368 = .2082
P a g e | 424
p̂
d) n = 80 P( > .30):
pˆ p .30 .25
pq (.25)(. 75)
n 80
z = = 1.03
p̂
P( > .30) = .5000 - .3485 = .1515
p̂
e) n = 800 P( > .30):
pˆ p .30 .25
pq (.25)(. 75)
n 800
z = = 3.27
p̂
P( > .30) = .5000 - .4995 = .0005
p̂
a) P( > .60):
pˆ p .60 .58
pq (.58)(. 42)
n 660
z = = 1.04
p̂
P( > .60) = .5000 - .3508 = .1492
p̂
b) P(.55 < < .65):
pˆ p .65 .58
pq (.58)(. 42)
n 660
z = = 3.64
pˆ p .55 .58
pq (.58)(. 42)
n 660
z = = 1.56
p̂
P(.55 < < .65) = .4998 + .4406 = .9404
p̂
c) P( > .57):
pˆ p .57 .58
pq (.58)(. 42)
n 660
z = = -0.52
p̂
P( > .57) = .1985 + .5000 = .6985
p̂
d) P(.53 < < .56):
p̂
P(.53 < < .56) = .4953 - .3508 = .1445
P a g e | 427
p̂
e) P( < .48):
pˆ p .48 .58
pq (.58)(. 42)
n 660
z = = -5.21
p̂
P( < .48) = .5000 - .5000 = .0000
P a g e | 428
p̂
7.24 p = .40 P( > .35) = .8000
p̂
P(.35 < < .40) = .8000 - .5000 = .3000
Solving for n:
pˆ p
pq
n
z =
0.84 .24
n
.05
n
8.23 =
n = 67.73 68
P a g e | 429
p̂ p̂0
7.25 p = .28 n = 140 P( < ) = .3000
p̂ p̂0
P( < < .28) = .5000 - .3000 = .2000
p̂0
Solving for :
pˆ 0 p
pq
n
z =
pˆ 0 .28
(.28)(. 72)
140
-0.52 =
p̂0
-.02 = - .28
p̂0
= .28 - .02 = .26
150
p̂ 600
= = .25
P a g e | 430
pˆ p .25 .21
pq (.21)(. 79)
n 600
z = = 2.41
90
p̂ 200
= = .45
pˆ p .45 .48
pq (.48)(. 52)
n 200
z = = -0.85
100
p̂ 200
= = .50
pˆ p .50 .48
pq (.48)(. 52)
n 200
z = = 0.57
80
p̂ 200
= = .40
pˆ p .40 .48
pq (.48)(. 52)
n 200
z = = -2.26
p̂
a) P( > .25):
pˆ p .25 .19
pq (.19)(.89)
n 950
z = = 4.71
p̂
P( > .25) = .5000 - .5000 = .0000
P a g e | 433
p̂
b) P(.15 < < .20):
pˆ p .15 .19
pq (.19)(.81)
n 950
z = = -3.14
pˆ p .20 .19
pq (.19)(.89)
n 950
z = = 0.79
p̂
P(.15 < < .20) = .4992 + .2852 = .7844
133 171
p̂1 950 p̂2 950
= = .14 = = .18
p̂
P(.14 < < .18):
P a g e | 434
7.29 µ = 76, = 14
x
a) n = 35, P( > 79):
x 79 76
14
n 35
z = = 1.27
x
P( > 79) = .5000 - .3980 = .1020
P a g e | 435
x
b) n = 140, P(74 < < 77):
x 74 76 x 77 76
14 14
n 140 n 140
z = = -1.69 z = = 0.85
x
P(74 < < 77) = .4545 + .3023 = .7568
P a g e | 436
x
c) n = 219, P( < 76.5):
x 76.5 76
14
n 219
z = = 0.53
x
P( < 76.5) = .5000 + .2019 = .7019
7.30 p = .46
a) n = 60
p̂
P(.41 < < .53):
pˆ p .53 .46
pq (.46)(. 54)
n 60
z = = 1.09
pˆ p .41 .46
pq (.46)(.54)
n 60
z = = -0.78
p̂
P(.41 < < .53) = .3621 + .2823 = .6444
p̂
b) n = 458 P( < .40):
p p .4 0 .4 6
p q ( .4 6 ) ( .5 4 )
n 458
z = = -2.58
p
P( < .40) = .5000 - .4951 = .0049
P a g e | 438
p̂
c) n = 1350 P( > .49):
pˆ p .49 .46
pq (.46)(. 54)
n 1350
z = = 2.21
p̂
P( > .49) = .5000 - .4864 = .0136
18 – 25 250(.18) = 45
26 – 50 250(.36) = 90
51 – 65 250(.10) = 25
over 65 250(.14) = 35
n = 250
x 298
p̂ n 600
= = .497
P a g e | 439
p̂
P( < .497):
pˆ p .497 .55
pq (.55)(. 45)
n 600
z = = -2.61
p̂
P( < .497) = .5000 - .4955 = .0045
x
P( < 17,500):
17,500 17,755
650 120 30
30 120 1
z = = -2.47
x
P( < 17,500) = .5000 - .4932 = .0068
P a g e | 441
7.35 Number the employees from 0001 to 1250. Randomly sample from the
random number table until 60 different usable numbers are obtained. You
cannot use numbers from 1251 to 9999.
x
7.36 µ = $125 n = 32 = $110 2 = $525
x
P( > $110):
x 110 125
525
n 32
z = = -3.70
x
P( > $110) = .5000 + .5000 = 1.0000
P a g e | 442
x
P( > $135):
x 135 125
525
n 32
z = = 2.47
x
P( > $135) = .5000 - .4932 = .0068
x
P($120 < < $130):
x 120 125
525
n 32
z = = -1.23
x 130 125
525
n 32
z = = 1.23
x
P($120 < < $130) = .3907 + .3907 = .7814
7.37 n = 1100
x 810
p̂ n 1100
=
pˆ p .7364 .73
pq (.73)(. 27)
n 1100
z = = 0.48
x 1030
p̂ n 1100
= = .9364
P a g e | 444
pˆ p .9364 .96
pq (.96)(.04)
n 1100
z = = -3.99
c) p = .85
p̂
P(.82 < < .84):
pˆ p .82 .85
pq (.85)(. 15)
n 1100
z = = -2.79
pˆ p .84 .85
pq (.85)(. 15)
n 1100
z = = -0.93
7.38 1) The managers from some of the companies you are interested in
7.39 Divide the factories into geographic regions and select a few factories to
represent those regional areas of the country. Take a random sample of
employees from each selected factory. Do the same for distribution centers
and retail outlets. Divide the United States into regions of areas. Select a
few areas. Take a random sample from each of the selected area distribution
centers and retail outlets.
x 339
p̂ n 565
= = .60
pˆ p .60 .54
pq (.54)(.46)
n 565
z = = 2.86
x 288
p̂ n 565
= = .5097
P a g e | 447
pˆ p .5097 .54
pq (.54)(.46)
n 565
z = = -1.45
p̂
c) P( < .50):
pˆ p .50 .54
pq (.54)(.46)
n 565
z = = -1.91
p̂
P( < .50) = .5000 - .4719 = .0281
x
P( < $530):
P a g e | 448
x 530 550
100
n 50
z = = -1.41
x
a) P( > 60):
x 60 56.8
12.3
n 51
z = = 1.86
x
P( > 60) = .5000 - .4686 = .0314
P a g e | 449
x
b) P( > 58):
x 58 56.8
12.3
n 51
z = = 0.70
x
P( > 58) = .5000 - .2580 = .2420
x
c) P(56 < < 57):
x 56 56.8 x 57 56.8
12.3 12.3
n 51 n 51
z = = -0.46 z = = 0.12
x
P(56 < < 57) = .1772 + .0478 = .2250
P a g e | 450
x
d) P( < 55):
x 55 56.8
12.3
n 51
z = = -1.05
x
P( < 55) = .5000 - .3531 = .1469
x
e) P( < 50):
x 50 56.8
12.3
n 51
z = = -3.95
x
P( < 50) = .5000 - .5000 = .0000
x 210 x 234
p̂1 n 300 p̂2 n 300
= = .70 = = .78
pˆ p .70 .73
pq (.73)(. 27)
n 300
z = = -1.17
pˆ p .78 .73
pq (.73)(. 27)
n 300
z = = 1.95
p̂
b) P( > .78):
pˆ p .78 .73
pq (.73)(. 27)
n 300
z = = 1.95
p̂
P( > .78) = .5000 - .4744 = .0256
P a g e | 452
p̂
c) p = .73 n = 800 P( > .78):
pˆ p .78 .73
pq (.73)(. 27)
n 800
z = = 3.19
p̂
P( > .78) = .5000 - .4993 = .0007
P a g e | 453
35
p̂ 140
= = .25 p = .22
pˆ p .25 .22
pq (.22)(.78)
n 140
z = = 0.86
21
p̂ 140
= = .15
pˆ p .15 .22
pq (.22)(.78)
n 140
z = = -2.00
n = 300 p = .20
p̂
P(.18 < < .25):
pˆ p .18 .20
pq (.20)(.80)
n 300
z = = -0.87
pˆ p .25 .20
pq (.20)(.80)
n 300
z = = 2.17
p̂
P(.18 < < .25) = .3078 + .4850 = .7928
7.47 By taking a sample, there is potential for obtaining more detailed information.
More time can be spent with each employee. Probing questions can
be asked. There is more time for trust to be built between employee and
With a census, data is usually more general and easier to analyze because it
is in a more standard format. Decision-makers are sometimes more
comfortable with a census because everyone is included and there is no
P a g e | 455
p̂
P( > .80):
pˆ p .80 .75
pq (.75)(. 25)
n 150
z = = 1.41
p̂
P( > .80) = .5000 - .4207 = .0793
x
P(21 < < 22):
x 21 21.24
3
n 40
z = = -0.51
P a g e | 456
x 22 21.24
3
n 40
z = = 1.60
x
P(21 < < 22) = .1950 + .4452 = .6402
P a g e | 457
Japan: n = 35 µ = $ 22.00 = $3
x
P( > 23):
x 23 22
3
n 35
z = = 1.97
x
P( > 23) = .5000 - .4756 = .0244
x
P( < 18.90):
x 18.90 19.86
3
n 50
z = = -2.26
x
P( < 18.90) = .5000 - .4881 = .0119
P a g e | 458
x
P( > $273):
x 273 281
47
n 65
z = = -1.37
x
P( > $273) = .5000 + .4147 = .9147
P a g e | 459
Chapter 8
Statistical Inference: Estimation for Single
Populations
LEARNING OBJECTIVES
One of the more common questions asked of statisticians is: "How large of a
size estimation gives the researcher a "ball park" figure as to how many to
sample.
CHAPTER OUTLINE
Mean
The t Distribution
Robustness
Statistic
Mean
Population Proportion
KEY WORDS
x
8.1 a) = 25 = 3.5 n = 60
3.5
xz
n 60
= 25 + 1.96 = 25 + 0.89 = 24.11 < µ < 25.89
x
b) = 119.6 = 23.89 n = 75
23.89
xz
n 75
= 119.6 + 2.33 = 119.6 ± 6.43 = 113.17 < µ <
126.03
x
c) = 3.419 = 0.974 n = 32
0.974
xz
n 32
= 3.419 + 1.645 = 3.419 ± .283 = 3.136 < µ <
3.702
x
d) = 56.7 = 12.1 N = 500 n = 47
N n 12.1 500 47
xz
n N 1 47 500 1
= 56.7 + 1.28 =
x
8.2 n = 36 = 211 = 23
95% C.I. z.025 = 1.96
23
xz
n 36
= 211 ± 1.96 = 211 ± 7.51 = 203.49 < µ <
218.51
x
8.3 n = 81 = 47 = 5.89
5.89
xz
n 81
= 47 ± 1.645 = 47 ± 1.08 = 45.92 < µ < 48.08
x
8.4 n = 70 2 = 49 = 90.4
x
= 90.4 Point Estimate
49
xz
n 70
= 90.4 ± 1.88 = 90.4 ± 1.57 = 88.83 < µ < 91.97
P a g e | 468
x
8.5 n = 39 N = 200 = 66 = 11
N n 11 200 39
xz
n N 1 39 200 1
= 66 ± 2.05 =
x
= 66 Point Estimate
x
8.6 n = 120 = 18.72 = 0.8735
x
= 18.72 Point Estimate
0.8735
xz
n 120
= 18.72 ± 2.575 = 8.72 ± .21 = 18.51 < µ < 18.93
x
8.7 N = 1500 n = 187 = 5.3 years = 1.28 years
x
= 5.3 years Point Estimate
P a g e | 469
x
8.8 n = 24 = 5.625 = 3.23
3.23
xz
n 24
= 5.625 ± 1.645 = 5.625 ± 1.085 = 4.540 < µ < 6.710
x
8.9 n = 36 = 3.306 = 1.17
1.17
xz
n 36
= 3.306 ± 2.33 = 3.306 ± .454 = 2.852 < µ < 3.760
x
8.10 n = 36 = 2.139 = .113
P a g e | 470
x
= 2.139 Point Estimate
(.113)
xz
n 36
= 2.139 ± 1.645 = 2.139 ± .031 = 2.108 < µ < 2.170
P a g e | 471
x
= 24.533 = 5.124 z = + 1.96
5.124
xz
n 45
= 24.533 + 1.96 =
x
8.13 n = 13 = 45.62 s = 5.694 df = 13 – 1 = 12
t.025,12 = 2.179
s 5.694
xt
n 13
= 45.62 ± 2.179 = 45.62 ± 3.44 = 42.18 < µ < 49.06
x
8.14 n = 12 = 319.17 s = 9.104 df = 12 - 1 = 11
s 9.104
xt
n 12
= 319.17 ± (1.796) = 319.17 ± 4.72 = 314.45 < µ <
323.89
x
8.15 n = 41 = 128.4 s = 20.6 df = 41 – 1 = 40
/2 = .01
t.01,40 = 2.423
P a g e | 473
s 20.6
xt
n 41
= 128.4 ± 2.423 = 128.4 ± 7.80 = 120.6 < µ < 136.2
x
= 128.4 Point Estimate
x
8.16 n = 15 = 2.364 s2 = 0.81 df = 15 – 1 = 14
/2 = .05
t.05,14 = 1.761
s 0.81
xt
n 15
= 2.364 ± 1.761 = 2.364 ± .409 = 1.955 < µ < 2.773
x
8.17 n = 25 = 16.088 s = .817 df = 25 – 1 = 24
/2 = .005
t.005,24 = 2.797
P a g e | 474
s (.817)
xt
n 25
= 16.088 ± 2.797 = 16.088 ± .457 = 15.631 < µ < 16.545
x
= 16.088 Point Estimate
P a g e | 475
x
8.18 n = 22 = 1,192 s = 279 df = n - 1 = 21
s 279
xt
n 22
= 1,192 + 2.518 = 1,192 + 149.78 = 1,042.22 < <
1,341.78
x
= 2.36116 s = 0.19721
0.1972
20
2.36116 + 2.093 = 2.36116 + 0.0923 = 2.26886 < <
2.45346
Error = 0.0923
P a g e | 476
x
8.20 n = 28 = 5.335 s = 2.016 df = 28 – 1 = 27
t.05,27 = 1.703
s 2.016
xt
n 28
= 5.335 ± 1.703 = 5.335 + .649 = 4.686 < µ < 5.984
x
8.21 n = 10 = 49.8 s = 18.22 df = 10 – 1 = 9
s 18.22
xt
n 10
= 49.8 ± 2.262 = 49.8 + 13.03 = 36.77 < µ < 62.83
t.01,13 = 2.650
P a g e | 477
x
from data: = 152.16 s = 14.42
s 14.42
xt
n 14
confidence interval: = 152.16 + 2.65 =
t.005,16 = 2.921
x
from data: = 8.06 s = 5.07
s 5.07
xt
n 17
confidence interval: = 8.06 + 2.921 =
x
8.24 The point estimate is which is 25.4134 hours. The sample size is 26 skiffs.
The confidence level is 98%. The confidence interval is:
s s
x t x t
n n
= 22.8124 < < 28.0145
The error of the confidence interval is 2.6011.
P a g e | 479
p̂
8.25 a) n = 44 =.51 99% C.I. z.005 = 2.575
pˆ qˆ (.51)(. 49)
pˆ z
n 44
= .51 ± 2.575 = .51 ± .194 = .316 < p< .704
p̂
b) n = 300 = .82 95% C.I. z.025 = 1.96
pˆ qˆ (.82)(. 18)
pˆ z
n 300
= .82 ± 1.96 = .82 ± .043 = .777 < p < .863
p̂
c) n = 1150 = .48 90% C.I. z.05 = 1.645
pˆ qˆ (.48)(.52)
pˆ z
n 1150
= .48 ± 1.645 = .48 ± .024 = .456 < p < .
504
P a g e | 480
p̂
d) n = 95 = .32 88% C.I. z.06 = 1.555
pˆ qˆ (.32)(. 68)
pˆ z
n 95
= .32 ± 1.555 = .32 ± .074 = .246 < p < .
394
x 57
p̂ n 116
= = .49
pˆ qˆ (.49)(. 51)
pˆ z
n 116
= .49 ± 2.575 = .49 ± .12 = .37 < p < .61
x 479
p̂ n 800
= = .60
P a g e | 481
pˆ qˆ (.60)(. 40)
pˆ z
n 800
= .60 ± 2.17 = .60 ± .038 = .562 < p < .
638
x 106
p̂ n 240
= = .44
pˆ qˆ (.44)(. 56)
pˆ z
n 240
= .44 ± 1.44 = .44 ± .046 = .394 < p < .
486
x 21
p̂ n 60
= = .35
pˆ qˆ (.35)(.65)
pˆ z
n 60
= .35 ± 1.645 = .35 ± .10 = .25 < p < .45
x 40
p̂ n 85
= = .47
pˆ qˆ (.47)(. 53)
pˆ z
n 85
= .47 ± 1.645 = .47 ± .09 = .38 < p < .56
pˆ qˆ (.47)(. 53)
pˆ z
n 85
= .47 ± 1.96 = .47 ± .11 = .36 < p < .58
pˆ qˆ (.47)(. 53)
pˆ z
n 85
= .47 ± 2.575 = .47 ± .14 = .33 < p < .61
All other things being constant, as the confidence increased, the width of the
interval increased.
p̂
8.28 n = 1003 = .255 99% CI z.005 = 2.575
P a g e | 483
pˆ qˆ (.255)(.745)
pˆ z
n 1003
= .255 + 2.575 = .255 + .035 = .220 < p < .
290
p̂
n = 10,000 = .255 99% CI z.005 = 2.575
pˆ qˆ (.255)(.745)
pˆ z
n 10,000
= .255 + 2.575 = .255 + .011 = .244 < p < .
266
interval constructed using n = 10,000. One might conclude that, all other
things
being constant, increasing the sample size reduces the width of the
confidence
interval.
p̂
8.29 n = 560 = .47 95% CI z.025 = 1.96
pˆ qˆ (.47)(.53)
pˆ z
n 560
= .47 + 1.96 = .47 + .0413 = .4287 < p < .5113
P a g e | 484
p̂
n = 560 = .28 90% CI z.05 = 1.645
pˆ qˆ (.28)(. 72)
pˆ z
n 560
= .28 + 1.645 = .28 + .0312 = .2488 < p < .
3112
x 997
p̂ n 1250
= = .80
pˆ qˆ (.80)(.20)
pˆ z
n 1250
= .80 ± 2.33 = .80 ± .026 = .774 < p < .826
P a g e | 485
x 927
p̂ n 3481
= = .266
p̂
a) = .266 Point Estimate
pˆ qˆ (.266)(.734)
pˆ z
n 3481
= .266 + 2.575 = .266 ± .019 =
x 48
p̂ n 89
= = .54
pˆ qˆ (.54)(.46)
pˆ z
n 89
= .54 ± 1.44 = .54 ± .076 = .464 < p < .616
p̂
8.33 = .63 n = 672 95% Confidence z.025 = + 1.96
P a g e | 486
pˆ qˆ (.63)(. 37)
pˆ z
n 672
= .63 + 1.96 = .63 + .0365 = .5935 < p < .
6665
x 121
p̂ n 275
= = .44
pˆ qˆ (.44)(.56)
pˆ z
n 275
= .44 ± 2.33 = .44 ± .07 = .37 < p < .51
P a g e | 487
x
8.35 a) n = 12 = 28.4 s2 = 44.9 99% C.I. df =
12 – 1 = 11
x
b) n=7 = 4.37 s = 1.24 s2 = 1.5376 95% C.I. df = 7 – 1 =
6
(7 1)(1.5376) (7 1)(1.5376)
14.4494 1.23734
< 2 <
x
c) n = 20 = 105 s = 32 s2 = 1024 90% C.I. df =
20 – 1 = 19
z 2 2 (1.96) 2 (36) 2
E2 52
n= = 199.15
Sample 200
P a g e | 491
z 2 2 ( 2.575) 2 (4.13) 2
E2 12
n= = 113.1
Sample 114
z 2 2 (1.645) 2 (105) 2
E2 10 2
n = = 298.3
Sample 299
z 2 2 (1.555) 2 (14.5) 2
E2 32
n = = 56.5
P a g e | 492
Sample 57
Sample 2522
z 2 p q (1.96) 2 (.50)(.50)
E2 (.04) 2
n = = 600.25
Sample 601
Sample 268
z 2 p q (2.575) 2 (.50)(.50)
E2 (.01) 2
n = = 16,576.6
Sample 16,577
z 2 2 (2.575) 2 (1000) 2
E2 200 2
n = = 165.77
Sample 166
P a g e | 494
z 2 2 (1.645) 2 (12.50) 2
E2 22
n = = 105.7
Sample 106
z 2 2 (1.645) 2 (475) 2
E2 100 2
n = = 61.05
Sample 62
Sample 1083
Sample 385
Sample 97
x
8.48 = 45.6 = 7.75 n = 35
7.75
xz 45.6 1.28
n 35
= 45.6 + 1.68
7.75
xz 45.6 1.88
n 35
= 45.6 + 2.46
7.75
xz 45.6 2.33
n 35
= 45.6 + 3.05
x
8.49 = 12.03 (point estimate) s = .4373 n = 10 df = 9
s (.4373)
xt 12.03 1.833
n 10
= 12.03 + .25
s (.4373)
xt 12.03 2.262
n 10
= 12.03 + .31
s (.4373)
xt 12.03 3.25
n 10
= 12.03 + .45
329
pˆ
715
= .46
pˆ qˆ (.46)(. 54)
pˆ z .46 1.96
n 715
= .46 + .0365
p̂
b) n = 284 = .71 90% confidence z.05 = 1.645
pˆ qˆ (.71)(.29)
pˆ z .71 1.645
n 284
= .71 + .0443
p̂
c) n = 1250 = .48 95% confidence z.025 = 1.96
pˆ qˆ (.48)(. 52)
pˆ z .48 1.96
n 1250
= .48 + .0277
270
pˆ
457
= .591
pˆ qˆ (.591)(.409)
pˆ z .591 2.33
n 457
= .591 + .0536
z 2 2 (1.96) 2 (44) 2
E2 32
n = = 826.4
Sample 827
b) E=2 Range = 88 - 20 = 68
z 2 2 (1.645) 2 (17) 2
E2 22
= 195.5
Sample 196
Sample 849
Sample 897
P a g e | 504
x
8.53 n = 17 = 10.765 s = 2.223 df = 17 - 1 = 16
s 2.223
xt 10.765 2.921
n 17
= 10.765 + 1.575
Sample 722
x
8.56 n = 45 = 213 = 48
48
xz 213 2.33
n 45
= 213 ± 16.67
x
8.57 n = 39 = 37.256 = 3.891
3.891
xz 37.256 1.645
n 39
= 37.256 ± 1.025
z 2 2 (2.33) 2 (6) 2
E2 12
n = = 195.44
Sample 196
714
pˆ
1255
= .569
P a g e | 507
pˆ qˆ (.569)(.431)
pˆ z .569 1.96
n 1,255
= .569 ± .027
x 128
8.60 n = 41 s = 21 98% C.I. df = 41 – 1 =
40
t.01,40 = 2.423
s 21
xt 128 2.423
n 41
= 128 + 7.947
x
8.61 n = 60 = 6.717 = 3.06 N = 300
s N n 3.06 300 60
xz 6.717 2.33
n N 1 60 300 1
=
P a g e | 509
6.717 ± 0.825
z 2 2 (1.96) 2 (142.50) 2
E2 20 2
n = = 195.02
Sample 196
P a g e | 510
x 189
pˆ
n 245
= .77
pˆ qˆ (.77)(. 23)
pˆ z .77 1.645
n 245
= .77 ± .044
x 30
pˆ
n 90
= .33
pˆ qˆ (.33)(.67)
pˆ z .33 1.96
n 90
= .33 ± .097
x
8.65 n = 12 = 43.7 s2 = 228 df = 12 – 1 = 11 95% C.I.
P a g e | 511
t.025,11 = 2.201
s 228
xt 43.7 2.201
n 12
= 43.7 + 9.59
x
8.66 n = 27 = 4.82 s = 0.37 df = 26
s 0.37
xt 4.82 2.056
n 27
= 4.82 + .1464
Since 4.50 is not in the interval, we are 95% confident that µ does
not
equal 4.50.
x
8.67 n = 77 = 2.48 = 12
12
xz 2.48 1.96
n 77
= 2.48 ± 2.68
The interval is inconclusive. It says that we are 95% confident that the
average arrival time is somewhere between .20 of a minute (12 seconds)
early and 5.16 minutes late. Since zero is in the interval, there is a possibility
that, on average, the flights are on time.
p̂
8.68 n = 560 =.33
pˆ qˆ (.33)(. 67)
pˆ z .33 2.575
n 560
= .33 ± .05
Sample 543
x
8.70 n = 27 = 2.10 s = 0.86 df = 27 - 1 = 26
s 0.86
xt 2.10 2.479
n 27
= 2.10 ± 0.41
x
8.72 n = 39 = 1.294 = 0.205 99% Confidence
z.005 = 2.575
0.205
xz 1.294 2.575
n 39
= 1.294 ± .085
8.73 The sample mean fill for the 58 cans is 11.9788 oz. with a standard deviation
of
.0536 oz. The 99% confidence interval for the population fill is 11.9607 oz. to
11.9969 oz. which does not include 12 oz. We are 99% confident that the
population mean is not 12 oz., indicating that the machine may be under
filling
the cans.
8.74 The point estimate for the average length of burn of the new bulb is 2198.217
can be constructed from the information given. The error of the confidence
interval is + 27.76691. Combining this with the point estimate yields the
90%
8.75 The point estimate for the average age of a first time buyer is 27.63 years.
The
confident that the actual population mean age of a first-time home buyer is
8.76 A poll of 781 American workers was taken. Of these, 506 drive their cars to
work. Thus, the point estimate for the population proportion is 506/781 = .
647887. A 95% confidence interval to estimate the population proportion
shows that we are 95% confident that the actual value lies between .61324
and .681413. The error of this interval is + .0340865.
Chapter 9
Statistical Inference:
Hypothesis Testing for Single Populations
LEARNING OBJECTIVES
The main objective of Chapter 9 is to help you to learn how to test hypotheses on single
populations, thereby enabling you to:
2. Understand Type I and Type II errors and know how to solve for Type II
errors.
3. Know how to implement the HTAB system to test hypotheses.
Hypothesis testing presents the logic in which ideas, theories, etc., are
scientifically
examined. The student can be made aware that much of the development of concepts to
this point including sampling, level of data measurement, descriptive tools such as mean
and standard deviation, probability, and distributions pave the way for testing hypotheses.
Often students (and instructors) will say "Why do we need to test this hypothesis when
we can make a decision by examining the data?" Sometimes it is true that examining the
data could allow hypothesis decisions to be made. However, by using the methodology
and structure of hypothesis testing even in "obvious" situations, the researcher has added
credibility and rigor to his/her findings. Some statisticians actually report findings in a
court of law as an expert witness. Others report their findings in a journal, to the public,
to the corporate board, to a client, or to their manager. In each case, by using the
hypothesis testing method rather than a "seat of the pants" judgment, the researcher
stands on a much firmer foundation by using the principles of hypothesis testing and
random sampling. Chapter 9 brings together many of the tools developed to this point
and formalizes a procedure for testing hypotheses.
two-tailed test always has = and in the null and alternative hypothesis.
One-tailed tests are presented with = in the null hypothesis and either > or <
in the alternative hypothesis. If in doubt, the researcher should use a two-
tailed test. Chapter 9 begins with a two-tailed test example. Often that
which the researcher wants to demonstrate true or prove true is set up as the
alternative hypothesis. The null hypothesis is that the new theory or idea is
not true, the status quo is still true, or that there is no difference. The null
hypothesis is assumed to be true before the process begins. Some
researchers liken this procedure to a court of law where the defendant is
presumed innocent (assume null is true - nothing has happened). Evidence is
brought before the judge or jury. If enough evidence is presented, then the
null hypothesis (defendant innocent) can no longer be accepted or assumed
true. The null hypothesis is rejected as not true and the alternate hypothesis
is accepted as true by default. Emphasize that the researcher needs to make
a decision after examining the observed statistic.
Page | 2
Some of the key concepts in this chapter are one-tailed and two-tailed
test and Type I and Type II error. In order for a one-tailed test to be
conducted, the problem must include some suggestion of a direction to be
tested. If the student sees such words as greater, less than, more than,
higher, younger, etc., then he/she knows to use a one-tail test. If no direction
is given (test to determine if there is a "difference"), then a two-tailed test is
called for. Ultimately, students will see that the only effect of using a one-
tailed test versus a two-tailed test is on the critical table value. A one-tailed
test uses all of the value of alpha in one tail. A two-tailed test splits alpha
and uses alpha/2 in each tail thus creating a critical value that is further out
in the distribution. The result is that (all things being the same) it is more
difficult to reject the null hypothesis with a two-tailed test. Many computer
packages such as MINITAB include in the results a p-value. If you
designate that the hypothesis test is a two-tailed test, the computer will
double the p-value so that it can be compared directly to alpha.
The student can be told that there are some widely accepted values for
alpha (probability of committing a Type I error) in the research world and that
Page | 3
a value is usually selected before the research begins. On the other hand,
since the value of Beta (probability of committing a Type II error) varies with
every possible alternate value of the parameter being tested, Beta is usually
examined and computed over a range of possible values of that parameter.
As you can see, the concepts of hypothesis testing are difficult and represent
higher levels of learning (logic, transfer, etc.). Student understanding of
these concepts will improve as you work your way through the techniques in
this chapter and in chapter 10.
CHAPTER OUTLINE
Types of Hypotheses
Research Hypotheses
Statistical Hypotheses
Substantive Hypotheses
the z Statistic
the t Test
KEY TERMS
Observed Value
Type II Error
9.1 a) Ho: µ = 25
Ha: µ 25
Page | 6
x
= 28.1 n = 57 = 8.46 = .01
x 28.1 25
8.46
n 57
z= = 2.77
Since the p-value of .0028 is less than /2 = .005, the decision is to:
xc
n
zc =
x c 25
8.46
57
± 2.575 =
x
c = 25 ± 2.885
x
c = 27.885 (upper value)
x
c = 22.115 (lower value)
x
= 6.91 n = 24 = 1.21 =.01
Page | 8
x 6.91 7.48
1.21
n 24
z = = -2.31
x
= 1,215 n = 113 = 100 = .10
x 1,215 1,200
100
n 113
z = = 1.59
xc
n
zc =
P a g e | 10
x c 1,200
100
113
1.28 =
x
c = 1,200 + 12.04
x x
Since the observed = 1,215 is greater than the critical = 1212.04,
the decision is to reject the null hypothesis.
P a g e | 11
9.4 Ho: µ = 82
Ha: µ < 82
x
= 78.125 n = 32 = 9.184 = .01
z.01 = -2.33
x 78.125 82
9.184
n 32
z= = -2.39
Ha: $424.20
x
= $432.69 n = 54 = $33.90 = .05
x 432.69 424.20
33.90
n 54
z = = 1.84
Since the observed z = 1.85 < z.025 = 1.96, the decision is to fail to
reject the
null hypothesis.
P a g e | 13
x
= $58,974 n = 18 = $7,810 = .01
x 58,974 62,600
7,810
n 18
z = = -1.97
Since the observed z = -1.97 > z.01 = -2.33, the decision is to fail to
reject the
null hypothesis.
9.7 H0: = 5
Ha: 5
x
= 5.0611 n = 42 N = 650 = 0.2803 =.
10
x 5.0611 5
N n 0.2803 650 42
n N 1 42 650 1
z = = 1.46
Since the observed z = 1.46 < z.05 = 1.645, the decision is to fail to
reject the
null hypothesis.
x
= 15.6 n = 32 = 2.3 = .10
x 15.6 18.2
2 .3
n 32
z = = -6.39
x
= $4,008 n = 55 = $386 = .01
x $4,008 $4,292
$386
n 55
z = = -5.46
x
= 132.36 s = 27.68
The average water usage per person is greater than 123 gallons.
P a g e | 17
x
9.11 n = 20 = 16.45 s = 3.59 df = 20 - 1 = 19 = .05
Ho: µ = 16
Ha: µ 16
x 16.45 16
s 3.59
n 20
t = = 0.56
x
9.12 n = 51 = 58.42 s2 = 25.68 df = 51 - 1 = 50
= .01
Ho: µ = 60
Ha: µ < 60
x 58.42 60
s 25.68
n 51
t = = -2.23
x
9.13 n = 11 = 1,235.36 s = 103.81 df = 11 - 1 =
10 = .05
Ho: µ = 1,160
x 1,236.36 1,160
s 103.81
n 11
t = = 2.44
x
9.14 n = 20 = 8.37 s = .1895 df = 20-1 = 19
= .01
Ho: µ = 8.3
Ha: µ 8.3
x 8.37 8.3
s .1895
n 20
t = = 1.65
x
9.15 n = 12 = 1.85083 s = .02353 df = 12 - 1 = 11
= .10
H0: µ = 1.84
Ha: µ 1.84
x 1.85083 1.84
s .02353
n 12
t = = 1.59
x
9.16 n = 25 = 3.1948 s = .0889 df = 25 - 1 = 24 = .01
P a g e | 21
Ho: µ = $3.16
x 3.1948 3.16
s .0889
n 25
t = = 1.96
x
9.17 n = 19 = $31.67 s = $1.29 df = 19 – 1 = 18
= .05
H0: = $32.28
Ha: $32.28
x 31.67 32.28
s 1.29
n 19
t = = -2.06
P a g e | 22
x
9.18 n = 61 = 3.72 s = 0.65 df = 61 – 1 = 60
= .01
H0: = 3.51
x 3.72 3.51
s 0.65
n 61
t = = 2.52
x
9.19 n = 22 = 1031.32 s = 240.37 df = 22 – 1 = 21
= .05
H0: = 1135
Ha: 1135
x 1031.32 1135
s 240.37
n 22
t = = -2.02
x
9.20 n = 12 = 42.167 s = 9.124 df = 12 – 1 = 11
= .01
H0: = 46
Ha: < 46
x 42.167 46
s 9.124
n 12
t = = -1.46
x
9.21 n = 26 = 19.534 minutes s = 4.100 minutes
= .05
H0: = 19
Ha: 19
Since the Excel p-value = .256 > /2 = .025 and MINITAB p-value
=.513 > .05, the decision is to fail to reject the null hypothesis.
She would not conclude that her city is any different from the
ones in the
national survey.
P a g e | 25
p̂
n = 310 = .465 = .05
pˆ p .465 .45
pq (.45)(. 55)
n 310
z = = 0.53
x 55
pˆ
n 100
n = 100 x = 55 = .55
pˆ p .55 .63
pq (.63)(. 37)
n 100
z = = -1.66
Ha: p .29
x 207
pˆ
n 740
n = 740 x = 207 = .28 = .05
pˆ p .28 .29
pq (.29)(. 71)
n 740
z = = -0.60
p-Value Method:
z = -0.60
Since the p-value = .2743 > /2 = .025, the decision is to Fail to
reject the null
hypothesis
pˆ c p
pq
n
z =
pˆ c .29
(.29)(. 71)
740
±1.96 =
p̂c
= .29 ± .033
p̂
Since = .28 is not outside critical values in tails, the decision is to Fail
to reject
Ha: p .48
x 164
pˆ
n 380
= .4316
pˆ p .4316 .48
pq (.48)(. 52)
n 380
z = = -1.89
Since the observed z = -1.89 is greater than z.005= -2.575, The decision
is to fail to reject the null hypothesis. There is not enough
evidence to declare that the proportion is any different than .48.
x 303
pˆ
n 415
= .7301
P a g e | 30
pˆ p 7301 .79
pq (.79)(.21)
n 415
z = = -3.00
Since the observed z = -3.00 is less than z.01= -2.33, The decision is to
reject the null hypothesis.
P a g e | 31
Ha: p .31
x 200
pˆ
n 600
= .3333
pˆ p .3333 .31
pq (.31)(. 69)
n 600
z = = 1.23
Since the observed z = 1.23 is less than z.005= 1.645, The decision is to
fail to reject the null hypothesis. There is not enough evidence to
declare that the proportion is any different than .31.
Ho: p = .24
x 130
pˆ
n 600
= .2167
P a g e | 32
pˆ p .2167 .24
pq (.24)(.76)
n 600
z = = -1.34
Since the observed z = -1.34 is greater than z.05= -1.645, The decision
is to fail to reject the null hypothesis. There is not enough
evidence to declare that the proportion is less than .24.
P a g e | 33
p̂
n = 376 = .22 = .01
pˆ p .22 .18
pq (.18)(. 82)
n 376
z = = 2.02
Since the observed z = 2.02 is less than z.01= 2.33, The decision is to
fail to reject the null hypothesis. There is not enough evidence to
declare that the proportion is greater than .18.
x 22
pˆ
n 118
n = 118 x = 22 = .1864 = .05
pˆ p .1864 .32
pq (.32)(.68)
n 118
z = = -3.11
Since the observed z = -3.11 is less than z.05= -1.645, the decision is to
reject the null hypothesis.
P a g e | 35
Ha: p .47
x 40
pˆ
n 67
= .597
pˆ p .597 .47
pq (.47)(.53)
n 67
z = = 2.08
Since the observed z = 2.08 is greater than z.025= 1.96, The decision is
to reject the null hypothesis.
Ha: 2 > 20
2.05,14 = 23.6848
P a g e | 36
(15 1)(32)
20
2 = = 22.4
hypothesis.
Ha: 2 8.5
2.05,21 = 32.6706
(22 1)(17)
8.5
2 = = 42
Ha: 2 < 45
2 .01,7 = 18.4753
P a g e | 37
(8 1)( 4.12) 2
45
2 = = 2.64
hypothesis.
Ha: 2 5
(11 1)(1.2)
5
2 = = 2.4
Since 2 = 2.4 < 2.975,10 = 3.24696, the decision is to reject the null
hypothesis.
Ha: 2 14
(12 1)(30.0833)
14
2 = = 23.64
hypothesis.
2.01,15 = 30.5780
hypothesis.
P a g e | 39
(13 1)(832,089,743.6)
199,996,164
2 = = 49.93
2.01,6 = 16.8119
(7 1)(.1156)
.04
2 = = 17.34
n = 48 µ = 99 = 14
xc
n
zc =
x c 100
14
48
-1.28 =
x
c = 97.4
xc 97.4 99
14
n 48
z = = = -0.79
xc
n
zc =
x c 100
14
48
-1.645 =
x
c = 96.68
xc 96.68 99
14
n 48
z = = = -1.15
xc
n
zc =
x c 100
14
48
-2.33 =
x
c = 95.29
xc 95.29 99
14
n 48
z = = = -1.84
a) µa = 98.5 zc = -1.645
xc
n
zc =
x c 100
14
48
-1.645 =
x
c = 96.68
xc 96.68 98.5
14
n 48
z = = = -0.90
b) µa = 98 zc = -1.645
x
c = 96.68
xc 96.68 98
14
n 48
zc = = = -0.65
c) µa = 97 z.05 = -1.645
x
c = 96.68
xc 96.68 97
14
n 48
z = = = -0.16
d) µa = 96 z.05 = -1.645
x
c = 96.68
xc 96.68 96
14
n 48
z = = = 0.34
e) As the alternative value gets farther from the null hypothesized value, the
probability of committing a Type II error reduces (all other variables being held
constant).
9.38 Ho: µ = 50
Ha: µ 50
µa = 53 n = 35 =7 = .01
xc
n
zc =
x c 50
7
35
±2.575 =
x
c = 50 ± 3.05
xc 53.05 53
7
n 35
z = = = 0.04
Other end:
xc 46.95 53
7
n 35
z = = = -5.11
P a g e | 48
pˆ c p
pq
n
zc =
pˆ c .65
(.65)(.35)
360
-1.645 =
p̂
c = .65 - .041 = .609
pˆ c p .609 .60
pq (.60)(. 40)
n 360
z = = = 0.35
p̂
b) pa = .55 z.05 = -1.645 c = .609
P a g e | 50
pˆ c P .609 .55
pq (.55)(.45)
n 360
z = = = 2.25
p̂
c) pa = .50 z.05 = -1.645 c = .609
pˆ c p .609 .50
pq (.50)(. 50)
n 360
z = = = -4.14
x
9.40 n = 58 = 45.1 = 8.7 = .05 /2 = .
025
H0: µ = 44
45.1 44
8 .7
58
z = = 0.96
Since z = 0.96 < zc = 1.96, the decision is to fail to reject the null
hypothesis.
x c 44
8.7
58
+ 1.96 =
x
± 2.239 = c - 44
x
c = 46.239 and 41.761
For 45 years:
46.239 45
8.7
58
z = = 1.08
For 46 years:
46.239 46
8.7
58
z = = 0.21
For 47 years:
P a g e | 53
46.239 47
8.7
58
z = = -0.67
For 48 years:
46.239 48
8.7
58
z = = 1.54
324
p̂ 463
n = 463 x = 324 = = .6998 = .10
P a g e | 55
z.10 = -1.28
pˆ p .6998 .71
pq (.71)(.29)
n 463
z = = -0.48
Since the observed z = -0.48 > z.10 = -1.28, the decision is to fail to
reject the null hypothesis.
Type II error:
p̂
Solving for the critical proportion, :
c
pˆ c p
pq
n
zc =
pˆ c .71
(.71)(. 29)
463
-1.28 =
p̂
= .683
P a g e | 56
For pa = .69
.683 .69
(.69)(.31)
463
z = = -0.33
For pa = .66
.683 .66
(.66)(. 34)
463
z = = 1.04
For pa = .60
P a g e | 57
.683 .60
(.60)(. 40)
463
z = = 3.65
1)Ho: µ = 36
Ha: µ 36
x
n
2)z =
3) = .01
If the observed value of z is greater than 2.575 or less than -2.575, the
decision will be to reject the null hypothesis.
x
5) n = 63, = 38.4, = 5.93
x 38.4 36
5.93
n 63
6) z = = = 3.21
7) Since the observed value of z = 3.21 is greater than z.005 = 2.575, the
decision is
1)Ho: µ = 7.82
x
s
n
t =
3) = .05
4) df = n - 1 = 16, t.05,16 = -1.746. If the observed value of t is less than -1.746, then
the decision will be to reject the null hypothesis.
x
5) n = 17 = 7.01 s = 1.69
7) Since the observed t = -1.98 is less than the table value of t = -1.746,
the decision
a. 1) Ho: p = .28
Ha: p > .28
pˆ p
pq
n
2) z =
3) = .10
5) n = 783 x = 230
230
pˆ
783
= .2937
.2937 .28
(.28)(. 72)
783
6) z = = 0.85
b. 1) Ho: p = .61
Ha: p .61
pˆ p
pq
n
2) z =
3) = .05
p̂
5) n = 401 = .56
.56 .61
(.61)(.39)
401
6) z = = -2.05
7) Since z = -2.05 is less than z.025 = -1.96, the decision is to reject the null
hypothesis.
1) H0: 2 = 15.4
( n 1) s 2
2
2) 2 =
3) = .01
2.01,17 = 33.4087
5) s2 = 29.6
(n 1) s 2 (17)( 29.6)
2 15.4
6) 2 = = = 32.675
x
Solving for c:
xc
n
zc =
x c 130
12
75
2.33 =
x
c = 133.23
133.23 135
12
75
z = = -1.28
b) H0: p = .44
Ha: p < .44
P a g e | 64
pˆ c p
pq
n
zc =
pˆ c .44
(.44)(.56)
1095
-1.645 =
p̂c
= .4153
.4153 .42
(.42)(. 58)
1095
z = = -0.32
p̂
n = 80 = .01 = .39
P a g e | 65
z.01 = 2.33
pˆ p .39 .32
pq (.32)(.68)
n 80
z = = 1.34
Since the observed z = 1.34 < z.01 = 2.33, the decision is to fail to
reject the null hypothesis.
x
9.48 = 3.45 n = 64 2 = 1.31 = .05
Ho: µ = 3.3
Ha: µ 3.3
x 93
pˆ
n 210
9.49 n = 210 x = 93 = .10 = .
443
Ho: p = .57
pˆ p .443 .57
pq (.57)(.43)
n 210
z = = -3.72
Since the observed z = -3.72 < zc = -1.28, the decision is to reject the
null hypothesis.
P a g e | 67
Ha: 2 > 16
2.05,11 = 19.6752
(12 1)(5.98544) 2
16
2 = = 24.63
Ha: µ 8.4
x
= 5.6 t.005,6 = + 3.707
5.6 8.4
1. 3
7
t = = -5.70
P a g e | 68
Since the observed t = - 5.70 < t.005,6 = -3.707, the decision is to reject
the null hypothesis.
x
9.52 = $26,650 n = 100 = $12,000
a) Ho: µ = $25,000
Since the observed z = 1.38 < z.05 = 1.645, the decision is to fail
to reject the null hypothesis.
b) µa = $30,000 zc = 1.645
x
Solving for :
c
P a g e | 69
xc
n
zc =
( x c 25,000)
12,000
100
1.645 =
x
c = 25,000 + 1,974 = 26,974
26,974 30,000
12,000
100
z = = -2.52
Ha: 2 > 4
2.10,7 = 12.0170
(8 1)(7.80) 2
4
2 = = 106.47
x 66
pˆ
n 125
n = 125 x = 66 = .05 = .528
pˆ p .528 .46
pq (.46)(.54)
n 125
z = = 1.53
Since the observed value of z = 1.53 < z.05 = 1.645, the decision is to
fail to reject the null hypothesis.
P a g e | 71
p̂c
Solving for :
pˆ c p
pq
n
zc =
pˆ c .46
(.46)(. 54)
125 p̂c
1.645 = and therefore, = .533
pˆ c p a .533 .50
pa qa (.50)(. 50)
n 125
z = = 0.74
x
9.55 n = 16 = 175 s = 14.28286 df = 16 - 1 = 15 =.
05
H0: µ = 185
t.05,15 = - 1.753
P a g e | 72
x 84
pˆ
n 428
n = 428 x = 84 = .01 = .1963
pˆ p .1963 .182
pq (.182)(.818)
n 428
z = = 0.77
Since the observed z = 0.77 < z.01 = 2.33, the decision is to fail to
reject the null hypothesis.
p̂c
Solving for :
pˆ c p
pq
n
zc =
. pˆ c .182
(.182)(.818)
428
2.33 =
p̂c
= .2255
pˆ c p a .2255 .21
pa qa (.21)(.79)
n 428
z = = 0.79
x
= $19.34 n = 35 = $4.52 = .10
x 19.34 15
s 4.52
n 35
z = = = 5.68
Since the observed z = 5.68 > zc = 1.28, the decision is to reject the
null hypothesis.
Ha: 2 > 16
2.05,21 = 32.6706
(22 1)(6) 2
16
2 = = 47.25
x
9.59 H 0: µ = 2.5 = 3.4 s = 0.6 = .01 n=9 df =
9–1=8
t.01,8 = 2.896
Since the observed t = 4.50 > t.01,8 = 2.896, the decision is to reject
the null hypothesis.
Ha: 23.58
x
n = 95 = 22.83 = 5.11 = .05
xc
n
b) zc =
xc 23.58
5.11
95
+ 1.96 =
xc
= 23.58 + 1.03
xc
= 22.55, 24.61
xc a 22.55 22.30
5.11
n 95
z = = = 0.48
P a g e | 77
xc a 24.61 22.30
5.11
n 95
z = = = 4.41
from Table A.5, the areas for z = 0.48 and z = 4.41 are .1844
and .5000
x
9.61 n = 12 = 12.333 s2 = 10.424
H0: 2 = 2.5
Ha: 2 2.5
2.025,11 = 21.9200
2..975,11 = 3.81574
P a g e | 78
( n 1) s 2 11(10.424)
2 2 .5
2 = = = 45.866
x
9.62 H0: µ = 23 = 18.5 s = 6.91 = .10 n = 16 df =
16 – 1 = 15
Ha: µ < 23
t.10,15 = -1.341
x 18.5 23
s 6.91
n 16
t = = = -2.60
Since the observed t = -2.60 < t.10,15 = -1.341, the decision is to reject
the null hypothesis.
x
9.63 The sample size is 22. is 3.969 s = 0.866 df = 21
x
s
n
t =
Ha: p .25
Since the p-value = .045 < = .05, the decision is to reject the null
hypothesis.
p̂
The sample proportion, = .205729 which is less than the
hypothesized p = .25.
This is a one-tailed test. The sample mean is 2.55 which is more than
the hypothesized value. The observed t value is 1.51 with an
associated
There is not enough evidence to conclude that beef prices are higher.
Since z = -2.09 < z.05 = -1.645, the decision is to reject the null
hypothesis at
However, the p-value of .018 also indicates that we would not reject
the null
hypothesis at = .01.
P a g e | 82
Chapter 10
Statistical Inferences about Two Populations
LEARNING OBJECTIVES
and 9. When the population variances are known, the z statistic can be used.
However, if the population variances are unknown and sample variances are
being used, then the
CHAPTER OUTLINE
Hypothesis Testing
Confidence Intervals
Hypothesis Testing
Test
Confidence Intervals
Hypothesis Testing
Populations
Confidence Intervals
P a g e | 86
Hypothesis Testing
Confidence Intervals
KEY TERMS
x x
1 = 51.3 2 = 53.2
s12 = 52 s22 = 60
n1 = 31 n2 = 32
a) Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 < 0
( x 1 x 2 ) c ( 1 2 )
1 2
2 2
n1 n2
zc =
( x 1 x 2 ) c ( 0)
52 60
31 32
-1.28 =
x x
( 1- 2)c = -2.41
n1 = 32 n2 = 31
x x
1 = 70.4 2 = 68.7
1 = 5.76 2 = 6.1
2 2
( x1 x 2 ) z 1 2
n1 n2
5.76 2 6.12
32 31
(70.4) – 68.7) + 1.645
1.7 ± 2.46
x x
1 = 88.23 2 = 81.2
P a g e | 90
n1 = 30 n2 = 30
Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 0
1 2
2 2
( x1 x 2 ) z
n1 n2
b)
22.74 26.65
30 30
(88.23 – 81.2) + 2.33
7.03 + 2.99
This supports the decision made in a) to reject the null hypothesis because zero is not in
the interval.
x x
1 = 1.96 2 = 3.02
n1 = 50 n2 = 50
Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 0
10.5 A B
n1 = 40 n2 = 37
x x
1 = 5.3 2 = 6.5
1 2
2 2
( x1 x 2 ) z
n1 n2
1.99 2.36
40 37
(5.3 – 6.5) + 1.96
n1 = 35 n2 = 41
P a g e | 94
x x
1 = 1.84 2 = 1.99
1 = .38 2 = .51
2 2
( x1 x 2 ) z 1 2
n1 n2
.382 .512
35 41
(1.84 - 1.99) ± 2.33
-.15 ± .2384
Hypothesis Test:
1) Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 0
P a g e | 95
( x 1 x 2 ) ( 1 2 )
1 2
2 2
n1 n2
2) z =
3) = .02
or less than -2.33, then the decision will be to reject the null
hypothesis.
7) Since z = -1.47 > z.01 = -2.33, the decision is to fail to reject the
null
hypothesis.
x x
1 = 190 2 = 198
1 = 18.50 2 = 15.60
n1 = 51 n2 = 47 = .01
H0: 1 - 2 = 0
Ha: 1 - 2 < 0
Since the observed z = -2.32 > z.01 = -2.33, the decision is to fail to
reject the null hypothesis.
n1 = 31 n2 = 31
x x
1 = 2.64 2 = 2.36
1 2
2 2
( x1 x 2 ) z
n1 n2
.03 .015
31 31
(2.64-2.36) ± 2.575
Between $ .18 and $ .38 difference with Seattle being more expensive.
x x
1 = 5.8 2 = 5.0
1 = 1.7 2 = 1.4
n1 = 36 n2 = 45
Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 0
Since the observed z = 2.27 > zc = 1.96, the decision is to reject the
null
hypothesis.
P a g e | 99
10.10 A B
x x
1 = 8.05 2 = 7.26
1 = 1.36 2 = 1.06
n1 = 50 n2 = 38
Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 > 0
Since the observed z = 3.06 > zc = 1.28, the decision is to reject the
null hypothesis.
Ha: µ1 - µ2 < 0 df = 8 + 11 - 2 = 17
Sample 1 Sample 2
n1 = 8 n2 = 11
P a g e | 100
x x
1 = 24.56 2 = 26.42
( x1 x 2 ) ( 1 2 )
(24.56 26.42) (0)
2 2
s1 ( n1 1) s 2 (n 2 1) 1 1 12.4(7) 15.8(10) 1 1
n1 n 2 2 n1 n2 8 11 2 8 11
t = = =
-1.05
Since the observed t = -1.05 > t.01,19 = -2.567, the decision is to fail to
reject the null hypothesis.
P a g e | 101
Ha: µ1 - µ2 0 df = 20 + 20 - 2 = 38
Sample 1 Sample 2
n1 = 20 n2 = 20
x x
1 = 118 2 = 113
s1 = 23.9 s2 = 21.6
( x1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n 2 1) 1 1
n1 n2 2 n1 n2
t = =
2 2
s1 (n1 1) s 2 (n 2 1) 1 1
( x1 x 2 ) t
n1 n2 2 n1 n 2
b) =
5 + 12.224
Ha: µ1 - µ2 > 0 df = n1 + n2 - 2 = 10 + 10 - 2 = 18
Sample 1 Sample 2
n1 = 10 n2 = 10
x x
1 = 45.38 2 = 40.49
s1 = 2.357 s2 = 2.355
( x1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n 2 1) 1 1
n1 n2 2 n1 n2
t = =
Ha: µ1 - µ2 0 df = 18 + 18 - 2 = 34
Sample 1 Sample 2
n1 = 18 n2 = 18
x x
1 = 5.333 2 = 9.444
( x 1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n2 1) 1 1
n1 n2 2 n1 n2
t = =
Since the observed t = -4.66 < t.005,34 = -2.75, reject the null
hypothesis.
2 2
s1 (n1 1) s 2 (n 2 1) 1 1
( x1 x 2 ) t
n1 n2 2 n1 n 2
=
(12)(17) (2.026)(17) 1 1
18 18 2 18 18
(5.333 – 9.444) + 2.457
-4.111 + 2.1689
n1 = 21 n2 = 26
x1 x2
= 116,900 = 114,000
s1 = 2,300 s2 = 1,750 df = 21 + 26 – 2
2 2
s (n 1) s 2 (n2 1) 1 1
( x1 x 2 ) t 1 1
n1 n 2 2 n1 n2
=
P a g e | 106
2,900 + 994.62
Ha: µ1 - µ2 0 df = 12 + 12 - 2 = 22
Co-op Interns
n1 = 12 n2 = 12
x x
1 = $15.645 2 = $15.439
s1 = $1.093 s2 = $0.958
( x 1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n2 1) 1 1
n1 n2 2 n1 n2
t = =
Since the observed t = 0.49 < t.05,22 = 1.717, the decision is to fail
reject the null hypothesis.
P a g e | 108
2 2
s (n 1) s 2 (n2 1) 1 1
( x1 x 2 ) t 1 1
n1 n 2 2 n1 n2
=
0.206 + 0.7204
1) Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 > 0
( x 1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n2 1) 1 1
n1 n2 2 n1 n2
2) t =
3) = .01
5) Boston Dallas
n1 = 8 n2 = 9
x x
1 = 47 2 = 44
s1 = 3 s2 = 3
7) Since t = 2.06 < t.01,15 = 2.602, the decision is to fail to reject the
null
P a g e | 110
hypothesis.
10.18 nm = 22 nno = 20
x x
m = 112 no = 122
sm = 11 sno = 12
df = nm + nno - 2 = 22 + 20 - 2 = 40
2 2
s (n 1) s 2 (n2 1) 1 1
( x1 x 2 ) t 1 1
n1 n 2 2 n1 n2
=
-10 ± 8.60
10.19 Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 0
df = n1 + n2 - 2 = 11 + 11 - 2 = 20
n1 = 11 n2 = 11
x x
1 = $67,381.82 2 = $63,481.82
s1 = $2,067.28 s2 = $1,594.25
( x 1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n2 1) 1 1
n1 n2 2 n1 n2
t = =
hypothesis.
10.20 Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 > 0
df = n1 + n2 - 2 = 9 + 10 - 2 = 17
Men Women
n1 = 9 n2 = 10
x x
1 = $110.92 2 = $75.48
s1 = $28.79 s2 = $30.51
( x 1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n2 1) 1 1
n1 n2 2 n1 n2
t = =
( 1 1 0 .9 2 7 5 .4 8 ) ( 0 )
( 2 8 . 7 9 ) 2 ( 8 ) ( 3 0 .5 1 ) 2 ( 9 ) 1 1
9 10 2 9 10
t = = 2.60
P a g e | 113
Since the observed t = 2.60 > t.01,17 = 2.567, the decision is to Reject
the null
hypothesis.
10.21 Ho: D = 0
Ha: D > 0
Sample 1 Sample 2 d
38 22 16
27 28 -1
30 21 9
41 38 3
36 38 -2
38 26 12
33 19 14
35 31 4
44 35 9
d
n=9 =7.11 sd=6.45 = .01
df = n - 1 = 9 - 1 = 8
P a g e | 114
d D 7.11 0
sd 6.45
n 9
t = = 3.31
Since the observed t = 3.31 > t.01,8 = 2.896, the decision is to reject
the null
hypothesis.
P a g e | 115
Ha: D0
Before After d
107 102 5
99 98 1
110 100 10
113 108 5
96 89 7
98 101 -3
100 99 1
102 102 0
107 105 2
109 110 -1
104 102 2
99 96 3
101 100 1
d
n = 13 = 2.5385 sd=3.4789 = .05
df = n - 1 = 13 - 1 = 12
d D 2.5385 0
sd 3.4789
n 13
t = = 2.63
Since the observed t = 2.63 > t.025,12 = 2.179, the decision is to reject
the null
hypothesis.
P a g e | 117
d
10.23 n = 22 = 40.56 sd = 26.58
t.01,21 = 2.518
sd
d t
n
26.58
22
40.56 ± (2.518)
40.56 ± 14.27
32 40 -8
28 25 3
35 36 -1
32 32 0
26 29 -3
P a g e | 118
25 31 -6
37 39 -2
16 30 -14
35 31 4
d
n=9 = -3 sd = 5.6347 = .025
df = n - 1 = 9 - 1 = 8
sd
d t
n
t =
5.6347
9
t = -3 + (1.86) = -3 ± 3.49
d
= 1302.82 sd = 4938.22 n = 11, df = 10
sd 4938.22
d t
n 11
= 1302.82 + 3.169 = 1302.82 + 4718.42
Ha: D<0
Before After d
2 4 -2
4 5 -1
1 3 -2
3 3 0
4 3 1
2 5 -3
2 6 -4
3 4 -1
1 5 -4
d
n=9 =-1.778 sd=1.716 = .05 df =
n-1=9-1=8
d D 1.778 0
sd 1.716
n 9
t = = -3.11
P a g e | 121
Since the observed t = -3.11 < t.05,8 = -1.86, the decision is to reject
the null hypothesis.
255 197 58
230 225 5
290 215 75
242 215 27
300 240 60
250 235 15
215 190 25
225 200 25
219 203 16
236 223 13
d
n = 11 = 28.09 sd=25.813 df =
n - 1 = 11 - 1 = 10
sd
d t
n
25.813
11
28.09 ± (2.764) = 28.09 ± 21.51
10.28 H0: D = 0
d
Ha: D > 0 n = 27 df = 27 – 1 = 26 = 3.17 sd = 5
d D 3.71 0
sd 5
n 27
t = = 3.86
Since the observed t = 3.86 > t.01,26 = 2.479, the decision is to reject
the null hypothesis.
d
10.29 n = 21 = 75 sd = 30 df = 21 - 1 = 20
sd
d t
n
30
21
75 + 1.725 = 75 ± 11.29
P a g e | 124
Ha: D0
d
n = 15 = -2.85 sd = 1.9 = .01 df
= 15 - 1 = 14
For a two-tail test, /2 = .005 and the critical t.005,14 = + 2.977
d D 2.85 0
sd 1.9
n 15
t = = -5.81
Since the observed t = -5.81 < t.005,14 = -2.977, the decision is to reject
the null
hypothesis.
n1 = 368 n2 = 405
x1 = 175 x2 = 182
P a g e | 125
x1 175 x 2 182
pˆ 1 pˆ 2
n1 368 n2 405
= .476 = .449
Ho: p1 - p2 = 0
Ha: p1 - p2 0
hypothesis.
b) Sample 1 Sample 2
p̂ p̂
1 = .38 2 = .25
n1 = 649 n2 = 558
P a g e | 126
n1 pˆ 1 n2 pˆ 2 649(.38) 558(.25)
p
n1 n 2 649 558
= .32
Ho: p1 - p2 = 0
Ha: p1 - p2 > 0
Since the observed z = 4.83 > zc = 1.28, the decision is to reject the
null
hypothesis.
p̂ p̂
10.32 a) n1 = 85 n2 = 90 1 = .75 2 = .67
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
P a g e | 127
(.75)(.25) (.67)(.33)
85 90
(.75 - .67) ± 1.645 = .08 ± .11
p̂ p̂
b) n1 = 1100 n2 = 1300 1 = .19 2 = .17
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
x1 275 x 2 275
pˆ 1 pˆ 2
n1 430 n2 399
= .64 = .69
P a g e | 128
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
x1 1050 x 2 1100
pˆ 1 pˆ 2
n1 1500 n 2 1500
= .70 = .733
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
10.33 H0: pm - pw = 0
p̂ p̂
Ha: pm - pw < 0 nm = 374 nw = 481
m = .59 w
= .70
n m pˆ m n w pˆ w 374(.59) 481(.70)
p
nm n w 374 481
= .652
Since the observed z = -3.35 < z.05 = -1.645, the decision is to reject
the null
hypothesis.
P a g e | 131
p̂1 p̂2
10.34 n1 = 210 n2 = 176 = .24 = .35
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
p̂ p̂
1 = .48 2 = .56
n1 = 56 n2 = 89
n1 pˆ 1 n 2 pˆ 2 56(.48) 89(.56)
p
n1 n2 56 89
= .529
Ho: p1 - p2 = 0
P a g e | 132
Ha: p1 - p2 0
hypothesis.
P a g e | 133
10.36 A B
n1 = 35 n2 = 35
x1 = 5 x2 = 7
x1 5 x2 7
pˆ 1 pˆ 2
n1 35 n 2 35
= .14 = .20
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
10.37 H0: p1 – p2 = 0
Ha: p1 – p2 0
P a g e | 134
p̂ p̂
= .10 = .09 = .06 n1 = 780 n2 = 915
1 2
n1 pˆ 1 n 2 pˆ 2 780(.09) 915(.06)
p
n1 n 2 780 915
= .0738
Since the observed z = 2.35 > z.05 = 1.645, the decision is to reject the null
hypothesis.
P a g e | 135
p̂ p̂
10.38 n1 = 850 n2 = 910 = .60 2 = .52
1
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
dfnum = 12 - 1 = 11 dfdenom = 10 - 1 = 9
Since the observed F = 1.80 < F.01,10,9 = 5.26, the decision is to fail to
reject the null hypothesis.
dfnum = 5 - 1 = 4 dfdenom = 19 - 1 = 18
2
s1 (4.68) 2
2
s2 (2.78) 2
F = = 2.83
Since the observed F = 2.83 < F.025,4,18 = 3.61, the decision is to fail to
reject the null hypothesis.
P a g e | 137
3.43 3.33
3.40 3.42
3.39 3.39
3.32 3.30
3.39 3.46
3.38 3.39
3.34 3.36
3.38 3.44
3.38 3.37
3.28 3.38
n1 = 10 df1 = 9 n2 = 10 df2 = 9
2
s1 .0018989
2
s2 .0023378
F = = 0.81
Since the observed F = 0.81 is greater than the lower tail critical value
of 0.314 and less than the upper tail critical value of 3.18, the decision
is to fail
2
s1
2
s2
2) F =
3) = .01
If the observed value is greater than 5.66 or less than .177, the
decision will be to reject the null hypothesis.
393.4
702.7
6) F = = 0.56
dfnum = 12 - 1 = 11 dfdenom = 15 - 1 = 14
2
s1 (7.52) 2
2
s2 (6.08) 2
F= = 1.53
Since the observed F = 1.53 < F.05,10,14 = 2.60, the decision is to fail to
reject the null hypothesis.
P a g e | 141
dfnum = 15 - 1 = 14 dfdenom = 15 - 1 = 14
2
s1 91.5
2
s2 67.3
F = = 1.36
Since the observed F = 1.36 < F.005,12,14 = 4.43 and > F.995,14,12 = .226,
the decision is to fail to reject the null hypothesis.
10.45 Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 0
For = .10 and a two-tailed test, /2 = .05 and z.05 = + 1.645
Sample 1 Sample 2
x x
1 = 138.4 2 = 142.5
P a g e | 142
1 = 6.71 2 = 8.92
n1 = 48 n2 = 39
Since the observed value of z = -2.38 is less than the critical value of z
= -1.645, the decision is to reject the null hypothesis. There is a
significant difference in the means of the two populations.
P a g e | 143
x x
1 = 34.9 2 = 27.6
n1 = 34 n2 = 31
2 2
s1 s
( x1 x 2 ) z 2
n1 n2
2.97 3.50
34 31
(34.9 – 27.6) + 2.33 = 7.3 + 1.04
10.47 Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 > 0
Sample 1 Sample 2
x x
= 2.06
1 2 = 1.93
P a g e | 144
n1 = 12 n2 = 15 = .05
t.05,25 = 1.708. If the observed value is greater than 1.708, the decision
will be to reject the null hypothesis.
( x 1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n2 1) 1 1
n1 n2 2 n1 n2
t =
Since the observed value of t = 0.85 is less than the critical value of t
= 1.708, the decision is to fail to reject the null hypothesis. The
mean for population one is not significantly greater than the mean for
population two.
x x
1 = 74.6 2 = 70.9
n1 = 18 n2 = 19
P a g e | 145
2 2
s1 (n1 1) s 2 (n2 1) 1 1
( x1 x 2 ) t
n1 n 2 2 n1 n2
(10.5)(17) (11.4)(18) 1 1
18 19 2 18 19
(74.6 – 70.9) + 2.042
3.7 + 2.22
Ha: D < 0
d
n = 21 df = 20 = -1.16 sd = 1.01
The critical t.01,20 = -2.528. If the observed t is less than -2.528, then
the decision will be to reject the null hypothesis.
P a g e | 146
d D 1.16 0
sd 1.01
n 21
t = = -5.26
Since the observed value of t = -5.26 is less than the critical t value of
-2.528, the decision is to reject the null hypothesis. The population
difference is less
than zero.
P a g e | 147
1 47 63 -16
2 33 35 -2
3 38 36 2
4 50 56 -6
5 39 44 -5
6 27 29 -2
7 35 32 3
8 46 54 -8
9 41 47 -6
d
= -4.44 sd = 5.703 df = 8
sd 5.703
d t
n 9
= -4.44 + 3.355 = -4.44 + 6.38
If the observed value of z is greater than 1.96 or less than -1.96, then
the decision will be to reject the null hypothesis.
Sample 1 Sample 2
x1 = 345 x2 = 421
n1 = 783 n2 = 896
x1 x 2 345 421
p
n1 n2 783 896
= .4562
x1 345 x 2 421
pˆ 1 pˆ 2
n1 783 n2 896
= .4406 = .4699
Since the observed value of z = -1.20 is greater than -1.96, the decision
is to fail to reject the null hypothesis. There is no significant difference.
n1 = 409 n2 = 378
p̂ p̂
1 = .71 2 = .67
P a g e | 149
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
dfnum = 8 - 1 = 7 dfdenom = 10 - 1 = 9
If the observed value of F is greater than 4.20 or less than .238, then
the decision will be to reject the null hypothesis.
P a g e | 150
2
s1 46
2
s2 37
F = = 1.24
Since the observed F = 1.24 is less than F.025,7,9 =4.20 and greater than
x x
t = $75,000 w = $45,000
st = $22,000 sw = $15,500
nt = 27 nw = 29
df = 27 + 29 - 2 = 54
For a 95% Confidence Level, /2 = .025 and t.025,50 = 2.009 (used
df=50)
2 2
s1 (n1 1) s 2 (n2 1) 1 1
( x1 x 2 ) t
n1 n 2 2 n1 n2
30,000 ± 10,160.11
43 41 2
P a g e | 152
51 49 2
37 44 -7
24 32 -8
47 46 1
44 42 2
50 47 3
55 51 4
46 49 -3
d
n=9 = -0.444 sd =4.447 df = 9 - 1 = 8
sd
d t
n
4.447
9
-0.444 + (1.86) = -0.444 ± 2.757
n1 = 400 n2 = 450
x1 = 220 x2 = 216
P a g e | 153
Ho: p1 - p2 = 0
220 216
p̂1 400 p̂2 450
= = .55 = = .48
x1 x 2 220 216
p
n1 n2 400 450
= .513
Since the observed z = 2.04 is less than z.01 = 2.33, the decision is to
fail to reject
n1 = 16 n2 = 14
x x
1 = 26,400 2 = 25,800
s1 = 1,200 s2 = 1,050
dfnum = 16 – 1 = 15 dfdenom = 14 – 1 = 13
2
s1 1,440,000
2
s2 1,102,500
F = = 1.31
Since the observed F = 1.31 is less than F.025,15,13 = 3.05 and greater
than
n1 = 60 n2 = 41
x x
1 = 631 2 = 848
1 = 100 2 = 100
P a g e | 155
2 2
s1 s
( x1 x 2 ) z 2
n1 n2
100 2 100 2
60 41
(631 – 848) + 1.96 = -217 ± 39.7
Ha: µ1 - µ2 0 df = 20 + 24 - 2 = 42
Detroit Charlotte
P a g e | 156
n1 = 20 n2 = 24
x x
1 = 17.53 2 = 14.89
s1 = 3.2 s2 = 2.7
For two-tail test, /2 = .005 and the critical t.005,40 = ±2.704 (used
df=40)
( x1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n 2 1) 1 1
n1 n2 2 n1 n2
t =
Since the observed t = 2.97 > t.005,40 = 2.704, the decision is to reject
the null
hypothesis.
x x
1 = 38.4 2 = 23.1
1 = 9.8 2 = 7.4
P a g e | 157
n1 = 35 n2 = 35
Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 > 0
Since the observed z = 7.37 > z.01 = 2.33, the decision is to reject the
null
hypothesis.
P a g e | 158
n1 = 350 n2 = 500
p̂ p̂
1 = .75 2 = .52
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
(.75)(.25) (.52)(.48)
350 500
(.75 - .52) + 1.645 = .23 ± .053
dfnum = 6 dfdenom = 7
2
s1 129,569
2
s2 72,909
F = = 1.78
Since F = 1.78 < F.005,6,7 = 9.16 but also > F.995,7,6 = .11, the decision is
to fail to reject the null hypothesis. There is no difference in the
variances of the shifts.
P a g e | 160
54 49 5
55 50 5
59 52 7
53 51 2
54 50 4
61 56 5
51 47 4
53 49 4
d
n=8 = 4.5 sd=1.414 df = 8 - 1 = 7
sd
d t
n
1.414
8
4.5 + 1.895 = 4.5 ± .947
Ha: µ1 - µ2 < 0 df = 23 + 19 - 2 = 40
Wisconsin Tennessee
n1 = 23 n2 = 19
x x
1 = 69.652 2 = 71.7368
( x 1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n2 1) 1 1
n1 n2 2 n1 n2
t =
Since the observed t = -2.44 < t.01,40 = -2.423, the decision is to reject
the null
hypothesis.
P a g e | 162
71 53 18
56 47 9
75 52 23
68 55 13
74 58 16
d
n=5 = 15.8 sd = 5.263 df = 5 - 1 = 4
Ho: D = 0 = .05
Ha: D > 0
d D 15.8 0
sd 5.263
n 5
t = = 6.71
Since the observed t = 6.71 > t.05,4 = 2.132, the decision is to reject
the null
hypothesis.
Ha: p1 - p2 0
P a g e | 163
Machine 1 Machine 2
x1 = 38 x2 = 21
n1 = 191 n2 = 202
x1 38 x2 21
pˆ 1 pˆ 2
n1 191 n2 202
= .199 = .104
For two-tail, /2 = .025 and the critical z values are: z.025 = ±1.96
Since the observed z = 2.64 > zc = 1.96, the decision is to reject the
null
hypothesis.
n1 = 338 n2 = 281
x1 = 297 x2 = 192
P a g e | 164
x1 297 x 2 192
pˆ 1 pˆ 2
n1 338 n2 281
= .879 = .683
pˆ 1 qˆ1 pˆ 2 qˆ 2
( pˆ 1 pˆ 2 ) z
n1 n2
n1 = 33 n2 = 35
x x
1 = 12.4 2 = 4.6
1 = 2.9 2 = 1.8
1 2
2 2
( x1 x 2 ) z
n1 n2
(2.9) 2 (1.8) 2
33 35
(12.4 – 4.6) + 2.575 = 7.8 ± 1.52
x x
1 = $47.20 2 = $27.40
1 = $12.45 2 = $9.82
n1 = 60 n2 = 40
Ho: µ1 - µ2 = 0 = .01
Ha: µ1 - µ2 0
Since the observed z = 8.86 > zc = 2.575, the decision is to reject the
null
P a g e | 166
hypothesis.
P a g e | 167
7 3 4
10 8 2
16 9 7
8 5 3
d
n=5 = 4.0 sd = 1.8708 df = 5 - 1 =
4
Ho: D = 0 = .01
Ha: D > 0
d D 4.0 0
sd 1.8708
n 5
t = = 4.78
Since the observed t = 4.78 > t.01,4 = 3.747, the decision is to reject
the null
hypothesis.
Ha: µ1 - µ2 0 df = 10 + 6 - 2 = 14
A B___
n1 = 10 n2 = 6
x x
1 = 18.3 2 = 9.667
For two-tail test, /2 = .005 and the critical t.005,14 = ±2.977
( x1 x 2 ) ( 1 2 )
2 2
s1 (n1 1) s 2 (n 2 1) 1 1
n1 n2 2 n1 n2
t =
Since the observed t = 4.52 > t.005,14 = 2.977, the decision is to reject
the null
hypothesis.
10.72 A t test was used to test to determine if Hong Kong has significantly
different
Ho: µ1 - µ2 = 0
Ha: µ1 - µ2 0
x x
n1 = 19 n2 = 23 1 = 130.4 2 = 128.4
10.73 H0: D = 0
Ha: D 0
10.74 The point estimates from the sample data indicate that in the
northern city the market share is .31078 and in the southern city the
market share is .27013. The point estimate for the difference in the
two proportions of market share are .04065. Since the 99% confidence
interval ranges from -.03936 to +.12067 and zero is in the interval, any
hypothesis testing decision based on this interval would result in failure
to reject the null hypothesis. Alpha is .01 with a two-tailed test. This is
underscored by an observed z value of 1.31 which has an associated p-
value of .191 which, of course, is not significant for any of the usual
values of .
P a g e | 171
Twenty-six pipes were measured for sample one and twenty-eight pipes
were measured for sample two. The observed F = 2.0575 is significant
at = .05 for a one-tailed test since the associated p-value is .034787.
The variance of pipe lengths for machine 1 is significantly greater than
the variance of pipe lengths for machine 2.
P a g e | 172
Chapter 11
Analysis of Variance and
Design of Experiments
LEARNING OBJECTIVES
The focus of this chapter is learning about the design of experiments and the
analysis of variance thereby enabling you to:
This important chapter opens the door for students to a broader view
of statistics than they have seen to this time. Through the topic of
experimental designs, the student begins to understand how they can
scientifically set up controlled experiments in which to test certain
hypotheses. They learn about independent and dependent variables. With
the completely randomized design, the student can see how the t test for two
independent samples can be expanded to include three or more samples by
using analysis of variance. This is something that some of the more curious
students were probably wondering about in chapter 10. Through the
randomized block design and the factorial designs, the student can
understand how we can analyze not only multiple categories of one variable,
but we can simultaneously analyze multiple variables with several categories
each. Thus, this chapter affords the instructor an opportunity to help the
student develop a structure for statistical analysis.
CHAPTER OUTLINE
Sizes
Applications
Interaction
P a g e | 176
KEY TERMS
a posteriori Factors
Factorial Design
11.1 a) Time Period, Market Condition, Day of the Week, Season of the Year
11.2 a) Type of 737, Age of the plane, Number of Landings per Week of the
plane,
maintenance
P a g e | 178
11.4 Average dollar expenditure per day/night, Age of adult registering the
family, Number of days stay (consecutive)
P a g e | 179
Total 16 36.24
Since the observed F = 11.07 > F.05,2,14 = 3.74, the decision is to reject
the null hypothesis.
Total 22 120.43
Since the observed F = 15.82 > F.01,4,18 = 4.58, the decision is to reject
the null
hypothesis.
P a g e | 180
11.7 Source df SS MS F_
Total 15 711.8
Since the observed F = 13.00 > F.01,3,12 = 5.95, the decision is to reject
the null hypothesis.
P a g e | 181
Total 13 107.71
Since the observed F = 17.76 > F.05,1,12 = 4.75, the decision is to reject
the null hypothesis.
1 2
n1 = 7 n2 = 7
x x
1 = 29 2 = 24.71429
F 17.76
Also, t = = 4.214
P a g e | 182
Total 1,555.57 54
11.10 Source SS df MS F _
Total 98.06 16
F.05,2,14 = 3.74
Since the observed F = 3.03 < F.05,2,14 = 3.74, the decision is to fail to
reject
Total 18 .010579
Since the observed F = 10.10 > F.01,3,15 = 5.42, the decision is to reject
the null
hypothesis.
Total 14 192400000
Since the observed F = 92.67 > F.01,2,12 = 6.93, the decision is to reject
the null
hypothesis.
P a g e | 184
Total 17 48.50
Since the observed F = 11.76 > F.05,2,15 = 3.68, the decison is to reject
the null
hypothesis.
P a g e | 185
Total 19 677400
Since the observed F = 11.03 > F.05,3,16 = 3.24, the decision is to reject
the null
hypothesis.
11.15 There are 4 treatment levels. The sample sizes are 18, 15, 21, and
11. The F
alpha of .05. The means are 226.73, 238.79, 232.58, and 239.82.
11.16 The independent variable for this study was plant with five
classification levels
(the five plants). There were a total of 43 workers who participated in
the study. The dependent variable was number of hours worked per
P a g e | 186
x x
q.05,6,40 = 4.23 n3 = 8 n6 = 7 3 = 15.85 6 = 17.2
.3352 1 1
2 8 7
HSD = 4.23 = 0.896
x 3 x 6 15.85 17.21
= 1.36
groups 3 and 6.
MSE 2.389
n 6
HSD = q = (3.96) = 2.50
x x
q.05,3,14 = 3.70 n1 = 6 n2 = 5 1 =2 2 = 4.6
1.002381 1 1
2 6 5
HSD = 3.70 = 1.586
x 1 x 2 2 4 .6
= 2.6
groups 1 and 2.
P a g e | 188
1.481481 1 1
2 5 5
HSD = 5.38 = 2.93
x x
2 = 10 4 = 16
x 3 x 6 10 16
= 6
groups 2 and 4.
P a g e | 189
MSE 13.95833
n 4
HSD = q = 5.50 = 10.27
x x x x
1 = 115.25 2 = 125.25 3 = 131.5 4 = 122.5
x x
1 and 3 are the only pair that are significantly different
using the
HSD test.
P a g e | 190
MSE 3.619048
n 7
HSD = q = 3.08 = 2.215
x x
1 = 29 and 2 = 24.71429
x x
Since 1 - 2 = 4.28571 > HSD = 2.215, the decision is to reject the
null
hypothesis.
P a g e | 191
q.01,4,15 = 5.25 n1 = 4 n2 = 6 n3 = 5 n4 = 4
x x x x
1 = 4.03, 2 = 4.001667, 3 = 3.974, 4 = 4.005
.000234 1 1
2 4 6
HSD1,2 = 5.25 = .0367
.000234 1 1
2 4 5
HSD1,3 = 5.25 = .0381
.000234 1 1
2 4 4
HSD1,4 = 5.25 = .0402
.000234 1 1
2 6 5
HSD2,3 = 5.25 = .0344
.000234 1 1
2 6 4
HSD2,4 = 5.25 = .0367
P a g e | 192
.000234 1 1
2 5 4
HSD3,4 = 5.25 = .0381
x1 x 3
= .056
MSE 975,000
n 5
HSD = q = 5.04 = 2,225.6
x x x
1 = 40,900 2 = 49,400 3 = 45,300
x1 x 2
= 8,500
x1 x 3
= 4,400
x 2 x3
= 4,100
q.05,3,15 = 3.67 n1 = 5 n2 = 7 n3 = 6
P a g e | 194
x x x
1 = 7.6 2 = 8.8571 3 = 5.8333
1.259365 1 1
2 5 7
HSD1,2 = 3.67 = 1.705
1.259365 1 1
2 5 6
HSD1,3 = 3.67 = 1.763
1.259365 1 1
2 7 6
HSD2,3 = 3.67 = 1.620
x1 x 3
= 1.767 (is significant)
x 2 x3
= 3.024 (is significant)
P a g e | 195
x x x x
1 = 591 2 = 350 3 = 776 4 = 563
MSE 13,798.13
n 5
HSD = q = 4.05 = 212.76
x1 x 2 x1 x 3 x1 x 4
= 241 = 185 = 28
x2 x3 x2 x4 x3 x4
= 426 = 213 = 213
11.27 = .05. There were five plants and ten pairwise comparisons. The
MINITAB
11.28 H0: µ1 = µ2 = µ3 = µ4
Source df SS MS F__
Total 19 365.75
For treatments, the observed F = 5.56 > F.05,3,12 = 3.49, the decision is
to
11.29 H0: µ1 = µ2 = µ3
Source df SS MS F_
Total 11 .082067
P a g e | 198
For treatments, the observed F = 1.48 < F.01,2,6 = 10.92 and the
decision is to
Total 59 17319.39
For treatments, the observed F = 1.91 < F.05,5,45 = 2.45 and decision is
to fail to
Total 27 771.31
For treatments, the observed F = 3.90 < F.01,3,18 = 5.09 and the decision
is to
Total 39 9027.5
For treatments, the observed F = 15.67 > F.05,3,27 = 2.96 and the
decision is to
Total 14 218.9300
For treatments, the observed F = 15.37 > F.01,2,8 = 8.65 and the
decision is to
11.35 The p value for Phone Type, .00018, indicates that there is an overall significant
difference in treatment means at alpha .001. The lengths of calls differ according to type
of telephone used. The p-value for managers, .00028, indicates that there is an overall
difference in block means at alpha .001. The lengths of calls differ according to Manager.
The significant blocking effects have improved the power of the F test for treatments.
11.36 This is a two-way factorial design with two independent variables and
one dependent variable. It is 2x4 in that there are two row treatment
levels and four column treatment levels. Since there are three
measurements per cell, interaction can be analyzed.
11.37 This is a two-way factorial design with two independent variables and
one dependent variable. It is 4x3 in that there are four treatment
P a g e | 202
levels and three column treatment levels. Since there are two
measurements per cell, interaction can be analyzed.
Total 79 1278.94
= .05
Critical F.05,3,60 = 2.76 for rows. For rows, the observed F = 3.46 >
F.05,3,60 = 2.76
Critical F.05,4,60 = 2.53 for columns. For columns, the observed F = 0.77
<
Total 23 12.632
= .05
Critical F.05,1,16 = 4.49 for rows. For rows, the observed F = 2.40 <
F.05,1,16 = 4.49
Critical F.05,3,16 = 3.24 for columns. For columns, the observed F = 2.94
<
Total 11 86.250
= .01
Critical F.01,1,6 = 13.75 for rows. For rows, the observed F = 38.37 >
Critical F.01,2,6 = 10.92 for columns. For columns, the observed F = 4.42
<
Total 31 6.92719
= .05
Total 23 107.9583
= .05
Critical F.05,3,12 = 3.49 for Age. For Age, the observed F = 14.77 >
Critical F.05,2,12 = 3.89 for No. Children. For No. Children, the observed
F = 25.61 > F.05,2,12 = 3.89 and the decision is to reject the null
hypothesis.
Total 35 3925.22
P a g e | 208
= .05
Critical F.05,2,24 = 3.40 for rows. For rows, the observed F = 34.31 >
11.44 This two-way design has 3 row treatments and 5 column treatments.
There are 45 total observations with 3 in each cell.
MS R 46.16
MS E 3.49
FR = = 13.23
p-value = .000 and the decision is to reject the null hypothesis for
rows.
MSC 249.70
MS E 3.49
FC = = 71.57
p-value = .000 and the decision is to reject the null hypothesis for
columns.
MS I 55.27
MS E 3.49
FI = = 15.84
p-value = .000 and the decision is to reject the null hypothesis for
interaction.
11.45 The null hypotheses are that there are no interaction effects, that there
are no significant differences in the means of the valve openings by
machine, and that there are no significant differences in the means of
the valve openings by shift. Since the p-value for interaction effects
is .876, there are no significant interaction effects and that is good
since significant interaction effects would confound that study. The p-
value for columns (shifts) is .008 indicating that column effects are
significant at alpha of .01. There is a significant difference in the mean
valve opening according to shift. No multiple comparisons are given in
the output. However, an examination of the shift means indicates that
the mean valve opening on shift one was the largest at 6.47 followed
by shift three with 6.3 and shift two with 6.25. The p-value for rows
(machines) is .937 and that is not significant.
P a g e | 211
11.46 This two-way factorial design has 3 rows and 3 columns with three
observations
per cell. The observed F value for rows is 0.19, for columns is 1.19,
and for
interaction is 1.40. Using an alpha of .05, the critical F value for rows
and columns (same df) is F2,18,.05 = 3.55. Neither the observed F value
for rows nor the observed F value for columns is significant. The
critical F value for interaction is F4,18,.05 = 2.93. There is no significant
interaction.
Since the treatment F = 8.82 > F.05,3,12 = 3.49, the decision is to reject
the null
hypothesis.
P a g e | 212
q.05,4,12 = 4.20
MSE 2.52
n 4
HSD = q = (4.20) = 3.33
x x x x
1 = 12 2 = 7.75 3 = 13.25 4 = 11.25
Total 25 317.80
P a g e | 213
Total 41 865
Since the observed F = 16.19 > F.01,2,22 = 5.72, the decision is to reject
the null
hypothesis.
x x x
1 = 9.200 2 = 14.250 3 = 8.714286
n1 = 10 n2 = 8 n3 = 7
P a g e | 214
4.66 1 1
2 10 8
HSD1,2 = 4.64 = 3.36
4.66 1 1
2 10 7
HSD1,3 = 4.64 = 3.49
4.66 1 1
2 8 7
HSD2,3 = 4.64 = 3.67
x1 x 2 x 2 x3
= 5.05 and = 5.5357 are significantly
different at = .01
P a g e | 215
of freedom block are five. The error degrees of freedom are ten. The
total
Total 39 71,338
Since the observed F = 5.58 > F.05,3,27 = 2.96 for treatments, the
decision is to
Total 23
Since for treatments the observed F = 31.51 > F.05,3,15 = 3.29, the
decision is to
Ignoring the blocking effects, the sum of squares blocking and sum of
squares error are combined together for a new SS error = 548.708 +
38.125 = 586.833. Combining the degrees of freedom error and
blocking yields a new dferror = 20. Using these new figures, we compute
a new mean square error, MSE = (586.833/20) = 29.34165.
MSE 29.34165
n 6
HSD = q = (3.96) = 8.757
x x x x
1 = 16.667 2 = 12.333 3 = 12.333 4 = 19.833
None of the pairs of means are significantly different using Tukey's HSD
= 8.757.
This may be due in part to the fact that we compared means by folding
the
blocking effects back into error and the blocking effects were highly
significant.
Total 39 225.67
= .05
P a g e | 218
should not be done in the usual manner. However, in this case, there
are no
Total 35 330.556
= .01
Critical F.01,3,24 = 4.72 for treatment 2. For the treatment 2 effects, the
observed
F = 38.21 > F.01,3,24 = 4.72 and the decision is to reject the null
hypothesis.
Critical F.01,2,24 = 5.61 for Treatment 1. For the treatment 1 effects, the
observed
F = 0.23 < F.01,2,24 = 5.61 and the decision is to fail to reject the null
hypothesis.
Critical F.01,6,24 = 3.67 for interaction. For the interaction effects, the
observed
F = 1.30 < F.01,6,24 = 3.67 and the decision is to fail to reject the null
hypothesis.
P a g e | 220
Total 35 67.2222
= .05
Critical F.05,2,24 = 3.40 for Age. For the age effects, the observed F =
38.65 >
Critical F.05,3,24 = 3.01 for Region. For the region effects, the observed F
= 0.64
< F.05,3,24 = 3.01 and the decision is to fail to reject the null
hypothesis.
F = 0.33 < F.05,6,24 = 2.51 and the decision is to fail to reject the null
hypothesis.
There are no significant interaction effects. Only the Age effects are
significant.
P a g e | 221
x x x
1 = 2.667 2 = 4.917 3 = 2.250
n = 12 C=3 N = 36 N - C = 33
q.05,3,33 = 3.49
MSE .5404
n 12
HSD = q = (3.49) = 0.7406
Shown below is a graph of the interaction using the cell means by Age.
P a g e | 222
P a g e | 223
Total 23 172000000
The treatment F = 7.38 > F.05,3,20 = 3.10 and the decision is to reject
the null
hypothesis.
Total 17 516,074
Total 20 194.6885
Since the treatment F = 0.46 > F.05,2,18 = 3.55, the decision is to fail to
reject the
null hypothesis.
Total 47 41.250
= .05
Critical F.05,2,36 = 3.32 for Years. For Years, the observed F = 5.16 >
Critical F.05,3,36 = 2.92 for Size. For Size, the observed F = 12.06 >
F.05,3,36 = 2.92
effects at = .05.
P a g e | 226
Total 39 97.900
For treatments, the observed F = 13.64 > F.05,4,28 = 2.71 and the
decision is to
11.62 This is a one-way ANOVA with four treatment levels. There are 36
observations
that the sample sizes are different with sizes of 8, 7, 11, and 10,
respectively. No
However, a study of sample means shows that the two most extreme
means are
from levels one and four. These two means would be the most likely
candidates
for multiple comparison tests. Note that the confidence intervals for
means one
this is a random block design. Neither the row nor the column p-values
are less
11.64 This is a two-way ANOVA with 5 rows and 2 columns. There are 2
observations
per cell. For rows, FR = 0.98 with a p-value of .461 which is not
significant. For columns, FC = 2.67 with a p-value of .134 which is not
significant. For interaction, FI = 4.65 with a p-value of .022 which is
significant at = .05. Thus, there are significant interaction effects
and the row and column effects are confounded. An examination of
the interaction plot reveals that most of the lines cross verifying the
finding of significant interaction.
11.65 This is a two-way ANOVA with 4 rows and 3 columns. There are 3
observations
11.66 This was a random block design with 5 treatment levels and 5 blocking
levels.
For both treatment and blocking effects, the critical value is F.05,4,16 =
3.01. The
is greater than the critical value. The null hypothesis for treatments is
rejected,
for blocking effects is MSR / MSE = 10.36 /7.36 = 1.41 which is less than
the
11.67 This one-way ANOVA has 4 treatment levels and 24 observations. The
F = 3.51
output shows that means 1 and 3 are the only pairs that are
significantly different
Chapter 12
Analysis of Categorical Data
LEARNING OBJECTIVES
CHAPTER OUTLINE
KEY TERMS
( fo fe )2
fe
12.1 f0 fe
53 68 3.309
37 42 0.595
32 33 0.030
28 22 1.636
18 10 6.400
15 8 6.125
( f0 fe )2
2
fe
Observed = 18.095
df = k - 1 = 6 - 1 = 5, = .05
2.05,5 = 11.0705
P a g e | 235
null hypothesis.
The observed frequencies are not distributed the same as the expected
frequencies.
P a g e | 236
( fo fe )2
fe
12.2 f0 fe
19 18 0.056
17 18 0.056
14 18 0.889
18 18 0.000
19 18 0.056
21 18 0.500
18 18 0.000
18 18 0.000
x
f 0
144
k 8
= 18
df = k – 1 = 8 – 1 = 7, = .01
2.01,7 = 18.4753
P a g e | 237
( f0 fe )2
2
fe
Observed = 1.557
distributed.
P a g e | 238
0 28 0
1 17 17
2 11 22
3 _5 15
61
54
54
61
= =0.9
Expected Expected
0 .4066 24.803
1 .3659 22.320
2 .1647 10.047
( f o fe )2
fe
Number fo fe
0 28 24.803 0.412
P a g e | 239
1 17 22.320 1.268
61 60.993 2.004
df = k - 2 = 3 - 2 = 1, = .05
2.05,1 = 3.8415
( f0 fe )2
2
fe
Observed = 2.001
Since the observed 2 = 2.001 < 2.05,1 = 3.8415, the decision is to fail
to reject
12.4
10-20 6 15 90 1,350
x
fm 5,715
f 129
= 44.3
( fM ) 2 (5,715) 2
fM 2 n
279,825
129
n 1 128
s= = 14.43
10 44.3
14.43
z = = -2.38 .4913
20 44.3
14.43
z = = -1.68 - .4535
30 44.3
14.43
z= = -0.99 -.3389
40 44.3
14.43
z = = -0.30 -.1179
50 44.3
14.43
z = = 0.40 +.1554
60 44.3
14.43
z = = 1.09 .3621
70 44.3
14.43
z = = 1.78 .4625
80 44.3
14.43
z = = 2.47 .4932
( fo fe )2
fe
Category fo fe
2.001
( f0 fe )2
2
fe
Calculated = 2.001
df = k - 3 = 7 - 3 = 4, = .05
2.05,4 = 9.4877
Since the observed 2 = 2.004 < 2.05,4 = 9.4877, the decision is to fail
to reject
( fo fe )2
fe
12.5 Definition fo Exp.Prop. fe
Achievement/
P a g e | 246
227 198.48
frequencies.
Ha: The observed frequencies are not distributed the same as the
expected
frequencies.
Observed 2 = 198.48
df = k – 1 = 4 – 1 = 3, = .05
2.05,3 = 7.8147
null hypothesis.
The observed frequencies for men are not distributed the same
as the
( fo fe )2
fe
212 1.13
expected frequencies.
expected frequencies.
= .01, df = k - 1 = 6 - 1 = 5
2.01,5 = 15.0863
Since the observed 2 = 1.13 < 2.01,5 = 15.0863, the decision is to fail
to reject
P a g e | 249
x
fM
9,155
n 231
= 39.63
( fM ) 2 (9,155) 2
fM 2 n
405,375
231
n 1 230
s = = 13.6
10 39.63
13.6
z = = -2.18 .4854
20 39.63
13.6
z = = -1.44 -.4251
P a g e | 251
30 39.63
13.6
z = = -0.71 -.2611
40 39.63
13.6
z = = 0.03 +.0120
50 39.63
13.6
z = = 0.76 .2764
P a g e | 252
60 39.63
13.6
z = = 1.50 .4332
70 39.63
13.6
z = = 2.23 .4871
Age Probability fe
( fo fe )2
fe
Age fo fe
2.45
df = k - 3 = 6 - 3 = 3, = .05
P a g e | 255
2.05,3 = 7.8147
Observed 2 = 2.45
hypothesis.
distributed.
P a g e | 256
0 18 0
1 28 28
2 47 94
3 21 63
4 16 64
5 11 55
6 or more 9 54
f number 358
f 150
= = 2.4
Number Probability fe
2 .2613 39.20
3 .2090 31.35
4 .1254 18.81
5 .0602 9.03
( f0 fe )2
f0
fo fe
18 13.61 1.42
28 32.66 0.66
47 39.20 1.55
21 31.35 3.42
16 18.81 0.42
11 9.03 0.43
9 5.36 2.47
10.37
Since the observed 2 = 10.37 < 2.01,5 = 15.0863, the decision is to fail
to reject
Ha: p .28
P a g e | 258
( fo f e )2
fe
fo fe
2.025,1 = 5.02389
( fo fe )2
fe
f0 fe
= .05 df = k - 1 = 2 - 1 = 1
2.05,1 = 3.8415
12.11
Variable Two
(271)(178) (436)(178)
707 707
e21 = = 68.23 e22 = = 109.77
Variable Two
(109.7
(68.23) 7) 178
68 110
P a g e | 261
271 707
436
(110 109.77) 2
109.77
=
2.01,1 = 6.6349
Since the observed 2 = 0.00 < 2.01,1 = 6.6349, the decision is to fail
to reject
12.12
Variable
Two
Variabl 24 13 47 58
e 142
93 59 187 244
One
583
117 72 234
302 725
(583)(117) (583)(72)
725 725
e21 = = 94.08 e22 = = 57.90
Variable Two
117 72 234
302 725
12.13
Social Class
Lower Middle
Upper
Numbe
r 0 7 18 6 31
of 1 9 38 23 70
Childre 2 or 34 97 58 189
n 3
47 31 30 108
>3
97 184 398
117
(31)(97) (189)(97)
398 398
e11 = = 7.56 e31 = = 46.06
(31)(184) (189)(184)
398 398
e12 = = 14.3 e32 = = 87.38
(31)(117) (189)(117)
398 398
e13 = = 9.11 e33 = = 55.56
P a g e | 265
(70)(97) (108)(97)
398 398
e21 = = 17.06 e41 = = 26.32
(70)(184) (108)(184)
398 398
e22 = = 32.36 e42 = = 49.93
(70)(117) (108)(117)
398 398
e23 = = 20.58 e43 = = 31.75
Social Class
Lower Middle
Upper
>3 34 97 58
108
(26.3 (49.9 (31.7
2) 3) 5)
47 31 30
97 184 398
117
P a g e | 266
.04 + .94 + 1.06 + 3.81 + .98 + .28 + 3.16 + 1.06 + .11 + 16.25
+
2.05,6 = 12.5916
null hypothesis.
12.14
NE 140 32 5 18 195
632
428 100 65
39
(195)(100) (235)(39)
632 632
e12 = = 30.85 e24 = = 14.50
P a g e | 268
(195)(39) (202)(100)
632 632
e14 = = 12.03 e32 = = 31.96
(235)(100) (202)(39)
632 632
e22 = = 37.18 e34 = = 12.47
134 41 52 8
154 27 8 13
P a g e | 269
428 100 65
39
2.01,6 = 16.8119
12.15
P a g e | 270
Transportation Mode
Industr Publishing 32 12 41 85
y
Comp.Har 5 6 24 35
d.
120
37 18 65
(85)(37) (35)(37)
120 120
e11 = = 26.21 e21 = = 10.79
(85)(18) (35)(18)
120 120
e12 = = 12.75 e22 = = 5.25
(85)(65) (35)(65)
120 120
e13 = = 46.04 e23 = = 18.96
Transportation Mode
2.05,2 = 5.9915
12.16
P a g e | 272
Number of
Bedrooms
<2 3 >4
2.10,2 = 4.6052
12.17
Mexican Citizens
Yes No
Type Dept. 24 17 41
of Disc. 20 15 35
Hard. 11 19 30
Store Shoe 32 28
60
87 79 166
(41)(87) (30)(87)
166 166
e11 = = 21.49 e31 = = 15.72
(41)(79) (30)(79)
166 166
e12 = = 19.51 e32 = = 14.28
(35)(87) (60)(87)
166 166
e21 = = 18.34 e41 = = 31.45
(35)(79) (60)(79)
166 166
e22 = = 16.66 e42 = = 28.55
P a g e | 275
Mexican Citizens
Yes No
20 15
11 19
60
Shoe (31.4 (28.5
5) 5)
32 28
87 79 166
2.05,3 = 7.8147
Since the observed 2 = 3.93 < 2.05,3 = 7.8147, the decision is to fail
to
12.18 = .01, k = 7, df = 6
Use:
( f0 fe )2
2
fe
( fo fe )2
fe
fo fe (f0-fe)2
3.100
( f0 fe )2
2
fe
= 3.100
Since the observed value of 2 = 3.1 < 2.01,6 = 16.8119, the decision
is to fail to
expected distribution.
12.19
Variable 2
12 23 21 56
P a g e | 278
Variable 8 17 20 45
1
7 11 18 36
27 51 59 137
(18 15.50) 2
15.50
=
2.05,4 = 9.4877
P a g e | 279
12.20
Location
NE W S
(413)(157) (417)(157)
830 830
e13 = = 78.12 e23 = = 78.88
Location
NE W S
2.10,2 = 4.6052
null hypothesis.
fo = 1,054
f 0
1,054
no.kinds 6
If cookie sales are uniformly distributed, then fe = =
175.67
( fo fe )2
fe
fo fe
14.91
P a g e | 283
= .05 df = k - 1 = 6 - 1 = 5
2.05,5 = 11.0705
null hypothesis.
12.22
Gender
M F
(272)(1,018) (27)(1,049)
2,067 2,067
e11 = = 133.96 e12 = = 138.04
(1,795)(1,018) (1,795)(1,049)
2,067 2,067
e21 = = 884.04 e22 = = 910.96
Gender
M F
N (884.0 (910.9
4) 6)
1,795
811 984
P a g e | 285
(984 910.96) 2
910.96
= 39.82 + 38.65 + 6.03 + 5.86 = 90.36
2.05,1 = 3.8415
null hypothesis.
0 26 0
1 40 40
2 57 114
3 32 96
4 17 68
5 12 60
6 8 48
( f )(arrivals ) 426
0
f 0 192
= = 2.2
Arrivals Probability fe
2 .2681 51.48
3 .1966 37.75
4 .1082 20.77
5 .0476 9.14
6 .0249 4.78
P a g e | 287
( fo fe )2
fe
fo fe
26 21.27 1.05
40 46.81 0.99
57 51.48 0.59
32 37.75 0.88
17 20.77 0.68
12 9.14 0.89
8 4.78 2.17
7.25
Observed 2 = 7.25
= .05 df = k - 2 = 7 - 2 = 5
2.05,5 = 11.0705
Since the observed 2 = 7.25 < 2.05,5 = 11.0705, the decision is to fail
to reject
expected frequencies.
( fo fe )2
fe
fo = 1,726
22.83
P a g e | 289
Observed 2 = 22.83
= .05 df = k - 1 = 8 - 1 = 7
2.05,7 = 14.0671
null hypothesis.
The observed frequencies are not distributed the same as the expected
frequencies
12.25
Position
Syste
ms
Manag Programm Operat Analys
er er or t
0-3 6 37 11 13 67
4-8 28 16 23 24 91
Years
>8 47 10 12 19 88
81 63 46 56 246
(67)(63) (91)(56)
246 246
e12 = = 17.16 e24 = = 20.72
(67)(56) (88)(63)
246 246
e14 = = 15.25 e32 = = 22.54
P a g e | 291
(91)(63) (88)(56)
246 246
e22 = = 23.30 e34 = = 20.03
Position
Syste
ms
Manag Programm Operat Analys
er er or t
0-3 (17.16) 67
(22.06) (12.53) (15.25
37 )
Years 6 11
13
4-8 (23.30) 91
(29.96) (17.02) (20.72
16 )
28 23
24
>8 (22.54) 88
(28.98) (16.46) (20.03
10 )
47 12
19
81 63 46 56 246
P a g e | 292
2.01,6 = 16.8119
( fo fe )2
fe
fo fe
More Work,
2.025,1 = 5.0239
hypothesis.
12.27
P a g e | 294
Type of College or
University
(234)(127) (93)(127)
571 571
e11 = = 52.05 e31 = = 20.68
(234)(387) (193)(387)
571 571
e12 = = 158.60 e32 = = 63.03
(234)(57) (93)(57)
571 571
e13 = = 23.36 e33 = = 9.28
P a g e | 295
(202)(127) (42)(127)
571 571
e21 = = 44.93 e41 = = 9.34
(202)(387) (42)(387)
571 571
e22 = = 136.91 e42 = = 28.47
(202)(57) (42)(57)
571 571
e23 = = 20.16 e43 = = 4.19
Type of College or
University
31 54 8
22 14 6
2.05,6 = 12.5916
null hypothesis.
12.28 The observed chi-square is 30.18 with a p-value of .0000043. The chi-
square
goodness-of-fit test indicates that there is a significant difference
between the observed frequencies and the expected frequencies. The
distribution of responses to the question is not the same for adults
between 21 and 30 years of age as they are for others. Marketing and
sales people might reorient their 21 to 30 year old efforts away from
home improvement and pay more attention to leisure travel/vacation,
clothing, and home entertainment.
12.29 The observed chi-square value for this test of independence is 5.366.
The associated p-value of .252 indicates failure to reject the null
hypothesis. There is not enough evidence here to say that color choice
is dependent upon gender. Automobile marketing people do not have
to worry about which colors especially appeal to men or to women
because car color is independent of gender. In addition, design and
production people can determine car color quotas based on other
variables.
Chapter 13
P a g e | 298
Nonparametric Statistics
LEARNING OBJECTIVES
3. Know when and how to use the Mann-Whitney U Test, the Wilcoxon
matched-pairs signed rank test, the Kruskal-Wallis test, and the
Friedman test.
Chapter 13 contains new six techniques for analysis. Only the first
technique, the runs test, is conceptually a different idea for the student to
consider than anything presented in the text to this point. The runs test is a
mechanism for testing to determine if a string of data are random. There is a
P a g e | 299
runs test for small samples that uses Table A.12 in the appendix and a test for
large samples, which utilizes a z test.
CHAPTER OUTLINE
P a g e | 300
Small-Sample Case
Large-Sample Case
KEY TERMS
Nonparametric Statistics
Wilcoxon Matched-Pairs Signed Rank Test
P a g e | 302
SOLUTIONS TO CHAPTER 13
= .05, The lower tail critical value is 6 and the upper tail critical value
is 16
n1 = 10 n2 = 10
R = 11
n1 = 26 n2 = 21 n = 47
R=9
R R 9 24.234
z
R 3.351
= -4.55
Since the observed value of z = -4.55 < z.025 = -1.96, the decision is to
reject the
13.3 n1 = 8 n2 = 52 = .05
This is a two-tailed test and /2 = .025. The p-value from the
printout is .0264.
Since the p-value is the lowest value of “alpha” for which the null
hypothesis can
(p-value = .0264 > .025). There is not enough evidence to reject that
the data are randomly generated.
is 14.333. The p-value for this test is .1452. Thus, the test is not
significant at alpha of .05 or .025 for a two-tailed test. The decision is to
fail to reject the
observed value of z is greater than 1.96 or less than -1.96, the decision
is to reject the null hypothesis.
R = 27 n1 = 40 n2 = 24
R R 27 31
z
R 3.716
= -1.08
Since the observed z of -1.08 is greater than the critical lower tail z
value
n1 = 5 n2 = 8 n = 13 = .05
R=4
Since R = 4 > than the lower critical value of 3 and less than the upper
critical
Use the small sample Mann-Whitney U test since both n1, n2 < 10, = .
05. Since this is a two-tailed test, /2 = .025. The p-value is obtained
using Table A.13.
11 1 1
P a g e | 307
13 2.5 1
13 2.5 2
14 4 2
15 5 1
17 6 1
18 7.5 1
18 7.5 2
21 9.5 1
21 9.5 2
22 11 1
23 12.5 2
23 12.5 2
24 14 2
26 15 1
29 16 1
n1 = 8
n2 = 8
n1 (n1 1) (8)(9)
U n1 n2 W1 (8)(8) 62.5
2 2
= 37.5
U ' n1 n2 U
= 64 – 37.5 = 26.5
P a g e | 308
Since this p-value is greater than /2 = .025, the decision is to fail to
reject the
null hypothesis.
13.8 Ho: Population 1 has values that are no greater than population 2
203 1 2
208 2 2
209 3 2
211 4 2
214 5 2
216 6 1
217 7 1
218 8 2
219 9 2
222 10 1
223 11 2
224 12 1
P a g e | 309
227 13 2
229 14 2
230 15.5 2
230 15.5 2
231 17 1
236 18 2
240 19 1
241 20 1
248 21 1
255 22 1
256 23 1
283 24 1
n1 = 11
n2 = 13
W1 = 6 + 7 + 10 + 12 + 17 + 19 + 20 + 21 + 22 + 23 + 24 =
W1 = 181
n1 n2 (11)(13)
2 2
= 71.5
n1 (n1 1) (11)(12)
U n1 n2 W1 (11)(13 181
2 2
= 28
U 28 71.5
z
17.26
= -2.52
2.52
Since z = = 2.52 > z = 2.33, the decision is to reject the
null
hypothesis.
6 1 1
8 2 1
9 3.5 1
9 3.5 2
10 5 2
11 6.5 1
11 6.5 1
P a g e | 311
12 8.5 1
12 8.5 2
13 11 1
13 11 2
13 11 2
14 13 2
15 14 2
16 15 2
17 16 2
W1 = 39
n1 ( n1 1) (7)(8)
U1 n1 n2 W1 (7)(9) 39
2 2
= 52
U 2 n1 n2 U1
= (7)(9) – 52 = 11
U = 11
From Table A.13, the p-value = .0156. Since this p-value is greater
than = .01,
1950 1 U
2050 2 R
2075 3 R
2110 4 U
2175 5 U
2200 6 U
2480 7 U
2490 8 R
2540 9 U
2585 10 R
2630 11 U
2655 12 U
2685 13 R
2710 14 U
2750 15 U
2770 16 R
2790 17 R
2800 18 R
2850 19.5 U
2850 19.5 U
2975 21 R
2995 22.5 R
P a g e | 313
2995 22.5 R
3100 24 R
n1 = 12
n2 = 12
W1 = 1 + 4 + 5 + 6 + 7 + 9 + 11 + 12 + 14 + 15 + 19.5 + 19.5 =
123
n1 n2 (12)(12)
2 2
= 72
n1 (n1 1) (12)(13)
U n1 n2 W1 (12)(12) 123
2 2
= 99
U 99 72
z
17.32
= 1.56
z.025 = +1.96
P a g e | 314
Since the observed z = 1.56 < z.025 = 1.96, the decision is to fail to
reject the
null hypothesis.
P a g e | 315
$28,900 1 F
31,400 2 F
36,600 3 F
40,000 4 F
40,500 5 F
41,200 6 F
42,300 7 F
42,500 8 F
44,500 9 F
45,000 10 M
47,500 11 F
47,800 12.5 F
47,800 12.5 M
48,000 14 F
50,100 15 M
51,000 16 M
51,500 17.5 M
51,500 17.5 M
53,850 19 M
55,000 20 M
57,800 21 M
P a g e | 316
61,100 22 M
63,900 23 M
n1 = 11 n2 = 12
n1 (n1 1) (11)(12)
U n1 n2 W1 (11)(12) 193.5
2 2
= 4.5
U 4.5 66
z
16.25
= -3.78
Since the observed z = 3.78 > z.01 = 2.33, the decision is to reject
the null
hypothesis.
Hartford
Hartford
132,405 1 D
134,127 2 H
134,157 3 D
134,514 4 H
135,062 5 D
135,238 6 H
135,940 7 D
136,333 8 H
136,419 9 H
136,981 10 D
137,016 11 D
137,359 12 H
137,741 13 H
137,867 14 H
138,057 15 D
139,114 16 H
139,638 17 D
140,031 18 H
140,102 19 D
140,479 20 D
P a g e | 318
141,408 21 D
141,730 22 D
141,861 23 D
142,012 24 H
142,136 25 H
143,947 26 H
143,968 27 H
144,500 28 H
n1 = 13
n2 = 15
W1 = 1 + 3 + 5 + 7 + 10 + 11 + 15 + 17 + 19 +
20 + 21 + 22 + 23 = 174
n1 ( n1 1) (13)(14)
U n1 n2 W1 (13)(15) 174
2 2
= 112
n1 n2 (13)(15)
2 2
= 97.5
U 112 97.5
z
21.708
= 0.67
z.025 = + 1.96.
Since the observed z = 0.67 < z.025 = 1.96, the decision is to fail
to reject
1 2 d Rank
212 179 33 15
234 184 50 16
194 189 5 6
234 212 22 11
225 221 4 5
218 217 1 1
234 208 26 12
196 198 -2 -2
213 201 12 10
n = 16
T- = 3.5 + 3.5 + 2 + 9 = 18
(n)( n 1) (16)(17)
4 4
= 68
T 18 68
z
19.34
= -2.59
Since the observed z = -2.59 < z.05 = -1.645, the decision is to reject
the null
hypothesis.
13.14 Ho: Md = 0
Ha: Md 0
49 43 6 +9
41 29 12 +12
47 30 17 +14
39 38 1 + 1.5
53 40 13 +13
51 43 8 +10
51 46 5 + 7.5
49 40 9 +11
38 42 -4 - 5.5
54 50 4 + 5.5
46 47 -1 - 1.5
50 47 3 +4
44 39 5 + 7.5
49 49 0
45 47 -2 -3
P a g e | 322
= .05, for two-tailed /2 = .025, and from Table A.14, T.025,14= 21
T- = 5.5 + 1.5 + 3 = 10
T+ = 3 + 6 = 9
P a g e | 324
T =9
(n)( n 1) (17)(18)
4 4
= 76.5
T 9 76.5
z
21.12
= -3.20
Since the observed z = -3.20 < z.05 = -1.645, the decision is to reject
the null
hypothesis.
13.16 Ho:Md = 0
Ha:Md < 0
506 510 -4 -2
427 414 13 7
438 439 -1 -1
512 517 -5 -3
n = 14
T+ = (+7)
Since the observed T = 7 < T.05,14 = 26, the decision is to reject the null
hypothesis.
The differences are significantly less than zero and the after scores are
P a g e | 326
significantly higher.
P a g e | 327
49 54 -5 -7.5
27 38 -11 -15
39 38 1 2
75 80 -5 -7.5
59 53 6 11
67 68 -1 -2
22 43 -21 -20
61 67 -6 -11
58 73 -15 -18
60 55 5 7.5
72 58 14 16.5
62 57 5 7.5
49 63 -14 -16.5
48 49 -1 -2
19 39 -20 -19
32 34 -2 -4.5
60 66 -6 -11
80 90 -10 -13.5
55 57 -2 -4.5
68 58 10 13.5
P a g e | 328
n = 20
T = 58
T 58 105
z
26.79
= -1.75
Since the observed z = -1.75 < z.10 = -1.28, the decision is to reject
the null
hypothesis.
April April
P a g e | 329
n = 16
T- = 8.5
T = 8.5
(n)( n 1) (16)(17)
4 4
= 68
P a g e | 330
T 8.5 68
z
19.339
= -3.08
3.08
Since the observed z = > z.05 = 1.645, the decision is to reject
the null
hypothesis.
P a g e | 331
1 2 3 4 5__
213
BY RANKS
1 2 3 4 5__
1 2 18 29 7.5
6 7.5 26 23.5 15
4 12 23.5 27 21
3 14 19 25 16.5
9 5 16.5 28 22
10.5 10.5 20
_ _ __ __ 13_
nj 6 5 6 5 7
P a g e | 332
2
Tj (33.5) 2 (40.5) 2 (113.5) 2 (132.5) (115) 2
n
6
5
6
5
7
j
= 8,062.67
n = 29
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
29(30)
(8,062.67) 3(30)
= 21.21
= .01 df = c - 1 = 5 - 1 = 4
2.01,4 = 13.2767
hypothesis.
P a g e | 333
19 30 39
21 38 32
29 35 41
22 24 44
37 29 30
42 27
33
By Ranks
1 8.5 15
2 14 10
6.5 12 16
3 4 18
13 6.5 8.5
17 5
__ _ 11_
Tj 42.5 45 83.5
nj 6 5 7
P a g e | 334
2
Tj (42.5) 2 (45) 2 (83.5) 2
n
6
5
7
j
= 1,702.08
n = 18
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
18(19)
(1,702.08) 3(19)
= 2.72
= .05, df = c - 1 = 3 - 1 = 2
2.05,2 = 5.9915
Since the observed K = 2.72 < 2.05,2 = 5.9915, the decision is to fail
to reject
200 680
425
By Ranks
23 5 15 21
12 19 13 20
2 1 22 17
18 9 7 3
10 6 14 8
4 16
_ _ _ 11
T j 69 40 71 96
nj 6 5 5 7
P a g e | 336
2
Tj (69) 2 (40) 2 (71) 2 (96) 2
n
6
5
5
7
j
= 3,438.27
n = 23
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
23(24)
(3,428.27) 3(24)
= 2.75
= .05 df = c - 1 = 4 - 1 = 3
2.05,3 = 7.8147
Since the observed K = 2.75 < 2.05,3 = 7.8147, the decision is to fail
to reject
22,550
By Ranks
4.5 11 8
12 6.5 14
3 6.5 16
10 15 9
2 4.5 1
__ __ 13
Tj 31.5 43.5 61
nj 5 5 6
2
Tj (31.5) 2 (43.5) 2 (61) 2
nj
5
5
6
= 1,197.07
n = 16
P a g e | 338
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
16(17)
(1,197.07) 3(17)
= 1.81
= .05 df = c - 1 = 3 - 1 = 2
2.05,2 = 5.9915
Since the observed K = 1.81 < 2.05,2 = 5.9915, the decision is to fail
to reject
0 3 2 2
1 2 2 4
1 3 3 3
0 5 2 4
2 4 3 3
1 4 2 5
0 3 3 4
5 3 4
2 1
By Ranks
2 33 11.5 28.5
33 20.5 28.5
11.5 5.5
__ __ 20.5 ____
Tj 34 207.5 154.0
199.5
nj 7 9 10 8
2
Tj (34) 2 (207.5) 2 (154) 2 (199.5)
n
7
9
10
8
j
= 12,295.80
n = 34
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
34(35)
(12,295.80) 3(35)
= 18.99
= .05 df = c - 1 = 4 - 1 = 3
2.05,3 = 7.8147
null hypothesis.
52 45 41
57 48 46
53 44 39
56 51 49
55 48 42
50 54 35
51 49 52
43
By Ranks
16.5 7 3
22 9.5 8
18 6 2
21 14.5 11.5
20 9.5 4
13 19 1
_ 5 _ ___
Tj 125 82 46
nj 7 8 7
2
Tj (125) 2 (82) 2 (46) 2
n
7
8
7
j
= 3,374.93
P a g e | 342
n = 22
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
22(23)
(3,374.93) 3(23)
= 11.04
= .05 df = c - 1 = 3 - 1 = 2
2.05,2 = 5.9915
null hypothesis.
P a g e | 343
Ha: At least one of the treatment populations yields larger values than
at least one
c = 5, b = 5, df = c - 1 = 4, 2.05,4 = 9.4877
If the observed value of 2 > 9.4877, then the decision will be to reject
the null
hypothesis.
1 2 3 4 5
1 1 4 3 5 2
2 1 3 4 5 2
3 2.5 1 4 5 2.5
4 3 2 4 5 1
5 4 2 3 5 1
Rj 11.5 12 18 25
8.5
P a g e | 344
Rj2 = 1,297.5
12 12
2
r
2
R j 3b(c 1) (1,297.5) 3(5)(6)
bc(c 1) (5)(5)(6)
= 13.8
Since the observed value of r2 = 13.8 > 4,.052 = 9.4877, the decision
is to
Ha: At least one of the treatment populations yields larger values than
at least one
c = 6, b = 9, df = c - 1 = 5, 2.05,5 = 11.0705
1 2 3 4 5 6
1 1 3 2 6 5 4
2 3 5 1 6 4 2
3 1 3 2 6 5 4
4 1 3 4 6 5 2
5 3 1 2 4 6 5
6 1 3 2 6 5 4
7 1 2 4 6 5 3
8 3 1 2 6 5 4
9 1 2 3 6 5 4
Rj 15 25 25 56 50 38
P a g e | 346
Rj2 = 8,555.5
12 12
2
r
2
R j 3b(c 1) (8,555) 3(9)(7)
bc(c 1) (9)(6)(7)
= 82.59
Since the observed value of r2 = 82.59 > 5,.052 = 11.0705, the
decision is to
c = 4, b = 6, df = c - 1 = 3, 2.01,3 = 11.3449
1 2 3 4
1 1 4 3 2
2 2 3 4 1
3 1 4 3 2
4 1 3 4 2
5 1 3 4 2
6 2 3 4 1
Rj 8 20 22 10
Rj2 = 1,048
12 12
2
r2 R j 3b(c 1) (1,048) 3(6)(5)
bc(c 1) (6)( 4)(5)
= 14.8
Since the observed value of r2 = 14.8 > 23,.01 = 11.3449, the decision
is to
Ha: At least one of the treatment populations yields larger values than
at least one
If the observed value of 2 > 5.9915, then the decision will be to reject
the null
hypothesis.
1 3 2 1
2 3 2 1
3 3
1
2
4 3 2 1
5 2 3 1
6 3 2 1
7 3 1 2
8 3 2 1
9 3 2 1
P a g e | 350
10 3 1 2
Rj 29 18 13
Rj2 = 1,334
12 12
2
r2 R j 3b(c 1) (1,334) 3(10)( 4)
bc (c 1) (10)(3)( 4)
= 13.4
Since the observed value of r2 = 13.4 > 2.05,2 = 5.9915, the decision
is to
Since the p-value of .564 > = .10, .05, or .01, the decision is to fail
to reject
13.30 The experimental design is a random block design that has been
analyzed using a Friedman test. There are five treatment levels and
seven blocks. Thus, the degrees of freedom are four. The observed
value of S = 13.71 is the equivalent of r2. The p-value is .009
indicating that this test is significant at alpha .01. The null hypothesis
is rejected. That is, at least one population yields larger values than at
least one other population. An examination of estimated medians
shows that treatment 1 has the lowest value and treatment 3 has the
highest value.
23 201 3 2 1 1
37 234 8 7 1 1
29 240 6 8 -2 4
P a g e | 352
25 231 4 6 -2 4
17 209 1 3 -2 4
33 229 7 5 2 4
40 248 9 10 -1 1
28 227 5 4 1 1
19 200 2 1 1 1
d2 = 23.5
n = 11
6 d 2 6(23.5)
rs 1 1
n(n 1)
2
11(120)
= .893
P a g e | 353
13.32 x y d d2
4 6 -2 4
5 8 -3 9
8 7 1 1
11 10 1 1
10 9 1 1
7 5 2 4
3 2 1 1
1 3 -2 4
2 1 1 1
9 11 -2 4
6 4 2 4
d2 = 34
n = 11
6 d 2 6(34)
rs 1 1
n(n 1)
2
11(120)
= .845
99 108 8 2 6 36
67 139 4 5 -1 1
82 117 6 3 3 9
46 168 1 8 -7 49
80 124 5 4 1 1
57 162 3 7 -4 16
49 145 2 6 -4 16
91 102 7 1 6 36
d2 = 164
n =8
6 d 2 6(164)
rs 1 1
n(n 1)
2
8(63)
= -.95
P a g e | 355
92 9.3 8 9 -1 1
96 9.0 9 8 1 1
89 8.0 5 3 2 4
88 8.4 4 6 -2 4
84 8.1 3 4 -1 1
81 7.9 1 2 -1 1
83 7.2 2 1 1 1
d2 = 15.5
n =9
6 d 2 6(15.5)
rs 1 1
n(n 1)
2
9(80)
= .871
2.33 1.66 13 6 7 49
P a g e | 356
2.54 1.77 10 3 7 49
2.18 1.47 14 10 4 16
3.34 1.75 3 4 -1 1
2.74 1.48 8 9 -1 1
3.11 1.30 5 13 -8 64
d2 = 636
n = 14
6 d 2 6(636)
rs 1 1
n(n 1)
2
14(142 1)
= -.398
1 12 12 0 0
P a g e | 357
2 11 10 1 1
3 3 5 -2 4
4 2 7 -5 25
5 4 6 -2 4
6 10 11 -1 1
7 9 9 0 0
8 8 8 0 0
9 7 4 3 9
10 1 3 -2 4
11 6 2 4 16
12 5 1 4 16
d2 = 80
6 d 2 6(80)
rs 1 1
n(n 1)
2
12(144 1)
= 0.72
1885 1055 10 2 8 64
2088 943 95 4 16
2361 1005 8 3 5 25
P a g e | 358
2570 981 7 4 3 9
2675 936 6 6 0 0
2907 896 4 7 -3 9
3047 893 2 8 -6 36
3114 862 19 -8 64
3025 769 3 10 -7 49
2862 765 5 11 -6
36 d2 = 162
d 2
= 408
n = 11
6 d 2 6( 408)
rs 1 1
n(n 1)
2
11(112 1)
= -0.855
on the NYSE and the number of equity issues on the American Stock
Exchange.
13.38 = .05
n1 = 13, n2 = 21
R = 10
Since this is a two-tailed test, use /2 = .025. The critical value is: z.025
= + 1.96
R R 10 17.06
z
R 2.707
= -2.61
Since the observed z = - 2.61 < z.025 = - 1.96, the decision is to reject
the null
573 547
532 566
544 551
565 538
540 557
548 560
536 557
523 547
x Rank Group
523 1 1
532 2 1
536
3 1
538 4 2
540 5 1
544 6 1
547 7.5 2
547 7.5 2
548 9 1
551 10 2
557 11.5 2
557 11.5 2
560 13 2
565 14 1
566 15 2
573 16 1
W1 = 1 + 2 + 3 + 5 + 6 + 9 + 14 + 16 = 56
n1 (n1 1) (8)(9)
U1 n1 n2 W1 (8)(8) 56
2 2
= 44
P a g e | 362
U 2 n1 n2 U1
= 8(8) - 44 = 20
From Table A.13, the p-value (1-tailed) is .1172, for 2-tailed, the p-value
is .2344. Since the p-value is > = .05, the decision is to fail to reject
the null hypothesis.
P a g e | 363
13.40 = .05, n = 9
H 0: Md = 0
Ha: Md 0
T+ = 2 + 5 + 2 = 9
T = min(T+, T-) = 9
P a g e | 364
13.41 nj = 7, n = 28, c = 4, df = 3
6 4 3 1
11 13 7 4
8 6 7 5
10 8 5 6
13 12 10 9
7 9 8 6
10 8 5 7
By Ranks:
9.5 3.5 2 1
23 17.5 6 9.5
27.5 26 23 20.5
23 17.5 6 13.5
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
28(29)
(6411.36) 3(29)
= 7.75
Since K = 7.75 < 23,.01 = 11.3449, the decision is to fail to reject the
null
hypothesis.
P a g e | 367
13.42 = .05, b = 7, c = 4, df = 3
2.05,3 = 7.8147
Ha: At least one treatment population yields larger values than at least
one other
treatment population
1 16 14 15 17
2 8 6 5 9
3 19 17 13 9
4 24 26 25 21
5 13 10 9 11
6 19 11 18 13
7 21 16
14 15
By Ranks:
1 3 1 2 4
2 3 2 1 4
3 4 3 2 1
4 2 4 3 1
P a g e | 368
5 4 2 1 3
6 4 1 3 2
7 4 3 1 2
Rj 24 16 13 17
12 12
2
r
2
R j 3b(C 1) (1,290) 3(7)(5)
bC (C 1) (7)( 4)(5)
= 5.57
Since r2 = 5.57 < 2.05,3 = 7.8147, the decision is to fail to reject
the null
13.43 Ranks
1 2 1 2 d d2
101 87 1 7 -6 36
129 89 2 8 -6 36
133 84 3 6 -3 9
147 79 4 5 -1 1
156 70 5 3 2 4
179 64 6 1 5 25
183 67 7 2 5 25
190 71 8 4 4 16
d2 = 152
n=8
6 d 2 6(152)
rs 1 1
n(n 1)
2
8(63)
= -.81
2.1 2.7
By Ranks
1 17.5 20
3.5 14 13
5 15 19
2 11 12
3.5 8 6.5
9.5 16
Tj 31 91 88
nj 7 7 6
2
Tj (31) 2 (91) 2 (88) 2
n
7
7
6
j
= 2,610.95
n = 20
P a g e | 371
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
20(21)
(2,610.95) 3(21)
= 11.60
= .01 df = c - 1 = 3 - 1 = 2
2.01,2 = 9.2104
hypothesis.
13.45 N = 40 n1 = 24 n2 = 16 = .05
Use the large sample runs test since both n1, n2 are not less than 20.
With a two-tailed test, /2 = .025, z.025 = + 1.96. If the observed z > .
196
R = 19
2n1n2 2(24)(16)
R 1 1
n1 n2 24 16
= 20.2
R R 19 20.2
z
R 2.993
= -0.40
Since z = -0.40 > z.025 = -1.96, the decision is to fail to reject the
null hypothesis.
c = 3 and b = 7
By ranks:
1 2 1 1
3 1 3 2
4 2 3 1
5 3 1 2
6 2 3 1
7 3 2 1
Rj 16 15 11
df = c - 1 = 2 2.05,2 = 5.99147.
If the observed 2r > 5.99147, the decision will be to reject the null
hypothesis.
P a g e | 374
12 12
2
r2 R j 3b(c 1) (602) 3(7)( 4)
bc(c 1) (7)(3)( 4)
= 2
Since 2r = 2 < 2.05,2 = 5.99147, the decision is to fail to reject the
null
hypothesis.
P a g e | 375
21 1 1
23 2 1
24 3 1
25 4 1
27 6 1
27 6 2
27 6 2
28 9 1
28 9 2
28 9 2
29 11 2
30 13 2
30 13 2
30 13 2
32 15 1
33 16.5 2
33 16.5 2
36 18.5 1
36 18.5 2
37 20 1
39 21 2
P a g e | 376
41 22 1
n1 = 10 n2 = 12
W1 = 1 + 2 + 3 + 4 + 6 + 9 + 15 + 18.5 + 20 + 22 = 100.5
n1 n2 (10)(12)
2 2
= 60
n1 (n1 1) (10)(11)
U n1 n2 W1 (10)(12) 100.5
2 2
= 74.5
U 74.5 60
z
15.17
= 0.96 with = .05, z.05 = -1.645
1.645
Since the observed z = 0.96 < z.05 = , the decision is to fail to
reject the
null hypothesis.
724 675 49 9
880 821 59 10
968 962 6 1
992 978 14 3
1046 973 73 11
852 841 11 2
n = 20
P a g e | 378
T- = 6 + 5 + 12.5 + 12.5 + 4 + 8 + 7 = 55
T = 55
T 55 105
z
26.79
= -1.87
Since the observed z = -1.87 > z.005 = -2.575, the decision is to fail to
reject the
null hypothesis.
350 1 J
P a g e | 379
430 2 M
460 3 J
470 4 J
490 5 M
500 6 M
510 7 M
520 8 J
530 9.5 M
530 9.5 J
540 11 M
550 12.5 M
550 12.5 J
560 14 M
570 15.5 M
570 15.5 J
590 17 J
600 18 M
610 19 J
630 20 J
n1 = 10 n2 = 10
n1 (n1 1) (10)(11)
U1 n1 n2 W1 (10)(10) 109.5
2 2
= 45.5
U 2 n1 n2 U1
= (10)(10) - 45.5 = 54.5
From Table A.13, the p-value for U = 45 is .3980 and for 44 is .3697.
For a
two-tailed test, double the p-value to at least .739. Using = .10, the
decision is
treatment population
Location
Brand 1 2 3 4
A 176 58 111 120
B 156 62 98 117
By ranks:
Location
Brand 1 2 3 4
A 4 1 2 3
B 4 1 2 3
C 4 1 3 2
D 4 1 3 2
P a g e | 382
E 4 1 2 3
F 4 1 2 3
Rj 24 6 14 16
Rj2 = 1,064
12 12
2
r2 R j 3b(c 1) (1,064) 3(6)(5)
bc(c 1) (6)( 4)(5)
= 16.4
Since r2 = 16.4 > 2.05,3 = 7.8147, the decision is to reject the null
hypothesis.
At least one treatment population yields larger values than at least one
other
185 170 15 11
109 112 -3 -3
92 90 2 2
105 87 18 13.5
60 51 9 7
45 49 -4 -4.5
25 11 14 10
58 40 18 13.5
108 82 26 15.5
89 94 -5 -6
34 21 13 8.5
68 55 13 8.5
59 60 -1 -1
78 52 26 15.5
n = 16
T- = 3 + 4.5 + 4.5 + 6 + 12 + 1 = 31
T = 31
P a g e | 384
(n)( n 1) (16)(17)
4 4
= 68
T 31 68
z
19.34
= -1.91
Since the observed z = -1.91 > z.025 = -1.96, the decision is to fail to
reject the
null hypothesis.
P a g e | 385
25 80 6 8 -2 4
41 85 9 9 0 0
16 35 4 3 1 1
0 45 1 5 -4 16
11 30 3 2 1 1
28 50 7 6 1 1
34 65 8 7 1 1
18 40 5 4 1 1
5 20 2 1 1 1
d2 = 26
n=9
6 d 2 6(26)
rs 1 1
n(n 1)
2
9(80)
= .783
= .05, /.025
P a g e | 386
R = 21
500 505 -5 -2
515 520 -5 -2
495 500 -5 -2
n = 12
From Table A.14, using n = 12, the critical T for = .01, one-tailed, is
10.
P a g e | 388
Since T = 19.5 is not less than or equal to the critical T = 10, the
decision is to fail
16 1 2
17 2 2
19 3.5 2
19 3.5 2
20 5 2
21 6.5 2
21 6.5 1
22 9 1
22 9 1
22 9 2
23 11.5 1
23 11.5 2
24 13 2
25 15.5 1
25 15.5 1
25 15.5 1
25 15.5 2
26 19 1
26 19 1
26 19 2
27 21 1
28 22 1
P a g e | 390
n1 = 11 n2 = 11
n1 n2 (11)(11)
2 2
= 60.5
n1 (n1 1) (11)(12)
U n1 n2 W1 (11)(11) 163.5
2 2
= 23.5
U 23.5 60.5
z
15.23
= -2.43 For = .05, z.05 = 1.645
2.43
Since the observed z = > z.05 = 1.645, the decision is to reject
the null
hypothesis.
92 1 M
105 2 M
106 3 M
110 4 A
114 5 M
117 6 M
118 7.5 A
118 7.5 M
125 9 M
126 10 M
128 11 A
129 12 M
137 13 A
143 14 A
144 15 A
152 16 A
153 17 A
168 18 A
n1 = 9 n2 = 9
W1 = 4 + 7.5 + 11 + 13 + 14 + 15 + 16 + 17 + 18 = 115.5
P a g e | 392
n1 (n1 1) (9)(10)
U1 n1 n2 W1 (9)(9) 115.5
2 2
= 10.5
U 2 n1 n2 U1
= 81 – 10.5 = 70.5
= .01
From Table A.13, the p-value = .0039. The decision is to reject the
null
45 55 70 85
By Ranks:
45 55 70 85
4 23 11.5 9
2 18.5 13 4
4 16 17 14.5
11.5 22 18.5 9
1 20.5 6.5
nj 6 6 5 6
P a g e | 394
2
Tj (31.5) 2 (120.5) 2 (74.5) 2 (49.5)
n
6
6
5
6
j
= 4,103.84
n = 23
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
23(24)
(4,103.84) 3(24)
= 17.21
= .01 df = c - 1 = 4 - 1 = 3
2.01,3 = 11.3449
null hypothesis.
P a g e | 395
13.58
Ranks Ranks
150,000 1,500 1 1 0 0
210,000 2,100 2 2 0 0
285,000 3,200 3 7 -4 16
301,000 2,400 4 4 0 0
335,000 2,200 5 3 2 4
390,000 2,500 6 5 1 1
400,000 3,300 7 8 -1 1
425,000 3,100 8 6 2 4
440,000 3,600 9 9 0 0
d2 = 26
n=9
6 d 2 6(26)
rs 1 1
n(n 1)
2
9(80)
= .783
P a g e | 396
9 27 16
11 38 21
17 25 18
10 40 28
22 31 29
15 19 20
6 35 31
By Ranks:
2 14 6
4 20 11
7 13 8
3 21 15
12 17.5 16
5 9 10
1 19 17.5
Tj 34 113.5 83.5
nj 7 7 7
P a g e | 397
2
Tj (34) 2 (113.5) 2 (83.5) 2
n
7
7
7
j
= 3,001.5
n = 21
2
12 Tj 12
K
n(n 1)
nj
3(n 1)
21(22)
(3,001.5) 3(22)
= 11.96
= .10 df = c - 1 = 3 - 1 = 2
2.10,2 = 4.6052
null hypothesis.
27 35 -8 -12
22 29 -7 -11
P a g e | 398
28 30 -2 -6.5
19 20 -1 -2.5
28 27 1 2.5
29 31 -2 -6.5
18 22 -4 -9.5
21 19 2 6.5
25 29 -4 -9.5
18 28 -10 -13.5
20 21 -1 -2.5
24 22 2 6.5
23 33 -10 -13.5
25 38 -13 -16.5
22 34 -12 -15
16 31 -15 -18
23 36 -13 -16.5
30 31 -1 -2.5
n = 18
T = 15.51
(n)( n 1) (18)(19)
4 4
= 85.5
P a g e | 399
T 15.5 85.5
z
22.96
= -3.05
Since the observed z = -3.05 < z.01 = -2.33, the decision is to reject
the null
hypothesis.
13.61 This problem uses a random block design, which is analyzed by the
Friedman
treatment two has a treatment median of 25.875. These two are the
farthest apart.
13.62 This is a Runs test for randomness. n1 = 21, n2 = 29. Because of the
size of the
n’s, this is a large sample Runs test. There are 28 runs, R = 28.
µR = 25.36 R = 3.34
28 25.36
z
3.34
= 0.79
The p-value for this statistic is .4387 for a two-tailed test. The decision
is to fail
The value of W is 191.5. The p-value for the test is .0066. The test is
significant
13.64 A Kruskal-Wallis test has been used to analyze the data. The null
hypothesis is
that the four populations are identical; and the alternate hypothesis is
that at least one of the four populations is different. The H statistic
(same as the K statistic) is 11.28 when adjusted for ties. The p-value
for this H value is .010 which indicates that there is a significant
difference in the four groups at = .05 and marginally so for = .01.
An examination of the medians reveals that all group medians are the
same (35) except for group 2 that has a median of 25.50. It is likely
that it is group 2 that differs from the other groups.
P a g e | 402
Chapter 14
Simple Regression Analysis
LEARNING OBJECTIVES
5. Test hypotheses about the slope of the regression model and interpret
the results.
variable into the regression model, predicted values can be determined. The
question then becomes: Are the predicted values good estimates of the actual
dependent values? One rule to emphasize is that the regression model should
not be used to predict for independent variable values that are outside the range
of values used to construct the model. MINITAB issues a warning for such
activity when attempted. There are many instances where the relationship
between x and y are linear over a given interval but outside the interval the
having been given, many forecasters use such regression models to extrapolate
to values of x outside the domain of those used to construct the model. Such
Forecasting Trend Line”. Whether the forecasts obtained under such conditions
P a g e | 404
are any better than "seat of the pants" or "crystal ball" estimates remains to be
seen.
It is my view that for many of these students, the most important facet
of this chapter lies in understanding the "buzz" words of regression such as
standard error of the estimate, coefficient of determination, etc. because they
may only interface regression again as some type of computer printout to be
deciphered. The concepts then may be more important than the calculations.
P a g e | 405
CHAPTER OUTLINE
14.6 Hypothesis Tests for the Slope of the Regression Model and Testing the
Overall
Model
14.7 Estimation
KEY TERMS
SOLUTIONS TO CHAPTER 14
14.1 x x
12 17
21 15
28 22
8 19
20 24
30
25
20
15
y
10
5 10 15 20 25 30
x = 89 y = 97 xy = 1,767
x y (89)(97)
SS xy xy n
1,767
5
SS x ( x) 2 (89) 2
x 2
n
1,833
5
b1 = = = 0.162
y b x 97 0.162 89
1
n n 5 5
b0 = = 16.51
ŷ
= 16.51 + 0.162 x
P a g e | 409
14.2 x_ _y_
140 25
119 29
103 46
91 70
65 88
29 112
24 128
x y (571)( 498)
SS xy xy n
30,099
7
SS x ( x) 2 (571) 2
x 2
n
58,293
7
b1 = = = -0.898
ŷ
= 144.414 – 0.898 x
P a g e | 411
12.5 148
3.7 55
21.6 338
60.0 994
37.6 541
6.1 89
16.8 126
41.2 379
x y (199.5)( 2,670)
SS xy xy n
107,610.4
8
SS x ( x) 2 (199.5) 2
x 2
n
7,667.15
8
b1 = = =
15.240
ŷ
= -46.292 + 15.240 x
P a g e | 412
16 5
6 12
8 9
4 15
7 7
x = 41 y = 48 xy = 333
x y ( 41)( 48)
SS xy xy n
333
5
SS x ( x) 2 (41) 2
x 2
n
421
5
b1 = = = -0.715
y b x 48 (0.715) 41
1
n n 5 5
b0 = = 15.460
ŷ
= 15.460 – 0.715 x
34.3 58.1
P a g e | 413
35.0 55.4
38.5 57.0
40.1 58.5
35.5 57.4
37.9 58.0
x y (344.4)( 221.3)
SS xy xy n
12,708.08
6
SS x ( x) 2 (344.4) 2
x 2
n
19,774.78
6
b1 = = =
b1 = 0.878
ŷ
= -13.503 + 0.878 x
P a g e | 414
5.65 213
4.65 258
3.96 297
3.36 340
2.95 374
2.52 420
2.44 426
2.29 441
2.15 460
2.07 469
2.17 434
2.10 444
x y (36.31)( 4,576)
SS xy xy n
12,766.71
12
SS x ( x) 2 (36.31) 2
x 2
n
124.7931
12
b1 = = = -72.328
P a g e | 415
ŷ
= 600.186 – 72.3281 x
P a g e | 416
99.9 2.74
97.9 2.87
98.9 2.93
87.9 2.87
92.9 2.98
97.9 3.09
100.6 3.36
104.9 3.61
105.3 3.75
108.6 3.95
x y (994.8)(32.15)
SS xy xy n
3,216.652
10
SS x ( x) 2 (994.8) 2
x 2
n
99,293.28
10
b1 = = =
0.05557
P a g e | 417
ŷ
= -2.31307 + 0.05557 x
P a g e | 418
14.8 x y
15 47
8 36
19 56
12 44
5 21
ŷ
= 13.625 + 2.303 x
Residuals:
ŷ ŷ
x y Residuals (y- )
15 47 48.1694 -1.1694
8 36 32.0489 3.9511
19 56 57.3811 -1.3811
12 44 41.2606 2.7394
5 21 25.1401 -4.1401
ŷ ŷ
14.9 x y Predicted ( ) Residuals (y- )
12 17 18.4582 -1.4582
P a g e | 419
21 15 19.9196 -4.9196
28 22 21.0563 0.9437
8 19 17.8087 1.1913
20 24 19.7572 4.2428
ŷ
= 16.51 + 0.162 x
ŷ ŷ
14.10 x y Predicted ( ) Residuals (y- )
91 70 62.6737 7.3263
65 88 86.0281 1.9720
ŷ
= 144.414 - 0.898 x
P a g e | 420
ŷ ŷ
14.11 x y Predicted ( ) Residuals (y- )
ŷ
= -46.292 + 15.240x
P a g e | 421
ŷ ŷ
14.12 x y Predicted ( ) Residuals (y- )
16 5 4.0259 0.9741
6 12 11.1722 0.8278
8 9 9.7429 -0.7429
4 15 12.6014 2.3986
7 7 10.4576 -3.4575
ŷ
= 15.460 - 0.715 x
ŷ ŷ
14.13 _ x_ _ y_ Predicted ( ) Residuals (y- )
The residual for x = 58.1 is relatively large, but the residual for x =
55.4 is quite
P a g e | 422
ŷ ŷ
small. 14.14 _x_ _ y_ Predicted ( ) Residuals (y- )
5 47 42.2756 4.7244
7 38 38.9836 -0.9836
11 32 32.3997 -0.3996
12 24 30.7537 -6.7537
19 22 19.2317 2.7683
25 10 9.3558 0.6442
ŷ
= 50.506 - 1.646 x
ŷ ŷ
14.15 Miles (x) Cost y ( ) (y- )
ŷ
= 2.2257 – 0.00025 x
14.16
14.17
SSE 46.5692
se
n2 3
= 3.94
SSE = 272.0
SSE 272.0
se
n2 5
= 7.376
SSE = 70,940
SSE 70,940
se
n2 6
= 108.7
Six out of eight (75%) of the sales estimates are within $108.7 million.
SSE 19.8885
se
n2 3
= 2.575
Four out of five (80%) of the estimates are within 2.575 of the actual
rate for
ŷ ŷ ( y yˆ ) 2
14.23 _ x_ _ y_ Predicted ( ) Residuals (y- )
( y yˆ ) 2
= 21.3355
( y yˆ ) 2
SSE = = 21.3355
SSE 21.3355
se
n2 4
= 2.3095
ŷ ŷ
14.24 (y- ) (y- )2
4.7244 22.3200
-0.9836 .9675
-0.3996 .1597
-6.7537 45.6125
2.7683 7.6635
0.6442 .4150
ŷ
(y- )2 = 77.1382
( y yˆ ) 2
SSE = = 77.1382
SSE 77.1382
se
n2 4
= 4.391
P a g e | 429
ŷ ŷ
14.25 (y- ) (y- )2
.1024 .0105
-.0222 .0005
-.1129 .0127
.0506 .0026
-.0996 .0099
.1076 .0116
-.0962 .0093
.0702 .0049
ŷ ( y yˆ ) 2
SSE .0620
se
n2 6
= .1017
time. However, the range of milk costs is only 45 cents for this data.
728.6 10.5
497.9 48.1
439.1 64.8
P a g e | 430
377.9 20.1
375.5 11.4
363.8 123.8
276.3 89.0
b1 = -.1504 b0 = 118.257
ŷ
= 118.257 - .1504x
SSE 8211.6245
se
n2 5
= 40.526
This is a relatively large standard error of the estimate given the sales
values
SSE 46.6399
1 1
( y ) 2
(97) 2
y n
2
1,935
5
14.27 r2 = = .123
SSE 272.121
1 1
( y ) 2
(498) 2
y n
2
45,154
7
14.28 r2 = = .972
SSE 70,940
1 1
( y ) 2
(2,670) 2
y 2
n
1,587,328
8
14.29 r2 = = .898
SSE 19.8885
1 1
( y ) 2
(48) 2
y n
2
524
5
14.30 r2 = = .685
P a g e | 432
SSE 21.33547
1 1
( y ) 2
(221.3) 2
y n
2
8,188.41
6
14.31 r2 = = .183
unaccounted for.
116.8 37.415
91.5 36.770
68.5 35.501
61.6 35.047
65.9 34.700
90.6 34.942
100.0 35.887
P a g e | 433
104.6 36.306
125.4 37.005
x y (323.573)(824.9)
SS xy xy n
29,804.4505
9
SS x ( x) 2 (323.573) 2
x 2
n
11,640.93413
9
b1 = = =
b1 = 19.2204
ŷ
= -599.3674 + 19.2204 x
SSE 1283.13435
se
n2 7
= 13.539
P a g e | 434
SSE 1283.13435
1 1
( y ) 2
(824.9) 2
y n
2
79,718.79
9
r2 = = .688
se 3.94
( x ) 2 (89) 2
x2 n
1.833
5
14.33 sb = = .2498
b1 = 0.162
Ho: = 0 = .05
Ha: 0
t.025,3 = ±3.182
b1 1 0.162 0
sb .2498
t = = 0.65
P a g e | 435
Since the observed t = 0.65 < t.025,3 = 3.182, the decision is to fail to
reject the
null hypothesis.
se 7.376
( x ) 2 (571) 2
x2 n
58,293
7
14.34 sb = = .068145
b1 = -0.898
Ho: = 0 = .01
Ha: 0
t.005,5 = ±4.032
b1 1 0.898 0
sb .068145
t= = -13.18
hypothesis.
P a g e | 436
se 108.7
( x ) 2 (199.5) 2
x2 n
7,667.15
8
14.35 sb = = 2.095
b1 = 15.240
Ho: = 0 = .10
Ha: 0
t.05,6 = 1.943
b1 1 15,240 0
sb 2.095
t = = 7.27
Since the observed t = 7.27 > t.05,6 = 1.943, the decision is to reject
the null
hypothesis.
P a g e | 437
se 2.575
( x ) 2
(41) 2
x 2
n
421
5
14.36 sb = = .27963
b1 = -0.715
Ho: = 0 = .05
Ha: 0
t.025,3 = ±3.182
b1 1 0.715 0
sb .27963
t = = -2.56
Since the observed t = -2.56 > t.025,3 = -3.182, the decision is to fail to
reject the
null hypothesis.
P a g e | 438
se 2.3095
( x ) 2
(344.4) 2
x 2
n
19,774.78
6
14.37 sb = = 0.926025
b1 = 0.878
Ho: = 0 = .05
Ha: 0
t.025,4 = ±2.776
b1 1 0.878 0
sb .926025
t = = 0.948
Since the observed t = 0.948 < t.025,4 = 2.776, the decision is to fail to
reject the
null hypothesis.
F
t = = 2.874
t.05,8 = 1.86 but t.01,8 = 2.896. The slope is significant at = .05 but not
at
= .01.
P a g e | 440
14.39 x0 = 25
df = n - 2 = 5 - 2 = 3 t.025,3 = ±3.182
x
x 89
n 5
= 17.8
x = 89 x2 = 1,833
se = 3.94
ŷ
= 16.5 + 0.162(25) = 20.55
1 ( x 0 x) 2
n ( x ) 2
ŷ x 2
n
± t /2,n-2 se
1 (25 17.8) 2
5 (89) 2
1,833
5
20.55 ± 3.182(3.94) = 20.55 ± 3.182(3.94)
(.63903) =
20.55 ± 8.01
P a g e | 441
df = n - 2 = 7 - 2 = 5 t.05,5 = ±2.015
x
x 571
n 7
= 81.57143
ŷ
= 144.414 - .0898(100) = 54.614
1 ( x0 x ) 2
1
n ( x ) 2
ŷ x n
2
± t /2,n-2 se =
1 (100 81.57143) 2
1
7 (571) 2
58,293
7
54.614 ± 2.015(7.377) =
ŷ
For x0 = 130, = 144.414 - 0.898(130) = 27.674
1 ( x0 x) 2
1
n ( x ) 2
x n
2
y ± t /2,n-2 se =
1 (130 81.57143) 2
1
7 (571) 2
58,293
7
27.674 ± 2.015(7.377) =
The width of this confidence interval of y for x0 = 130 is wider that the
df = n - 2 = 8 - 2 = 6 t.01,6 = 3.143
x
x 199.5
n 8
= 24.9375
P a g e | 444
ŷ
= -46.29 + 15.24(20) = 258.51
1 ( x 0 x) 2
n ( x ) 2
ŷ x 2
n
± t /2,n-2 se
1 (20 24.9375) 2
8 (199.5) 2
7,667.15
8
258.51 ± (3.143)(108.8)
1 ( x0 x) 2
1
n ( x ) 2
ŷ x n
2
± t /2,n-2 se
P a g e | 445
1 (20 24.9375) 2
1
8 (199.5) 2
7,667.15
8
258.51 ± (3.143)(108.8)
The confidence interval for the single value of y is wider than the
confidence
middle and individual values of y can vary more than values of the
average.
P a g e | 446
df = n - 2 = 5 - 2 = 3 t.005,3 = 5.841
x
x 41
n 5
= 8.20
ŷ
= 15.46 - 0.715(10) = 8.31
1 ( x 0 x) 2
n ( x ) 2
ŷ x n
2
± t /2,n-2 se
1 (10 8.2) 2
5 (41) 2
421
5
8.31 ± 5.841(2.575) =
If the prime interest rate is 10%, we are 99% confident that the
average bond rate
2001 11.9
2002 17.9
2003 22.0
2004 21.8
2005 26.0
x y (10,015)(99.6)
SS xy xy n
199,530.9
5
SS x ( x) 2 (10,015) 2
x 2
n
20,060,055
5
b1 = = = 3.21
y b x 99.6 3.2110,015
1
n n 5 5
b0 = = -6,409.71
ŷ
= -6,409.71 + 3.21 x
ŷ
(2008) = -6,409.71 + 3.21(2008) = 35.97
P a g e | 449
1998 5860
1999 6632
2000 7125
2001 6000
2002 4380
2003 3326
2004 2642
x y (14,007)(35,965)
SS xy xy n
71,946,954
7
SS x ( x) 2 (14,007) 2
x 2
n
28,028,035
7
b1 = = =
-678.9643
ŷ
= 1,363,745.39 + -678.9643 x
ŷ
(2007) = 1,363,745.39 + -678.9643(2007) = 1,064.04
P a g e | 450
2003 1 1 11.93
2 2 12.46
3 3 13.28
4 4 15.08
2004 1 5 16.08
2 6 16.82
3 7 17.60
4 8 18.66
2005 1 9 19.73
2 10 21.11
3 11 22.21
4 12 22.94
x y (78)( 207.9)
SS xy xy n
1,499.07
12
SS x ( x) 2 (78) 2
x 2
n
650
12
b1 = = = 1.033
P a g e | 451
y b x 207.9 1.033 78
1
n n 12 12
b0 = = 10.6105
ŷ
= 10.6105 + 1.033 x
sales for the third quarter of year 2007, we must convert this time
frame to
scheme.
ŷ
(19) = 10.6105 + 1.033(19) = 30.2375
P a g e | 452
14.46 x y
5 8
7 9
3 11
16 27
12 15
9 13
x = 52 x2 = 564
ŷ
a) = 2.6941 + 1.2853 x
ŷ ŷ
b) (Predicted Values) (y- ) residuals
9.1206 -1.1206
11.6912 -2.6912
6.5500 4.4500
23.2588 3.7412
18.1177 -3.1176
14.2618 -1.2618
P a g e | 453
ŷ
c) (y- )2
1.2557
7.2426
19.8025
13.9966
9.7194
1.5921
SSE = 53.6089
SSE 53.6089
se
n2 4
= 3.661
SSE 53.6089
1 1
( y ) 2
(83) 2
y n
2
1,389
6
d) r2 = = .777
P a g e | 454
e) Ho: = 0 = .01
Ha: 0
t.005,4 = ±4.604
se 3.661
( x ) 2 (52) 2
x 2
n
564
6
sb = = .34389
b1 1 1.2853 0
sb .34389
t = = 3.74
14.47 x y
53 5
47 5
41 7
50 4
58 10
62 12
45 3
60 11
ŷ
a) = -11.335 + 0.355 x
P a g e | 456
ŷ ŷ
b) (Predicted Values) (y- ) residuals
7.48 -2.48
5.35 -0.35
3.22 3.78
6.415 -2.415
9.255 0.745
10.675 1.325
4.64 -1.64
9.965 1.035
ŷ
c) (y- )2
6.1504
0.1225
14.2884
5.8322
0.5550
1.7556
2.6896
1.0712
SSE = 32.4649
SSE 32.4649
n2 6
d) se = = 2.3261
P a g e | 457
SSE 32.4649
1 1
( y ) 2
(57) 2
y n
2
489
8
e) r2 = = .608
f) Ho: = 0 = .05
Ha: 0
t.025,6 = ±2.447
se 2.3261
( x ) 2 (416) 2
x2 n
22,032
8
sb = = 0.116305
b1 1 0.3555 0
sb .116305
t = = 3.05
null hypothesis.
se is not small.
b0 = 25.42778 b1 = 0.209369
SSE 46.845468
1 1
( y ) 2
153.5
y2 n
r2 = = .695
P a g e | 459
14.49a) x0 = 60
df = n - 2 = 8 - 2 = 6
t.025,6 = ±2.447
ŷ
= -9.026 + 0.5481(60) = 23.86
x
x 524
n 8
= 65.5
P a g e | 460
1 ( x0 x) 2
n ( x ) 2
ŷ x n
2
± t /2,n-2 se
1 (60 65.5) 2
8 (524) 2
36,224
8
23.86 + 2.447(3.201)
b) x0 = 70
ŷ
70 = -9.026 + 0.5481(70) = 29.341
1 ( x 0 x) 2
1
n ( x ) 2
ŷ x n
2
+ t/2,n-2 se
1 (70 65.5) 2
1
8 (524) 2
36,224
8
29.341 + 2.447(3.201)
P a g e | 461
c) The confidence interval for (b) is much wider because part (b) is for a
single value
part (b) is slightly closer to the mean (x) than x0 = 60. However, the
width of the
part (a).
P a g e | 462
1 56
2 54
3 49
4 46
5 45
x y (15)( 250)
SS xy xy n
720
5
SS x ( x) 2 (15) 2
x 2
n
55
5
b1 = = = -3
y b x 250 (3) 15
1
n n 5 5
b0 = = 59
ŷ
= 59 - 3 x
ŷ
(7) = 59 - 3(7) = 38
P a g e | 463
x = 21 x2 = 101
xy = 1,256 n =5
b0 = 9.234375 b1 = 10.515625
SSE 297.7969
1 1
( y ) 2
1,713.2
y2 n
r2 = = .826
same would hold true for a model to predict number of sales people by
the
b1 = 10.626383 b0 = 9.728511
ŷ
= 9.728511 + 10.626383 x
SSE 94,337.9762
se
n2 10
= 97.1277
SSE 94,337.9762
1 1
( y ) 2
271,246
y2 n
r2 = = .652
10.626383 0
97.1277
(548) 2
26,592
12
t = = 4.33
If = .01, then t.005,10 = 3.169. Since the observed t = 4.33 > t.005,10 =
3.169, the
17.1 12.4
7.9 7.5
4.8 6.8
4.7 8.7
4.6 4.6
4.0 5.1
2.9 11.2
2.7 5.1
2.7 2.9
b1 = 0.92025 b0 = -0.863565
ŷ
= -0.863565 + 0.92025 x
SSE 99.153926
1 1
( y ) 2
166.55
y 2
n
r2 = = .405
P a g e | 468
1995 11,152
1996 10,935
1997 11,050
1998 10,845
1999 10,776
2000 10,764
2001 10,697
2002 9,234
2003 9,223
2004 9,158
x y (19,995)(103,834)
SS xy xy n
207,596,350
10
SS x ( x) 2 (19,995) 2
x 2
n
39,980,085
10
b1 = = = -239.188
ŷ
= 488,639.564 + -239.188 x
ŷ
(2008) = 488,639.564 + -239.188(2008) = 8,350.30
P a g e | 470
581 666
213 214
668 496
345 204
1476 1600
1776 6278
x y (5059)(9358)
SS xy xy n
13,593,272
6
SS x ( x) 2 (5059) 2
x 2
n
6,280,931
6
b1 = = = 3.1612
ŷ
= -1089.0712 + 3.1612 x
P a g e | 471
for x = 700:
ŷ
= 1076.6044
1 ( x 0 x) 2
n ( x ) 2
ŷ x n
2
+ t/2,n-2se
x0 = 700, n = 6
x
= 843.167
SSE 7,701,506.49
se
n2 4
= 1387.58
Confidence Interval =
P a g e | 472
1 (700 843.167) 2
6 (5059) 2
6,280,931
6
1123.757 + (2.776)(1387.58) =
1123.757 + 1619.81
-496.05 to 2743.57
H0: 1 = 0
Ha: 1 0
= .05 df = 4
b1 0 3.1612 0 2.9736
sb 1387.58 .8231614
2,015,350.833
t = = 3.234
Since the observed t = 3.234 > t.025,4 = 2.776, the decision is to reject
the null
hypothesis.
ŷ
= -39.0071 + 66.36277 x
SSE 2,232.343
se
n2 5
= 21.13
SSE 2,232.343
1 1
( y ) 2
(516.8) 2
y 2
n
61,899.06
7
r2 = = 1 - .094 = .906
x y (44,754)(17,314)
SS xy xy n
59,852,571
13
SS x ( x) 2 (44,754) 2
x 2
n
167,540,610
13
b1 = = = .01835
P a g e | 474
ŷ
= 1268.685 + .01835 x
x y (91)( 44,754)
SS xy xy n
304,797
13
SS x ( x) 2 (91) 2
x 2
n
819
13
b1 = = = -46.5989
y b x 44,754 (46.5989) 91
1
n n 13 13
b0 = = 3,768.81
ŷ
= 3,768.81 – 46.5989 x
P a g e | 475
ŷ
(2007) = 3,768.81 - 46.5989(15) = 3,069.83
P a g e | 476
xy = 339,342.76 n = 7
x y (323.3)(6765.8)
SS xy xy n
339,342.76
7
SS x ( x) 2 (323.3) 2
x 2
n
29,629.13
7
b1 = = = 1.82751
ŷ
= 882.138 + 1.82751 x
SSE 994,623.07
se
n2 5
= 446.01
P a g e | 477
SSE 994,623.07
1 1
( y ) 2
(6765.8) 2
y n
2
7,583,144 .64
7
r2 = = 1 - .953 = .
047
H 0: = 0
x 2
(323.3) 2
x 2
n
29,629.13
7
SSxx = = 14,697.29
b1 0 1.82751 0
se 446.01
SS xx 14,697.29
t = = 0.50
Since the observed t = 0.50 < t.025,5 = 2.571, the decision is to fail to
reject the
null hypothesis.
P a g e | 478
ŷ
= -54.35604 + 2.40107 x
ŷ
100 = -54.35604 + 2.40107(100) = 185.751
SSE 1,919.5146
se
n2 6
= 17.886
SSE 1,919.5145
1 1
( y ) 2
(1025) 2
y n
2
152,711
8
r2 = = 1 - .09 = .91
Ho: = 0
Ha: 0 = .01
df = n - 2 = 8 - 2 = 6
t.005,6 = ±3.707
se 17.886
( x) 2 (608) 2
x 2
n
49,584
8
sb = = .30783
b1 1 2.40107 0
sb .30783
t = = 7.80
Since the observed t = 7.80 < t.005,6 = 3.707, the decision is to reject
the null
hypothesis.
ŷ
14.60 a) The regression equation is: = 67.2 – 0.0565 x
decrease by -.0565.
c) The t ratio for the slope is –5.50 with an associated p-value of .000.
This is
the numerator of the t ratio formula equals the slope minus zero.
data.
taking the square root of r2 which is .627 yields .7906 which is the
magnitude
14.61 The F value for overall predictability is 7.12 with an associated p-value
of .0205 which is significant at = .05. It is not significant at alpha of .
01. The coefficient of determination is .372 with an adjusted r2 of .
32. This represents very modest predictability. The standard error of
the estimate is 982.219, which in units of 1,000 laborers means that
about 68% of the predictions are within 982,219 of the actual figures.
The regression model is:
14.62 The Residual Model Diagnostics from MINITAB indicate a relatively healthy
set
of residuals. The Histogram indicates that the error terms are
generally normally
Chapter 15
Multiple Regression Analysis
LEARNING OBJECTIVES
CHAPTER OUTLINE
Residuals
Adjusted R2
KEY TERMS
Adjusted R2 R2
Dependent Variable
Response Plane
Independent Variable Response Surface
ŷ
= 25.03 - 0.0497 x1 + 1.928 x2
P a g e | 487
ŷ
= 25.03 - 0.0497(200) + 1.928(7) = 28.586
ŷ
= 118.56 - 0.0794 x1 - 0.88428 x2 + 0.3769 x3
ŷ
= 118.56 - 0.0794(33) - 0.88428(29) + 0.3769(13) = 95.19538
P a g e | 488
ŷ
= 121.62 - 0.174 x1 + 6.02 x2 + 0.00026 x3 + 0.0041 x4
There are four independent variables. If x2, x3, and x4 are held
constant, the predicted y will decrease by - 0.174 for every unit
increase in x1. Predicted y will increase by 6.02 for every unit increase
in x2 as x1, x3, and x4 are held constant. Predicted y will increase by
0.00026 for every unit increase in x3 holding x1, x2, and x4 constant. If
x4 is increased by one unit, the predicted y will increase by 0.0041 if x1,
x2, and x3 are held constant.
ŷ
= 31,409.5 + 0.08425 x1 + 289.62 x2 - 0.0947 x3
For every unit increase in paper consumption, the predicted per capita
consumption increases by 116.2549 if fish and gasoline consumption
are held constant. For every unit increase in fish consumption, the
predicted per capita consumption decreases by 120.0904 if paper and
gasoline consumption are held constant. For every unit increase in
gasoline consumption, the predicted per capita consumption increases
by 45.73328 if paper and fish consumption are held constant.
P a g e | 490
Insider Ownership =
The coefficients mean that for every unit of increase in debt ratio there
is a predicted decrease of - 0.0594 in insider ownership if dividend
payout is held constant. On the other hand, if dividend payout is
increased by one unit, then there is a predicted drop of insider
ownership by 0.118 with debt ratio is held constant.
15.7 There are 9 predictors in this model. The F test for overall significance
of the model is 1.99 with a probability of .0825. This model is not
significant at = .05. Only one of the t values is statistically
significant. Predictor x1 has a t of 2.73 which has an associated
probability of .011 and this is significant at = .05.
Insider Ownership =
The overall value of F is only 0.02 with p-value of .982. This model is
not significant. Neither of the t values are significant (tDebt = -0.19 with
a p-value of .855 and tDividend = -0.11 with a p-value of .913).
ŷ
= 3.981 + 0.07322 x1 - 0.03232 x2 - 0.003886 x3
(x2: t = -1.55, p-value of .15 and x3: t = -1.01, p-value of .332). Since
x2 and x3 are non significant predictors, the researcher should consider
the using a simple regression model with only x1 as a predictor. The R2
would drop some but the model would be much more parsimonious.
15.12 The regression equation for the model using both x1 and x2 is:
P a g e | 493
ŷ
= 243.44 - 16.608 x1 - 0.0732 x2
ˆ
Yˆ
= 235.143 - 16.7678 x1
ŷ
= 657.053 + 5.7103 x1 – 0.4169 x2 –3.4715 x3
15.14 The standard error of the estimate is 3.503. R2 is .408 and the
adjusted R2 is only .203. This indicates that there are a lot of
insignificant predictors in the model. That is underscored by the fact
that eight of the nine predictors have non significant t values.
15.15 se = 9.722, R2 = .515 but the adjusted R2 is only .404. The difference
in the two is due to the fact that two of the three predictors in the
model are non-significant. The model fits the data only modestly. The
adjusted R2 indicates that 40.4% of the variance of y is accounted for
by this model and 59.6% is unaccounted for by the model.
15.16 The standard error of the estimate of 14,660.57 indicates that this
model predicts Per Capita Personal Consumption to within + 14,660.57
about 68% of the time. The entire range of Personal Per Capita for the
data is slightly less than 110,000. Relative to this range, the standard
error of the estimate is modest. R2 = .85988 and the adjusted value of
R2 is .799828 indicating that there are potentially some non significant
variables in the model. An examination of the t statistics reveals that
two of the three predictors are not significant. The model has
relatively good predictability.
P a g e | 495
15.19 For the regression equation for the model using both x1 and x2, se =
6.333,
moderately strong.
P a g e | 497
15.21 The Histogram indicates that there may be some problem with the
error
terms being normally distributed. The Residuals vs. Fits plot reveals
that there may be some lack of homogeneity of error variance.
15.22 There are four predictors. The equation of the regression model is:
ŷ
= -55.9 + 0.0105 x1 – 0.107 x2 + 0.579 x3 – 0.870 x4
x2, x3, and x4. The R2 value of 80.2% indicates strong predictability for
the model. The value of the adjusted R2 (78.8%) is close to R2 and se is
9.025.
15.23 There are two predictors in this model. The equation of the regression
model is:
ŷ
= 203.3937 + 1.1151 x1 – 2.2115 x2
ŷ
= 137.27 + 0.0025 x1 + 29.206 x2
Yˆ
= 362.3054 – 4.745518 x1 - 13.89972 x2 + 1.874297 x3
P a g e | 499
15.26 The overall F for this model was 12.19 with a p-value of .002 which is
significant at = .01. The t test for Silver is significant at = .01 ( t =
4.94, p = .001). The t test for Aluminum yields a t = 3.03 with a p-
value of .016 which is significant at = .05. The t test for Copper was
insignificant with a p-value of .939. The value of R2 was 82.1%
compared
The low value of adjusted R2 indicates that the model has very low
predictability. Both t values are not significant (tNavalVessels = 0.67 with
Only one of the four predictors has a significant t ratio and that is
Utility with t = 2.57 and p = .018. The ratios and their respective
probabilities are:
This model is very weak. Only the predictor, Utility, shows much
promise in accounting for the grocery variability.
ŷ
= 87.89 – 0.256 x1 – 2.714 x2 + 0.0706 x3
(x2 and x3) significant at .01 and the other, x1 significant at = .05.
This is
15.32 Two of the diagnostic charts indicate that there may be a problem
with the
error terms being normally distributed. The histogram indicates that
the error term distribution might be skewed to the right and the normal
probability plot is somewhat nonlinear. In addition, the residuals vs.
fits chart indicates a potential heteroscadasticity problem with
P a g e | 504
Chapter 16
Building Multiple Regression Models
LEARNING OBJECTIVES
CHAPTER OUTLINE
Polynomial Regression
Model Transformation
Search Procedures
Stepwise Regression
Forward Selection
Backward Elimination
16.4 Multicollinearity
P a g e | 507
KEY TERMS
Backward Elimination
Search Procedures
Dummy Variable Stepwise Regression
Multicollinearity Transformations
ŷ
= - 147.27 + 27.128 x
ŷ
= - 22.01 + 3.385 X + 0.9373 x2
t = 0.75 with p = .483, and for x2: t = 5.33 with p = .002. The
quadratic model is also very strong with an even higher R2 value.
However, in this model only the x2 term is a significant predictor.
P a g e | 509
ŷ
= b0b1x
The regression model is solved for in the computer using the values of
x and the values of log y. The resulting regression equation is:
ˆ
Yˆ
= - 1456.6 + 71.017 x
ŷ
= 1012 - 14.06 x + 0.6115 x2
7000
6000
5000
Ad Exp
4000
3000
2000
1000
30 40 50 60 70 80 90 100 110
Eq & Sup Exp
ŷ
= b0b1x
The regression model is solved for in the computer using the values of
x and the values of log y where x is failures and y is liabilities. The
resulting regression equation is:
ŷ
= - 28.61 - 2.68 x1 + 18.25 x2 - 0.2135 x12 - 1.533 x22 + 1.226
x1*x2
None of the t ratios for this model are significant. They are t(x1) = -
0.25 with p = .805, t(x2) = 0.91 with p = .378, t(x12) = - 0.33 with .745,
t(x22) = - 0.68 with .506, and t(x1*x2) = 0.52 with p = .613. This model
has a high R2 yet none of the predictors are individually significant.
The same thing occurs when the interaction term is not in the model.
None of the t tests are significant. The R2 remains high at .957
indicating
16.6 The F value shows very strong overall significance with a p-value of .
00000073. This is reinforced by the high R2 of .910 and adjusted R2 of .
878. An examination of the t values reveals that only one of the
regression coefficients is significant at = 05 and that is the
interaction term with a p-value of .039. Thus, this model with both
variables, the square of both variables, and the interaction term
contains only one significant t test and that is for interaction.
= .01.
ŷ
= 13.619 - 0.01201 x1 + 2.998 x2
The t ratio for the x1 variable is only t = -0.14 with p = .893. However
the t ratio for the dummy variable, x2 is t = 3.88 with p = .004. The
indicator variable is the significant predictor in this regression model
that has some predictability (adjusted R2 = .575).
ŷ
= 7.909 + 0.581 x1 + 1.458 x2 - 5.881 x3 - 4.108 x4
For the predictors, t = 0.56 with p = .585 for the x1 variable (not
significant), t = 1.32 with p = .208 for the first indicator variable (x2)
and is non significant, t = -5.32 with p = .000 for x3 the second
indicator variable and this is significant at = .001, t = -3.21 with p = .
007 for the third indicator variable (x4) which is significant at = .01.
This model has strong predictability and the only significant predictor
variables are the two dummy variables, x3 and x4.
ŷ
= 41.225 + 1.081 x1 – 18.404 x2
ŷ
becomes = 41.225 + 1.081x1. With x2 = 1, the regression model
ŷ
becomes = 22.821 + 1.081x1. The presence of x2 causes the y
the dummy variable and with the dummy variable equal to one) is
shown
below:
P a g e | 517
three for occupation, two for industry, and one for marital status. The
dependent variable is job satisfaction. In total, there will be seven
variables in this analysis.
P a g e | 518
ŷ
The model is = 26.40 - 0.101 x2 + 0.116 x3
ŷ
The model is = 133.53 - 0.78 x4
ŷ
The model is = 91.01 - 0.60 x4 + 0.51 x2
P a g e | 519
16.15 The output shows that the final model had four predictor variables, x4,
x2, x5, and x7. The variables, x3 and x6 did not enter the stepwise
analysis. The procedure took four steps. The final model was:
The R2 for this model was .5929, and se was 3.36. The t ratios were:
16.16 The output indicates that the stepwise process only went two steps.
Variable x3 entered at step one. However, at step two, x3 dropped out
of the analysis and x2 and x4 entered as the predictors. x1 was the
dependent variable. x5 never entered the procedure and was not
included in the final model as x3 was not. The final regression model
was:
Yˆ
= 22.30 + 12.38 x2 + 0.0047 x4.
16.17 The output indicates that the procedure went through two steps. At
step 1, dividends entered the process yielding an r2 of .833 by itself.
ŷ
The t value was 6.69 and the model was = - 11.062 + 61.1 x1. At
step 2, net income entered the procedure and dividends remained in
the model. The R2 for this two-predictor model was .897 which is a
modest increase from the simple regression model shown in step one.
The step 2 model was:
correlation matrix
Premiums 1
Income 0.808236 1
P a g e | 521
16.18 This stepwise regression procedure only went one step. The only
significant predictor was natural gas. No other predictors entered the
model. The regression model is:
For this model, R2 = .9295 and se = 0.490. The t value for natural gas
was 11.48.
P a g e | 522
Chapter 18
Statistical Quality Control
LEARNING OBJECTIVES
x
4. Learn how to construct charts, R charts, p charts, and c charts.
In this chapter, four types of control charts are presented. Two of the
charts, the x bar chart and the R chart, deal with measurements of product
attributes such as weight, length, temperature and others. The other two
charts deal with whether or not items are in compliance with specifications (p
chart) or the number of noncompliances per item (c chart). The c chart is
less widely known and used than the other three. As part of the material on
control charts, a discussion on variation is presented. Variation is one of the
main concerns of quality control. A discussion on various types of variation
that can occur in a business setting can be profitable in helping the student
understand why particular measurements are charted and controlled.
CHAPTER OUTLINE
Benchmarking
Just-in-Time Systems
Reengineering
Poka-Yoke
Six Sigma
Lean Manufacturing
Team Building
Flowcharts
Pareto Analysis
Control Charts
Check Sheets
Histogram
Scatter Chart
Variation
x
Chart
R Charts
p Charts
P a g e | 526
c Charts
Double-Sample Plan
Multiple-Sample Plan
KEY TERMS
c Chart Poka-Yoke
x
Multiple-Sample Plan Chart
P a g e | 528
x1 x2 x3 x4 x5
18.4 = 27.00, = 24.29, = 25.29, = 27.71, = 25.86
R1 = 8, R2 = 8, R3 = 9, R4 = 7, R5 = 6
x R
= 26.03 = 7.6
x
For Chart: Since n = 7, A2 = 0.419
x
Centerline: = 26.03
x R
UCL: + A2 = 26.03 + (0.419)(7.6) = 29.21
x R
LCL: - A2 = 26.03 – (0.419)(7.6) = 22.85
R
Centerline: = 7.6
P a g e | 531
R
UCL: D4 = (1.924)(7.6) = 14.62
R
LCL: D3 = (0.076)(7.6) = 0.58
x
Chart:
R Chart:
P a g e | 532
P a g e | 533
x1 x2 x3 x4
18.5 = 4.55, = 4.10, = 4.80, = 4.70,
x5 x6 x7
= 4.30, = 4.73, = 4.38
x R
= 4.51 = 0.90
x
For Chart: Since n = 4, A2 = 0.729
x
Centerline: = 4.51
x R
UCL: + A2 = 4.51 + (0.729)(0.90) = 5.17
x R
LCL: - A2 = 4.51 – (0.729)(0.90) = 3.85
R
Centerline: = 0.90
P a g e | 534
R
UCL: D4 = (2.282)(0.90) = 2.05
R
LCL: D3 = 0
x
Chart:
P a g e | 535
R Chart:
p = .033
Centerline: p = .033
(.033)(. 967)
100
UCL: .033 + 3 = .033 + .054 = .087
P a g e | 536
(.033)(.967)
100
LCL: .033 - 3 = .033 - .054 = .000
pChart:
P a g e | 537
p = .050
Centerline: p = .050
(.05)(.95)
40
UCL: .05 + 3 = .05 + .1034 = .1534
(.05)(. 95)
40
LCL: .05 - 3 = .05 - .1034 = .000
p Chart:
P a g e | 538
22
c 35
18.8 = = 0.62857
c
Centerline: = 0.62857
c3 c 0.62857
UCL: = 0.62857 + 3 =
c3 c 0.62857
LCL: = 0.62857 - 3 =
c Chart:
P a g e | 539
43
c 32
18.9 = = 1.34375
c
Centerline: = 1.34375
c3 c 1.34375
UCL: = 1.34375 + 3 =
c3 c 1.34375
LCL: = 1.34375 - 3 =
c Chart:
18.10 a.) Six or more consecutive points are decreasing. Two of three
consecutive points are in the outer one-third (near LCL). Four out of
c.) One point is above the UCL. Two out of three consecutive points are
in
the outer one-third (both near LCL and near UCL). There are six
18.11 While there are no points outside the limits, the first chart exhibits
some
p1 = .14
P(x = 0) = C0(.14)0(.86)10 =
15 .2213
1 - [12C0(.04)0(.96)12 + 12 C1(.04)1(.96)11] =
p1 = .15
p Probability
.01 .9227
.02 .8506
.03 .7837
.06 .6096
.07 .5596
.09 .4703
.10 .4305
.11 .3937
.12 .3596
.13 .3282
.14 .2992
.15 .2725
OC Chart:
P a g e | 545
P a g e | 546
p Probability
.02 .9805
.04 .9308
.06 .8618
.10 .6974
.12 .6127
.14 .5311
.18 .3849
.20 .3221
.22 .2667
.24 .2186
OC Chart:
P a g e | 547
P a g e | 548
18.17
Stop
(no)
D K L M (yes) Stop
Stop
(no) (no)
Start A B (yes) C E F G
(yes)
H(no) J Stop
(yes)
I
P a g e | 549
1 673 26.96
2 29 1.16
3 108 4.33
4 379 15.18
5 73 2.92
6 564 22.60
7 12 0.48
8 402 16.11
9 54 2.16
10 202 8.09
2496
P a g e | 550
Pareto Chart:
13 C0(.05)0(.95)13 + 13 C1(.05)1(.95)12 =
(1)(1)(.51334) + (13)(.05)(.54036) =
13C0(.12)0(.88)13 + C1(.12)1(.88)12 =
13
(1)(1)(.18979) + (13)(.12)(.21567) =
b) n = 20, c = 2, p0 = .03
52
500
p= = .104
Centerline: p = .104
(. 104)(. 896)
50
UCL: .104 + 3 = .104 + .130 = .234
(.104)(.896)
50
LCL: .104 - 3 = .104 - .130 = .000
P a g e | 554
p Chart:
x1 x2 x3 x4
18.22 = 24.022, = 24.048, = 23.996, = 24.000,
x5 x6 x7 x8
= 23.998, = 24.018, = 24.000, = 24.034,
x9 x 10 x 11 x12
= 24.014, = 24.002, = 24.012, = 24.022
x R
= 24.01383 = 0.05167
x
For Chart: Since n = 12, A2 = .266
x
Centerline: = 24.01383
x R
UCL: + A2 = 24.01383 + (0.266)(.05167) =
x R
LCL: - A2 = 24.01383 - (0.266)(.05167) =
R
Centerline: = .05167
R
UCL: D4 = (1.716)(.05167) = .08866
R
LCL: D3 = (.284)(.05167) = .01467
P a g e | 556
x
Chart:
R Chart:
P a g e | 557
p Probability
.01 .8601
.02 .7386
.04 .5421
.06 .3953
.08 .2863
.10 .2059
.12 .1470
.14 .1041
OC Curve:
P a g e | 558
P a g e | 559
77
c 36
18.24 = = 2.13889
c
Centerline: = 2.13889
c3 c 2.13889
UCL: = 2.13889 + 3 =
c3 c 2.13889
LCL: = 2.13889 - 3 =
c Chart:
P a g e | 560
x1 x2 x3 x4
18.25 = 1.2100, = 1.2050, = 1.1900, = 1.1725,
x5 x6 x7 x8
= 1.2075, = 1.2025, = 1.1950, = 1.1950,
x9
= 1.1850
x R
= 1.19583 = 0.04667
x
For Chart: Since n = 4, A2 = .729
x
Centerline: = 1.19583
x R
UCL: + A2 = 1.19583 + .729(.04667) =
x R
LCL: - A2 = 1.19583 - .729(.04667) =
R
Centerline: = .04667
R
UCL: D4 = (1.816)(.04667) = .08475
R
LCL: D3 = (.184)(.04667) = .00859
x
Chart:
P a g e | 562
R chart:
x1 x2 x3 x4
18.26 = 14.99333, = 15.00000, = 14.97833, = 14.97833,
x5 x6 x7 x8
= 15.01333, = 15.00000, = 15.01667, = 14.99667,
x R
= 14.99854 = 0.05
x
For Chart: Since n = 6, A2 = .483
P a g e | 563
x
Centerline: = 14.99854
x R
UCL: + A2 = 14.99854 + .483(.05) =
x R
LCL: - A2 = 14.99854 - .483(.05) =
R
Centerline: = .05
R
UCL: D4 = 2.004(.05) = .1002
R
LCL: D3 = 0(.05) = .0000
x
Chart:
P a g e | 564
R chart:
70
900
p= = .07778 Centerline: p = .07778
(.07778)(.92222)
75
UCL: .07778 + 3 = .07778 + .09278 = .17056
(.07778)(.92222)
75
LCL: .07778 - 3 = .07778 - .09278 = .00000
p Chart:
P a g e | 566
16
c 25
18.28 = = 0.64
c
Centerline: = 0.64
c3 c 0.64
UCL: = 0.64 + 3 =
c3 c 0.64
LCL: = 0.64 - 3 =
c Chart:
P a g e | 567
P a g e | 568
p Probability
.05 .9885
.10 .9298
.15 .8202
.20 .6778
.25 .5256
.30 .3828
.35 .2616
.40 .1673
.45 .0996
.50 .0547
81
c 40
18.30 = = 2.025
c
Centerline: = 2.025
c3 c 2.025
UCL: = 2.025 + 3 =
c3 c 2.025
LCL: = 2.025 - 3 =
c Chart:
P a g e | 571
36
600
p= = .06
Centerline: p = .06
(.06)(.94)
40
UCL: .06 + 3 = .06 + .11265 = .17265
P a g e | 572
(.06)(. 94)
40
LCL: .06 - 3 = .06 - .112658 = .00000
p Chart:
18.33 There are some items to be concerned about with this chart. Only one
sample range is above the upper control limit. However, near the
beginning of the chart there are eight sample ranges in a row below
the centerline. Later in the run, there are nine sample ranges in a row
above the centerline. The quality manager or operator might want to
determine if there is some systematic reason why there is a string of
ranges below the centerline and, perhaps more importantly, why there
are a string of ranges above the centerline.
18.34 This p chart reveals that two of the sixty samples (about 3%) produce
proportions that are too large. Nine of the sixty samples (15%)
p̂
produce proportions large enough to be greater than 1 above the
centerline. In general, this chart indicates a process that is under
control.
18.35 The centerline of the c chart indicates that the process is averaging
0.74 nonconformances per part. Twenty-five of the fifty sampled items
have zero nonconformances. None of the samples exceed the upper
control limit for nonconformances. However, the upper control limit is
3.321 nonconformances, which in and of itself, may be too many.
Indeed, three of the fifty (6%) samples actually had three
nonconformances. An additional six samples (12%) had two
nonconformances. One matter of concern may be that there is a run of
ten samples in which nine of the samples exceed the centerline
(samples 12 through 21). The question raised by this phenomenon is
P a g e | 574
Chapter 19
Decision Analysis
LEARNING OBJECTIVES
The notion of contemporary decision making is built into the title of the
text as a statement of the importance of recognizing that statistical analysis
is primarily done as a decision-making tool. For the vast majority of students,
statistics take on importance only in as much as they aid decision-makers in
weighing various alternative pathways and helping the manager make the
best possible determination. It has been an underlying theme from chapter 1
that the techniques presented should be considered in a decision-making
context. This chapter focuses on analyzing the decision-making situation and
presents several alternative techniques for analyzing decisions under varying
conditions.
CHAPTER OUTLINE
Decision Table
Maximax Criterion
Maximin Criterion
Hurwicz Criterion
Minimax Regret
Decision Trees
Utility
KEY TERMS
EMV'er Risk-Avoider
c.) For = .3
decision: Select d2
For = .8
decision: Select d3
Comparing the results for the two different values of alpha, with a
more pessimist point-of-view ( = .3), the decision is to select d2 and
the payoff is 82. Selecting by using a more optimistic point-of-view (
= .8) results in choosing d3 with a higher payoff of 296.
S1 S2 S3 Max
d1 140 0 95 140
P a g e | 581
d2 280 75 0 280
d3 0 35 150 150
d2 80 20 75 100 100 20
d3 20 45 30 60 60 20
d5 0 -10 65 80 80 -10
Decision: Select d1
Decision: Select d1
c.) = .5
Decision: Select d1
P a g e | 583
S1 S2 S3 S4 Max
d1 50 15 0 0 50
d2 20 65 45 10 65
d3 80 40 90 50 90
d5 100 95 55 30 100
Decision: Select d1
P a g e | 584
A 60 15 -25 60 -25
B 10 25 30 30 10
C -10 40 15 40 -10
D 20 25 5 25 5
Decision: Select A
Decision: Select B
P a g e | 585
19.5, 19.6
P a g e | 588
19.10 EMV
No Layoff -960
d1 350 -100 35
F(S1) = .445
F(S2) = .555
19.13
S -225 425 35
M 125 -150 15
L 350 -400 50
F(Dec) = .51
F(Inc) = .49
P a g e | 597
P(Fdec) = .156
P a g e | 600
P(Fsame) = .375
P(Finc) = .469
P a g e | 602
Don't Drill 0 0 0
Actual
Oil No Oil
Forecast
Forecast Oil:
P(FOil) = .111
Forecast No Oil:
d1 50 100 100 50
d3 25 40 40 25
d4 75 10 75 10
Decision: Select d2
Decision: Select d1
Decision: Select d2
S1 S2 Maximum
d1 25 100 100
d2 150 0 150
d3 50 160 160
d4 0 190 190
Decision: Select d1
P a g e | 607
19.17
Decision: Select d1
Select d2
Forecast S1:
S1 .4 .9 .36 .667
S2 .6 .3 .18 .333
P(FS1) = .54
Forecast S2:
P a g e | 610
S1 .4 .1 .04 .087
S2 .6 .7 .42 .913
P(FS2) = .46
P a g e | 611
a.) = .1:
b.) = .5:
c.) = .8:
P a g e | 616
d.) Two of the three alpha values (.5 and .8) lead to a decision of pricing
low.
Decision: Based on Max EMV = Max {-235, 90} = 90, select Produce.
Forecast Reduction:
Forecast Constant:
Forecast Increase:
P(FInc) = .400
P a g e | 623
Forecast Chosen:
P(FC) = .410
P(FC) = .590
16.19 y x1 x2 x3
16.20 y x1 x2 x3 x4
Net
Income
Natural Fuel
P a g e | 631
Natural
ŷ
= 564 - 27.99 x1 - 6.155 x2 - 15.90 x3
ŷ
= 1540 + 48.2 x1.
The R2 at this step is .9112 and the t ratio is 11.55. At step 2, x12
entered the procedure and x1 remained in the analysis. The stepwise
regression procedure stopped at this step and did not proceed. The
final model was:
ŷ
= 1237 + 136.1 x1 - 5.9 x12.
P a g e | 633
The R2 at this step was .9723, the t ratio for x1 was 7.89, and the t ratio
for x12 was - 5.14.
16.25 In this model with x1 and the log of x1 as predictors, only the log x1 was
a significant predictor of y. The stepwise procedure only went to step
1. The regression model was:
ŷ
= - 13.20 + 11.64 Log x1. R2 = .9617 and the t ratio of Log x1
was 17.36. This model has very strong predictability using only the
log of the x1 variable.
P a g e | 634
toilseed = 3.74 with p = .006 and tlivestock = 3.78 with p = .005. Both
predictors are significant at = .01. This is a model with strong
predictability.
16.27 The stepwise regression procedure only used two steps. At step
2
1, Silver was the lone predictor. The value of R was .5244. At
step 2, Aluminum entered the model and Silver remained in the
model. However, the R2 jumped to .8204. The final model at step 2
was:
The low value of adjusted R2 indicates that the model has very low
predictability. Both t values are not significant (tNavalVessels = 0.67 with
The t ratios were: tfood = 8.32, tfuel oil = 2.81, tshelter = 2.56.
16.30 The stepwise regression process with these two independent variables
only went one step. At step 1, Soybeans entered in producing the
model,
Corn = - 2,962 + 5.4 Soybeans. The R2 for this model was .7868.
The t ratio for Soybeans was 5.43. Wheat did not enter in to the
analysis.
Only one of the four predictors has a significant t ratio and that is
Utility with t = 2.57 and p = .018. The ratios and their respective
probabilities are:
This model is very weak. Only the predictor, Utility, shows much
promise in accounting for the grocery variability.
P a g e | 638
16.32 The output suggests that the procedure only went two steps.
ŷ
= 124.5 - 43.4 x1 + 1.36 x2
ŷ
= 87.89 + 0.071 x3 - 2.71 x2 - 0.256 x1
P a g e | 639
16.34 The R2 for the full model is .321. After dropping out variable, x3, the R2
is still .321. Variable x3 added virtually no information to the model.
This is underscored by the fact that the p-value for the t test of the
slope for x3 is .878 indicating that there is no significance. The
standard error of the estimate actually drops slightly after x3 is
removed from the model.
Chapter 17
Time-Series Forecasting and Index Numbers
LEARNING OBJECTIVES
The first section of the chapter contains a general discussion about the
various possible components of time-series data. It creates the setting
against which the chapter later proceeds into trend analysis and seasonal
effects. In addition, two measurements of forecasting error are presented so
P a g e | 641
that students can measure the error of forecasts produced by the various
techniques and begin to compare the merits of each.
Trend is solved for next using the time periods as the predictor
variable. In this chapter both linear and quadratic trends are explored and
compared. There is a brief introduction to Holt’s two-parameter exponential
smoothing method that includes trend. A more detailed explanation of Holt’s
method is available on WileyPLUS. The trend analysis section is placed
earlier in the chapter than seasonal effects because finding seasonal effects
makes more sense when there are no trend effects in the data or the trend
effect has been removed.
CHAPTER OUTLINE
Error
Averaging Models
Simple Averages
Moving Averages
Exponential Smoothing
Decomposition
Autocorrelation
Transforming Variables
Autoregression
KEY TERMS
Forecasting Trend
e
17.1 Period e e2
e
12.30
no. forecasts 9
MAD = = 1.367
e 2
20.43
no. forecasts 9
MSE = = 2.27
P a g e | 647
e
17.2 Period Value F e e2
1 202 -
5 171 174 -3 3 9
6 175 172 3 3 9
7 182 174 8 8 64
8 196 179 17 17 289
e
125.00
no. forecasts 10
MAD = = 12.5
e 2
1,919
no. forecasts 10
MSE = = 191.9
P a g e | 648
e
17.3 Period Value F e e2
e
2 1 .5
N o .F o r e c a s t s 6
MAD = = 3.583
e2
9 4 .5 9
N o .F o r e c a s t s 6
MSE = = 15.765
e
17.4 Year Acres Forecast e e2
1 140,000 - - - -
e
4 9 ,0 4 7
N o .F o r e c a s t s 10
MAD = = 4,904.7
e2
3 6 1 , 3 3 1 ,8 4 7
N o .F o r e c a s t s 10
MSE = = 36,133,184.7
44.75 14.25
52.75 13.25
61.50 9.50
64.75 21.25
P a g e | 650
70.50 30.50
81.00 16.00
53.25 5.75
56.375 9.625
62.875 8.125
67.25 18.75
76.375 24.625
89.125 7.875
P a g e | 651
3.626
1.375
2.5
5.875
8.125
1 211
and 7), the forecasts using alpha = .1 produced smaller errors. Each
exponential smoothing model produced nearly the same amount of
error in forecasting the value for period 8. There is no strong argument
in favor of either model.
17.7 Period Value =.3 Error =.7 Error 3-mo.avg.
Error
1 9.4
1 2512.7
2 2739.9
3 2874.9
4 2934.1
5 2865.7
e e
17.9 Year No.Issues F(=.2) F(=.9)
1 332 -
e e
= 2289.9 =2023.0
2289.9
12
For = .2, MAD = = 190.8
2023.0
12
For = .9, MAD = = 168.6
ŷ
= 37,969 + 9899.1 Period
P a g e | 655
ŷ
= 35,769 + 10,473 Period - 26.08 Period2
17500
17000
16500
16000
Union Members
15500
15000
14500
Year
P a g e | 657
17.12
1200
1000
800
600
Loans
400
200
0 1 2 3 4 5 6 7 8 9 10
Year
Trend Model:
Quadratic Model:
F = 28.95 (p = .0008)
The graph indicates a quadratic fit rather than a linear fit. The
quadratic model produced an R2 = 90.6 compared to R2 = 33.0 for
linear trend indicating a much better fit for the quadratic model. In
addition, the standard error of the estimate drops from 96.80 to 39.13
with the quadratic model along with the t values becoming significant.
P a g e | 659
17.13
Jan.(yr. 1) 132.5
Feb. 164.8
Mar. 141.2
Apr. 133.8
May 138.4
June 150.9
1655.2
1627.6
1562.1
1522.9
1511.5
1485.2
1442.7
1415.1
1387.2
1363.4
1341.4
1340.7
1332.0
July 119.0
Aug. 119.0
Sept. 114.9
Oct. 106.0
Nov. 111.7
Dec. 112.3
P a g e | 661
17.14
23822
23867
23985
24124
24268
24431
24695
24926
25063
25245
25485
25647
25863
26110
P a g e | 663
26236
26332
26520
26726
28193
28346
P a g e | 664
28590
28914
29161
29265
29308
29377
29438
29368
29425
29495
29523
29576
29565
29541
29392
29387
29307
29275
29268
P a g e | 666
29308
29279
29276
29182
29170
29088
28834
28824
28723
28677
28714
28694
28652
28683
28679
28745
July 2341
Aug 2491
Sept 2452
Oct 2561
Nov 2377
Dec 2277
Seasonal Indexing:
Total 1199.69
Month Index
Jan 96.85
Feb 100.52
Mar 100.67
Apr 100.74
May 101.17
P a g e | 669
June 105.01
July 95.30
Aug 103.09
Sept 103.86
Oct 103.82
Nov 96.07
Dec 92.92
Ŷ
Regression Output for Trend Line: = 2035.58 + 7.1481 X
R2 = .682, se = 102.9
Note: Trend Line was determined after seasonal effects were removed (based on
TCI column).
Yˆ
Food Housing e e2
Total 21.1603
P a g e | 671
(e e t t 1 )2
= 1.873 + 2.503 + 0.173 + 0.505 + 1.825 + 2.228 +
0.490 +
e 2
= 21.160
( e t e t 1 ) 2
2 4 .5 6 1
e 2
2 1 .1 6 0
D = = 1.16
Since D = 1.16 is less than dL, the decision is to reject the null
hypothesis.
8.5 15.7 -
Since D = 1.04 is less than dL, the decision is to reject the null
hypothesis.
ŷ
for x= 150: = 21,881 (million $)
D = 2.49
P a g e | 674
ŷ
Failed Bank Assets Number of Failures e e2
27 1 1,516.0 - 1,489.0
2,217,144
P a g e | 676
17.18
8,189 11 -
104 7 -4
1,862 34 27
4,137 45 11
36,394 79 34
3,034 118 39
7,609 144 26
7,538 201 57
56,620 221 20
16,915 100 -8
2,588 42 -58
825 11 -31
753 6 -5
186 5 -1
27 1 -4
D = 1.57
333.0 * *
P a g e | 678
270.4 333.0 *
The model with 1 lag is the best model with a strong R2 = 89.2%. The
model
The F value for this model is 27.0 which is significant at alpha = .001.
17.23 Year
9 .0 4
(1 0 0 )
9 .0 4
Index1993 = = 100.0
9 .3 3
(1 0 0 )
9 .0 4
Index1999 = = 103.2
1 1 .2 8
(1 0 0 )
9 .0 4
Index2005 = = 124.8
17.24 Year
P a g e | 684
1 2 .4 3
(1 0 0 )
1 3 .2 3
Index1998 = = 94.0
1 2 .5 5
(1 0 0 )
1 3 .2 3
Index1999 = = 94.9
1 3 .2 3
(1 0 0 )
1 3 .2 3
Index2000 = = 100.0
1 3 .1 7
(1 0 0 )
1 3 .2 3
Index2001 = = 99.5
P a g e | 685
1 2 .8 2
(1 0 0 )
1 3 .2 3
Index2002 = = 100.0
1 3 .2 2
(1 0 0 )
1 3 .2 3
Index2003 = = 101.0
1 3 .7 9
(1 0 0 )
1 3 .2 3
Index2004 = = 106.4
1 4 .6 8
(1 0 0 )
1 3 .2 3
Index2005 = = 111.0
1 5 .1 7
(1 0 0 )
1 3 .2 3
Index2006 = = 114.7
P 2005Q 2005
(1 0 0 ) 519.15
P 2000 Q 2005 422.85
(100)
Index2005 = = = 122.8
P 2006 Q 2006
(1 0 0 ) 546.72
P 2000 Q 2006 438.60
(100)
Index2006 = = = 124.7
b) x F e
10.08 - -
10.05 - -
9.24 -
-
9.23 - -
e
= 6.77
6.77
20
MAD = = .3385
c)
= .3 = .7
e e
x F F
10.08 - - - -
8.59
8.98 .39
8.50 .09
7.99 8.86 .87 8.56 .57
e e
= 10.06 = 5.97
10.06 5.97
23 23
MAD=.3 = = .4374 MAD=.7 = = .2596
d). MAD for b) .3385, c) .4374 and .2596. Exponential smoothing with
= .7
e) 4 period 8 period
10.08
10.05
38.60
P a g e | 692
38.21
37.71
37.84
37.16
35.83
34.87
33.49
33.06
32.61
31.75
31.15
30.51
30.12
29.87
29.83
29.39
28.75
28.15
27.97
28.15
7.17
7.22
The highs and lows of each period (underlined) are eliminated and the
others are
2nd 101.05
3rd 98.64
4th 98.67
total 398.18
Since the total is not 400, adjust each seasonal index by multiplying by
400
398.18
= 1.004571 resulting in the final seasonal indexes of:
1st 100.28
2nd 101.51
3rd 99.09
4th 99.12
P 2002
(1 0 0 )
8 .1 4
(1 0 0 )
P 2002 8 .1 4
Index2002 = = 100.0
P 2003
(1 0 0 )
8 .5 3
(1 0 0 )
P 2002 8 .1 4
Index2003 = = 104.8
P a g e | 696
P 2004
(1 0 0 )
9 .3 2
(1 0 0 )
P 2002 8 .1 4
Index2004 = = 114.5
P 2005
(1 0 0 )
9 .4 0
(1 0 0 )
P 2002 8 .1 4
Index2005 = = 115.5
P 2006
(1 0 0 )
9 .2 9
(1 0 0 )
P 2002 8 .1 4
Index2006 = = 114.1
Item P Q P Q P Q P Q
33.00 38.52
39.95 46.06
26.60 28.00
P 2006 Q 2003
(1 0 0 ) 112.58
P 2003Q 2003 99.55
(100)
Laspeyres Index2006 = = = 113.1
24.75 27.90
51.85 57.95
33.25 34.00
P 2005Q 2005
(1 0 0 ) 119.85
P 2003Q 2005 109.85
(100)
Paasche Index2005 = = = 109.1
P a g e | 698
P a g e | 699
e e
Year Quantity F F
1980 6559
e e
=11,889.67 =18,621.46
e 1 1 ,8 8 9 .6 7
numberforecasts 22
MADmoving average = = = 540.44
e 1 8 , 6 2 1 .4 6
numberforecasts 22
MAD=.2 = = = 846.43
17.32-17.34
Jan(1) 23.701
Feb 24.189
Mar 24.200
Apr 24.971
May 24.560
June 24.992
288.00
287.65
287.58
288.66
289.12
289.74
291.24
292.76
293.39
294.42
294.63
295.42
297.21
298.07
299.72
P a g e | 703
302.03
303.56
304.29
306.27
307.00
307.89
309.03
310.36
313.12
314.89
316.64
318.67
321.17
322.86
324.79
328.19
331.76
334.70
337.87
341.52
345.14
349.16
352.18
354.11
356.43
359.07
361.67
P a g e | 706
363.47
364.32
368.01
370.08
July 29.342
Aug 30.765
Sept 31.637
Oct 30.206
Nov 30.842
Dec 31.090
Seasonal Indexing:
Total 1199.88
Month Index
Jan 95.35
Feb 99.69
Mar 106.75
Apr 103.99
May 100.99
June 106.96
July 94.53
Aug 99.60
Sept 104.16
Oct 97.04
Nov 95.75
Dec 95.19
ŷ
= 22.4233 + 0.144974 x
R2 = .913
ŷ
= 23.8158 + 0.01554 x + .000247 x2
R2 = .964
In this model, the linear term yields a t = 0.66 with p = .513 but the
squared term predictor yields a t = 8.94 with p = .000.
ŷ
= 23.9339 + 0.00236647 x2
R2 = .964
Note: The trend model derived using only the squared predictor was used in computing T
(trend) in the decomposition process.
P a g e | 710
17.35
P 2004
(1 0 0 )
7 .4 9
(1 0 0 )
P 2004 7 .4 9
Index2004 = = 100.0
P 2005
(1 0 0 )
7 .7 3
(1 0 0 )
P 2004 7 .4 9
Index2005 = = 103.2
P 2006
(1 0 0 )
8 .3 7
(1 0 0 )
P 2004 7 .4 9
Index2006 = = 111.8
P 2005Q 2004
(1 0 0 ) 2 3 1 .2 3
P 2004 Q 2004 2 1 9 .7 3
(1 0 0 )
IndexLaspeyres2005 = = = 105.2
P 2006 Q 2004
(1 0 0 ) 2 4 3 .9 2
P 2004 Q 2004 2 1 9 .7 3
(1 0 0 )
IndexLaspeyres2006 = = = 111.0
P 2005Q 2005
(1 0 0 ) 2 3 5 .5 5
P 2004 Q 2005 2 2 4 .1 8
(1 0 0 )
IndexPaasche2005 = = = 105.1
P 2006 Q 2006
(1 0 0 ) 2 3 7 .8 3
P 2004 Q 2006 2 1 4 .6 6
(1 0 0 )
IndexPaasche2006 = = = 110.8
ŷ
17.36 = 9.5382 – 0.2716 x
ŷ
(7) = 7.637
se = 0.264862
Durbin-Watson:
n = 21 k=1 = .05
D = 0.44
P a g e | 713
1988 118.5
1989 123.0
1990 128.5
1991 133.6
SE = 1,727.60 1,111.40
SE 1 7 2 7 .6 0
N o .F o r e c a sts 14
MSEma = = 123.4
P a g e | 715
SE 1 1 1 1 .4
N o .F o r e c a sts 14
MSEwma = = 79.39
The model with the four-month lag does not have overall significance
and has an
17.39
P a g e | 717
Year1 1 54.019
2 56.495
213.574
211.470
210.076
213.326
217.671
222.819
230.206
237.160
248.918
254.810
P a g e | 718
257.693
260.805
263.527
263.158
263.147
263.573
257.842
253.421
248.264
3 58.088
4 61.443
Total 399.26
400
399.26
Adjust the seasonal indexes by: = 1.00185343
Quarter Index
1 98.07
2 103.84
3 97.04
4 101.05
Total 400.00
P a g e | 720
Q1(yr1)
55.082
Q2 54.406
Q3 51.699
Q4 52.341
Q1(yr2) 52.937
Q2 53.063
Q3 55.048
Q4 56.641
Q1(yr3) 58.186
Q2 60.177
Q3 62.215
Q4 62.676
Q1(yr4) 63.957
Q2 65.851
Q3 65.185
Q4 65.756
Q1(yr5) 66.733
Q2 65.496
Q3 65.174
Q4 66.177
Q1(yr6) 60.889
Q2 61.238
Q3 59.860
Q4 60.805
P a g e | 721
ŷ
17.41 Linear Model: = 53.41032 + 0.532488 x
se = 3.43
ŷ
Quadratic Model: = 47.68663 + 1.853339 x –0.052834 x2
se = 2.55
In the quadratic regression model, both the linear and squared terms
have significant t statistics at alpha .001 indicating that both are
contributing. In addition, the R2 for the quadratic model is
considerably higher than the R2 for the linear model. Also, se is smaller
for the quadratic model. All of these indicate that the quadratic model
is a stronger model.
P a g e | 722
se = 50.18
This model with a lag of one year has modest predictability. The
overall F is
R2 = 88.2% se = 582.685
D = 0.84
e e e
Year PurPwr F F F
1 6.04
e e e
= 10.26 . = 4.83 = 3.97
e 10.26
N 17
MAD1 = = = .60
e 4.83
N 17
MAD2 = = = .28
e 3.97
N 17
MAD3 = = = .23
- 1,338.58 1,791,796
(e t et 1 ) 2 e t
2
=921,525,945 =936,737,358
P a g e | 726
(e e
t t 1 )2
921,525,945
e
2
t
936,737,358
D = = 0.98