Solution Manual Stats
Solution Manual Stats
Solution Manual Stats
= 0.113
1000
0.032
P( A B)
(e) P ( A | B ) =
=
= 0.2207
(25 + 63 + 15 + 7 + 35) / 1000
P( B)
5
(f) P =
= 0.005
1000
(d) P ( A'B ) =
2-167.
(a) Let A denote that a part conforms to specifications and let B denote for a part a simple
component.
For supplier 1,
P(A)=P(A|B)P(B)+ P(A|B)P(B)
= [(1000-2)/1000](1000/2000)+ [(1000-10)/1000](1000/2000)
= 0.994
For supplier 2,
P(A)= P(A|B)P(B)+ P(A|B)P(B)
= [(1600-4)/1600](1600/2000)+ [(400-6)/400](400/2000)
= 0.995
(b)
For supplier 1,
For supplier 2,
(c)
For supplier 1,
For supplier 2,
P(A|B)=0.99
P(A|B)=0.985
P(A|B)=0.998
P(A|B)=0.9975
(d) The unusual result is that for both a simple component and for a complex assembly, supplier 1
has a greater probability that a part conforms to specifications. However, for overall parts,
supplier 1 has a lower probability. The overall conforming probability is dependent on both the
conforming probability conditioned on the part type and the probability of each part type. Supplier
1 produces more of the compelx parts so that overall conformance from supplier 1 is lower.
Mind-Expanding Exercises
2-168.
2-169.
Let E denote a read error and let S, O, B, P denote skewed, off-center, both, and proper alignments,
respectively.
P(E) = P(E|S)P(S) + P(E|O)P(O) + P(E|B)P(B) + P(E|P)P(P)
= 0.01(0.10) + 0.02(0.05) + 0.06(0.01) + 0.001(0.84) = 0.00344
Let n denote the number of washers selected.
a) The probability that non are thicker than, that is, all are less than the target is 0.4 n , by independence.
n
0.4 n
1
0.4
2
0.16
3
0.064
Therefore, n = 3
b) The requested probability is the complement of the probability requested in part a. Therefore, n = 3
2-29
2-170.
Mean Profit
Maximum Profit
74.75 x
$ 3737.50 at x=50
32.75 x + 2100
$ 5375 at x=100
1.25 x + 5250
$ 5500 at x=200
Therefore, profit is maximized at 200 kits. However, the difference in profit over 100 kits is small.
0 x 50
50 x 100
100 x 200
2-171.
Let E denote the probability that none of the bolts are identified as incorrectly torqued. The requested
probability is P(E'). Let X denote the number of bolts in the sample that are incorrect. Then,
P(E) = P(E|X=0)P(X=0) + P(E|X=1) P(X=1) + P(E|X=2) P(X=2) + P(E|X=3) P(X=3) + P(E|X=4)P(X=4)
and P(X=0) = (15/20)(14/19)(13/18)(12/17) = 0.2817. The remaining probability for x can be determined
from the counting methods in Appendix B-1. Then,
( )( ) = 4 ! 1! 3 ! 12 ! = 5 ! 15 ! 4 ! 16 ! = 0.4696
P( X = 1) =
( ) 420! 16! ! 4 ! 3 ! 12 ! 20 !
5 15
1 3
5 ! 15 !
20
4
( )( ) = 3 ! 2 ! 2 ! 13 ! = 0.2167
P( X = 2) =
20 !
( )
4 ! 16 !
5 15
2 2
5 ! 15 !
20
4
( )( ) = 3 ! 2 ! 1! 14 ! = 0.0309
P( X = 3) =
( ) 420! 16! !
5 15
3 1
5 ! 15 !
20
4
2-30
2-173.
2-31
CHAPTER 2
Section 2-1
2-1.
2-2.
S=
2-4.
S = {0,12
, ,...} = set of nonnegative integers
2-5.
, ,...,24}
If only the number of tracks with errors is of interest, then S = {0,12
2-6.
A vector with three components can describe the three digits of the ammeter. Each digit can be
0,1,2,...,9. Then S is a sample space of 1000 possible three digit integers, S = {000,001,...,999}
2-7.
2-8.
2-9.
2-10.
2-11.
2-12.
2-13
2-14.
S = {0,1,2,...,} in milliseconds
S = {1.0,1.1,1.2, 14.0}
S = {0,1,2,...,} in milliseconds.
automatic
transmission
with
air
standard
transmission
with
air
without
air
2-1
without
air
2-15.
PRESS
CAVITY
2-16.
memory
12
disk storage
200
300
400
200
300
2-17.
2-18.
2-19
a)
b)
c)
2-2
400
200
300
400
d)
e)
2.20
a)
2-3
b)
c)
d)
e)
2-4
2-21.
a) S = nonnegative integers from 0 to the largest integer that can be displayed by the scale.
Let X represent weight.
A is the event that X > 11
B is the event that X 15
C is the event that 8 X <12
S = {0, 1, 2, 3, }
b) S
c) 11 < X 15 or {12, 13, 14, 15}
d) X 11 or {0, 1, 2, , 11}
e) S
f) A C would contain the values of X such that: X 8
Thus (A C) would contain the values of X such that: X < 8 or {0, 1, 2, , 7}
g)
h) B would contain the values of X such that X > 15. Therefore, B C would be the empty set. They
have no outcomes in common or
i) B C is the event 8 X <12. Therefore, A (B C) is the event X 8 or {8, 9, 10, }
2-22.
a)
C
b)
C
c)
d)
2-5
e)
2-23.
If the events are mutually exclusive, then AB is equal to zero. Therefore, the process does not
produce product parts with X=50 cm and Y=10 cm. The process would not be successful.
Let "d" denoted a distorted bit and let "o" denote a bit that is not distorted.
dddd, dodd, oddd, oodd,
c)
dddd , dodd ,
dddo, dodo
A1 =
ddod , dood
ddoo, dooo
d)
oddd , oodd ,
oddo, oodo,
A1 =
odod , oood ,
odoo, oooo
e) A1 A2 A3 A4 = {dddd }
f) ( A1 A2 ) ( A3 A4 ) = {dddd , dodd , dddo, oddd , ddod , oodd , ddoo}
2-24
Let w denote the wavelength. The sample space is {w | w = 0, 1, 2, }
(a) A={w | w = 675, 676, , 700 nm}
(b) B={ w | w = 450, 451, , 500 nm}
(c) A B =
(d) A B = {w | w = 450, 451, , 500, 675, 676, , 700 nm}
2-25
Let P denote being positive and let N denote being negative.
The sample space is {PPP, PPN, PNP, NPP, PNN, NPN, NNP, NNN}.
(a) A={ PPP }
(b) B={ NNN }
(c) A B =
(d) A B = { PPP , NNN }
2-26.A B = 70, A = 14, A B = 95
2-27.
a) A B = 10, B =10, A B = 92
b)
2-6
Surface 1
Edge 1
G
E
G
Surface 2
E
E
Edge 2
E
G
E
E
E
E
G
G
2-28.
2-29.
A B = 55, B =23, A B = 85
a) A = {x | x 72.5}
b) B = {x | x 52.5}
c) A B = {x | 52.5 < x < 72.5}
d) A B = {x | x > 0}
2.30
a) {ab, ac, ad, bc, bd, cd, ba, ca, da, cb, db, dc}
b) {ab, ac, ad, ae, af, ag, bc, bd, be, bf, bg, cd, ce, cf, cg, ef, eg, fg, ba, ca, da, ea, fa, ga, cb, db,
eb, fb, gb, dc, ec, fc, gc, fe, ge, gf}
2.31
b) S={gg,gm,gj,mg,mm,mj,jg,jm}
2-32.a.) The sample space contains all points in the positive X-Y plane.
b)
10
c)
20
B
d)
2-7
20
10
10
e)
B
20
2-33
a)
b)
c)
2-8
d)
2-34.212 = 4096
2-35.
From the multiplication rule, the answer is 5 3 4 2 = 120
2-36.
2-37.
2-38.
2-39.
From the multiplication rule and equation 2-1, the answer is 5!5! = 14,400
2-40.
2-41.
7!
= 35 sequences are possible
3! 4!
!
( ) = 5140
= 416,965,528
!135!
140
5
b) There are 10 ways of selecting one nonconforming chip and there are
!
( ) = 4130
= 11,358,880
!126!
130
4
ways of selecting four conforming chips. Therefore, the number of samples that contain exactly one
( )
= 113,588,800
nonconforming chip is 10 4
c) The number of samples that contain at least one nonconforming chip is the total number of samples
130
130
5
2-9
( ) ( )
140! 130!
= 130,721,752
5!135! 5!125!
That is 140
- 130
=
5
5
2-42.
a) If the chips are of different types, then every arrangement of 5 locations selected from the 12 results in a
different layout. Therefore,
P512 =
12!
= 95,040
7!
b) If the chips are of the same type, then every subset of 5 locations chosen from the 12 results in a different
12!
=
= 792 layouts are possible.
layout. Therefore, 12
5
5! 7!
( )
2-43.
a)
b)
7!
= 21 sequences are possible.
2!5!
7!
= 2520 sequences are possible.
1!1!1!1!1!2!
2-44.
12!
= 792
5!7!
locations for the second component are selected from the nine remaining locations in
ways. From the multiplication rule, the number of designs is
designs are
( ) = 49!5!! = 126
9
4
2-45.
2-46.
2-47.
!
( ) = 424
= 10,626. The number of samples in which
!20!
24
4
!
( )( ) = 16!5!! 318
= 4896 . Therefore, the probability is
!15!
6
1
18
3
4896
= 0.461
10626
2-10
3060
= 0.712 .
10626
requested probability is 1
2-48.
18
4
!
( )( )( ) = 16!5!! 14!3!! 214
= 2184 .
!12!
6
1
4
1
14
2
2184
= 0.206
10626
( )
12 !
a) The total number of samples is 12
3 = 3 ! 9 ! = 220. The number of samples that result in one
nonconforming part is
10
2
90/220 = 0.409.
b) The number of samples with no nonconforming part is
nonconforming part is 1
2-49.
!
( ) = 418
= 3060. Therefore, the
!14!
120
= 0.455 .
220
5 4
= 0.0082
50 49
50! 50 49
50
=
. The number of samples with two defective
b) The total number of samples is 2 =
2!48!
2
5 4
5 4
5! 5 4
5
2
. Therefore, the probability is
=
= 0.0082 .
parts is 2 =
=
50 49
50 49
2!3!
2
2
a) The probability that both parts are defective is
( )
()
2-11
(b) There are 36 experiments that use all three steps. The probability the best result uses all three
steps is 36/52 = 0.6923.
(a) No, it will not change. With k amounts in the first step the number of experiments is k + 3k +
9k = 13k. The number of experiments that complete all three steps is 9k out of 13k. The
probability is 9/13 = 0.6923.
2-62.
2-63.
2-64.
P ( A) =
44 + 4 + 4
= 0.52
100
100 5
= 0.95
100
44 + 4 + 4
(c) P ( A B ) =
= 0.52
100
(d) P ( A B ' ) = 0
100 5
= 0.95
(e) P ( A B ) =
100
(b) P ( B ) =
2-65.
Section 2-3
2-66.
2-67.
a) P(
A B C ) = P(A) + P(B) + P(C), because the events are mutually exclusive. Therefore,
P( A B C ) = 0.2+0.3+0.4 = 0.9
b) P ( A B C ) = 0, because A B C =
2-13
c) P( A B ) = 0 , because A B =
d) P( ( A B ) C ) = 0, because ( A B ) C = ( A C ) ( B C )
e) P( A B C ) =1-[ P(A) + P(B) + P(C)] = 1-(0.2+0.3+0.4) = 0.1
2-68.
2-70.
2-71.
a) 350/370
345 + 5 + 12 362
b)
=
370
370
345 + 5 + 8 358
c)
=
370
370
d) 345/370
2-74.
(a) 5/36
(b) 5/36
(c) P ( A B )
= P( A) P( B) = 0.01929
(d) P ( A B ) = P( A) + P ( B ) = 0.2585
Section 2-4
2-75.
a) P(A) = 86/100
b) P(B) = 79/100
2-14
c) P( A B ) = 0 , because A B =
d) P( ( A B ) C ) = 0, because ( A B ) C = ( A C ) ( B C )
e) P( A B C ) =1-[ P(A) + P(B) + P(C)] = 1-(0.2+0.3+0.4) = 0.1
2-68.
2-70.
2-71.
a) 350/370
345 + 5 + 12 362
b)
=
370
370
345 + 5 + 8 358
c)
=
370
370
d) 345/370
2-74.
(a) 5/36
(b) 5/36
(c) P ( A B )
= P( A) P( B) = 0.01929
(d) P ( A B ) = P( A) + P ( B ) = 0.2585
Section 2-4
2-75.
a) P(A) = 86/100
b) P(B) = 79/100
2-14
2-76.
c) P( A B ) =
P ( A B) 70 / 100 70
=
=
P( B)
79 / 100 79
d) P( B A ) =
P( A B) 70 / 100 70
=
=
P( A)
86 / 100 86
7 + 32
= 0.39
100
13 + 7
= 0.2
(b) P ( B ) =
100
P( A B) 7 / 100
=
= 0.35
(c) P ( A | B ) =
20 / 100
P( B)
P( A B) 7 / 100
=
= 0.1795
(d) P ( B | A) =
39 / 100
P( A)
(a) P ( A) =
2-77.
Let A denote the event that a leaf completes the color transformation and let B denote the event
that a leaf completes the textural transformation. The total number of experiments is 300.
P( A B)
243 / 300
=
= 0.903
(243 + 26) / 300
P( A)
P( A B' )
26 / 300
=
= 0.591
(b) P ( A | B ' ) =
(18 + 26) / 300
P( B' )
(a) P ( B | A) =
2-78.a) 0.82
b) 0.90
c) 8/9 = 0.889
d) 80/82 = 0.9756
e) 80/82 = 0.9756
f) 2/10 = 0.20
2-79.
a) 12/100
b) 12/28 c) 34/122
2-80.
a) P(A) = 0.05 + 0.10 = 0.15
P( A B ) 0.04 + 0.07
b) P(A|B) =
=
= 0.153
0.72
P( B)
c) P(B) = 0.72
d) P(B|A) = P( A B) = 0.04 + 0.07 = 0.733
P( A)
0.15
e) P(A B) = 0.04 +0.07 = 0.11
f) P(A B) = 0.15 + 0.72 0.11 = 0.76
2-81. Let A denote the event that autolysis is high and let B denote the event that putrefaction is high. The
total number of experiments is 100.
P( A B' )
18 / 100
=
= 0.5625
(14 + 18) / 100
P( A)
P( A B)
14 / 100
=
= 0.1918
(b) P ( A | B ) =
(14 + 59) / 100
P( B)
(a) P ( B ' | A) =
2-15
P ( A' B' )
9 / 100
=
= 0.333
(18 + 9) / 100
P( B' )
2-83.
a) 20/100
b) 19/99
c) (20/100)(19/99) = 0.038
d) If the chips are replaced, the probability would be (20/100) = 0.2
2-84.
a) 4/499 = 0.0080
b) (5/500)(4/499) = 0.000080
c) (495/500)(494/499) = 0.98
d) 3/498 = 0.0060
e) 4/498 = 0.0080
f) 5 4 3 = 4.82x10 7
500 499 498
2-85.
2-86.
2-87.
a)P=(8-1)/(350-1)=0.020
b)P=(8/350) [(8-1)/(350-1)]=0.000458
c)P=(342/350) [(342-1)/(350-1)]=0.9547
1
36 7
1
(b)
5(36 6 )
1
(c)
5(365 )5
(a)
P( A B) P(B)
=
=1
P(B)
P(B)
A
B
2-88.
Section 2-5
2-16
P ( A' B' )
9 / 100
=
= 0.333
(18 + 9) / 100
P( B' )
2-83.
a) 20/100
b) 19/99
c) (20/100)(19/99) = 0.038
d) If the chips are replaced, the probability would be (20/100) = 0.2
2-84.
a) 4/499 = 0.0080
b) (5/500)(4/499) = 0.000080
c) (495/500)(494/499) = 0.98
d) 3/498 = 0.0060
e) 4/498 = 0.0080
f) 5 4 3 = 4.82x10 7
500 499 498
2-85.
2-86.
2-87.
a)P=(8-1)/(350-1)=0.020
b)P=(8/350) [(8-1)/(350-1)]=0.000458
c)P=(342/350) [(342-1)/(350-1)]=0.9547
1
36 7
1
(b)
5(36 6 )
1
(c)
5(365 )5
(a)
P( A B) P(B)
=
=1
P(B)
P(B)
A
B
2-88.
Section 2-5
2-16
2-89.
2-90.
P( A ) = P( A B) + P( A B )
= P( A B)P(B) + P( A B )P(B )
= (0.2)(0.8) + (0.3)(0.2)
= 0.16 + 0.06 = 0.22
2-91.
2-92.
P ( F) = P ( F C ) P ( C ) + P ( F C ) P ( C )
= ( 0. 02 )( 0. 70) + ( 0. 03)( 0. 30) = 0. 023
2-93.
Let R denote the event that a product exhibits surface roughness. Let N,A, and W denote the events that the
blades are new, average, and worn, respectively. Then,
P(R)= P(R|N)P(N) + P(R|A)P(A) + P(R|W)P(W)
= (0.01)(0.25) + (0.03) (0.60) + (0.05)(0.15)
= 0.028
2-94.
Let A denote the event that a respondent is a college graduate and let B denote the event that a
voter votes for Bush.
P(B)=P(A)P(B|A)+ P(A)P(B|A)=0.38 0.53+0.62 0.5=0.5114
2-95.a) (0.88)(0.27) = 0.2376
b) (0.12)(0.13+0.52) = 0.0.078
2-96.
2-97.
a)P=0.13 0.73=0.0949
b)P=0.87 (0.27+0.17)=0.3828
Let A denote a event that the first part selected has excessive shrinkage.
Let B denote the event that the second part selected has excessive shrinkage.
a) P(B)= P( B A )P(A) + P( B A ')P(A')
= (4/24)(5/25) + (5/24)(20/25) = 0.20
b) Let C denote the event that the third part selected has excessive shrinkage.
2-98.
Let A and B denote the events that the first and second chips selected are defective, respectively.
2-17
b) Let C denote the event that the third chip selected is defective.
P( A B C ) = P(C A B) P( A B) = P(C A B) P( B A) P( A)
18 19 20
98 99 100
= 0.00705
=
2-99.
Open surgery
large stone
small stone
overall summary
success
192
81
273
failure
71
6
77
sample
size
263
87
350
sample
percentage
0.751428571
0.248571429
1
conditional success
rate
0.73%
0.93%
0.78
success
55
234
289
failure
25
36
61
sample
size
80
270
350
sample
percentage
0.228571429
0.771428571
1
conditional success
rate
69%
83%
0.83
PN
large stone
small stone
overall summary
The reason is that the overall success rate is dependent on both the success rates conditioned on the two
groups and the probability of the groups. It is the weighted average of the group success rate weighted by
the group size; instead of the direct average.
P(overall success)=P(large stone)P(success| large stone)+ P(small stone)P(success| small stone).
For open surgery, the dominant group (large stone) has a smaller success rate while for PN, the dominant
group (small stone) has a larger success rate.
Section 2-6
2-100.
Because P( A B )
2-101.
2-103.
a) P( B A ) = 4/499 and
P( B) = P( B A) P( A) + P( B A' ) P( A' ) = (4 / 499)(5 / 500) + (5 / 499)(495 / 500) = 5 / 500
2-105.
2-18
b) Let C denote the event that the third chip selected is defective.
P( A B C ) = P(C A B) P( A B) = P(C A B) P( B A) P( A)
18 19 20
98 99 100
= 0.00705
=
2-99.
Open surgery
large stone
small stone
overall summary
success
192
81
273
failure
71
6
77
sample
size
263
87
350
sample
percentage
0.751428571
0.248571429
1
conditional success
rate
0.73%
0.93%
0.78
success
55
234
289
failure
25
36
61
sample
size
80
270
350
sample
percentage
0.228571429
0.771428571
1
conditional success
rate
69%
83%
0.83
PN
large stone
small stone
overall summary
The reason is that the overall success rate is dependent on both the success rates conditioned on the two
groups and the probability of the groups. It is the weighted average of the group success rate weighted by
the group size; instead of the direct average.
P(overall success)=P(large stone)P(success| large stone)+ P(small stone)P(success| small stone).
For open surgery, the dominant group (large stone) has a smaller success rate while for PN, the dominant
group (small stone) has a larger success rate.
Section 2-6
2-100.
Because P( A B )
2-101.
2-103.
a) P( B A ) = 4/499 and
P( B) = P( B A) P( A) + P( B A' ) P( A' ) = (4 / 499)(5 / 500) + (5 / 499)(495 / 500) = 5 / 500
2-105.
2-18
2-106.
(a) P = (0.001) = 10
2
2-107.
It is useful to work one of these exercises with care to illustrate the laws of probability. Let Hi denote the
event that the ith sample contains high levels of contamination.
a) P(H1' H2' H3' H4' H5' ) = P(H1' )P(H2' )P(H3' )P(H4' )P(H5' )
by independence. Also, P(Hi' ) = 0.9 . Therefore, the answer is 0.9 5 = 0.59
b) A1 = (H1 H'2 H'3 H4' H5' )
1
'
c
b) By independence, P ( A 1' A '2 ... A 10
) = P ( A 1' ) P ( A '2 )... P ( A 10
) = ( ) 10 = 0. 000976
2
( )
10
10 !
sequences consisting of five "1"'s, and five "0"'s is 10
=
= 252 . The answer is 252 1 = 0.246
5
5! 5!
2
2-109.
(a) The probability that one technician obtains equivalence at 100 mL is 0.1.
So the probability that both technicians obtain equivalence at 100 mL is 0.1 = 0.01 .
(b) The probability that one technician obtains equivalence between 98 and 104 mL is 0.7.
So the probability that both technicians obtain equivalence between 98 and 104 mL is
2
0.7 2 = 0.49 .
(c) The probability that the average volume at equivalence from the technician is 100 mL is
9(0.12 ) = 0.09 .
2-19
2-112.
2-113.
2-114.
2-115.
10 6
= 10 10
16
10
1
(b) P = 0.25 ( ) = 0.020833
12
(a) P =
Let A denote the event that a sample is produced in cavity one of the mold.
1
a) By independence, P( A1 A 2 A 3 A 4 A 5 ) = ( )5 = 0.00003
8
b) Let Bi be the event that all five samples are produced in cavity i. Because the B's are mutually
exclusive, P(B1 B 2 ...B 8 ) = P(B1) + P(B 2 )+...+P(B 8 )
1
1
From part a., P(Bi ) = ( )5 . Therefore, the answer is 8( )5 = 0.00024
8
8
1 4 7
'
c) By independence, P( A 1 A 2 A 3 A 4 A 5 ) = ( ) ( ) . The number of sequences in
8
8
1
7
which four out of five samples are from cavity one is 5. Therefore, the answer is 5( ) 4 ( ) = 0.00107 .
8
8
Let A denote the upper devices function. Let B denote the lower devices function.
P(A) = (0.9)(0.8)(0.7) = 0.504
P(B) = (0.95)(0.95)(0.95) = 0.8574
P(AB) = (0.504)(0.8574) = 0.4321
Therefore, the probability that the circuit operates = P(AB) = P(A) +P(B) P(AB) = 0.9293
[1-(0.1)(0.05)][1-(0.1)(0.05)][1-(0.2)(0.1)] = 0.9702
Section 2-7
2-116.
P( B A) =
P( A B) P( B)
P( B A) =
P( A B) P( B)
2-117.
P( A)
P( A)
0.7(0.2)
= 0.28
0.5
P( A B) P( B )
P ( A B) P( B) + P( A B ' ) P( B ' )
0.4 0.8
= 0.89
0.4 0.8 + 0.2 0.2
2-118.
Let F denote a fraudulent user and let T denote a user that originates calls from two or more
metropolitan areas in a day. Then,
P(T F ) P( F )
0.30(0.0001)
=
= 0.003
P( F T ) =
P(T F ) P( F ) + P(T F ' ) P( F ' ) 0.30(0.0001) + 0.01(.9999)
2-119.
(a) P=(0.31)(0.978)+(0.27)(0.981)+(0.21)(0.965)+(0.13)(0.992)+(0.08)(0.959)
= 0.97638
(b) P =
(0.21)(0.965)
= 0.207552
0.97638
2-120.
Let A denote the event that a respondent is a college graduate and let B denote the event that a
voter votes for Bush.
2-20
2-112.
2-113.
2-114.
2-115.
10 6
= 10 10
16
10
1
(b) P = 0.25 ( ) = 0.020833
12
(a) P =
Let A denote the event that a sample is produced in cavity one of the mold.
1
a) By independence, P( A1 A 2 A 3 A 4 A 5 ) = ( )5 = 0.00003
8
b) Let Bi be the event that all five samples are produced in cavity i. Because the B's are mutually
exclusive, P(B1 B 2 ...B 8 ) = P(B1) + P(B 2 )+...+P(B 8 )
1
1
From part a., P(Bi ) = ( )5 . Therefore, the answer is 8( )5 = 0.00024
8
8
1 4 7
'
c) By independence, P( A 1 A 2 A 3 A 4 A 5 ) = ( ) ( ) . The number of sequences in
8
8
1
7
which four out of five samples are from cavity one is 5. Therefore, the answer is 5( ) 4 ( ) = 0.00107 .
8
8
Let A denote the upper devices function. Let B denote the lower devices function.
P(A) = (0.9)(0.8)(0.7) = 0.504
P(B) = (0.95)(0.95)(0.95) = 0.8574
P(AB) = (0.504)(0.8574) = 0.4321
Therefore, the probability that the circuit operates = P(AB) = P(A) +P(B) P(AB) = 0.9293
[1-(0.1)(0.05)][1-(0.1)(0.05)][1-(0.2)(0.1)] = 0.9702
Section 2-7
2-116.
P( B A) =
P( A B) P( B)
P( B A) =
P( A B) P( B)
2-117.
P( A)
P( A)
0.7(0.2)
= 0.28
0.5
P( A B) P( B )
P ( A B) P( B) + P( A B ' ) P( B ' )
0.4 0.8
= 0.89
0.4 0.8 + 0.2 0.2
2-118.
Let F denote a fraudulent user and let T denote a user that originates calls from two or more
metropolitan areas in a day. Then,
P(T F ) P( F )
0.30(0.0001)
=
= 0.003
P( F T ) =
P(T F ) P( F ) + P(T F ' ) P( F ' ) 0.30(0.0001) + 0.01(.9999)
2-119.
(a) P=(0.31)(0.978)+(0.27)(0.981)+(0.21)(0.965)+(0.13)(0.992)+(0.08)(0.959)
= 0.97638
(b) P =
(0.21)(0.965)
= 0.207552
0.97638
2-120.
Let A denote the event that a respondent is a college graduate and let B denote the event that a
voter votes for Bush.
2-20
P( A | B) =
2-121.
(0.38)(0.53)
P( A B)
P( A) P( B | A)
=
=
= 39.3821%
P( B)
P( A) P( B | A) + P ( A' ) P( B | A' ) (0.38)(0.53) + (0.62)(0.5)
Let G denote a product that received a good review. Let H, M, and P denote products that were high,
moderate, and poor performers, respectively.
a)
a) P(D)=P(D|G)P(G)+P(D|G)P(G)=(.005)(.991)+(.99)(.009)=0.013865
b) P(G|D)=P(GD)/P(D)=P(D|G)P(G)/P(D)=(.995)(.991)/(1-.013865)=0.9999
Supplemental Exercises
Let "d" denote a defective calculator and let "a" denote an acceptable calculator
d) A B = {ddd , dda}
2-21
P( A | B) =
2-121.
(0.38)(0.53)
P( A B)
P( A) P( B | A)
=
=
= 39.3821%
P( B)
P( A) P( B | A) + P ( A' ) P( B | A' ) (0.38)(0.53) + (0.62)(0.5)
Let G denote a product that received a good review. Let H, M, and P denote products that were high,
moderate, and poor performers, respectively.
a)
a) P(D)=P(D|G)P(G)+P(D|G)P(G)=(.005)(.991)+(.99)(.009)=0.013865
b) P(G|D)=P(GD)/P(D)=P(D|G)P(G)/P(D)=(.995)(.991)/(1-.013865)=0.9999
Supplemental Exercises
Let "d" denote a defective calculator and let "a" denote an acceptable calculator
d) A B = {ddd , dda}
2-21
P( A | B) =
2-121.
(0.38)(0.53)
P( A B)
P( A) P( B | A)
=
=
= 39.3821%
P( B)
P( A) P( B | A) + P ( A' ) P( B | A' ) (0.38)(0.53) + (0.62)(0.5)
Let G denote a product that received a good review. Let H, M, and P denote products that were high,
moderate, and poor performers, respectively.
a)
a) P(D)=P(D|G)P(G)+P(D|G)P(G)=(.005)(.991)+(.99)(.009)=0.013865
b) P(G|D)=P(GD)/P(D)=P(D|G)P(G)/P(D)=(.995)(.991)/(1-.013865)=0.9999
Supplemental Exercises
Let "d" denote a defective calculator and let "a" denote an acceptable calculator
d) A B = {ddd , dda}
2-21
If A,B,C are mutually exclusive, then P( A B C ) = P(A) + P(B) + P(C) = 0.3 + 0.4 + 0.5 =
1.2, which greater than 1. Therefore, P(A), P(B),and P(C) cannot equal the given values.
2-130. a) 345/357
b) 5/13
25 24
1 = 0.615
40 39
2-22
= 0.18143
(b) P = C (0.998 )0.002 = 0.005976
1
3
100
] = 0.86494
100 10
main-storage
backup
0.75
0.25
life > 5 yrs
0.05(0.25)=0.0125 0.995(0.75)=0.74625
0.005(0.75)=0.00375
a) P(B) = 0.25
b) P( A B ) = 0.95
c) P( A B ') = 0.995
d) P( A B ) = P( A B )P(B) = 0.95(0.25) = 0.2375
e) P( A B ') = P( A B ')P(B') = 0.995(0.75) = 0.74625
f) P(A) = P( A B ) + P( A B ') = 0.95(0.25) + 0.995(0.75) = 0.98375
g) 0.95(0.25) + 0.995(0.75) = 0.98375.
h)
P ( B A' ) =
P ( A' B ) P ( B )
P( A' B) P( B ) + P( A' B' ) P( B' )
0.05(0.25)
= 0.769
0.05(0.25) + 0.005(0.75)
Let Di denote the event that the primary failure mode is type i and let A denote the event that a board passes
the test.
2-23
a) 20/200
b) 135/200
c) 65/200
2-24
2-144.
Let A i denote the event that the ith order is shipped on time.
a) By independence,
P ( A1 A2 A3 ) = P( A1 ) P( A2 ) P ( A3 ) = (0.95) 3 = 0.857
b) Let
B1 = A 1' A 2 A 3
B 2 = A 1 A 2' A 3
B 3 = A 1 A 2 A '3
Then, because the B's are mutually exclusive,
P(B1 B 2 B 3 ) = P(B1 ) + P(B 2 ) + P(B 3 )
= 3(0.95) 2 (0.05)
= 0.135
c) Let
B1 = A 1' A '2 A 3
B 2 = A 1' A 2 A '3
B 3 = A 1 A '2 A '3
B 4 = A 1' A '2 A '3
Because the B's are mutually exclusive,
P(B1 B 2 B 3 B 4 ) = P(B1) + P(B 2 ) + P(B 3 ) + P(B 4 )
2-25
2-147.
Let Ai denote the event that the ith bolt selected is not torqued to the proper limit.
a) Then,
P ( A1 A2 A3 A4 ) = P( A4 A1 A2 A3 ) P( A1 A2 A3 )
= P( A4 A1 A2 A3 ) P( A3 A1 A2 ) P ( A2 A1 ) P ( A1 )
12 13 14 15
= = 0.282
17 18 19 20
2-148.
b) Let B denote the event that at least one of the selected bolts are not properly torqued. Thus, B' is the
event that all bolts are properly torqued. Then,
15 14 13 12
P(B) = 1 - P(B') = 1 = 0.718
20 19 18 17
Let A,B denote the event that the first, second portion of the circuit operates. Then, P(A) =
(0.99)(0.99)+0.9-(0.99)(0.99)(0.9) = 0.998
P(B) = 0.9+0.9-(0.9)(0.9) = 0.99 and
P( A B ) = P(A) P(B) = (0.998) (0.99) = 0.988
Let D denote the event that a container is incorrectly filled and let H denote the event that a container is
filled under high-speed operation. Then,
a) P(D) = P( D H )P(H) + P( D H ')P(H') = 0.01(0.30) + 0.001(0.70) = 0.0037
b) P ( H D ) = P ( D H ) P ( H ) = 0.01(0.30) = 0.8108
P( D)
0.0037
b)
c)
2 73 72 73 2 72 73 72 2
P(D = 1) = + + = 0.0778
75 74 73 75 74 73 75 74 73
2 1 73 2 73 1 73 2 1
+ + = 0.00108
75 74 73 75 74 73 75 74 73
Let A represent the event that the two items NOT inspected are not defective. Then,
P(A)=(73/75)(72/74)=0.947.
P(D = 2) =
2-154.
The tool fails if any component fails. Let F denote the event that the tool fails. Then, P(F') = 0. 9910 by
independence and P(F) = 1 - 0. 9910 = 0.0956
2-155.
P( E )
1 0.9764
2-26
2-156.
2-157.
Let A i denote the event that the ith washer selected is thicker than target.
a)
30 29 28
= 0.207
50 49 48
b) 30/48 = 0.625
c) The requested probability can be written in terms of whether or not the first and second washer selected
are thicker than the target. That is,
+
+
+
48 50 49 48 50 49 48 50 49 48 50 49
= 0.60
2-158. a) If n washers are selected, then the probability they are all less than the target is
20 19
20 n + 1
.
...
50 49
50 n + 1
n
probability all selected washers are less than target
1
20/50 = 0.4
2
(20/50)(19/49) = 0.155
3
(20/50)(19/49)(18/48) = 0.058
Therefore, the answer is n = 3
b) Then event E that one or more washers is thicker than target is the complement of the event that all are
less than target. Therefore, P(E) equals one minus the probability in part a. Therefore, n = 3.
2-27
2-159.
a)
b)
c)
d)
. e)
f)
2-160.
112 + 68 + 246
= 0.453
940
246
P( A B) =
= 0.262
940
514 + 68 + 246
P( A' B) =
= 0.881
940
514
P( A' B' ) =
= 0.547
940
P( A B) =
P( A B) 246 / 940
=
= 0.783
P(B)
314 / 940
P( B A ) = P ( B A ) = 246 / 940 = 0. 687
P(A )
358 / 940
P( A B ) =
Let E denote a read error and let S,O,P denote skewed, off-center, and proper alignments, respectively.
Then,
a) P(E) = P(E|S) P(S) + P(E|O) P (O) + P(E|P) P(P)
= 0.01(0.10) + 0.02(0.05) + 0.001(0.85)
= 0.00285
b) P(S|E) =
P ( E S) P (S)
P( E)
2-161.
0. 01( 0. 10)
= 0. 351
0. 00285
Let A i denote the event that the ith row operates. Then,
P ( A1 ) = 0. 98, P ( A 2 ) = ( 0. 99)( 0. 99) = 0. 9801, P ( A 3 ) = 0. 9801, P ( A 4 ) = 0. 98.
The probability the circuit does not operate is
(a) P=(24/36)(23/35)(22/34)(21/33)(20/32)(19/31)=0.069
(b) P=1-0.069=0.931
2-165.
(a) 367
(b) Number of permutations of six letters is 266. Number of ways to select one number = 10.
Number of positions among the six letters to place the one number = 7. Number of passwords =
266 10 7
(c) 265102
2-166.
5 + 25 + 30 + 7 + 20
= 0.087
1000
25 + 7
(b) P ( A B ) =
= 0.032
1000
800
(c) P ( A B ) = 1
= 0.20
1000
(a) P ( A) =
2-28
63 + 35 + 15
= 0.113
1000
0.032
P( A B)
(e) P ( A | B ) =
=
= 0.2207
(25 + 63 + 15 + 7 + 35) / 1000
P( B)
5
(f) P =
= 0.005
1000
(d) P ( A'B ) =
2-167.
(a) Let A denote that a part conforms to specifications and let B denote for a part a simple
component.
For supplier 1,
P(A)=P(A|B)P(B)+ P(A|B)P(B)
= [(1000-2)/1000](1000/2000)+ [(1000-10)/1000](1000/2000)
= 0.994
For supplier 2,
P(A)= P(A|B)P(B)+ P(A|B)P(B)
= [(1600-4)/1600](1600/2000)+ [(400-6)/400](400/2000)
= 0.995
(b)
For supplier 1,
For supplier 2,
(c)
For supplier 1,
For supplier 2,
P(A|B)=0.99
P(A|B)=0.985
P(A|B)=0.998
P(A|B)=0.9975
(d) The unusual result is that for both a simple component and for a complex assembly, supplier 1
has a greater probability that a part conforms to specifications. However, for overall parts,
supplier 1 has a lower probability. The overall conforming probability is dependent on both the
conforming probability conditioned on the part type and the probability of each part type. Supplier
1 produces more of the compelx parts so that overall conformance from supplier 1 is lower.
Mind-Expanding Exercises
2-168.
2-169.
Let E denote a read error and let S, O, B, P denote skewed, off-center, both, and proper alignments,
respectively.
P(E) = P(E|S)P(S) + P(E|O)P(O) + P(E|B)P(B) + P(E|P)P(P)
= 0.01(0.10) + 0.02(0.05) + 0.06(0.01) + 0.001(0.84) = 0.00344
Let n denote the number of washers selected.
a) The probability that non are thicker than, that is, all are less than the target is 0.4 n , by independence.
n
0.4 n
1
0.4
2
0.16
3
0.064
Therefore, n = 3
b) The requested probability is the complement of the probability requested in part a. Therefore, n = 3
2-29
125 375
5 0 2.3453E8
=
= 0.00092
f X (5) =
2.5524 E11
500
5
b)
x
f(x)
3-150.
0
1
2
3
4
5
6
7
8
9
10
0.0546 0.1866 0.2837 0.2528 0.1463 0.0574 0.0155 0.0028 0.0003 0.0000 0.0000
Let X denote the number of totes in the sample that exceed the moisture content. Then X is a binomial
random variable with n = 30. We are to determine p.
If P(X 1) = 0.9, then P(X = 0) = 0.1. Then
30 0
( p) (1 p )30 = 0.1 , giving 30ln(1p)=ln(0.1),
0
Let t denote an interval of time in hours and let X denote the number of messages that arrive in time t.
Then, X is a Poisson random variable with = 10t.
Then, P(X=0) = 0.9 and e-10t = 0.9, resulting in t = 0.0105 hours = 37.8 seconds
3-152.
a) Let X denote the number of flaws in 50 panels. Then, X is a Poisson random variable with
= 50(0.02) = 1. P(X = 0) = e-1 = 0.3679.
b) Let Y denote the number of flaws in one panel, then
P(Y 1) = 1 P(Y=0) = 1 e-0.02 = 0.0198. Let W denote the number of panels that need to be
inspected before a flaw is found. Then W is a geometric random variable with p = 0.0198 and
E(W) = 1/0.0198 = 50.51 panels.
0.02
c) P (Y 1) = 1 P (Y = 0) = 1 e
= 0.0198
Let V denote the number of panels with 1 or more flaws. Then V is a binomial random
variable with n=50 and p=0.0198
50
50
P(V 2) = 0.0198 0 (.9802) 50 + 0.01981 (0.9802) 49
0
1
50
+ 0.0198 2 (0.9802) 48 = 0.9234
2
Mind Expanding Exercises
3-153. The binomial distribution
P(X=x) =
n!
px(1-p)n-x
x!(n x )!
If n is large, then the probability of the event could be expressed as /n, that is =np. We
could re-write the probability mass function as:
P(X=x) =
n!
[/n]x[1 (/n)]n-x
x!(n x )!
Where p = /n.
P(X=x)
n (n 1) (n 2) (n 3)...... (n x + 1) x
(1 (/n))n-x
x
x!
n
3-29
((1 - n ) )
n/
So:
P(X=x)=
n (n 1) (n 2) (n 3)...... (n x + 1) x
(1 - n )n/
x
x!
n
[1 (/n)]-x
Now:
In the limit as n
n (n 1) (n 2) (n 3)...... (n x + 1)
In the limit as n
[1 (/n)]-x 1
Thus:
x
P(X=x) =
n/
(
1- n)
x!
1
Limit z 1 + = e 2.7183
z
n/
= e. Thus,
In our case above n/ = z, so (1 - n )
x
P(X=x) =
e
x!
n!
e x
px(1 p)n-x =
x!(n x )!
x!
The distribution of the probability associated with this process is known as the Poisson distribution.
The pmf can be expressed as:
f(x) =
e x
x!
i =1
i =1
i =1
as
p (1 p)i 1 =
i =1
p
p
= =1
1 (1 p ) p
3-30
3-155.
i
i
i =1
= i =1
(b a + 1)
(b a + b + a )
=
2
b(b + 1) (a 1)a
2
2
(b a + 1)
(b a + 1)
(b + a)(b a + 1)
(b a + 1)
(b + a)
2
b
b 2
(b a + 1)(b + a ) 2
+
+
(
)
i
b
a
i
[i ]
4
i =a
i =a
i =a
=
V (X ) =
b + a 1
b + a 1
2
b(b + 1)(2b + 1) (a 1)a(2a 1)
b(b + 1) (a 1)a (b a + 1)(b + a )
(b + a )
+
6
6
2
4
=
b a +1
2
(b a + 1) 1
=
12
b
b+ a 2
2
3-156. Let X denote the number of nonconforming products in the sample. Then, X is approximately binomial with
p = 0.01 and n is to be determined.
If P ( X 1) 0.90 , then P ( X
Now, P(X = 0) =
( )p (1 p)
n
0
= 0) 0.10 .
= (1 p ) n . Consequently,
ln 0.10
= 229.11 . Therefore, n = 230 is required
ln(1 p )
3-157. If the lot size is small, 10% of the lot might be insufficient to detect nonconforming product. For example, if
the lot size is 10, then a sample of size one has a probability of only 0.2 of detecting a nonconforming
product in a lot that is 20% nonconforming.
If the lot size is large, 10% of the lot might be a larger sample size than is practical or necessary. For
example, if the lot size is 5000, then a sample of 500 is required. Furthermore, the binomial
approximation to the hypergeometric distribution can be used to show the following. If 5% of the lot of size
5000 is nonconforming, then the probability of zero nonconforming product in the sample is approximately
7 10 12 . Using a sample of 100, the same probability is still only 0.0059. The sample of size 500 might
be much larger than is needed.
3-31
3-158. Let X denote the number of panels with flaws. Then, X is a binomial random variable with n =100 and p is
the probability of one or more flaws in a panel. That is, p = 1 e
0.1
= 0.095.
P( X < 5) = P( X 4) = P( X = 0) + P( X = 1) + P( X = 2) + P( X = 3) + P( X = 4)
0
100
1
99
2
98
= (100
+ (100
+ (100
0 ) p (1 p )
1 ) p (1 p )
2 ) p (1 p )
3
97
4
96
+ (100
+ (100
3 ) p (1 p )
4 ) p (1 p )
= 0.034
3-159. Let X denote the number of rolls produced.
Revenue at each demand
1000
2000
3000
0.3x
0.3x
0.3x
0 x 1000
mean profit = 0.05x(0.3) + 0.3x(0.7) - 0.1x
0.05x
0.3(1000) +
0.3x
0.3x
1000 x 2000
0.05(x-1000)
mean profit = 0.05x(0.3) + [0.3(1000) + 0.05(x-1000)](0.2) + 0.3x(0.5) - 0.1x
0.05x
0.3(1000) +
0.3(2000) +
0.3x
2000 x 3000
0.05(x-1000)
0.05(x-2000)
mean profit = 0.05x(0.3) + [0.3(1000)+0.05(x-1000)](0.2) + [0.3(2000) + 0.05(x-2000)](0.3) + 0.3x(0.2) - 0.1x
0.05x
0.3(1000) +
0.3(2000) +
0.3(3000)+
3000 x
0.05(x-1000)
0.05(x-2000)
0.05(x-3000)
mean profit = 0.05x(0.3) + [0.3(1000)+0.05(x-1000)](0.2) + [0.3(2000)+0.05(x-2000)]0.3 + [0.3(3000)+0.05(x3000)]0.2 - 0.1x
0
0.05x
0 x 1000
1000 x 2000
2000 x 3000
3000 x
Profit
0.125 x
0.075 x + 50
200
-0.05 x + 350
Max. profit
$ 125 at x = 1000
$ 200 at x = 2000
$200 at x = 3000
$200 at x = 3000
The bakery can make anywhere from 2000 to 3000 and earn the same profit.
3-160.Let X denote the number of acceptable components. Then, X has a binomial distribution with p = 0.98 and
n is to be determined such that P( X 100 ) 0.95 .
P( X 100 )
102
0.666
103
0.848
104
0.942
105
0.981
Therefore, 105 components are needed.
3-32
CHAPTER 3
Section 3-1
{0,1,2,...,1000}
3-1.
The range of X is
3-2.
3-3.
3-4.
3-5.
The range of X is 1,2,...,491 . Because 490 parts are conforming, a nonconforming part must be
selected in 491 selections.
3-6.
exceed 10%.
3-7.
The range of X is conveniently modeled as all nonnegative integers. That is, the range of X is
, ,...}
{0,12
3-8.
The range of X is conveniently modeled as all nonnegative integers. That is, the range of X is
, ,...}
{0,12
3-9.
The range of X is
3-10.
The possible totals for two orders are 1/8 + 1/8 = 1/4, 1/8 + 1/4 = 3/8, 1/8 + 3/8 = 1/2, 1/4 + 1/4 = 1/2,
1/4 + 3/8 = 5/8, 3/8 + 3/8 = 6/8.
1 3 1 5 6
Therefore the range of X is , , , ,
4 8 2 8 8
3-11.
The range of X is
{0,1,2,...,15}
{0,1,2, ,10000}
f X (0) = P ( X = 0) = 1 / 6 + 1 / 6 = 1 / 3
f X (1.5) = P ( X = 1.5) = 1 / 3
f X (2) = 1 / 6
f X (3) = 1 / 6
a) P(X = 1.5) = 1/3
b) P(0.5< X < 2.7) = P(X = 1.5) +P(X = 2) = 1/3 + 1/6 = 1/2
c) P(X > 3) = 0
d) P (0 X < 2) = P ( X = 0) + P ( X = 1.5) = 1/ 3 + 1/ 3 =
e) P(X = 0 or X = 2) = 1/3 + 1/6 = 1/2
3-15.
All probabilities are greater than or equal to zero and sum to one.
3-1
2/3
CHAPTER 3
Section 3-1
{0,1,2,...,1000}
3-1.
The range of X is
3-2.
3-3.
3-4.
3-5.
The range of X is 1,2,...,491 . Because 490 parts are conforming, a nonconforming part must be
selected in 491 selections.
3-6.
exceed 10%.
3-7.
The range of X is conveniently modeled as all nonnegative integers. That is, the range of X is
, ,...}
{0,12
3-8.
The range of X is conveniently modeled as all nonnegative integers. That is, the range of X is
, ,...}
{0,12
3-9.
The range of X is
3-10.
The possible totals for two orders are 1/8 + 1/8 = 1/4, 1/8 + 1/4 = 3/8, 1/8 + 3/8 = 1/2, 1/4 + 1/4 = 1/2,
1/4 + 3/8 = 5/8, 3/8 + 3/8 = 6/8.
1 3 1 5 6
Therefore the range of X is , , , ,
4 8 2 8 8
3-11.
The range of X is
{0,1,2,...,15}
{0,1,2, ,10000}
f X (0) = P ( X = 0) = 1 / 6 + 1 / 6 = 1 / 3
f X (1.5) = P ( X = 1.5) = 1 / 3
f X (2) = 1 / 6
f X (3) = 1 / 6
a) P(X = 1.5) = 1/3
b) P(0.5< X < 2.7) = P(X = 1.5) +P(X = 2) = 1/3 + 1/6 = 1/2
c) P(X > 3) = 0
d) P (0 X < 2) = P ( X = 0) + P ( X = 1.5) = 1/ 3 + 1/ 3 =
e) P(X = 0 or X = 2) = 1/3 + 1/6 = 1/2
3-15.
All probabilities are greater than or equal to zero and sum to one.
3-1
2/3
All probabilities are greater than or equal to zero and sum to one.
a) P(X 1)=P(X=1)=0.5714
b) P(X>1)= 1-P(X=1)=1-0.5714=0.4286
c) P(2<X<6)=P(X=3)=0.1429
d) P(X1 or X>1)= P(X=1)+ P(X=2)+P(X=3)=1
3-17.
3-18.
3-19.
3-20.
3-21.
3-22.
X: the number of computers that vote for a left roll when a right roll is appropriate.
p=0.0001.
P(X=0)=(1-p)4=0.99994=0.9996
P(X=1)=4*(1-p)3p=4*0.999930.0001=0.0003999
P(X=2)=C42(1-p)2p2=5.999*10-8
P(X=3)=C43(1-p)1p3=3.9996*10-12
P(X=4)=C40(1-p)0p4=1*10-16
3-23.
3-24.
3-25.
P(X = 15 million) = 0.6, P(X = 5 million) = 0.3, P(X = -0.5 million) = 0.1
3-26.
3-2
3-27.
Section 3-3
3-28.
x<0
0,
1 / 3 0 x < 1.5
F ( x) = 2 / 3 1.5 x < 2
5 / 6 2 x < 3
3 x
1
f X (0) = P ( X = 0) = 1 / 6 + 1 / 6 = 1 / 3
where
f X (1.5) = P ( X = 1.5) = 1 / 3
f X (2) = 1 / 6
f X (3) = 1 / 6
3-29.
x < 2
0,
1 / 8 2 x < 1
3 / 8 1 x < 0
F ( x) =
0 x <1
5 / 8
7 / 8
1 x < 2
1
2 x
f X (2) = 1 / 8
f X (1) = 2 / 8
where
f X ( 0) = 2 / 8
f X (1) = 2 / 8
f X ( 2) = 1 / 8
3-30.
0,
1 / 25
4 / 25
F ( x) =
9 / 25
16 / 25
x<0
0 x < 1
1 x < 2
2 x < 3
3 x < 4
4 x
f X (0) = 1 / 25
f X (1) = 3 / 25
f X (2) = 5 / 25
where
f X (3) = 7 / 25
f X (4) = 9 / 25
3-3
3-27.
Section 3-3
3-28.
x<0
0,
1 / 3 0 x < 1.5
F ( x) = 2 / 3 1.5 x < 2
5 / 6 2 x < 3
3 x
1
f X (0) = P ( X = 0) = 1 / 6 + 1 / 6 = 1 / 3
where
f X (1.5) = P ( X = 1.5) = 1 / 3
f X (2) = 1 / 6
f X (3) = 1 / 6
3-29.
x < 2
0,
1 / 8 2 x < 1
3 / 8 1 x < 0
F ( x) =
0 x <1
5 / 8
7 / 8
1 x < 2
1
2 x
f X (2) = 1 / 8
f X (1) = 2 / 8
where
f X ( 0) = 2 / 8
f X (1) = 2 / 8
f X ( 2) = 1 / 8
3-30.
0,
1 / 25
4 / 25
F ( x) =
9 / 25
16 / 25
x<0
0 x < 1
1 x < 2
2 x < 3
3 x < 4
4 x
f X (0) = 1 / 25
f X (1) = 3 / 25
f X (2) = 5 / 25
where
f X (3) = 7 / 25
f X (4) = 9 / 25
3-3
x<0
0,
0.008, 0 x < 1
F ( x) = 0.104, 1 x < 2
0.488, 2 x < 3
1,
3 x
.
f (0) = 0.2 3 = 0.008,
where
3-32.
x<0
0,
0.9996, 0 x < 1
F ( x ) = 0.9999, 1 x < 3
0.99999, 3 x < 4
4 x
1,
.
where
3-33.
x < 10
0,
0.2, 10 x < 25
F ( x) =
0.5, 25 x < 50
1,
50 x
where P(X = 50 million) = 0.5, P(X = 25 million) = 0.3, P(X = 10 million) = 0.2
3-4
3-34.
x <1
0,
0.1, 1 x < 5
F ( x) =
0.7, 5 x < 10
1,
10 x
where
3-35.
The sum of the probabilities is 1 and all probabilities are greater than or equal to zero;
pmf: f(1) = 0.5, f(3) = 0.5
a) P(X 3) = 1
b) P(X 2) = 0.5
c) P(1 X 2) = P(X=1) = 0.5
d) P(X>2) = 1 P(X2) = 0.5
3-36.
The sum of the probabilities is 1 and all probabilities are greater than or equal to zero;
pmf: f(1) = 0.7, f(4) = 0.2, f(7) = 0.1
a) P(X 4) = 0.9
b) P(X > 7) = 0
c) P(X 5) = 0.9
d) P(X>4) = 0.1
e) P(X2) = 0.7
3-37.
The sum of the probabilities is 1 and all probabilities are greater than or equal to zero;
pmf: f(-10) = 0.25, f(30) = 0.5, f(50) = 0.25
a) P(X50) = 1
b) P(X40) = 0.75
c) P(40 X 60) = P(X=50)=0.25
d) P(X<0) = 0.25
e) P(0X<10) = 0
f) P(10<X<10) = 0
3-5
3-38.
The sum of the probabilities is 1 and all probabilities are greater than or equal to zero;
pmf: f1/8) = 0.2, f(1/4) = 0.7, f(3/8) = 0.1
a) P(X1/18) = 0
b) P(X1/4) = 0.9
c) P(X5/16) = 0.9
d) P(X>1/4) = 0.1
e) P(X1/2) = 1
Section 3-4
3-39.
3-6
3-38.
The sum of the probabilities is 1 and all probabilities are greater than or equal to zero;
pmf: f1/8) = 0.2, f(1/4) = 0.7, f(3/8) = 0.1
a) P(X1/18) = 0
b) P(X1/4) = 0.9
c) P(X5/16) = 0.9
d) P(X>1/4) = 0.1
e) P(X1/2) = 1
Section 3-4
3-39.
3-6
3-44.
E( X ) =
3 1
3 1
1
x
=
x =
4 x =0 4
4 x =1 4
3
The result uses a formula for the sum of an infinite series. The formula can be derived from the fact that the
h( a ) = a x =
x =1
a
1 a
with respect to a.
For the variance, another formula can be derived from the second derivative of h(a) with respect to a.
Calculate from this formula
x
3 1
3 1
5
E( X ) = x 2 = x 2 =
4 x =0 4
4 x =1 4
9
5 1 4
2
2
Then V ( X ) = E ( X ) [E ( X ) ] = =
9 9 9
2
3-45.
3-47.
3-48.
(a) F(0)=0.17
Nickel Charge: X
0
2
3
4
CDF
0.17
0.17+0.35=0.52
0.17+0.35+0.33=0.85
0.17+0.35+0.33+0.15=1
(b)E(X) = 0*0.17+2*0.35+3*0.33+4*0.15=2.29
4
V(X) =
f ( x )( x )
i
i =1
3-49.
f ( x )( x )
i =1
= 0.00039996
=E(X)=350*0.06+450*0.1+550*0.47+650*0.37=565
4
V(X)=
f ( x )( x )
i =1
=
3-51.
= 1.5259
X = number of computers that vote for a left roll when a right roll is appropriate.
= E(X)=0*f(0)+1*f(1)+2*f(2)+3*f(3)+4*f(4)
= 0+0.0003999+2*5.999*10-8+3*3.9996*10-12+4*1*10-16= 0.0004
V(X)=
3-50.
=6875
V (X ) =82.92
(a)
Transaction
New order
Payment
Order status
Delivery
Stock level
total
Frequency
43
44
4
5
4
100
Selects: X
23
4.2
11.4
130
0
f(X)
0.43
0.44
0.04
0.05
0.04
V(X) =
f ( x )( x )
i =1
(b)
Transaction
New order
Payment
Order status
Delivery
Stock level
total
= 735.964
Frequency
43
44
4
5
4
100
= V ( X ) = 27.1287
All operation: X
23+11+12=46
4.2+3+1+0.6=8.8
11.4+0.6=12
130+120+10=260
0+1=1
= E(X) = 46*0.43+8.8*0.44+12*0.04+260*0.05+1*0.04=37.172
3-8
f(X)
0.43
0.44
0.04
0.05
0.04
V(X) =
f ( x )( x )
i =1
=2947.996
= V ( X ) = 54.2955
Section 3-5
3-52.
3-53.
3-54.
1 15 + 19
= 0.17 mm
100 2
2
1 (19 15 + 1) 1
2
V (X ) =
= 0.0002 mm
100
12
3-55.
1 1 1
E ( X ) = 2 + 3 + 4 = 3
3 3 3
in 100 codes the expected number of letters is 300
2
21
21
21
2
V ( X ) = (2 ) + (3) + (4 ) (3) =
3
3
3
3
in 100 codes the variance is 6666.67
3-56.
0+9
590 + 0.1
= 590.45 mm
2
2
2 (9 0 + 1) 1
V ( X ) = (0.1)
= 0.0825
12
E(X) =
mm2
3-57.
a = 675, b = 700
(a) = E(X) = (a+b)/2= 687.5
V(X) = [(b a +1)2 1]/12= 56.25
(b) a = 75, b = 100
= E(X) = (a+b)/2 = 87.5
V(X) = [(b a + 1)2 1]/12= 56.25
The range of values is the same, so the mean shifts by the difference in the two minimums (or
maximums) whereas the variance does not change.
3-58.
X is a discrete random variable. X is discrete because it is the number of fields out of 28 that has an error.
However, X is not uniform because P(X=0) P(X=1).
3-59.
3-9
V(X) =
f ( x )( x )
i =1
=2947.996
= V ( X ) = 54.2955
Section 3-5
3-52.
3-53.
3-54.
1 15 + 19
= 0.17 mm
100 2
2
1 (19 15 + 1) 1
2
V (X ) =
= 0.0002 mm
100
12
3-55.
1 1 1
E ( X ) = 2 + 3 + 4 = 3
3 3 3
in 100 codes the expected number of letters is 300
2
21
21
21
2
V ( X ) = (2 ) + (3) + (4 ) (3) =
3
3
3
3
in 100 codes the variance is 6666.67
3-56.
0+9
590 + 0.1
= 590.45 mm
2
2
2 (9 0 + 1) 1
V ( X ) = (0.1)
= 0.0825
12
E(X) =
mm2
3-57.
a = 675, b = 700
(a) = E(X) = (a+b)/2= 687.5
V(X) = [(b a +1)2 1]/12= 56.25
(b) a = 75, b = 100
= E(X) = (a+b)/2 = 87.5
V(X) = [(b a + 1)2 1]/12= 56.25
The range of values is the same, so the mean shifts by the difference in the two minimums (or
maximums) whereas the variance does not change.
3-58.
X is a discrete random variable. X is discrete because it is the number of fields out of 28 that has an error.
However, X is not uniform because P(X=0) P(X=1).
3-59.
3-9
3-60.
E (cX ) = cxf ( x) = c xf ( x) = cE ( X ) ,
x
V (cX ) = (cx c ) f ( x) = c 2 ( x ) 2 f ( x) = cV ( X )
2
Section 3-6
3-61.
A binomial distribution is based on independent trials with two outcomes and a constant probability of
success on each trial.
a) reasonable
b) independence assumption not reasonable
c) The probability that the second component fails depends on the failure time of the first component. The
binomial distribution is not reasonable.
d) not independent trials with constant probability
e) probability of a correct answer not constant.
f) reasonable
g) probability of finding a defect not constant.
h) if the fills are independent with a constant probability of an underfill, then the binomial distribution for
the number packages underfilled is reasonable.
i) because of the bursts, each trial (that consists of sending a bit) is not independent
j) not independent trials with constant probability
3-62.
3-63.
3-64.
10
= 5) = 0.5 5 (0.5) 5 = 0.2461
5
10 0 10 10 1 9 10 2 8
b) P( X 2) =
0 0.5 0.5 + 1 0.5 0.5 + 2 0.5 0.5
10
10
10
= 0.5 + 10(0.5) + 45(0.5) = 0.0547
10 10
10 9
1
d)
3-65.
10
10
P(3 X < 5) = 0.530.57 + 0.540.56
4
3
= 120(0.5)10 + 210(0.5)10 = 0.3223
a) P ( X
10
5
= 5) = 0.015 (0.99) = 2.40 108
5
3-10
3-79.
The probability is 0.651 that at least one sample from the next five will contain more than one defective
Let X denote the passengers with tickets that do not show up for the flight. Then, X is binomial
with n = 125 and p = 0.1.
a) P( X 5) = 1 P( X 4)
125 0
125 1
125 2
0.1 (0.9 )125 +
0.1 (0.9 )124 +
0.1 (0.9 )123
0
1
2
= 1
125
125 4
+
4
= 0.9961
b) P( X > 5) = 1 P( X 5) = 0.9886
3-80.
a)
0
100
P( X = 0) = (100
= 0.133
0 )0.02 0.98
b)
0
102
1
101
2
100
P( X 2) = (102
+ (102
+ (102
= 0.666
0 )0.02 0.98
1 )0.02 0.98
2 )0.02 0.98
c)
P( X 5) = 0.981
Section 3-7
3-81 .
3-82.
a)
P( X
P( X
P( X
P( X
a)
3-83.
3-84.
a) E(X) = 4/0.2 = 20
b) P(X=20) =
19
(0.80)16 0.24 = 0.0436
3
3-15
18
(0.80)15 0.2 4 = 0.0459
3
20
17
4
d) P(X=21) =
3 (0.80) 0.2 = 0.0411
c) P(X=19) =
e) The most likely value for X should be near X. By trying several cases, the most likely value is x = 19.
3-85.
Let X denote the number of trials to obtain the first successful alignment. Then X is a geometric random
variable with p = 0.8
a)
b)
c)
3-86.
Let X denote the number of people who carry the gene. Then X is a negative binomial random variable with
r=2 and p = 0.1
a) P ( X 4) = 1 P ( X < 4) = 1 [ P ( X = 2) + P ( X = 3)]
b)
3-87.
2
= 1 (1 0.1) 0 0.12 + (1 0.1)1 0.12 = 1 (0.01 + 0.018) = 0.972
1
1
E ( X ) = r / p = 2 / 0.1 = 20
Let X denote the number of calls needed to obtain a connection. Then, X is a geometric random variable
with p = 0.02.
a)
b) P ( X > 5) = 1 P ( X 4) = 1 [ P ( X = 1) + P ( X = 2) + P ( X = 3) + P ( X = 4) + P ( X = 5)]
= 1 [0.02 + 0.98(0.02) + 0.982 (0.02) + 0.983 (0.02) + 0.983 (0.02) + 0.984 (0.02)]
= 1 0.0961 = 0.9039
May also use the fact that P(X > 5) is the probability of no connections in 5 trials. That is,
5
P( X > 5) = 0.020 0.985 = 0.9039
0
c) E(X) = 1/0.02 = 50
3-88.
3-16
p=0.13
(a) P(X=1) = (1-0.13)1-1*0.13=0.13.
(b) P(X=3)=(1-0.13)3-1*0.13 =0.098
(c) =E(X)= 1/p=7.698
3-90.
V (X ) =222222
(b) p=100/363=0.00214
=E(X) = 1/p= 467
V(X)= (1-p)/p2 = 217892.39
=
V (X ) =466.78
Based on the answers to (a) and (b) above, it is clearly more secure to use a 6 character password.
3-91 .
p = 0.005 , r = 8
a.)
b).
P( X = 8) = 0.0058 = 3.91x10 19
1
= E( X ) =
= 200 days
0.005
c) Mean number of days until all 8 computers fail. Now we use p=3.91x10-19
= E (Y ) =
1
= 2.56 x1018 days
3.91x10 91
3-92.
Let Y denote the number of samples needed to exceed 1 in Exercise 3-66. Then Y has a geometric
distribution with p = 0.0169.
a) P(Y = 10) = (1 0.0169)9(0.0169) = 0.0145
b) Y is a geometric random variable with p = 0.1897 from Exercise 3-66.
P(Y = 10) = (1 0.1897)9(0.1897) = 0.0286
c) E(Y) = 1/0.1897 = 5.27
3-93.
Let X denote the number of transactions until all computers have failed. Then, X is negative binomial
random variable with p = 10-8 and r = 3.
a) E(X) = 3 x 108
b) V(X) = [3(110-80]/(10-16) = 3.0 x 1016
3-94.
3-95 .
x 1
(1 p) x r p r .
r 1
When r = 1, this reduces to f(x; p, r) = (1p)x-1p, which is the pdf of a geometric random variable.
Also, E(X) = r/p and V(X) = [r(1p)]/p2 reduce to E(X) = 1/p and V(X) = (1p)/p2, respectively.
3-17
3-96.
1 1
E ( X ) = x (1 p ) x 1 p = p xq x 1 = p 2 =
x =1
x =1
p p
V ( X ) = ( x 1p ) 2 (1 p) x 1 p = px 2 2 x +
x =1
x =1
= p x 2 q x 1 2 xq x 1 +
x =1
x =1
= p x 2 q x 1
x =1
2
p2
1
p
1
p
)(1 p)
x 1
x =1
1
p2
= p x 2 q x 1
x =1
1
p2
= p dqd q + 2q 2 + 3q 3 + ...
1
p2
1
p2
[ 2(1 p) + p 1] = (1 p) = q
=
p2
p2
p2
Section 3-8
3-97.
( )( ) = 20(82160) = 0.4191
a.) P ( X = 1) =
( ) 3921225
20 80
1
3
100
4
b.)
( )( ) = 4845(1) = 0.001236
c.) P ( X = 4) =
( ) 3921225
20 80
4
0
100
4
K
20
= 4
= 0 .8
N
100
96
N n
V ( X ) = np (1 p )
= 4(0.2)(0.8) = 0.6206
99
N 1
d.) E ( X )
= np = n
3-18
1
p2
x 1
3-96.
1 1
E ( X ) = x (1 p ) x 1 p = p xq x 1 = p 2 =
x =1
x =1
p p
V ( X ) = ( x 1p ) 2 (1 p) x 1 p = px 2 2 x +
x =1
x =1
= p x 2 q x 1 2 xq x 1 +
x =1
x =1
= p x 2 q x 1
x =1
2
p2
1
p
1
p
)(1 p)
x 1
x =1
1
p2
= p x 2 q x 1
x =1
1
p2
= p dqd q + 2q 2 + 3q 3 + ...
1
p2
1
p2
[ 2(1 p) + p 1] = (1 p) = q
=
p2
p2
p2
Section 3-8
3-97.
( )( ) = 20(82160) = 0.4191
a.) P ( X = 1) =
( ) 3921225
20 80
1
3
100
4
b.)
( )( ) = 4845(1) = 0.001236
c.) P ( X = 4) =
( ) 3921225
20 80
4
0
100
4
K
20
= 4
= 0 .8
N
100
96
N n
V ( X ) = np (1 p )
= 4(0.2)(0.8) = 0.6206
99
N 1
d.) E ( X )
= np = n
3-18
1
p2
x 1
3-98.
( )( ) = (4 16 15 14) / 6 = 0.4623
( ) (20 19 18 17) / 24
( )( ) =
1
b) P ( X = 4) =
= 0.00021
( ) (20 19 18 17) / 24
a)
c)
P ( X = 1) =
4
1
16
3
20
4
4 16
4 0
20
4
P( X 2) = P ( X = 0) + P ( X = 1) + P ( X = 2)
( )( ) + ( )( ) + ( )( )
=
( ) ( ) ( )
4
0
16
4
20
4
4
1
16
3
20
4
4
2
16
2
20
4
24
6
2
20191817
24
= 0.9866
0.5
0.4
P(x)
0.3
0.2
0.1
0.0
0
3-100.
24 12
/
x 3 x
(a) f(x) =
36
3
3-19
3-101.
Let X denote the number of men who carry the marker on the male chromosome for an increased risk for high
blood pressure. N=800, K=240 n=10
a) n=10
( )( ) = ( )( ) = 0.1201
P( X = 1) =
( )
240 560
1
9
800
10
240!
560!
1!239! 9!551!
800!
10!790!
b) n=10
P ( X > 1) = 1 P ( X 1) = 1 [ P ( X = 0) + P ( X = 1)]
P( X = 0) =
( )( ) = (
( )
240 560
0
10
800
10
)(
240!
560!
0!240! 10!550!
800!
10!790!
= 0.0276
3-102.
Let X denote the number of cards in the sample that are defective.
a)
P( X 1) = 1 P( X = 0)
( )( ) =
P( X = 0) =
( )
20
0
120
20
140
20
120!
20!100!
140!
20!120!
= 0.0356
P( X 1) = 1 0.0356 = 0.9644
b)
P ( X 1) = 1 P( X = 0)
P ( X = 0) =
( )( ) =
( )
5
0
135
20
140
20
135!
20!115!
140!
20!120!
135!120!
= 0.4571
115!140!
P ( X 1) = 1 0.4571 = 0.5429
3-103.
N=300
(a) K=243, n=3, P(X=1)=0.087
(b) P(X1)=0.9934
(c) K=26+13=39
P(X=1)=0.297
(d) K=300-18=282
P(X1)=0.9998
3-104.
Let X denote the count of the numbers in the state's sample that match those in the player's
sample. Then, X has a hypergeometric distribution with N = 40, n = 6, and K = 6.
( )( ) = 40! = 2.61 10
( ) 6!34!
( )( ) = 6 34 = 5.31 10
b) P ( X = 5) =
( ) ( )
( )( ) = 0.00219
c) P ( X = 4) =
( )
a)
P( X = 6) =
6
6
6
5
6
4
34
0
40
6
34
1
40
6
34
2
40
6
40
6
d) Let Y denote the number of weeks needed to match all six numbers. Then, Y has a geometric distribution with p =
3-20
1
3,838,380
3-105.
and E(Y) = 1/p = 3,838,380 weeks. This is more than 738 centuries!
Let X denote the number of blades in the sample that are dull.
a)
P ( X 1) = 1 P( X = 0)
P ( X = 0) =
( )( ) =
( )
10
0
38
5
48
5
38!
5!33!
48!
5!43!
38!43!
= 0.2931
48!33!
P ( X 1) = 1 P( X = 0) = 0.7069
b) Let Y denote the number of days needed to replace the assembly.
P(Y = 3) =
( )( ) =
c) On the first day, P ( X = 0) =
( )
( )( ) =
On the second day, P ( X = 0) =
( )
2
0
46
5
48
5
6
0
46!
5!41!
48!
5!43!
42
5
48
5
42!
5!37!
48!
5!43!
46!43!
= 0.8005
48!41!
=
42!43!
= 0.4968
48!37!
4
1
4
4
The results from the binomial approximation are close to the probabilities obtained in Exercise
3-97.
c) Assume X has a binomial distribution with n = 4 and p = 0.2. Consequently, P(X = 1) and
P(X = 4) are the same as computed in part b. of this exercise. This binomial approximation is
not as close to the true answer as the results obtained in part b. of this exercise.
d) From Exercise 3-102, X is approximately binomial with n = 20 and p = 20/140 = 1/7.
P( X 1) = 1 P( X = 0) =
( )( ) ( )
20
0
1 0 6 20
7
7
= 1 0.0458 = 0.9542
P( X 1) = 1 P( X = 0) = ( 020 )( 281 )
( )
0 27 20
28
3-21
= 1 0.4832 = 0.5168
Section 3-9
3-107.
e 4 4 0
= e 4 = 0.0183
0!
b) P( X 2) = P( X = 0) + P( X = 1) + P( X = 2)
a) P( X = 0) =
e 4 41 e 4 42
+
1!
2!
= 0.2381
= e 4 +
e 4 4 4
= 01954
.
4!
e 4 4 8
= 0.0298
d) P( X = 8) =
8!
c) P( X = 4) =
3-108
a) P ( X = 0) = e 0.4 = 0.6703
3-109.
3-110
a) Let X denote the number of calls in one hour. Then, X is a Poisson random variable with = 10.
e 10 105
= 0.0378 .
5!
e 10 10
e 10 102 e 10 103
+
+
= 0.0103
b) P( X 3) = e 10 +
1!
2!
3!
c) Let Y denote the number of calls in two hours. Then, Y is a Poisson random variable with
e 20 2015
= 0.0516
= 20. P( Y = 15) =
15!
d) Let W denote the number of calls in 30 minutes. Then W is a Poisson random variable with
e 5 55
= 01755
.
= 5. P( W = 5) =
5!
P( X = 5) =
3-111.
3-112.
3-113.
3-22
3-114.
a) Let X denote the number of flaws in one square meter of cloth. Then, X is a Poisson random variable
with = 0.1.
P( X = 2) =
e 0.1 (0.1) 2
= 0.0045
2!
b) Let Y denote the number of flaws in 10 square meters of cloth. Then, Y is a Poisson random variable
with = 1.
P (Y = 1) =
e111
= e1 = 0.3679
1!
c) Let W denote the number of flaws in 20 square meters of cloth. Then, W is a Poisson random variable
P (W = 0) = e 2 = 0.1353
P (Y 2) = 1 P (Y 1) = 1 P (Y = 0) P (Y = 1)
with = 2.
d)
= 1 e 1 e 1
= 0.2642
3-115.a)
2) = e
0.2
e 0.2 0.2
e 0.2 (0.2) 2
+
+
= 0.9989
1!
2!
a) Let X denote the number of cracks in 5 miles of highway. Then, X is a Poisson random variable with
10
= 10. P ( X = 0) = e
= 4.54 10
b) Let Y denote the number of cracks in a half mile of highway. Then, Y is a Poisson random variable with
1
= 1. P (Y 1) = 1 P (Y = 0) = 1 e = 0.6321
c) The assumptions of a Poisson process require that the probability of a count is constant for all intervals.
If the probability of a count depends on traffic load and the load varies, then the assumptions of a Poisson
process are not valid. Separate Poisson random variables might be appropriate for the heavy and light
load sections of the highway.
3-117.
a) Let X denote the number of flaws in 10 square feet of plastic panel. Then, X is a Poisson random
0.5
0.6065
10
P (Y = 10) = (0.6065)10 (0.3935) 0 = 0.0067
10
c) Let W denote the number of cars with surface flaws. Because the number of flaws has a
Poisson distribution, the occurrences of surface flaws in cars are independent events with
constant probability. From part a., the probability a car contains surface flaws is 10.6065 =
0.3935. Consequently, W is binomial with n = 10 and p = 0.3935.
10
P (W = 0) = (0.3935)0 (0.6065)10 = 0.0067
0
10
P (W = 1) = (0.3935)1 (0.6065)9 = 0.0437
1
P (W 1) = 0.0067 + 0.0437 = 0.0504
3-23
3-118.
a) Let X denote the failures in 8 hours. Then, X has a Poisson distribution with = 0.16.
P( X = 0) = e 0.16 = 0.8521
b) Let Y denote the number of failure in 24 hours. Then, Y has a Poisson distribution with
= 0.48.
Supplemental Exercises
3-119.
E( X ) =
1 1 1 1 3 1 1
+ + = ,
8 3 4 3 8 3 4
2
3-120.
1 1 1 1 3 1 1
V ( X ) = + + = 0.0104
8 3 4 3 8 3 4
1000
1
999
a) P ( X = 1) =
1 0.001 (0.999) = 0.3681
1000
999
0.0010 (0.999) = 0.6319
b) P( X 1) = 1 P( X = 0) = 1
0
1000
1000
1000
1000
999
0.0012 0.999998
0.0011 (0.999) +
0.0010 (0.999) +
c) P( X 2) =
2
1
0
= 0.9198
d) E ( X ) = 1000(0.001) = 1
V ( X ) = 1000(0.001)(0.999) = 0.999
3-121.
50
50
50
50
49
48
2) = 0.10 (0.9) + 0.11 (0.9) + 0.12 (0.9) = 0.112
2
1
0
50 50
50 49
1
0.1 (0.9 )0 = 4.51 10 48
(
)
c) P ( X 49) =
0
.
1
0
.
9
+
49
50
b) P( X
3-122.
12 0
12
p (1 p)12 - p 1 (1 p )14 =0.0062
0
1
(b) P(X>1)=1-P(X1)= 1-
3-123.
3-124.
V (X ) = 0.3447
(b)
(0.5)12 = 0.000244
C126 (0.5)6 (0.5)6 = 0.2256
(c)
(a)
3-24
3-118.
a) Let X denote the failures in 8 hours. Then, X has a Poisson distribution with = 0.16.
P( X = 0) = e 0.16 = 0.8521
b) Let Y denote the number of failure in 24 hours. Then, Y has a Poisson distribution with
= 0.48.
Supplemental Exercises
3-119.
E( X ) =
1 1 1 1 3 1 1
+ + = ,
8 3 4 3 8 3 4
2
3-120.
1 1 1 1 3 1 1
V ( X ) = + + = 0.0104
8 3 4 3 8 3 4
1000
1
999
a) P ( X = 1) =
1 0.001 (0.999) = 0.3681
1000
999
0.0010 (0.999) = 0.6319
b) P( X 1) = 1 P( X = 0) = 1
0
1000
1000
1000
1000
999
0.0012 0.999998
0.0011 (0.999) +
0.0010 (0.999) +
c) P( X 2) =
2
1
0
= 0.9198
d) E ( X ) = 1000(0.001) = 1
V ( X ) = 1000(0.001)(0.999) = 0.999
3-121.
50
50
50
50
49
48
2) = 0.10 (0.9) + 0.11 (0.9) + 0.12 (0.9) = 0.112
2
1
0
50 50
50 49
1
0.1 (0.9 )0 = 4.51 10 48
(
)
c) P ( X 49) =
0
.
1
0
.
9
+
49
50
b) P( X
3-122.
12 0
12
p (1 p)12 - p 1 (1 p )14 =0.0062
0
1
(b) P(X>1)=1-P(X1)= 1-
3-123.
3-124.
V (X ) = 0.3447
(b)
(0.5)12 = 0.000244
C126 (0.5)6 (0.5)6 = 0.2256
(c)
(a)
3-24
V (X ) =0.995
Let X denote the number of mornings needed to obtain a green light. Then X is a geometric random
variable with p = 0.20.
a) P(X = 4) = (1-0.2)30.2= 0.1024
b) By independence, (0.8)10 = 0.1074. (Also, P(X > 10) = 0.1074)
3-126
Let X denote the number of attempts needed to obtain a calibration that conforms to specifications. Then, X
is geometric with p = 0.6.
P(X 3) = P(X=1) + P(X=2) + P(X=3) = 0.6 + 0.4(0.6) + 0.42(0.6) = 0.936.
3-127.
Let X denote the number of fills needed to detect three underweight packages. Then X is a negative
binomial random variable with p = 0.001 and r = 3.
a) E(X) = 3/0.001 = 3000
b) V(X) = [3(0.999)/0.0012] = 2997000. Therefore, X = 1731.18
3-128.
3-129.
(a) =6*0.5=3.
P(X=0) = 0.0498
(b) P(X3)=0.5768
(c) P(Xx) 0.9, x=5
(d) 2= =6. Not appropriate.
3-130.
Let X denote the number of totes in the sample that do not conform to purity requirements. Then, X has a
hypergeometric distribution with N = 15, n = 3, and K = 2.
2 13
0 3
13!12!
= 0.3714
P( X 1) = 1 P( X = 0) = 1 = 1
10!15!
15
3
3-131.
Let X denote the number of calls that are answered in 30 seconds or less. Then, X is a binomial random
variable with p = 0.75.
a) P(X = 9) =
10
(0.75)9 (0.25)1 = 0.1877
9
3-25
20
20
20
= (0.75) 16 (0.25) 4 + (0.75) 17 (0.25) 3 + (0.75) 18 (0.25) 2
16
17
18
20
20
+ (0.75) 19 (0.25) 1 + (0.75) 20 (0.25) 0 = 0.4148
19
20
c) E(X) = 20(0.75) = 15
3-132.
Let Y denote the number of calls needed to obtain an answer in less than 30 seconds.
a) P (Y = 4) = (1 0.75)
b) E(Y) = 1/p = 1/0.75 = 4/3
3-133.
Let W denote the number of calls needed to obtain two answers in less than 30 seconds. Then, W has a
negative binomial distribution with p = 0.75.
5
a) P(W=6) = (0.25)4 (0.75)2 = 0.0110
1
b) E(W) = r/p = 2/0.75 = 8/3
3-134.
P( X = 5) =
e 5 5 5
= 0.1755
5!
b) Let Y denote the number of messages sent in 1.5 hours. Then, Y is a Poisson random variable with
=7.5.
P(Y = 10) =
e 7.5 (7.5)10
= 0.0858
10!
c) Let W denote the number of messages sent in one-half hour. Then, W is a Poisson random variable with
= 2.5. P (W < 2) = P (W = 0) + P (W = 1) = 0.2873
3-135X is a negative binomial with r=4 and p=0.0001
3-136.
P (Y 3) = e 1 +
3-137.
Let X denote the number of individuals that recover in one week. Assume the individuals are independent.
Then, X is a binomial random variable with n = 20 and p = 0.1. P(X 4) = 1 P(X 3) = 1 0.8670 =
0.1330.
3-138.
a.) P(X=1) = 0 , P(X=2) = 0.0025, P(X=3) = 0.01, P(X=4) = 0.03, P(X=5) = 0.065
P(X=6) = 0.13, P(X=7) = 0.18, P(X=8) = 0.2225, P(X=9) = 0.2, P(X=10) = 0.16
b.) P(X=1) = 0.0025, P(X=1.5) = 0.01, P(X=2) = 0.03, P(X=2.5) = 0.065, P(X=3) = 0.13
P(X=3.5) = 0.18, P(X=4) = 0.2225, P(X=4.5) = 0.2, P(X=5) = 0.16
3-139.
Let X denote the number of assemblies needed to obtain 5 defectives. Then, X is a negative binomial
random variable with p = 0.01 and r=5.
a) E(X) = r/p = 500.
b) V(X) =(5* 0.99)/0.012 = 49500 and X = 222.49
3-140.
Here n assemblies are checked. Let X denote the number of defective assemblies. If P(X 1) 0.95, then
3-26
n
(0.01) 0 (0.99) n = 99 n
0
n(ln(0.99)) ln(0.05)
ln(0.05)
n
= 298.07
ln(0.95)
P(X=0) =
3-142.
Let X denote the number of products that fail during the warranty period. Assume the units are
independent. Then, X is a binomial random variable with n = 500 and p = 0.02.
a) P(X = 0) =
500
b) E(X) = 500(0.02) = 10
c) P(X >2) = 1 P(X 2) = 0.9995
3-143.
3-144.
3-145.
x
f(x)
3-146.
2
0.2
5.7
0.3
6.5
0.3
8.5
0.2
Let X denote the number of bolts in the sample from supplier 1 and let Y denote the number of bolts in the
sample from supplier 2. Then, x is a hypergeometric random variable with N = 100, n = 4, and K = 30.
Also, Y is a hypergeometric random variable with N = 100, n = 4, and K = 70.
a) P(X=4 or Y=4) = P(X = 4) + P(Y = 4)
30 70 30 70
4 0
0 4
= +
100
100
4
4
= 0.2408
3-27
30 70 30 70
+
3 1
1 3
= = 0.4913
100
3-147.
Let X denote the number of errors in a sector. Then, X is a Poisson random variable with = 0.32768.
a) P(X>1) = 1 P(X1) = 1 e-0.32768 e-0.32768(0.32768) = 0.0433
b) Let Y denote the number of sectors until an error is found. Then, Y is a geometric random variable and
P = P(X 1) = 1 P(X=0) = 1 e-0.32768 = 0.2794
E(Y) = 1/p = 3.58
3-148.
Let X denote the number of orders placed in a week in a city of 800,000 people. Then X is a Poisson
random variable with = 0.25(8) = 2.
a) P(X 3) = 1 P(X 2) = 1 [e-2 + e-2(2) + (e-222)/2!] = 1 0.6767 = 0.3233.
b) Let Y denote the number of orders in 2 weeks. Then, Y is a Poisson random variable with = 4, and
P(Y>2) =1- P(Y 2) = e-4 + (e-441)/1!+ (e-442)/2! =1 - [0.01832+0.07326+0.1465] = 0.7619.
125 375
0 5 6.0164 E10
f X (0) =
=
= 0.2357
2.5524 E11
500
5
125 375
1 4 125(8.10855E8)
f X (1) =
=
= 0.3971
2.5525E11
500
5
125 375
2 3 7750(8718875)
=
= 0.2647
f X (2) =
2.5524 E11
500
5
125 375
3 2 317750(70125)
=
= 0.0873
f X (3) =
2.5524 E11
500
5
125 375
4 1 9691375(375)
=
= 0.01424
f X (4) =
2.5524 E11
500
3-28
125 375
5 0 2.3453E8
=
= 0.00092
f X (5) =
2.5524 E11
500
5
b)
x
f(x)
3-150.
0
1
2
3
4
5
6
7
8
9
10
0.0546 0.1866 0.2837 0.2528 0.1463 0.0574 0.0155 0.0028 0.0003 0.0000 0.0000
Let X denote the number of totes in the sample that exceed the moisture content. Then X is a binomial
random variable with n = 30. We are to determine p.
If P(X 1) = 0.9, then P(X = 0) = 0.1. Then
30 0
( p) (1 p )30 = 0.1 , giving 30ln(1p)=ln(0.1),
0
Let t denote an interval of time in hours and let X denote the number of messages that arrive in time t.
Then, X is a Poisson random variable with = 10t.
Then, P(X=0) = 0.9 and e-10t = 0.9, resulting in t = 0.0105 hours = 37.8 seconds
3-152.
a) Let X denote the number of flaws in 50 panels. Then, X is a Poisson random variable with
= 50(0.02) = 1. P(X = 0) = e-1 = 0.3679.
b) Let Y denote the number of flaws in one panel, then
P(Y 1) = 1 P(Y=0) = 1 e-0.02 = 0.0198. Let W denote the number of panels that need to be
inspected before a flaw is found. Then W is a geometric random variable with p = 0.0198 and
E(W) = 1/0.0198 = 50.51 panels.
0.02
c) P (Y 1) = 1 P (Y = 0) = 1 e
= 0.0198
Let V denote the number of panels with 1 or more flaws. Then V is a binomial random
variable with n=50 and p=0.0198
50
50
P(V 2) = 0.0198 0 (.9802) 50 + 0.01981 (0.9802) 49
0
1
50
+ 0.0198 2 (0.9802) 48 = 0.9234
2
Mind Expanding Exercises
3-153. The binomial distribution
P(X=x) =
n!
px(1-p)n-x
x!(n x )!
If n is large, then the probability of the event could be expressed as /n, that is =np. We
could re-write the probability mass function as:
P(X=x) =
n!
[/n]x[1 (/n)]n-x
x!(n x )!
Where p = /n.
P(X=x)
n (n 1) (n 2) (n 3)...... (n x + 1) x
(1 (/n))n-x
x
x!
n
3-29
b) P (T < 1) = P Z <
1 1.5
= P ( Z < 1.79) = 0.0367
0.078
1.5 3
= P( Z < 6.10) 0
0.246
Let X and Y denote the percentage returns for security one and two respectively.
If of the total dollars is invested in each then X+ Y is the percentage return.
E(X+ Y) = 0.05 (or 5 if given in terms of percent)
V(X+ Y) = 1/4 V(X)+1/4V(Y)+2(1/2)(1/2)Cov(X,Y)
where Cov(XY)=XY=-0.5(2)(4) = -4
V(X+ Y) = 1/4(4)+1/4(6)-2 = 3
Also, E(X) = 5 and V(X) = 4. Therefore, the strategy that splits between the securities
has a lower standard deviation of percentage return than investing 2million in the first
security.
Mind-Expanding Exercises
5-94
By the independence,
P( X 1 A1 , X 2 A2 ,..., X p A p ) =
...
A1
A2
Ap
= f X 1 ( x1 )dx1 f X 2 ( x 2 )dx 2 ... f X p ( x p )dx p
A1
A2
A p
= P( X 1 A1 ) P( X 2 A2 )...P( X p A p )
5-95
E (Y ) = c11 + c2 2 + ... + c p p .
Also,
5-47
[c x + c x + ... + c
= [c ( x ) + ... + c
V (Y ) =
1 1
2 2
c c ( x )( x ) f ( x ) f
= c c [ ( x ) f ( x )dx ][ ( x
1
1 2
X1
X1
X2
2 ) f X 2 ( x2 )dx2 = 0
from the definition of the mean. Therefore, each cross-term in the last integral for V(Y)
is zero and
V (Y ) =
[ c ( x )
2
1
][
5-96
f XY ( x, y )dydx =
0
f X ( x) = cdy =
0
1
a
f XY (x,y)=f X (x)f Y (y) for all x and y and X and Y are independent.
5-97
=
g
(
x
)
h
(
y
)
dydx
g
(
v
)
dv
5-48
CHAPTER 5
Section 5-1
5-1.
V(X)=E(X2)-[E(X)]2=[12(1/4)+1.52(3/8)+2.52(1/4)+32(1/8)]-1.81252=0.4961
V(Y)= E(Y2)-[E(Y)]2=[12(1/4)+22(1/8)+32(1/4)+42(1/4)+52(1/8)]-2.8752=1.8594
f) marginal distribution of X
x
f(x)
1
1.5
3/8
2.5
3
1/8
g) fY 1.5 ( y ) =
h) f X 2 ( x) =
f XY (1.5, y )
and f X (1.5) = 3/8. Then,
f X (1.5)
y
fY 1.5 ( y )
2
3
(1/8)/(3/8)=1/3
(1/4)/(3/8)=2/3
f XY ( x,2)
and fY (2) = 1/8. Then,
fY (2)
x
f X 2 ( y)
1.5
(1/8)/(1/8)=1
f ( x , y ) = c ( 2 + 3 + 4 + 3 + 4 + 5 + 4 + 5 + 6) = 1 ,
1
36
(2) = 1 / 18
5-1
e)
) + (3 1536 ) = 13 / 6 = 2.167
= (1 369 ) + (2 12
36
V ( X ) = (1 136 ) 2
E (Y ) = 2.167
+ (2 136 ) 2
9
36
12
36
+ (3 136 ) 2
15
36
= 0.639
V (Y ) = 0.639
f) marginal distribution of X
x
f X ( x ) = f XY ( x,1) + f XY ( x,2) + f XY ( x,3)
1
1/4
2
1/3
3
5/12
g) fY X ( y ) =
f XY (1, y )
f X (1)
f Y X ( y)
1
2
3
(2/36)/(1/4)=2/9
(3/36)/(1/4)=1/3
(4/36)/(1/4)=4/9
h) f X Y ( x) =
f XY ( x,2)
and f Y (2) = f XY (1,2) + f XY (2,2) + f XY (3,2) =
fY (2)
f X Y ( x)
1
2
3
(3/36)/(1/3)=1/4
(4/36)/(1/3)=1/3
(5/36)/(1/3)=5/12
f ( x, y ) 0 and
f ( x, y ) = 1
R
3
8
5-2
1
4
5
8
7
8
3
8
12
36
= 1/ 3
e)
E ( X ) = 1( 18 ) 0.5( 14 ) + 0.5( 12 ) + 1( 18 ) =
E (Y ) = 2( 18 ) 1( 14 ) + 1( 12 ) + 2( 18 ) =
1
8
1
4
V(X)=(-1-1/8)2(1/8)+(-0.5-1/8)2(1/4)+(0.5-1/8)2(1/2)+(1-1/8)2(1/8)=0.4219
V(Y)=(-2-1/4)2(1/8)+(-1-1/4)2(1/4)+(1-1/4)2(1/2)+(2-1/4)2(1/8)=1.6875
f) marginal distribution of X
x
f X (x)
-1
-0.5
0.5
1
1/8
1/8
g) fY X ( y ) =
f XY (1, y )
f X (1)
f Y X ( y)
1/8/(1/8)=1
h) f X Y ( x) =
f XY ( x,1)
f Y (1)
f X Y ( x)
0.5
/(1/2)=1
i) E(X|Y=1) = 0.5
j) no, X and Y are not independent
5-4.
5-5.
X 0, Y 0 and X + Y = 4
3
y 2
1
0
2
x
The problem needs probabilities to total one. Modify so that the probability of moderate
distortion is 0.04.
5-3
x,y
0,0
0,1
0,2
0,3
1,0
1,1
1,2
2,0
2,1
3,0
fxy(x,y)
x
0
1
2
3
fx(x)
0.970299
0.029403
0.000297
0.000001
0.857375
0.1083
0.00456
0.000064
0.027075
0.00228
0.000048
0.000285
0.000012
0.000001
b)
c) E(X)=0(0.970299)+1(0.029403)+2(0.000297)+3*(0.000001)=0.03
(or np=3*0.01)
d) fY 1 ( y ) =
f XY (1, y )
, fx(1) = 0.029403
f X (1)
y
0
1
2
fY|1(x)
0.920824
0.077543
0.001632
e) E(Y|X=1)=0(.920824)+1(0.077543)+2(0.001632)=0.080807
g) No, X and Y are not independent because, for example, fY(0)fY|1(0).
5-6.
y=4
y=3
y=2
y=1
y=0
x=0
x=1
x=2
x=3
x=4
0
0
0
0
5.35x10-05
0.00184
0.00092
0
0
0
0.02031
0.02066
0.00499
0
0
0.08727
0.13542
0.06656
0.01035
0
0.12436
0.26181
0.19635
0.06212
0.00699
b)
x=0
f(x)
x=1
0.2338
x=2
0.4188
x=3
0.2679
5-4
x=4
0.0725
0.0070
c)
E(X)=
4
d) f Y 3 ( y ) =
f XY (3, y )
, fx(3) = 0.0725
f X (3)
Y
0
1
2
3
4
fY|3(y)
0.857
0.143
0
0
0
items found with inspection device 1 and Y is the number of defective items found with
inspection device 2.
y=0
y=1
y=2
y=3
y=4
x=0
x=1
x=2
x=3
x=4
1.10x10-16
2.35x10-14
2.22x10-12 7.88x10-11
1.94x10-19
2.59x10-16
1.47x10-13
3.12x10-11
2.95x10-9 1.05x10-7
-13
-11
-8
1.29x10
7.31x10
1.56x10
1.47x10-6 5.22x10-5
-11
-8
-6
2.86x10
1.62x10
3.45x10
3.26x10-4
0.0116
-9
-6
-4
2.37x10
1.35x10
2.86x10
0.0271
0.961
4
4
x=1
x=2
f(x)
2.40 x 10-9
1.36 x 10-6
2.899 x 10-4
x=3
x = 1,2,3,4
0.0274
d) f Y |2 ( y ) =
f XY (2, y )
= f ( y ) , fx(2) = 2.899 x 10-4
f X (2)
y
fY|1(y)=f(y
5-5
x=4
0.972
0
1
2
3
4
)
8.1 x 10-11
1.08 x 10-7
5.37 x 10-5
0.0119
0.988
g)
h)
P( X = 1, Y = 1)
0.05 + 0.10
=
= 0.3
P(Y = 1)
0.15 + 0.2 + 0.1 + 0.05
P( X = 1, Y = 1, Z = 2)
0.1
P ( X = 1, Y = 1 | Z = 2) ==
=
= 0.2
P( Z = 2)
0.1 + 0.2 + 0.15 + 0.05
P ( X = 1, Y = 1, Z = 2)
0.10
P ( X = 1 | Y = 1, Z = 2) =
=
= 0.4
0.10 + 0.15
P (Y = 1, Z = 2)
P( X = 1 | Y = 1) =
i) f X YZ ( x ) =
5-9.
f XYZ ( x,1,2)
and f YZ (1,2) = f XYZ (1,1,2) + f XYZ (2,1,2) = 0.25
f YZ (1,2)
f X YZ (x)
1
2
0.10/0.25=0.4
0.15/0.25=0.6
5-6
fXY(x,y)
0.1296
0.0648
0.0324
0.0162
0.0081
0.0216
0.0108
0.0054
0.0027
0.0036
0.0018
0.0009
0.0006
0.0003
0.0001
(b) fX(x)= P(X=x) =
X +Y 4
XY
0
0
0
0
0
1
1
1
1
2
2
2
3
3
4
( x, y ) .
fX(x)
0.2511
0.0405
0.0063
0.0009
0.0001
(c) E(X)=
xf
0
1
2
3
4
0
1
2
3
0
1
2
0
1
0
x
0
1
2
3
4
x
0
1
2
3
4
3
3
3
3
3
(e) E(Y|X=3)=0(0.6667)+1(0.3333)=0.3333
(f) V(Y|X=3)=(0-0.3333)2(0.6667)+(1-0.3333)2(0.3333)=0.0741
(g) fX(0)= 0.2511, fY(0)=0.1555, fX(0) * fY(0)= 0.039046 fXY(0,0) = 0.1296
X and Y are not independent.
5-10.
5-7
5-11
( )
( )0.05 0.95
20
1
19
= 0.7358
P( X = 2, Y = 17)
. Now, because x+y+z = 20,
P(Y = 17)
20!
P(X=2, Y=17) = P(X=2, Y=17, Z=1) =
0.05 2 0.8517 0.101 = 0.0540
2!17!1!
P( X = 2, Y = 17) 0.0540
P( X = 2 | Y = 17) =
=
= 0.2224
P(Y = 17)
0.2428
j) P ( X = 2 | Y = 17) =
k)
P( X = 0, Y = 17) P( X = 1, Y = 17)
+ 1
E ( X | Y = 17) = 0
P(Y = 17)
P (Y = 17)
P ( X = 2, Y = 17) P ( X = 3, Y = 17)
+ 3
+ 2
P (Y = 17)
P(Y = 17)
5-8
5-12.
c)
f XY ( x , 2 )
fY (2)
P ( X = x | Y = 2) =
4
0
5
2
15
4
6
2
4
1
4
0
4
1
4
1
5
2
15
4
5
2
15
4
5
2
15
4
5
2
15
4
f XY (x,2)
0
1
2
0.1098/0.3296=0.3331
0.1758/0.3296=0.5334
0.0440/0.3296=0.1335
4
2
6
1
5
2
15
4
6
0
6
2
6
1
6
1
d)
P(X=x, Y=y, Z=z) is the number of subsets of size 4 that contain x printers with graphics
enhancements, y printers with extra memory, and z printers with both features divided by
the number of subsets of size 4.
( )( )( )
P ( X = x, Y = y , Z = z ) =
for x+y+z = 4.
( )
( )( )( ) = 0.1758
P( X = 1, Y = 2, Z = 1) =
( )
( )( )( ) = 0.2198
e) P ( X = 1, Y = 1) = P( X = 1, Y = 1, Z = 2) =
( )
4
x
5
y
6
z
15
4
4
1
5
2
15
4
6
1
4
1
5
1
15
4
6
2
P ( X = 1, Y = 2 | Z = 1) = P ( X = 1, Y = 2, Z = 1) / P ( Z = 1)
=
h)
5-9
5
2
15
4
6
1
6
1
9
3
15
4
P ( X = 2 | Y = 2 ) = P ( X = 2 , Y = 2 ) / P (Y = 2 )
4 5
6
= ( 2 )((152 )() 0 )
4
] [( ()( ) )] = 0 .1334
5
2
10
2
15
4
5-13
a) The probability distribution is multinomial because the result of each trial (a dropped
oven) results in either a major, minor or no defect with probability 0.6, 0.3 and 0.1
respectively. Also, the trials are independent
b) Let X, Y, and Z denote the number of ovens in the sample of four with major, minor,
and no defects, respectively.
4!
0.6 2 0.3 2 0.10 = 0.1944
2!2!0!
4!
c) P( X = 0, Y = 0, Z = 4) =
0.6 0 0.3 0 0.14 = 0.0001
0!0!4!
P( X = 2, Y = 2, Z = 0) =
d) fXY ( x, y) = fXYZ ( x, y, z) where R is the set of values for z such that x+y+z = 4. That is,
R
f XY ( x, y ) =
4!
0.6 x 0.3 y 0.14 x y
x! y!(4 x y )!
for x + y 4.
P( X = 2, Y = 2) 0.1944
=
= 0.7347
P(Y = 2)
0.2646
4
P(Y = 2) = 0.3 2 0.7 4 = 0.2646
2
P( X = 0, Y = 2) 4!
P( X = 1, Y = 2) 4!
P ( X = 2, Y = 2) 4!
P ( X = 0 | Y = 2) =
5-10
5-14
Let X, Y, and Z denote the number of bits with high, moderate, and low distortion. Then,
the joint distribution of X, Y, and Z is multinomial with n =3 and
p1 = 0.01, p2 = 0.04, and p3 = 0.95 .
a)
P ( X = 2, Y = 1) = P( X = 2, Y = 1, Z = 0)
3!
=
0.0120.0410.950 = 1.2 10 5
2!1!0!
3!
b) P ( X = 0, Y = 0, Z = 3) =
0.0100.0400.953 = 0.8574
0!0!3!
c) X has a binomial distribution with n = 3 and p = 0.01. Then, E(X) = 3(0.01) = 0.03
and V(X) = 3(0.01)(0.99) = 0.0297.
d) First find P( X | Y = 2)
P (Y = 2) = P( X = 1, Y = 2, Z = 0) + P( X = 0, Y = 2, Z = 1)
3!
3!
=
0.01(0.04) 2 0.95 0 +
0.010 (0.04) 2 0.951 = 0.0046
1!2!0!
0!2!1!
P( X = 0, Y = 2) 3!
P ( X = 1 | Y = 2) =
P( X = 1, Y = 2) 3!
(a) fXYZ(x,y,z)
fXYZ(x,y,z)
0.43
0.44
0.04
0.05
0.04
Selects(X
)
23
4.2
11.4
130
0
Updates(Y)
11
3
0
120
0
Inserts(Z
)
12
1
0
0
0
(b)PXY|Z=0
PXY|Z=0(x,y)
4/13 = 0.3077
5/13 = 0.3846
4/13= 0.3077
Selects(X
)
11.4
130
0
updates(Y)
0
120
0
5-11
Inserts(Z
)
0
0
0
5-16
a) Let X, Y, and Z denote the risk of new competitors as no risk, moderate risk,
and very high risk. Then, the joint distribution of X, Y, and Z is multinomial with n =12
and p1 = 0.13, p 2 = 0.72, and p 3 = 0.15 . X, Y and Z 0 and x+y+z=12
b) P ( X = 1, Y = 3, Z = 1) = 0, not possible since x+y+z12
c)
12
12
12
P( Z 2) = 0.15 0 0.8512 + 0.151 0.8511 + 0.15 2 0.8510
0
1
2
= 0.1422 + 0.3012 + 0.2924 = 0.7358
d) P ( Z = 2 | Y = 1, X = 10) = 0
e)
f)
P (Y 1, Z 1 | X = 10) =
P( Z = 1, Y = 1, X = 10)
P( X = 10)
12!
0.13100.7210.151 6.89 x108
10!1!1!
= 0.2852
=
g)
5-12
Section 5-2
3 3
5-17
xydxdy = c y x2
0 0
dy = c(4.5
y2
2
)=
81
4
c.
Therefore, c = 4/81.
3 2
a) P ( X < 2, Y < 3) =
4
81
xydxdy =
4
81
0 0
b) P(X < 2.5) = P(X < 2.5, Y < 3) because the range of Y is from 0 to 3.
3 2.5
4
81
xydxdy =
4
81
0 0
2.5 3
4
81
xydxdy =
4
81
1 0
(4.5) ydy = 18
81
y2
2
2.5
=0.5833
d)
2.5 3
4
81
xydxdy =
2.5
4
81
1 1.8
3 3
e) E ( X ) =
4
81
ydxdy =
9 ydy =
4
81
0 0
4 0
f) P ( X < 0, Y < 4) =
4
81
g) f X ( x) =
XY
xydxdy = 0 ydy = 0
0
2x
9
f XY (1.5, y )
=
f X (1.5)
4
81
2
9
y (1.5)
(1.5)
2 2
2y3
2
i) E(Y|X=1.5) = y y dy = y dy =
9
90
27
0
2
f XY ( x,2)
=
fY (2)
4
81
2
9
x(2)
(2)
=2
0
9 ydy = 9 y
0
k) f X 2 ( x) =
=2
= 92 x
2
0
5-13
h) fY 1.5 ( y ) =
4
81
0 0
2
4 y
9 2
(2.88) ydy =
4
4
0 =
9
9
5-18.
3 x+2
( x + y)dydx = xy +
x
x+2
y2
2
dx
x
= x( x + 2) + ( x +22 ) x 2
2
x2
2
]dx
= c (4 x + 2)dx = 2 x 2 + 2 x
= 24c
Therefore, c = 1/24.
a) P(X < 1, Y < 2) equals the integral of f XY ( x, y ) over the following region.
y
2
0
1 2
0
Then,
1 2
P( X < 1, Y < 2) =
1
1
( x + y )dydx =
xy +
24 0 x
24 0
1 2
x + 2x
24
x3
2
y2
2
dx =
2
1 3
2 x + 2 3 x2 dx =
24 0
= 0.10417
0
b) P(1 < X < 2) equals the integral of f XY ( x, y ) over the following region.
y
2
0
0
2 x+2
1
24 1
( x + y)dydx =
x
1
xy +
24 1
y2
2
x+2
dx
x
2
1
1 2
1
=
(
4
x
+
2
)
dx
=
2
x
+
2
x
= 6.
24
24 0
1
5-14
1 1
1
24
1
y2
1
0 x ( x + y)dydx = 1 24 0 ( xy + 2 )
x
1
1
1
1 3 2
1 x2 1 1 3
= 1
+ x
x + x dx = 1
24 0
2 2
24 2 2 2 0
= 1 0.02083 = 0.9792
d) P(X < 2, Y < 2) is the integral of fXY ( x, y) over the following region.
y
2
0
2
0
3 x+2
E( X ) =
1
24 0
x( x + y)dydx =
x
1
x2 y +
24 0
xy 2
2
x+2
dx
x
3
1
1 4x3
15
2
2
+
=
+
(
4
2
)
x
x
dx
x
24 0
24 3
8
0
3
=
e)
3 x+2
1
E( X ) =
24 0
=
1
2
x x( x + y)dydx = 24 0 x y +
xy 2
2
x+2
dx
x
3
3
1
1 4x3
15
2
2
(
4
2
)
x
+
x
dx
=
+
x
24 0
24 3
8
0
5-15
f)
3 x+2
V (X ) =
1
24 0
1
15
x 2 ( x + y )dydx =
x3 y +
24 0
8
x4
1
15
3
2
+
+
x
x
x
(
3
4
4
)dx
24 0
4
8
x2 y 2
2
x+2
x
15
dx
8
x 5 3 15
1 3x 4 4 x 3
31707
2
x
=
+
+
=
24 4
3
20 0 8
320
g) f X (x) is the integral of f XY ( x, y ) over the interval from x to x+2. That is,
1
f X ( x) =
24
h) f Y 1 ( y ) =
x+2
1
xy +
24
( x + y)dy =
x
f XY (1, y )
f X (1)
1
(1+ y )
24
1 1
+
6 12
1+ y
6
y2
2
x+2
x
x 1
= 6 + 12 for 0 < x < 3.
y
f
Y|1
1
0
1 2
3
1 y2 y3
1
1+ y
2
= 2.111
i) E(Y|X=1) = y
+
dy = ( y + y )dy =
6
2
3
6
6
1
1
1
3
3
y2
1
1
1+ y
=0.5833
j) P (Y > 2 | X = 1) =
dy = (1 + y )dy = y +
6
2
6
6
2
2
2
3
k)
f X 2 ( x) =
f XY ( x , 2 )
.
fY ( 2)
three regions of integration. For 0 < y 2 the integration is from 0 to y. For 2 < y 3
the integration is from y-2 to y. For 3 < y < 5 the integration is from y to 3. Because
the condition is y=2, only the first integration is needed.
y
1 x2
1
f Y ( y) =
x
+
y
dx
=
+
xy
(
)
2
24
24 0
0
y
5-16
y2
16
for 0 < y 2 .
y
f X|2 (x) defined over this line segment
2
1
0
1 2
1
( x + 2)
x+2
24
for 0 < x < 3
=
Therefore, fY (2) = 1 / 4 and f X 2 ( x) =
1/ 4
6
3 x
5-19
3
y2
x3
x 4 81
c xydyd x = c x
dx = c
dx
= c. Therefore, c = 8/81
2 0
8
8
0 0
0
0 2
1 x
8
8 x3
8 1 1
xydyd
x
=
dx = = .
a) P(X<1,Y<2)=
81 0 0
81 0 2
81 8 81
2 x
4
8
8
x2
8 x
b) P(1<X<2) =
=
=
xydyd
x
x
dx
81 1 0
81 1 2
81 8
4
8 (2 1) 5
=
=
.
27
81 8
c)
3
3
3 x
3
3
8
8 x2 1
8 x
8 x4 x2
x
8 3 4 3 2 14 12 1
= 0.01235
=
81 8
4 8 4 81
2 x
2
8
8 x3
8 2 4 16
= .
d) P(X<2,Y<2) =
xydyd
x
=
dx
=
81 0 0
81 0 2
81 8 81
e)
3 x
3 x
3 x
8
8
8 x2 2
8 x4
E ( X ) = x( xy )dyd x = x 2 ydyd x =
x dx = dx
81 0 0
81 0 0
81 0 2
81 0 2
5
8 3 12
= =
81 10 5
f)
3 x
x3
8
8
8
E (Y ) = y ( xy )dyd x = xy 2 dyd x = x
dx
81 0 0
81 0 0
81 0 3
3
5
8 x4
8 3 8
=
dx = =
81 0 3
81 15 5
5-17
8
4x 3
xydy
=
81 0
81
x
g) f ( x) =
0< x<3
8
(1) y
f (1, y ) 81
f
y
(
)
=
=
= 2y
h) Y | x =1
f (1)
4(1) 3
81
1
i) E (Y | X = 1) = 2 ydy = y 2
0
1
0
0 < y <1
=1
j) P (Y > 2 | X = 1) = 0 this isnt possible since the values of y are 0< y < x.
k) f ( y ) =
3
4y
8
, therefore
xydx =
81 0
9
8
x ( 2)
f ( x,2) 81
2x
(
)
f X |Y = 2 x =
=
=
4( 2)
f ( 2)
9
9
5-20.
0< x<3
Solve for c
x
c e 2 x 3 y dyd x =
0 0
c 2 x
c
e 1 e 3 x d x = e 2 x e 5 x d x =
30
30
c 1 1 1
= c. c = 10
3 2 5 10
a)
1 x
10 e 5 x e 2 x
= 0.77893
=
3 5
2 0
2 x
b)
10 2 x
e e 5 x d x
3 1
10 e 5 x e 2 x
= 0.19057
=
3 5
2 1
c)
x
P (Y > 3) = 10 e
3 3
2 x 3 y
10 2 x 9
dyd x =
e (e e 3 x )dy
3 3
10 e 5 x e 9 e 2 x
=
3 5
2
10 2 x
10
e (1 e 3 x )dy = e 2 x e 5 x dy
3 0
3 0
= 3.059 x10 7
3
5-18
d)
2
2 x
P ( X < 2, Y < 2) = 10 e
2 x 3 y
0 0
2
10 2 x
10 e 10 e 4
3 x
(
1
)
=
dyd x =
e
e
dx
3 0
3 5
2 0
= 0.9695
x
e) E(X) = 10
2 x 3 y
dyd x =
7
10
2 x 3 y
dyd x =
1
5
xe
0 0
f) E(Y) = 10
ye
0 0
2 x 3 y
dy =
g) f ( x ) = 10 e
0
h) f Y \ X =1 ( y ) =
f X ,Y (1, y )
f X (1)
10e 2 z
10
(1 e 3 x ) = (e 2 x e 5 x ) for 0 < x
3
3
10e 2 3 y
10 2
(e e 5 )
3
j) f X |Y = 2 ( x) =
f X ,Y (x,2 )
10e 2 x 6
=
= 2e 2 x + 4 for 2 < x,
10
fY ( 2 )
5e
5-21
c e
0 x
2 x
3 y
c
c
1
dydx = e 2 x (e 3 x )dx = e 5 x dx = c
30
30
15
c = 15
a)
1 2
P ( X < 1, Y < 2) = 15 e
2 x 3 y
0 x
dyd x = 5 e 2 x (e 3 x e 6 )d x
0
5
= 5 e 5 x dx 5e 6 e 2 x dx = 1 e 5 + e 6 (e 2 1) = 0.9879
2
0
0
2
e
1 x
2 x 3 y
c)
3
3 2 x 3 y
2 x 3 y
9 2 x
P (Y > 3) = 15 e
dydx + e
dydx = 5 e e dx + 5 e 5 x dx
3 x
0
3
0 3
3
5
= e 15 + e 9 = 0.000308
2
2
5-19
d)
2 2
5
= 5 e 5 x dx 5e 6 e 2 x dx = 1 e 10 + e 6 e 4 1 = 0.9939
2
0
0
e) E(X) = 15
0 x
2 x 3 y
5 x
xe dyd x = 5 xe dx =
1
= 0.04
52
f)
E (Y ) = 15 ye 2 x 3 y dyd x =
0 x
5
3
5 ye 5 y dy + 3 ye 3 y dy
2 0
20
3 5 8
+ =
10 6 15
g) f ( x) = 15 e 2 x 3 y dy =
x
15 2 z 3 x
(e
) = 5e 5 x for x > 0
3
15e 2 3 y
f Y | X =1 ( y ) =
= 3e 3 3 y for 1 <y
5
5e
i) E (Y | X = 1) =
3 ye 33 y dy = y e 33 y
j)
+ e 33 y dy = 4 / 3
1
k) For y > 0 f X |Y = 2 ( y ) =
5-22
15 6
e
2
15e 2 x 6
= 2e 2 x for 0 < x < 2
15 6
e
2
a) fY|X=x(y), for x = 2, 4, 6, 8
5-20
f(y2)
4
3
2
1
0
0
b) P (Y < 2 | X = 2) =
d) E (Y | X = x) =
c) E (Y | X = 2) =
e) Use fX(x) =
2e
2 y
dy = 0.9817
1
1
f ( x, y )
xy
=
, fY | X ( x, y ) = xe , and the relationship fY | X ( x, y ) = XY
b a 10
f X ( x)
f XY ( x, y )
xe xy
and
=
f XY ( x, y ) =
Therefore, xe
1 / 10
10
xy
10 y
10 y
10 xe
1 10 ye
e
f) fY(y) =
dx =
(using integration by parts)
2
0
10
10 y
xy
5-23.
5-21
1 x +1
4 x +1
0 0
1 x 1
cdydx +
cdydx = 1
= c ( x + 1)dx + 2c dx
= c + 6c = 7.5c = 1
3
2
Therefore, c = 1/7.5=2/15
0.50.5
1
7.5
dydx =
1
30
0 0
0.5x +1
b) P ( X < 0.5) =
1
7.5
0.5
dydx =
( x + 1)dx =
1
7.5
0 0
2
15
( 85 ) = 121
c)
1 x +1
E( X ) =
4 x +1
dydx +
x
7.5
x
7.5
1 x 1
0 0
dydx
2
( x + x)dx +
1
7.5
2
7.5
( x)dx =
12
15
( 56 ) +
2
7.5
(7.5) =
19
9
d)
1 x +1
E (Y ) =
1
7.5
4 x +1
ydydx + ydydx
1
7.5
1 x 1
0 0
1
7.5
( x +1) 2
2
dx + 71.5 ( x +1)
(x
+ 2 x + 1)dx +
1
15
( ) + (30) =
7
3
dx
1
15
( x 1) 2
2
1
15
1
15
4 xdx
1
97
45
e)
x +1
f ( x) =
1
x +1
dy =
for
7.5
7.5
0 < x < 1,
x +1
f ( x) =
1
x + 1 ( x 1) 2
dy =
for 1 < x < 4
=
7.5
7.5
7.5
x 1
f)
f Y | X =1 ( y ) =
f XY (1, y ) 1 / 7.5
=
= 0.5
f X (1)
2 / 7.5
y
y2
g) E (Y | X = 1) = dy =
2
4
0
=1
0
5-22
0.5
0.5
h) P (Y
5-24
= 0.25
0
Let X, Y, and Z denote the time until a problem on line 1, 2, and 3, respectively.
a)
P( X > 40) =
(e )
1
40
e x / 40 dx = e x / 40
40
40
1 3
= e 3 = 0.0498 .
b) P (20 < X < 40,20 < Y < 40,20 < Z < 40) = [ P (20 < X < 40)]3 and
P (20 < X < 40) = e x / 40
40
= e 0.5 e 1 = 0.2387 .
20
c) The joint density is not needed because the process is represented by three independent
exponential distributions. Therefore, the probabilities may be multiplied.
5-25
=3.2 =1/3.2
x
y
3 .2 3 .2
5 5
= e
3 .2
3 .2
= e
x
3 .2
x
y
3.2 3.2
10 10
10
3.2
dydx = 3 .2 e
35.2
e
dx
= 0 . 0439
3.2
dydx = 3.2 e
10
x
3.2
310.2
e
dx
= 0.0019
b) Let X denote the number of orders in a 5-minute interval. Then X is a Poisson random
variable with = 5/3.2 = 1.5625.
e 1.5625 (1.5625) 2
= 0.256
P ( X = 2) =
2!
For both systems, P ( X = 2) P(Y = 2) = 0.256 2 = 0.0655
c) The joint probability distribution is not necessary because the two processes are
independent and we can just multiply the probabilities.
5-26.
5-23
f(x) = 3e-3x
f(y) = 4e-4y
P(X5, Y5)=P(X5)P(Y5)=e-3(5)e-4(5)0
(b) P(X>t, Y>t)=e-3te-4t=e-7t=0.95t=ln(.95)/-7=0.00733
5-27.
a) P ( X < 0.5) =
0.5 1 1
0.5 1
0.5
0.5
0 0 0
0 0
(8 xyz)dzdydx =
2
(4 xy)dydx = (2 x)dx = x
= 0.25
b)
0.5 0.5 1
(8xyz)dzdydx
0 0 0
0.5
0.5 0.5
x2
4
0 0
0.5
= 0.0625
1 1 1
0 0 0
2
2
(8 x yz)dzdydx = (2 x )dx =
1
2 x3
3 0
= 2/3
f) P ( X < 0.5 Y = 0.5) is the integral of the conditional density f X Y (x) . Now,
f X 0.5 ( x) =
f XY ( x,0.5)
f Y (0.5)
and
< 1 and
1 1
(8 xyz)dzdx = 2 y
0 0
Therefore, f X
0.5
( x) =
2x
= 2 x for 0 < x < 1.
1
0.5
2 xdx = 0.25 .
0
g) P ( X < 0.5, Y < 0.5 Z = 0.8) is the integral of the conditional density of X and Y.
Now, f Z ( z ) = 2 z for
f XY Z ( x, y ) =
f XYZ ( x, y, z ) 8 xy(0.8)
=
= 4 xy for 0 < x < 1 and 0 < y < 1.
f Z ( z)
2(0.8)
0.50.5
0.5
(4 xy)dydx = ( x / 2)dx =
0 0
1
16
= 0.0625
Then, f X YZ ( x) =
f XYZ ( x, y, z ) 8 x(0.5)(0.8)
=
= 2 x for 0 < x < 1.
fYZ ( y, z )
4(0.5)(0.8)
5-24
0.5
2 xdx = 0.25
0
5-28
0
x2 + y 2 4
( 22 )4 = 16 . Therefore, c =
1
16
2 and a height of 4 (
8
= 1 / 2.
16
b) P(Z < 2) equals half the volume of the region where f XYZ ( x, y, z ) is positive times
1/c. Therefore, the answer is 0.5.
2
c E( X ) =
4 x 2 4
x
c
2 4 x 2 0
2
2
4 x 2
2
dzdydx = c 4 xy
dx = c (8 x 4 x )dx . Using
2
4 x
2
2
4
c
1
4
4 2
c 3
(4 x 2 ) 2
for x 2 + y 2 < 4 .
4 y 2 4
Also, f Y ( y ) = c
dzdx = 8c
4 y 2 0
Then, f ( x) =
X y
4c
8c 4 y 2
1
2 3
for
3<x< 3.
1
1
2 3
dx =
1+ 3
= 0.7887
2 3
2
2 4 x
2
f XYZ ( x, y,1)
e) f XY 1 ( x, y ) =
and f Z ( z ) = cdydx = 2c 4 x 2 dx
f Z (1)
2 4 x 2
2
Because f Z (z ) is a density over the range 0 < z < 4 that does not depend on Z,
f Z (z ) =1/4 for
0 < z < 4.
Then,
c
1
for x 2 + y 2 < 4 .
=
1 / 4 4
area in x 2 + y 2 < 1
= 1/ 4 .
P ( X 2 + Y 2 < 1 | Z = 1) =
4
Then, f XY 1 ( x, y ) =
5-25
= 0.
f) f Z
xy
f XYZ ( x, y, z )
and from part 5-59 a., f XY ( x, y ) =
f XY ( x, y )
( z) =
Therefore, f Z
5-29
xy
( z) =
1
16
1
4
for x 2 + y 2 < 4 .
1
4
Determine c such that f ( xyz ) = c is a joint density probability over the region x>0, y>0
and z>0 with x+y+z<1
1 1 x 1 x y
1 1 x
0 0
0 0
f ( xyz ) = c
2 1 x
c( y xy y ) dx
2 0
1
(1 x ) 2
dx = c
2
(1 x) 2
= c (1 x) x(1 x)
2
0
1
= c . Therefore, c = 6.
6
1
1
x2 x3
dx = c x
+
2
2
6 0
1 1 x 1 x y
x<0.5, y<0.5,
z<0.5 and x+y+z<1 make a space that is a cube with a volume of 0.125. Therefore the
probability of P( X < 0.5, Y < 0.5, Z < 0.5) = 6(0.125) = 0.75
b)
0 . 50 . 5
6(1 x y ) dydx =
0 0
0 .5
(6 y 6 xy 3 y )
2
0 .5
0
dx
0 .5
0 .5
3
9
9
= 3 x dx = x x 2 = 3 / 4
4
2 0
4
0
c)
0.51 x 1 x y
P ( X < 0.5) = 6
dzdydx =
0 0
0.5
= 6(
0
0.51 1 x
5-26
1 x
y2
6
(
1
x
y
)
dydx
6
(
y
xy
)
0
0
2 0
x2
1
x + )dx = x 3 3x 2 + 3 x
2
2
d)
0.5
0.5
0
= 0.875
1 1 x 1 x y
1 1 x
0 0
E ( X ) = 6
xdzdydx =
0
1 x
0.5
y2
0 6 x(1 x y)dydx = 0 6x( y xy 2 )
0
3x 4
3x 2
x3
x
= 6( x 2 + )dx =
2x3 +
2
2
2
4
0
1
= 0.25
0
e)
1 x 1 x y
y2
dzdy
=
6
(
1
y
)
dy
=
6
y
xy
0
0
2 0
f ( x) = 6
0
= 6(
1 x
1 x
x2
1
x + ) = 3( x 1) 2 for 0 < x < 1
2
2
f)
1 x y
f ( x, y ) = 6
dz = 6(1 x y)
0
f ( x | y = 0.5, z = 0.5) =
f ( x, y = 0.5, z = 0,5) 6
= = 1 for x > 0
f ( y = 0.5, z = 0.5)
6
f X |Y ( x | 0.5) =
5-30
f ( x,0.5) 6(0.5 x)
=
= 4(1 2 x) for x < 0.5
f Y (0.5)
3(0.25)
1400 1500
10000
) = P( Z > 1) = 0.84134 .
a) Let Y denote the number of days out of five such that the yield exceeds 1400. Then, by
independence, Y has a binomial distribution with n = 5 and p = 0.8413. Therefore, the
answer is P (Y = 5) = 55 0.84135 (1 0.8413) 0 = 0.4215 .
b) As in part (a), the answer is
()
P (Y 4) = P(Y = 4) + P(Y = 5)
=
( )0.8413 (1 0.8413)
5
4
+ 0.4215 = 0.8190
5-31
a) Let X denote the weight of a brick. Then,
2.75 3
0.25
) = P( Z > 1) = 0.84134 .
Let Y denote the number of bricks in the sample of 20 that exceed 2.75 pounds. Then, by
independence, Y has a binomial distribution with n = 20 and p = 0.84134. Therefore, the
20
= 0.032 .
answer is P(Y = 20) = 20
20 0.84134
( )
5-27
b) Let A denote the event that the heaviest brick in the sample exceeds 3.75 pounds.
Then, P(A) = 1 - P(A') and A' is the event that all bricks weigh less than 3.75 pounds. As
in part a., P(X < 3.75) = P(Z < 3) and
P ( A) = 1 [ P( Z < 3)] 20 = 1 0.99865 20 = 0.0267 .
5-32
Let Y denote the number of bulbs in the sample of 25 that have less than 1.14 grams.
Then, by independence, Y has a binomial distribution with n = 25 and p = 0.022750.
Therefore, the answer is
P(Y 1) = 1 P(Y = 0) = 025 0.02275 0 (0.97725) 25 = 1 0.5625 = 0.4375 .
b)
( )
2
23
= ( 025 )0.02275 0 (0.97725) 25 + ( 125 )0.022751 (0.97725) 24 + ( 25
2 )0.02275 (0.97725)
4
21
+ ( 325 )0.02275 3 (0.97725) 22 + ( 25
+ ( 525 )0.02275 5 (0.97725) 20
4 )0.02275 (0.97725)
( )0.02275
25
0
(0.97725) 25 = 0.5625
d) The lamps are normally and independently distributed, therefore, the probabilities can
be multiplied.
Section 5-3
5-33
XY =
5-34
XY
0.703125
=
= 0.8851
XY
(0.8594)(0.7344)
5-28
b) Let A denote the event that the heaviest brick in the sample exceeds 3.75 pounds.
Then, P(A) = 1 - P(A') and A' is the event that all bricks weigh less than 3.75 pounds. As
in part a., P(X < 3.75) = P(Z < 3) and
P ( A) = 1 [ P( Z < 3)] 20 = 1 0.99865 20 = 0.0267 .
5-32
Let Y denote the number of bulbs in the sample of 25 that have less than 1.14 grams.
Then, by independence, Y has a binomial distribution with n = 25 and p = 0.022750.
Therefore, the answer is
P(Y 1) = 1 P(Y = 0) = 025 0.02275 0 (0.97725) 25 = 1 0.5625 = 0.4375 .
b)
( )
2
23
= ( 025 )0.02275 0 (0.97725) 25 + ( 125 )0.022751 (0.97725) 24 + ( 25
2 )0.02275 (0.97725)
4
21
+ ( 325 )0.02275 3 (0.97725) 22 + ( 25
+ ( 525 )0.02275 5 (0.97725) 20
4 )0.02275 (0.97725)
( )0.02275
25
0
(0.97725) 25 = 0.5625
d) The lamps are normally and independently distributed, therefore, the probabilities can
be multiplied.
Section 5-3
5-33
XY =
5-34
XY
0.703125
=
= 0.8851
XY
(0.8594)(0.7344)
5-28
0.8438
XY
0.4219 1.6875
=1
5-35
3
c( x + y) = 36c,
c = 1 / 36
x =1 y = 1
13
13
E( X ) =
E (Y ) =
6
6
16
16
E( X 2 ) =
E (Y 2 ) =
3
3
1
36
=
= 0.0435
23 23
36 36
5-36
14
E ( XY ) =
3
xy
V ( X ) = V (Y ) =
1
14 13
= =
3 6
36
23
36
XY = 0.9702 (0.99)(0.98) = 0
XY = =
XY
0.0099 0.0196
=0
5-37
E(X1) = np1 = 3(1/3)=1
E(X2) = np2= 3(1/3)= 1
V(X1) = 3p1(1-p1)=3(1/3)(2/3)=2/3
V(X2) = 3p2(1-p2)=3(1/3)(2/3)=2/3
E(X1X2) =n(n-1)p1p2 =3(2)(1/3)(1/3)=2/3
XY = 2 / 3 12 = 1 / 3 and XY =
1/ 3
(2 / 3)(2 / 3)
= 0.5
5-29
2.267
(4.44)(4.44)
= 0.51
5-38.
Transaction
New Order
Payment
Order Status
Delivery
Stock Level
Mean Value
Frequency
43
44
4
5
4
Selects(X
)
23
4.2
11.4
130
0
18.694
Updates(Y
)
11
3
0
120
0
12.05
Inserts(Z
)
12
1
0
0
0
5.6
5-39
3 x
8
8 x3 2
8 x5
8
E ( XY ) = xy ( xy )dyd x = x 2 y 2 dyd x =
x dx =
dx
81 0 0
81 0 3
81 0 3
81 0 0
xy
6
8 3
= = 4
81 18
12 8
= 4 = 0.16
5 5
E( X 2 ) = 6
E (Y 2 ) = 3
V ( x) = 0.24,
V (Y ) = 0.44
0.16
=
= 0.4924
0.24 0.44
5-40
E( X ) =
2
19 0
xdydx +
0
1 x +1
2
E (Y ) =
19 0
5 x +1
2
xdydx = 2.614
19 1 x1
5 x +1
2
0 ydydx + 19 1 x1ydydx = 2.649
1 x +1
5 x +1
2
2
xydydx + xydydx = 8.7763
19 0 0
19 1 x 1
= 8.7763 (2.614)(2.649) = 1.85181
Now, E ( XY ) =
xy
E ( X 2 ) = 8.7632
E (Y 2 ) = 9.11403
V ( x) = 1.930,
V (Y ) = 2.0968
1.852
1.930 2.097
= 0.9206
5-30
5-41
a) E(X) = 1 E(Y) = 1
E ( XY ) = xye x y dxdy
0 0
= xe dx ye y dy
x
= E ( X ) E (Y )
Therefore, XY = XY = 0 .
5-42
xye
.001 x .002 y
dydx = 388,888.9
0 x
111,115.01
111113.31 361117.11
= 0.5547
E ( X ) = 1(1 / 4) + 1(1 / 4) = 0
E (Y ) = 1(1 / 4) + 1(1 / 4) = 0
E(XY) = [-1 0 (1/4)] + [-1 0 (1/4)] + [1 0 (1/4)] + [0 1 (1/4)] = 0
V(X) = 1/2
V(Y) = 1/2
XY = 0 (0)(0) = 0
XY = =
XY
0
1/ 2 1/ 2
=0
The correlation is zero, but X and Y are not independent, since, for example, if y = 0, X
must be 1 or 1.
5-44
hence XY=0
5-45
Suppose the correlation between X and Y is . for constants a, b, c, and d, what is the
correlation between the random variables U = aX+b and V = cY+d?
Now, E(U) = a E(X) + b and E(V) = c E(Y) + d.
Also, U - E(U) = a[ X - E(X) ] and V - E(V) = c[ Y - E(Y) ]. Then,
5-31
UV =
ac XY
a 2 X2
c 2 Y2
Section 5-4
5-46 a)
0.04
0.03
0.02
z(0)
0.01
10
0.00
-2
0
-1
-10
4
b)
5-32
Then,
UV =
ac XY
a 2 X2
c 2 Y2
Section 5-4
5-46 a)
0.04
0.03
0.02
z(0)
0.01
10
0.00
-2
0
-1
-10
4
b)
5-32
Then,
0.07
0.06
0.05
0.04
z(.8)
0.03
0.02
0.01
10
0.00
-2
-1
-10
4
c)
0.07
0.06
0.05
0.04
z(-.8)
0.03
0.02
0.01
10
0.00
-2
-1
5-47
-10
4
5-33
3.05 3
= 0.7887
0.04
7.80 7.70
= 0.7887
0.08
3.05 3
0.04
7.80 7.70
0.08
) = 0.7887 2 = 0.6220
5-49
0.100465 0.1
0.099535 0.1
P(0.099535 < X < 0.100465) = P
<Z<
0.00031
0.00031
0.23034 0.23
0.22966 0.23
P(0.22966 < X < 0.23034) = P
<Z<
0.00017
0.00017
= P(2 < Z < 2) = P( Z < 2) P( Z < 2) = 0.9545
Probability that a randomly selected lamp is within specification limits is
(0.8664)(0.9594) = 0.8270
5-50 a) By completing the square in the numerator of the exponent of the bivariate normal PDF, the joint
PDF can be written as
1
fY|X =x
2
f XY ( x , y ) 2 x y 1
=
=
f X ( x)
2
1
x x
( x x )) + (1 2 )
( y ( Y +
Y2
X
2
2 (1 )
1
2 x
Also, fx(x) =
1
2 x
xx
x x
By definition,
5-34
1
fY | X = x
2
f ( x, y ) 2 x y 1
= XY
=
f X ( x)
2
1
x x
y
x
+ (1 2 )
(
(
(
))
Y
x
Y2
X
2
2(1 )
2 x
1
2 y 1 2
Y
1
( x x ))
( y ( Y +
X
Y2
2(1 2 )
x x
2
2(1 )
y
x
( x x ) . This answer can be seen from part a). Since the PDF is in the
form of a normal distribution, then the mean can be obtained from the exponent.
c) V(Y|X=x) = 2y (1 2 ) . This answer can be seen from part a). Since the PDF is in the form of a
normal distribution, then the variance can be obtained from the exponent.
5-51
( x X ) 2 ( y Y ) 2
12
+
2
Y 2 dxdy =
X
1
f
x
y
dxdy
=
e
(
,
)
XY
2 X Y
e
2 X
( x X )2
12
2
X
dx
e
2 Y
( y Y ) 2
12
2
Y
dy
and each of the last two integrals is recognized as the integral of a normal probability
density function from to . That is, each integral equals one. Since
f XY( x, y ) = f ( x) f ( y ) then X and Y are independent.
5-52
Let f XY ( x, y ) =
1
2 x y 1
X
X
2( X X )(Y X ) Y Y
+
X Y
2(1 2 )
2 ( X X )(Y X ) Y Y
+
XY
= Y Y
X X
X
2 X
X
+ (1 )
X
But,
Y
Y
Y
X X
= 1
(Y Y ) Y ( X x ) =
Y
5-35
(Y ( Y + Y ( X x ))
f XY ( x, y ) =
1
e
2 x
2 x y 1 2
1 xx
2 x
2
2
1
xx
y ( Y + Y ( x x )) + (1 2 )
2
X
Y
2(1 2 )
dx
2 y 1 2
y ( y + y ( x x ))
x
2 x2 (1 2 )
dydx
dy
The integrand in the second integral above is in the form of a normally distributed random
variable. By definition of the integral over this function, the second integral is equal to 1:
2 x
1 xx
2 x
dx
1
e
2 x
1 xx
2 x
2 y 1 2
y ( y + y ( x x ))
x
2 x2 (1 2 )
dy
dx 1
The remaining integral is also the integral of a normally distributed random variable and therefore,
it also integrates to 1, by definition. Therefore,
f XY ( x, y ) =1
5-53
1
f X ( x) =
2 X Y
1 2
0.5 ( x X )
1
2
1 2
2
X
e
2 X
2
2
0.5 ( x X ) 2 ( x X )( y Y ) + ( y Y )
2
2
2
1
X
Y
X
Y
0.5
( x X
2
X
)2
1
Y
1 2
1
Y
1 2
dy
0.5 ( y Y ) ( x X ) ( x X )
1 2
Y
X
X
0 . 5 ( y Y ) ( x X )
2
1
X
Y
dy
dy
The last integral is recognized as the integral of a normal probability density with mean
Y + Y ( xX X ) and variance Y 2 (1 2 ) . Therefore, the last integral equals one and
the requested result is obtained.
Section 5-5
5-36
5-54
30 30
97
) = P( Z < 0) = 0.5
40 30
97
5-55
(a) E(3X+2Y)= 3*2+2*6=18
(b) V(3X+2Y) = 9*5+4*8 =77
(c) 3X+2Y ~ N( 18, 77)
P(3X+2Y<18) = P(Z<(18-18)/770.5)=0.5
(d) P(3X+2Y<28) = P(Z<(28-18)/770.5)=P(Z<1.1396) =0.873
Section 5-5
5-56
5-57
4.3 4
5-59
a) XN(0.1, 0.00031) and YN(0.23, 0.00017) Let T denote the total thickness.
Then, T = X + Y and E(T) = 0.33 mm,
V(T) = 0.000312 + 0.00017 2 = 1.25 x10 7 mm 2 , and T = 0.000354 mm.
0.2337 0.33
b)
P ( T > 0 . 2405
5-60
0 . 2405 0 . 33
) = P Z >
= P ( Z > 253 ) = 1 P ( Z < 253 ) 1
0 . 000345
Let D denote the width of the casing minus the width of the door. Then, D is normally
distributed.
a) E(D) = 1/8
V(D) = ( 18 ) 2 + ( 161 ) 2 =
5
256
5-37
T = 5 256 = 0.1398
5-54
30 30
97
) = P( Z < 0) = 0.5
40 30
97
5-55
(a) E(3X+2Y)= 3*2+2*6=18
(b) V(3X+2Y) = 9*5+4*8 =77
(c) 3X+2Y ~ N( 18, 77)
P(3X+2Y<18) = P(Z<(18-18)/770.5)=0.5
(d) P(3X+2Y<28) = P(Z<(28-18)/770.5)=P(Z<1.1396) =0.873
Section 5-5
5-56
5-57
4.3 4
5-59
a) XN(0.1, 0.00031) and YN(0.23, 0.00017) Let T denote the total thickness.
Then, T = X + Y and E(T) = 0.33 mm,
V(T) = 0.000312 + 0.00017 2 = 1.25 x10 7 mm 2 , and T = 0.000354 mm.
0.2337 0.33
b)
P ( T > 0 . 2405
5-60
0 . 2405 0 . 33
) = P Z >
= P ( Z > 253 ) = 1 P ( Z < 253 ) 1
0 . 000345
Let D denote the width of the casing minus the width of the door. Then, D is normally
distributed.
a) E(D) = 1/8
V(D) = ( 18 ) 2 + ( 161 ) 2 =
5
256
5-37
T = 5 256 = 0.1398
b) P ( D > 14 ) = P( Z >
c) P ( D < 0) = P ( Z <
1 1
4 8
5
0 18
5
256
256
5-61
5-62
0.5 2
100
= 0.05 .
12 12.1
= P( Z < 2) = 0.023
0.05
12
12
Then 0.05 = -2.58 and = 12.129 .
12 12.1
/ 100
12 12.1
= 0.005.
/ 100
12 12.1
= 0.01.
0.5 / n
1212.1
0.5 / n
1
n
1110
1
n
) = 0.01 ,
11 10
1
rounded up to 6.
c) P ( X > 11) = 0.0005 and X =
Therefore, P( X > 11) = P( Z >
1110
10
) = 0.0005 ,
10
1110
= 3.29
10
= 10 / 3.29 = 0.9612
5-64
X~N(160, 900)
a) Let Y = X1 + X2 + ... + X25, E(Y) = 25E(X) = 4000, V(Y) = 252(900) = 22500
P(Y > 4300) =
4300 4000
P Z >
= P( Z > 2) = 1 P( Z < 2) = 1 0.9773 = 0.0227
22500
5-38
Then
x 4000
150
x 4000
= 0.0001.
22500
5-65
W: weights of parts
E: measurement error.
W~ N(w, w2) , E ~ N(0, e2) ,W+E ~ N(w, w2+e2) .
Wsp = weights of the specification P
(a) P(W > w + 3w) + P(W < w 3w) = P(Z > 3) + P(Z < -3) = 0.0027
(b) P(W+E > w + 3w) + P( W+E < w 3w)
= P (Z > 3w / (w2+e2)1/2) + P (Z < -3w / (w2+e2)1/2)
Because e2 = 0.5w2 the probability is
= P (Z > 3w / (1.5w2)1/2) + P (Z < -3w / (1.5w2)1/2)
= P (Z > 2.45) + P (Z < -2.45) = 2(0.0072) = 0.014
No.
(c) P( E + w + 2w > w + 3w) = P(E > w) = P(Z > w/(0.5w2)1/2)
= P(Z > 1.41) = 0.079
Also, P( E + w + 2w < w 3w) = P( E < 5w)
= P(Z < -5w/(0.5w2)1/2) = P(Z < -7.07) 0
5-66
D=A-B-C
a) E(D) = 10 - 2 - 2 = 6 mm
5-68
f Y ( y) =
1
4
( )p (1 p)
( y ) = ( )(0.25) (0.75)
f Y ( y) = f X ( y ) =
If p = 0.25,
5-69
a)
b)
fY
3 y
and
x=
for y = 0, 1, 4, 9.
3 y
for y = 0, 1, 4, 9.
y 10 1 y 10
fY ( y ) = f X
for 10 y 22
=
72
2 2
22 2
y 10 y
1 y 3 10 y 2 22
E (Y ) =
dy =
2
= 18
10
72
72 3
10
5-70
y = x2
Because y = -2 ln x,
= x . Then, f Y ( y ) = f X (e 2 ) 12 e 2 = 12 e
y
2
for
0 e 2 1 or
y 0 , which is an exponential distribution with =1/2 (which equals a chi-square distribution with k = 2
degrees of freedom).
5-39
Then
x 4000
150
x 4000
= 0.0001.
22500
5-65
W: weights of parts
E: measurement error.
W~ N(w, w2) , E ~ N(0, e2) ,W+E ~ N(w, w2+e2) .
Wsp = weights of the specification P
(a) P(W > w + 3w) + P(W < w 3w) = P(Z > 3) + P(Z < -3) = 0.0027
(b) P(W+E > w + 3w) + P( W+E < w 3w)
= P (Z > 3w / (w2+e2)1/2) + P (Z < -3w / (w2+e2)1/2)
Because e2 = 0.5w2 the probability is
= P (Z > 3w / (1.5w2)1/2) + P (Z < -3w / (1.5w2)1/2)
= P (Z > 2.45) + P (Z < -2.45) = 2(0.0072) = 0.014
No.
(c) P( E + w + 2w > w + 3w) = P(E > w) = P(Z > w/(0.5w2)1/2)
= P(Z > 1.41) = 0.079
Also, P( E + w + 2w < w 3w) = P( E < 5w)
= P(Z < -5w/(0.5w2)1/2) = P(Z < -7.07) 0
5-66
D=A-B-C
a) E(D) = 10 - 2 - 2 = 6 mm
5-68
f Y ( y) =
1
4
( )p (1 p)
( y ) = ( )(0.25) (0.75)
f Y ( y) = f X ( y ) =
If p = 0.25,
5-69
a)
b)
fY
3 y
and
x=
for y = 0, 1, 4, 9.
3 y
for y = 0, 1, 4, 9.
y 10 1 y 10
fY ( y ) = f X
for 10 y 22
=
72
2 2
22 2
y 10 y
1 y 3 10 y 2 22
E (Y ) =
dy =
2
= 18
10
72
72 3
10
5-70
y = x2
Because y = -2 ln x,
= x . Then, f Y ( y ) = f X (e 2 ) 12 e 2 = 12 e
y
2
for
0 e 2 1 or
y 0 , which is an exponential distribution with =1/2 (which equals a chi-square distribution with k = 2
degrees of freedom).
5-39
5-71
y = x 2 , then x =
a) If
f Y ( y ) = f X ( y ) 12 y
12
for
2 y
y > 0.
y = x 1 / 2 , then x = y 2
b) If
0.
c) If y = ln x, then x = e y for
< y < .
x0
for
and
y 0 . Thus, fY ( y ) = f X ( y 2 )2 y = 2 ye y
x 0 . Thus, f Y ( y ) = f X (e y )e y = e y e e = e y e
y
5-72
a) Now,
2 bv
av e dv
b) If
, then
v=
f W ( w) = f V
for
5-73
If
5-74
du = a u 2 e u du.
b b 3 0
b3
a
2a
a
=
.
Therefore,
(
3
)
=
2
b3
b3
w1 / 2 e
ln y 2 .
If y =
That is,
fY ( y ) =
( x 2) 2 , then x = 2 y
fY ( y ) = f X (2 y ) | 12 y
=
2 y
16 y
= ( 14 )y
1 / 2
1
y
for
The sum of
for
1 / 2
0 x2
and x =
2+ y
| + f X (2 + y ) | 12 y
1 / 2
for
2 x 4 . Thus,
2+ y
16 y
for 0 y 4
f ( x, y ) = 1 , ( 1 4 ) + ( 1 8 ) + ( 1 8 ) + ( 1 4 ) + ( 1 4 ) = 1
x
and
1 1
=
y y
e y e2 .
Supplemental Exercises
5-75
From
2w
m
for
( )
2
w 0.
for y >
2w
for v 0, w 0 .
m
b 3 2 w b 2mw
2 w dv
e
=
(2mw) 1 / 2
m
dw
2m
b 3 m 3 / 2
mv 2
w=
2
a ( ) e
u 2
b
f XY ( x, y ) 0
5-40
for
5-71
y = x 2 , then x =
a) If
f Y ( y ) = f X ( y ) 12 y
12
for
2 y
y > 0.
y = x 1 / 2 , then x = y 2
b) If
0.
c) If y = ln x, then x = e y for
< y < .
x0
for
and
y 0 . Thus, fY ( y ) = f X ( y 2 )2 y = 2 ye y
x 0 . Thus, f Y ( y ) = f X (e y )e y = e y e e = e y e
y
5-72
a) Now,
2 bv
av e dv
b) If
, then
v=
f W ( w) = f V
for
5-73
If
5-74
du = a u 2 e u du.
b b 3 0
b3
a
2a
a
=
.
Therefore,
(
3
)
=
2
b3
b3
w1 / 2 e
ln y 2 .
If y =
That is,
fY ( y ) =
( x 2) 2 , then x = 2 y
fY ( y ) = f X (2 y ) | 12 y
=
2 y
16 y
= ( 14 )y
1 / 2
1
y
for
The sum of
for
1 / 2
0 x2
and x =
2+ y
| + f X (2 + y ) | 12 y
1 / 2
for
2 x 4 . Thus,
2+ y
16 y
for 0 y 4
f ( x, y ) = 1 , ( 1 4 ) + ( 1 8 ) + ( 1 8 ) + ( 1 4 ) + ( 1 4 ) = 1
x
and
1 1
=
y y
e y e2 .
Supplemental Exercises
5-75
From
2w
m
for
( )
2
w 0.
for y >
2w
for v 0, w 0 .
m
b 3 2 w b 2mw
2 w dv
e
=
(2mw) 1 / 2
m
dw
2m
b 3 m 3 / 2
mv 2
w=
2
a ( ) e
u 2
b
f XY ( x, y ) 0
5-40
for
XY
g) fY 1 ( y ) =
f XY (1, y )
and fY 1 (0) =
f X (1)
h) E (Y | X = 1) =
yf
x =1
Y | X =1
1/ 8
3/8
= 1 / 3, fY 1 (1) =
1/ 4
3/8
= 2/3.
( y ) =0(1 / 3) + 1(2 / 3) = 2 / 3
i) As is discussed after Example 5-19, because the range of (X, Y) is not rectangular, X
and Y are not independent.
j) E(XY) = 1.25, E(X) = E(Y)= 0.875 V(X) = V(Y) = 0.6094
COV(X,Y)=E(XY)-E(X)E(Y)= 1.25-0.8752=0.4844
0.4844
= 0.7949
0.6094 0.6094
20!
a) P ( X = 2, Y = 4, Z = 14) =
0.10 2 0.20 4 0.7014 = 0.0631
2!4!14!
0
20
b) P ( X = 0) = 0.10 0.90 = 0.1216
c) E ( X ) = np1 = 20(0.10) = 2
V ( X ) = np1 (1 p1 ) = 20(0.10)(0.9) = 1.8
f ( x, z )
d) f X | Z = z ( X | Z = 19) XZ
f Z ( z)
20!
f XZ ( xz ) =
0.1x 0.2 20 x z 0.7 z
x! z!(20 x z )!
20!
f Z ( z) =
0.3 20 z 0.7 z
z! (20 z )!
XY =
5-76
f ( x, z )
(20 z )! 0.1 x 0.2 20 x z
(20 z )! 1 2
=
=
f X | Z = z ( X | Z = 19) XZ
20 z
f Z ( z)
x! (20 x z )! 0.3
x! (20 x z )! 3 3
x
Therefore, X is a binomial random variable with n=20-z and p=1/3. When z=19,
2
1
and f X |19 (1) = .
3
3
2 1 1
e) E ( X | Z = 19) = 0 + 1 =
3 3 3
f X |19 (0) =
5-77
Let X, Y, and Z denote the number of bolts rated high, moderate, and low. Then, X, Y,
and Z have a multinomial distribution.
5-41
20 x z
a) P ( X = 12, Y = 6, Z = 2) =
20!
0.6120.360.12 = 0.0560 .
12!6!2!
f XZ (16, z )
20!
0.6 x 0.3 ( 20 x z ) 0.1 z for
and f XZ ( x, z ) =
x! z! (20 x z )!
f X (16)
x + z 20 and 0 x,0 z . Then,
f Z 16 ( z ) =
20!
16!z!( 4 z )!
20!
16!4!
16
0.6 0.4
( ) ( )
4!
0.3 4 z 0.1 z
z!( 4 z )! 0.4
0.4
Let X, Y, and Z denote the number of calls answered in two rings or less, three or
four rings, and five rings or more, respectively.
10!
0.780.2510.051 = 0.0649
8!1!1!
a) P( X = 8, Y = 1, Z = 1) =
b) Let W denote the number of calls answered in four rings or less. Then, W is a binomial
random variable with n = 10 and p = 0.95.
10
0
Therefore, P(W = 10) = 10
10 0.95 0.05 = 0.5987 .
c) E(W) = 10(0.95) = 9.5.
( )
d) f Z 8 ( z ) =
f XZ (8, z )
10!
0.70 x 0.25(10 x z )0.05 z for
and f XZ ( x, z ) =
x! z!(10 x z )!
f X (8)
10!
8!z!( 2 z )!
10!
8!2!
0.70 0.30
( ) ( )
2!
0.25 2 z 0.05 z
z!( 2 z )! 0.30
0.30
5-79
2
2
cx ydydx = cx
0 0
y2
2
2
0
1 1
a) P ( X < 1, Y < 1) =
0 0
dx = 2c x3
1
18
x ydydx = 181 x 2
2
5-42
y2
2
1
0
dx =
1 x3
36 3
1
0
1
108
2.5 2
b) P ( X < 2.5) =
2.5
1
18
x 2 ydydx =
0 0
y2
2
x2
1
18
0 1
x ydydx = 181 x 2
2
dx =
3 2
1
18
y2
2
1 x3
9 3
dx =
2.5
= 0.5787
1 x3
12 3
3
4
d)
3
1.5
95
432
3 2
e) E ( X ) =
x 3 ydydx =
0 0
x 2 ydydx = 181 x 2
2
1
18
x 3 2dx =
1 x4
9 4
0 0
x 2 y 2 dydx = 181 x 2 83 dx =
1.5
dx =
5 x3
144 3
3
2
9
4
1
18
y2
2
= 0.2199
1
18
3 2
f) E (Y ) =
1
18
4 x3
27 3
4
3
g) f X ( x) =
1
18
h) fY X ( y ) =
1
18
1
9
2
1
18
f XY (1, y )
=
f X (1)
y
for 0 < y < 2.
2
x
f ( x,1)
=
and fY ( y ) =
i) f X 1 ( x) = XY
fY (1)
fY (1)
Therefore, f X 1 ( x) =
5-80
1
18
x 2 ydx =
y
2
x2 1 2
= x for 0 < x < 3.
1/ 2 9
1
18
The region x2 + y 2 1 and 0 < z < 4 is a cylinder of radius 1 ( and base area ) and
height 4. Therefore, the volume of the cylinder is 4 and fXYZ ( x, y, z) =
1
4
for x2 + y 2 1
P ( X 2 + Y 2 0.5) =
4 ( 0.5 )
4
= 1/ 2 .
b) The region X 2 + Y 2 0.5 and 0 < z < 2 is a cylinder of radius 0.5 and height 2.
Therefore,
1 / 4
=
1/ 4
5-43
for x 2 + y 2 1 .
1 x 2
f X ( x) =
d)
1 x
1
4
dydz = 21 1 x 2 dz = 2 1 x 2
4
f XYZ (0,0, z )
2
2
1
and f XY ( x, y ) = 4 dz = 1 / for x + y 1 . Then,
f XY (0,0)
0
1 / 4
f Z 0, 0 ( z ) =
= 1 / 4 for 0 < z < 4 and Z 0, 0 = 2 .
1/
f ( x, y, z ) 1 / 4
=
= 1 / 4 for 0 < z < 4. Then, E(Z) given X = x and Y = y is
f) f Z xy ( z ) = XYZ
f XY ( x, y )
1/
e) f Z 0, 0 ( z ) =
z
4
dz = 2 .
5-81
cdxdy = 1 and c = 1.
0
Because
That is,
5-82
0.5
P ( X Y < 0.5)
= 3/4.
a) Let X1, X 2 ,..., X 6 denote the lifetimes of the six components, respectively. Because of
independence,
P( X1 > 5000, X 2 > 5000,..., X 6 > 5000) = P( X1 > 5000)P( X 2 > 5000)... P( X 6 > 5000)
P ( X > x) =
e t / dt = e t /
0.5
0.5
0.25
0.25
and
5 / 8
0.2
2.325
e e e e e e
=e
= 0.0978 .
b) The probability that at least one component lifetime exceeds 25,000 hours is
the same as 1 minus the probability that none of the component lifetimes exceed
25,000 hours. Thus,
1-P(Xa<25,000, X2<25,000, , X6<25,000)=1-P(X1<25,000)P(X6<25,000)
=1-(1-e-25/8)(1-e-2.5)(1-e-2.5)(1-e-1.25)(1-e-1.25)(1-e-1)=1-.2592=0.7408
5-44
5-83
Let X, Y, and Z denote the number of problems that result in functional, minor, and no
defects, respectively.
!
a) P ( X = 2, Y = 5) = P( X = 2, Y = 5, Z = 3) = 210
0.2 2 0.5 5 0.3 3 = 0.085
!5!3!
b) Z is binomial with n = 10 and p = 0.3.
c) E(Z) = 10(0.3) = 3.
5-84
a) Let X denote the mean weight of the 25 bricks in the sample. Then, E( X ) = 3 and
25
3 ) = P (Z < -1) = 0.159.
X = 0.25
= 0.05 . Then, P( X < 2.95) = P(Z < 2 .095
. 05
b)
5-85
P ( X > x) = P( Z >
x3
x3
) = 0.99 . So,
= -2.33 and x=2.8835.
0.05
.05
a)
18.25
5.25
17.75
4.75
Because
cdydx = 0.25c, c = 4. The area of a panel is XY and P(XY > 90) is the
5.25
4.75
18.25
17.25
18.25
5.25
17.75
90 / x
That is,
18.25
18.25
17.75
17.75
) = 0.499
5.25
17.75
23 x
18.25
18.25
x2
= 4 (17.75 + x)dx = 4(17.75 x +
2
17.75
18.25
) = 0.5
17.75
5-86
a) Let X denote the weight of a piece of candy and XN(0.1, 0.01). Each package has 16
candies, then P is the total weight of the package with 16 pieces and E( P ) = 16(0.1)=1.6
ounces and V(P) = 162(0.012)=0.0256 ounces2
b) P( P < 1.6) = P( Z < 1.06.161.6 ) = P( Z < 0) = 0.5 .
c) Let Y equal the total weight of the package with 17 pieces, E(Y) = 17(0.1)=1.7 ounces
and V(Y) = 172(0.012)=0.0289 ounces2
1.7
P(Y < 1.6) = P( Z < 1.06.0289
) = P( Z < 0.59) = 0.2776 .
5-87
Let X denote the average time to locate 10 parts. Then, E( X ) =45 and X =
a) P ( X > 60) = P ( Z >
60 45
30 / 10
5-45
30
10
b) Let Y denote the total time to locate 10 parts. Then, Y > 600 if and only if X > 60.
Therefore, the answer is the same as part a.
5-88
a) Let Y denote the weight of an assembly. Then, E(Y) = 4 + 5.5 + 10 + 8 = 27.5 and
V(Y)= 0.4 2 + 0.5 2 + 0.2 2 + 0.5 2 = 0.7 .
29.5 27.5
0.7
b) Let X denote the mean weight of 8 independent assemblies. Then, E( X ) = 27.5 and
27.5
) = P( Z > 5.07) = 0 .
V( X ) = 0.7/8 = 0.0875. Also, P ( X > 29) = P( Z > 290.0875
5-89
0.07
0.06
0.05
0.04
z(-.8)
0.03
0.02
0.01
10
0.00
-2
0
-1
-10
4
5-90
1
{( x 1)
1
f XY ( x , y ) =
e 0.72
1.2
f XY ( x , y ) =
f XY ( x , y ) =
1
2 .36
1.6 ( x 1)( y 2 ) + ( y 2 ) 2 }
1
{( x 1) 2 1.6 ( x 1)( y 2 ) + ( y 2 ) 2 }
2 ( 0 .36 )
1
2 1 .8 2
1
{( x 1) 2 2 (. 8 )( x 1)( y 2 ) + ( y 2 ) 2 }
2
2 (1 0 .8 )
E ( X ) = 1 , E (Y ) = 2 V ( X ) = 1 V (Y ) = 1 and = 0.8
5-91 Let T denote the total thickness. Then, T = X1 + X2 and
a) E(T) = 0.5+1=1.5 mm
5-46
b) P (T < 1) = P Z <
1 1.5
= P ( Z < 1.79) = 0.0367
0.078
1.5 3
= P( Z < 6.10) 0
0.246
Let X and Y denote the percentage returns for security one and two respectively.
If of the total dollars is invested in each then X+ Y is the percentage return.
E(X+ Y) = 0.05 (or 5 if given in terms of percent)
V(X+ Y) = 1/4 V(X)+1/4V(Y)+2(1/2)(1/2)Cov(X,Y)
where Cov(XY)=XY=-0.5(2)(4) = -4
V(X+ Y) = 1/4(4)+1/4(6)-2 = 3
Also, E(X) = 5 and V(X) = 4. Therefore, the strategy that splits between the securities
has a lower standard deviation of percentage return than investing 2million in the first
security.
Mind-Expanding Exercises
5-94
By the independence,
P( X 1 A1 , X 2 A2 ,..., X p A p ) =
...
A1
A2
Ap
= f X 1 ( x1 )dx1 f X 2 ( x 2 )dx 2 ... f X p ( x p )dx p
A1
A2
A p
= P( X 1 A1 ) P( X 2 A2 )...P( X p A p )
5-95
E (Y ) = c11 + c2 2 + ... + c p p .
Also,
5-47
6-94.
a) Descriptive Statistics
Variable
N N*
Density
29
0
Variable
Density
Mean
5.4197
Minimum
4.0700
SE Mean
0.0629
Q1 Median
5.2950 5.4600
StDev
0.3389
Variance
0.1148
Q3 Maximum
5.6150
5.8600
Probability Plot of C8
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
5.420
0.3389
29
1.355
<0.005
Percent
80
70
60
50
40
30
20
10
5
4.0
4.5
5.0
5.5
6.0
6.5
C8
c) Due to the very low data point at 4.07, the mean may be lower than it should be. Therefore, the
median would be a better estimate of the density of the earth. The median is not affected by a few
outliers.
9
xi = 559504
i =1
6-95
xi2 = 62572
i =1
n=9
i
9
i =1
559504
2
xi
62572
n
9
s 2 = i =1
=
= 50.61
n 1
9 1
s = 50.61 = 7.11
Subtract 30 and multiply by 10
2
x
i =1
2
i
= 2579200
9
xi = 22848400
i =1
6-58
n=9
xi
9
22848400
xi2 i =1
2579200
n
9
s 2 = i =1
=
= 5061.1
n 1
9 1
s = 5061.1 = 71.14
Yes, the rescaling is by a factor of 10. Therefore, s2 and s would be rescaled by multiplying s2 by 102
(resulting in 5061.1) and s by 10 (71.14). Subtracting 30 from each value has no affect on the variance or
standard deviation. This is because
6-96
i =1
i =1
V (aX + b) = a 2V ( X ) .
( x i a ) 2 = ( xi x ) 2 + n( x a ) 2 ;
minimized when a = x .
6-97
i =1
given that
x.
(x
, the quantity
i =1
This is because
i =1
xi s.
x)
will be smaller
6-98
yi = a + bxi
x=
xi
i =1
(x
i =1
y=
and
sx =
(a + bxi )
i =1
6-99
x = 835.00 F
i =1
= a + bx
x)
and
n 1
sx = sx
(a + bxi a bx ) 2
i =1
n 1
Therefore,
na + b xi
(bxi bx ) 2
i =1
s y = bs x
s x = 10.5
sy =
The results in C:
6-59
n 1
b 2 ( xi x ) 2
i =1
n 1
= b2 sx
y = 32 + 5 / 9 x = 32 + 5 / 9(835.00) = 431.89 C
s y = b 2 s x = (5 / 9) 2 (10.5) 2 = 34.028 C
2
x
and b = 1/s, the mean and standard deviation
s
6-101. Yes, in this case, since no upper bound on the last electronic component is available, use a
measure of central location that is not dependent on this value. That measure is the median.
Sample Median =
x ( 4 ) + x (5)
2
63 + 75
= 69 hours
2
6-60
n +1
6-102
a)
x n +1 =
x n +1
x n +1
b)
xi
i =1
i =1
+ x n +1
=
n +1
n +1
nx n + x n +1
=
n +1
x
n
=
x n + n +1
n +1
n +1
ns n2+1
x
+
x
i
n
+
1
n
i =1
2
2
= xi + x n +1
n +1
i =1
n
n
x
2
x
i
n +1 xi
n
x2
i =1
2
2
i =1
n +1
= xi + x n +1
n +1
n +1
n +1
i =1
n
xi
n
n
n 2
2
2x n +1 x n i =1
x n +1
= xi +
n +1
n +1
n +1
i =1
2
n
(
xi )
n
2
+
x n2+1 2 x n +1 x n
= xi
1
1
n
n
+
+
i =1
n
( xi )2 ( xi )2 ( xi )2
n
2
x n2+1 2 x n x n
= xi +
+
n
n +1
n +1
n
i =1
= x
i =1
( x )
(n + 1)( xi ) n( xi )
2
i
( x )
+
n(n + 1)
= (n 1) s
2
n
n(n + 1)
n
x n2+1 2 x n x n
n +1
nx 2
n
x n2+1 2 x n x n
+
n +1 n +1
n
x 2 n +1 2 x n x n + x n2
= (n 1) s n2 +
n +1
2
n
x n2+ 1 x n
= (n 1) s n2 +
n +1
= (n 1) s n2 +
(
(
6-61
n
x n2+1 2 x n x n
n +1
c)
xn = 65.811 inches
xn+1 = 64
s n = 4.435
n = 37 s n = 2.106
37(65.81) + 64
x n +1 =
= 65.76
37 + 1
37
(37 1)4.435 +
(64 65.811) 2
37
+
1
s n +1 =
37
= 2.098
2
6-103.
6-104.
If nT/100 is not an integer, calculate the two surrounding integer values and interpolate between the two.
For example, if nT/100 = 2/3, one could calculate the mean after trimming 2 and 3 observations from each
end and then interpolate between these two means.
6-62
CHAPTER 6
Section 6-1
6-1.
Sample average:
n
x=
x
i =1
592.035
= 74.0044 mm
8
Sample variance:
8
x
i =1
= 592.035
x
i =1
2
i
= 43813.18031
n
xi
n
2
i =1
(
592.035)
2
x
43813.18031
i
n
8
s 2 = i =1
=
n 1
8 1
0.0001569
=
= 0.000022414 (mm) 2
7
Sample standard deviation:
s = 0.000022414 = 0.00473 mm
The sample standard deviation could also be found using
n
( xi x)
s=
i =1
n 1
where
( xi x)
= 0.0001569
i =1
Dot Diagram:
.. ...:
.
-------+---------+---------+---------+---------+---------diameter
73.9920
74.0000
74.0080
74.0160
74.0240
74.0320
6-1
6-2.
Sample average:
19
x=
x
i =1
19
272.82
= 14.359 min
19
Sample variance:
19
x
i =1
= 272.82
19
x
i =1
2
i
=10333.8964
2
n
xi
n
2
i =1
(
272.82 )
2
x
10333.8964
i
n
19
s 2 = i =1
=
n 1
19 1
6416.49
=
= 356.47 (min) 2
18
Sample standard deviation:
( xi x)
s=
i =1
n 1
where
19
(x
i =1
x ) = 6416.49
2
6-2
6-3.
Sample average:
x=
84817
= 7068.1 yards
12
Sample variance:
12
x
i =1
= 84817
19
x
i =1
=600057949
2
i
n
xi
n
2
i =1
2
(
84817 )
x
600057949
i
n
12
s 2 = i =1
=
n 1
12 1
564324.92
2
=
= 51302.265 ( yards )
11
Sample standard deviation:
s=
(x
i =1
x)
n 1
where
12
(x
i =1
x ) = 564324.92
2
6-3
6-4.
Sample mean:
18
x=
x
i =1
18
2272
= 126.22 kN
18
Sample variance:
18
x
i =1
= 2272
18
x
i =1
= 298392
2
i
n
xi
n
(2272)2
2
xi i =1
298392
n
18
s 2 = i =1
=
n 1
18 1
11615.11
=
= 683.24 (kN ) 2
17
Sample standard deviation:
s = 683.24 = 26.14 kN
The sample standard deviation could also be found using
n
s=
(x
i =1
x)
n 1
where
18
(x
i =1
x ) = 11615.11
2
Dot Diagram:
.
:
:: .
::
.
. : .
.
+---------+---------+---------+---------+---------+-------yield
90
105
120
135
150
165
6-4
6-5.
Sample average:
x=
351.8
= 43.975
8
Sample variance:
8
x
i =1
= 351.8
19
x
i =1
=16528.403
2
i
n
xi
n
2
i =1
(
351.8)
2
x
16528.043
i
n
8
s 2 = i =1
=
8 1
n 1
1057.998
=
= 151.143
7
Sample standard deviation:
s = 151.143 = 12.294
The sample standard deviation could also be found using
n
s=
(x
i =1
x)
n 1
where
8
(x
i =1
x ) = 1057.998
2
Dot Diagram:
.
. ..
.
..
.
+---------+---------+---------+---------+---------+------24.0
32.0
40.0
48.0
56.0
64.0
6-5
6-6.
Sample average:
35
x=
x
i =1
35
28368
= 810.514 watts / m 2
35
Sample variance:
19
x
i =1
= 28368
19
x
i =1
= 23552500
2
i
n
xi
n
(28368)2
2
xi i =1
23552500
n
35
s 2 = i =1
=
n 1
35 1
2 2
= 16465.61 ( watts / m )
559830.743
34
s=
(x
i =1
x)
n 1
where
35
(x
i =1
x ) = 559830.743
2
Dot Diagram (rounding of the data is used to create the dot diagram)
x
. .
.
:
.: ... . . .: :. . .:. .:: :::
-----+---------+---------+---------+---------+---------+-x
500
600
700
800
900
1000
The sample mean is the point at which the data would balance if it were on a scale.
6-7.
6905
= 5.44 ; The value 5.44 is the population mean since the actual physical population
1270
6-8.
Sample average:
6-6
x=
x
i =1
19.56
= 2.173 mm
9
Sample variance:
9
x
i =1
= 19.56
x
i =1
= 45.953
2
i
n
xi
n
2
xi i =1
n
s 2 = i =1
n 1
= 0.4303 (mm) 2
45.953
(19.56)2
9 1
3.443
8
s = 0.4303 = 0.6560 mm
Dot Diagram
.
.
.
. .
. .
..
-------+---------+---------+---------+---------+---------crack length
1.40
1.75
2.10
2.45
2.80
3.15
6-9.
x=
x
i =1
3454.68
= 287.89
12
x
i =1
x
i =1
2
i
= 3454.68
= 1118521.54
2
n
xi
n
i =1
(3454.68) 2
2
x
1118521.54
i
n
12
s 2 = i =1
=
n 1
12 1
123953.71
=
= 11268.52
11
6-7
s = 11268.52 = 106.15
Dot Diagram of exercise group:
D o tp lo t o f 6
1 5 0
2 0 0
2 5 0
6
h o u r s
h o u r s
o f E x e r c is e
3 0 0
o f E x e r c is e
3 5 0
x=
x
i =1
2600.08
= 325.010
8
x
i =1
x
i =1
2
i
= 2600.08
= 947873.4
2
n
xi
n
i =1
(2600.08) 2
2
x
947873.4
i
n
8
=
s 2 = i =1
n 1
8 1
102821.4
=
= 14688.77
7
s = 14688.77 = 121.20
6-8
4 0 0
4 5 0
D o tp lo t o f N o E x e r c is e
1 50
6-10.
20 0
2 50
300
No
35 0
E x e r c is e
4 00
450
500
Dot Diagram of CRT data in exercise 6-5 (Data were rounded for the plot)
Dotplot for Exp 1-Exp 2
Exp 2
Exp 1
10
20
30
40
50
60
70
The data are centered a lot lower in the second experiment. The lower CRT resolution reduces the
visual accommodation.
n
6-11.
x=
x
i =1
57.47
= 7.184
8
2
n
xi
n
2
i =1
(
57.47 )
2
x
412.853
i
0.00299
n
8
2
i =1
=
=
= 0.000427
s =
8 1
7
n 1
s = 0.000427 = 0.02066
Examples: repeatability of the test equipment, time lag between samples, during which the pH of
the solution could change, and operator skill in drawing the sample or using the instrument.
n
6-12.
sample mean x =
x
i =1
748.0
= 83.11 drag counts
9
6-9
n
xi
n
(748.0)2
2
xi i =1
62572
n
9
sample variance s 2 = i =1
=
n 1
9 1
404.89
=
= 50.61 drag counts 2
8
sample standard deviation s = 50.61 = 7.11 drag counts
Dot Diagram
.
. . . . . . .
.
---+---------+---------+---------+---------+---------+---drag counts
75.0
80.0
85.0
90.0
95.0
100.0
6-13.
a)
x = 65.86 F
s = 12.16 F
Dot Diagram
: :
.
.
.
. ..
.: .: . .:..: .. :: .... ..
-+---------+---------+---------+---------+---------+-----temp
30
40
50
60
70
80
b) Removing the smallest observation (31), the sample mean and standard deviation become
x = 66.86 F
s = 10.74 F
Section 6-2
6-14.
NOTE: The data set contains one additional measurement of 91.2 that should be included in the
book. Solutions use this additional measurement.
Stem-and-leaf display of octane rating
Leaf Unit = 0.10
= 82
1
83|4
3
84|33
4
85|3
7
86|777
13
87|456789
24
88|23334556679
34
89|0233678899
(13) 90|0111344456789
36
91|00011122256688
22
92|22236777
14
93|023347
8
94|2247
4
95|
4
96|15
2
97|
2
98|8
1
99|
1 100|3
6-10
n
xi
n
(748.0)2
2
xi i =1
62572
n
9
sample variance s 2 = i =1
=
n 1
9 1
404.89
=
= 50.61 drag counts 2
8
sample standard deviation s = 50.61 = 7.11 drag counts
Dot Diagram
.
. . . . . . .
.
---+---------+---------+---------+---------+---------+---drag counts
75.0
80.0
85.0
90.0
95.0
100.0
6-13.
a)
x = 65.86 F
s = 12.16 F
Dot Diagram
: :
.
.
.
. ..
.: .: . .:..: .. :: .... ..
-+---------+---------+---------+---------+---------+-----temp
30
40
50
60
70
80
b) Removing the smallest observation (31), the sample mean and standard deviation become
x = 66.86 F
s = 10.74 F
Section 6-2
6-14.
NOTE: The data set contains one additional measurement of 91.2 that should be included in the
book. Solutions use this additional measurement.
Stem-and-leaf display of octane rating
Leaf Unit = 0.10
= 82
1
83|4
3
84|33
4
85|3
7
86|777
13
87|456789
24
88|23334556679
34
89|0233678899
(13) 90|0111344456789
36
91|00011122256688
22
92|22236777
14
93|023347
8
94|2247
4
95|
4
96|15
2
97|
2
98|8
1
99|
1 100|3
6-10
0T|3
0F|
0S|7777
0o|88899
1*|000000011111
1T|22222223333
1F|444445555555555
1S|66667777777
1o|888899
2*|011
2T|22
32|1
32|56789
33|114
33|56666688
34|0111223
34|55666667777779
35|001112344
35|56789
36|234
36|6888
37|13
37|689
1|2 represents 12
7o|8
8*|
8T|223333
8F|44444444555555
8S|66666666667777777
8o|88888999999
9*|00000000001111
9T|22233333
9F|444444445555
9S|666677
9o|8
6-11
Sample mean: x =
6-19.
Sample median is at
x
i =1
7514.3
= 90.534
83
Modes are 1102, 1315, and 1750 which are the most frequent data.
n
Sample mean: x =
6-20.
Sample median is at
x
i =1
98259
= 1403.7
70
(64 + 1) = 32.5th
2
The 32nd is 34.7 and the 33rd is 34.7, so the median is 34.7.
Sample mean: x =
x
i =1
2227.1
= 34.798
64
6-21.
Stem-and-leaf of Billion of kilowatt hours
Leaf Unit = 100
(18)
5
3
2
2
1
1
1
1
0
0
0
0
0
1
1
1
1
000000000000000111
23
5
9
Sample mean: x =
x
i =1
4398.8
= 191.0
23
6-12
= 23
6-22.
sample mean:
6-23.
N = 37
61|0
62|00
63|00
64|0000
65|00000000
66|0000
67|00000000
68|00000
69|00
70|0
Stem-and-leaf display for Problem 6-23. Strength: unit = 1.0 1|2 represents 12
1
1
2
4
5
9
20
26
37
46
(13)
41
33
22
13
8
5
532|9
533|
534|2
535|47
536|6
537|5678
538|12345778888
539|016999
540|11166677889
541|123666688
542|0011222357899
543|011112556
544|00012455678
545|2334457899
546|23569
547|357
548|11257
i 0 .5
100 = 95 i = 95.5 95 th percentile is 5479
100
6-24.
2|9
3|1
3|9
4|22223
4|5689
5|01223444
5|5666777899999
6|11244
6|556677789
7|022333
7|6777
8|01
8|9
The data have a symmetrical bell-shaped distribution, and therefore may be normally distributed.
6-13
Sample Mean x =
x
i =1
3592.0
= 59.87
60
x
i =1
60
= 3592 .0
and
x
i =1
2
i
= 224257
xi
n
(3592 .0 )2
x i2 i =1
224257
9215 .93
n
60
s 2 = i =1
=
=
n 1
60 1
59
= 156 .20
and
s = 156 .20 = 12 .50
Sample Median
Variable
concentration
~
x = 59.45
N
60
Median
59.45
i 0 .5
100 = 90 i = 54.5 90 th percentile is 76.85
60
6-14
6-25.
1
5
8
16
20
33
46
(15)
39
31
12
4
1
22 | 6
23 | 2334
23 | 677
24 | 00112444
24 | 5578
25 | 0111122334444
25 | 5555556677899
26 | 000011123334444
26 | 56677888
27 | 0000112222233333444
27 | 66788999
28 | 003
28 | 5
100
Sample Mean x =
x
i =1
x
i =1
100
26030 .2
= 260 .3 yards
100
100
x i = 26030 .2
and
i =1
x
i =1
2
i
= 6793512
xi
n
2
i =1
( (26030 .2 )
2
x
6793512
i
17798 .42
n
100
s 2 = i =1
=
=
n 1
100 1
99
2
= 179 .782 yards
and
s = 179 .782 = 13 .41 yards
Sample Median
Variable
yards
N
100
Median
260.85
i 0 .5
100 = 90 i = 90.5 90 th percentile is 277.2
100
6-15
6-26.
63|47
64|24899
65|223566899
66|0000001233455788899
67|0022455567899
68|00001111233333458
69|0000112345555677889
70|011223444556
71|0057889
72|000012234447
73|59
74|68
75|
76|3
Sample Mean x =
x
i =1
120
82413
= 686.78 mhz
120
120
xi = 82413
and
i =1
x
i =1
2
i
= 56677591
xi
n
2
i =1
(
82413 )
2
x
56677591
i
78402 .925
n
120
s 2 = i =1
=
=
n 1
120 1
119
2
= 658 .85 mhz
and
s = 658 .85 = 25 .67 mhz
Sample Median
Variable
speed
~
x = 683.0 mhz
N
120
Median
683.00
6-16
6-27
Stem-and-leaf display of Problem 6-27. Rating: unit = 0.10 1|2 represents 1.2
1
2
5
7
9
12
18
(7)
15
8
4
3
2
83|0
84|0
85|000
86|00
87|00
88|000
89|000000
90|0000000
91|0000000
92|0000
93|0
94|0
95|00
Sample Mean
40
x =
x
i =1
x
i =1
40
3578
= 89 . 45
40
x
i =1
40
= 3578
and
i =1
2
i
= 320366
xi
n
(3578 )2
x i2 i =1
320366
313 .9
n
40
s 2 = i =1
=
=
n 1
40 1
39
= 8 .05
and
s = 8 .05 = 2 .8
Sample Median
Variable
rating
N
40
Median
90.000
22/40 or 55% of the taste testers considered this particular Pinot Noir truly exceptional.
6-17
6-28.
0|444444
0|5
1|001122233
1|5679
2|
2|5677
3|124
27
Sample mean x =
x
i =1
27
41553
= 1539 gram mole/liter x10 3
27
x
i =1
27
= 41553
and
i =1
2
i
=87792869
xi
n
(41553 )2
x i2 i =1
87792869
23842802
n
27
s 2 = i =1
=
=
n 1
26
27 1
-3
and s = 917030 .85 = 957 .62 gram - mole/liter x 10
~
Sample Median x = 1256
Variable
NbOCl3
6-29.
= 917030 .85
Median
1256
The male engineering students are taller than the female engineering students. Also there is a
slightly wider range in the heights of the male students.
Section 6-3
6-18
6-28.
0|444444
0|5
1|001122233
1|5679
2|
2|5677
3|124
27
Sample mean x =
x
i =1
27
41553
= 1539 gram mole/liter x10 3
27
x
i =1
27
= 41553
and
i =1
2
i
=87792869
xi
n
(41553 )2
x i2 i =1
87792869
23842802
n
27
s 2 = i =1
=
=
n 1
26
27 1
-3
and s = 917030 .85 = 957 .62 gram - mole/liter x 10
~
Sample Median x = 1256
Variable
NbOCl3
6-29.
= 917030 .85
Median
1256
The male engineering students are taller than the female engineering students. Also there is a
slightly wider range in the heights of the male students.
Section 6-3
6-18
6-30.
Solution uses the n = 83 observations from the data set.
Frequency Tabulation for Exercise 6-14.Octane Data
-------------------------------------------------------------------------------Lower
Upper
Relative
Cumulative Cum. Rel.
Class
Limit
Limit
Midpoint
Frequency Frequency Frequency
Frequency
-------------------------------------------------------------------------------at or below
81.75
0
.0000
0
.0000
1
81.75
84.25
83.0
1
.0120
1
.0120
2
84.25
86.75
85.5
6
.0723
7
.0843
3
86.75
89.25
88.0
19
.2289
26
.3133
4
89.25
91.75
90.5
33
.3976
59
.7108
5
91.75
94.25
93.0
18
.2169
77
.9277
6
94.25
96.75
95.5
4
.0482
81
.9759
7
96.75
99.25
98.0
1
.0120
82
.9880
8
99.25
101.75
100.5
1
.0120
83
1.0000
above
101.75
0
.0000
83
1.0000
-------------------------------------------------------------------------------Standard Deviation = 2.888
Median = 90.400
Mean = 90.534
Frequency
30
20
10
0
83.0
85.5
88.0
90.5
93.0
octane data
6-19
95.5
98.0
100.5
6-31.
Frequency Tabulation for Exercise 6-15.Cycles
-------------------------------------------------------------------------------Lower
Upper
Relative
Cumulative Cum. Rel.
Class
Limit
Limit
Midpoint
Frequency Frequency Frequency
Frequency
-------------------------------------------------------------------------------at or below
.000
0
.0000
0
.0000
1
.000
266.667
133.333
0
.0000
0
.0000
2
266.667
533.333
400.000
1
.0143
1
.0143
3
533.333
800.000
666.667
4
.0571
5
.0714
4
800.000 1066.667
933.333
11
.1571
16
.2286
5 1066.667 1333.333 1200.000
17
.2429
33
.4714
6 1333.333 1600.000 1466.667
15
.2143
48
.6857
7 1600.000 1866.667 1733.333
12
.1714
60
.8571
8 1866.667 2133.333 2000.000
8
.1143
68
.9714
9 2133.333 2400.000 2266.667
2
.0286
70
1.0000
above 2400.000
0
.0000
70
1.0000
-------------------------------------------------------------------------------Mean = 1403.66
Standard Deviation = 402.385
Median = 1436.5
Frequency
15
10
0
500
750
1000
1250
1500
1750
6-20
2000
2250
6-32.
Frequency Tabulation for Exercise 6-16.Cotton content
-------------------------------------------------------------------------------Lower
Upper
Relative
Cumulative Cum. Rel.
Class
Limit
Limit
Midpoint
Frequency Frequency Frequency
Frequency
-------------------------------------------------------------------------------at or below
31.0
0
.0000
0
.0000
1
31.0
32.0
31.5
0
.0000
0
.0000
2
32.0
33.0
32.5
6
.0938
6
.0938
3
33.0
34.0
33.5
11
.1719
17
.2656
4
34.0
35.0
34.5
21
.3281
38
.5938
5
35.0
36.0
35.5
14
.2188
52
.8125
6
36.0
37.0
36.5
7
.1094
59
.9219
7
37.0
38.0
37.5
5
.0781
64
1.0000
8
38.0
39.0
38.5
0
.0000
64
1.0000
above
39.0
0
.0000
64
1.0000
-------------------------------------------------------------------------------Mean = 34.798 Standard Deviation = 1.364 Median = 34.700
Frequency
20
10
0
31.5
32.5
33.5
34.5
35.5
36.5
cotton percentage
6-21
37.5
38.5
6-33.
Frequency Tabulation for Exercise 6-17.Yield
-------------------------------------------------------------------------------Lower
Upper
Relative
Cumulative Cum. Rel.
Class
Limit
Limit
Midpoint
Frequency Frequency Frequency
Frequency
-------------------------------------------------------------------------------at or below
77.000
0
.0000
0
.0000
1
77.000
79.400
78.200
1
.0111
1
.0111
2
79.400
81.800
80.600
0
.0000
1
.0111
3
81.800
84.200
83.000
11
.1222
12
.1333
4
84.200
86.600
85.400
18
.2000
30
.3333
5
86.600
89.000
87.800
13
.1444
43
.4778
6
89.000
91.400
90.200
19
.2111
62
.6889
7
91.400
93.800
92.600
9
.1000
71
.7889
8
93.800
96.200
95.000
13
.1444
84
.9333
9
96.200
98.600
97.400
6
.0667
90
1.0000
10
98.600
101.000
99.800
0
.0000
90
1.0000
above 101.000
0
.0000
90
1.0000
-------------------------------------------------------------------------------Mean = 89.3756
Standard Deviation = 4.31591
Median = 89.25
6-22
Histogram 8 bins:
Histogram of Fuel (8 Bins)
40
Frequency
30
20
10
81
84
87
90
93
96
99
102
Fuel
Histogram 16 Bins:
Histogram of Fuel (16 Bins)
18
16
14
12
Frequency
6-34.
10
8
6
4
2
0
84.0
86.4
88.8
91.2
93.6
Fuel
96.0
6-23
98.4
100.8
Frequency
12
10
8
6
4
2
0
500
750
1000
1250
1500
1750
2000
Cycles to failure of aluminum test coupons
2250
Histogram 16 bins:
Histogram of Cycles to failure (16 bins)
14
12
Frequency
10
8
6
4
2
0
200
500
800
1100
1400
1700
2000
Cycles to failure of aluminum test coupons
2300
Frequency
15
10
32.0
32.8
33.6
34.4
35.2
36.0
36.8
37.6
Percentage of cotton in in material used to manufacture men's shirts
6-24
Histogram 16 Bins:
Histogram of Percentage of cotton (8 Bins)
12
Frequency
10
8
6
4
2
0
32.0
32.8
33.6
34.4
35.2
36.0
36.8
37.6
Percentage of cotton in material used to manufacture men's shirts
Frequency
15
10
1000
2000
3000
Energy consumption
6-25
4000
6-38.
Frequency Tabulation for Problem 6-22. Height Data
-------------------------------------------------------------------------------Lower
Upper
Relative
Cumulative Cum. Rel.
Class
Limit
Limit
Midpoint
Frequency Frequency Frequency
Frequency
-------------------------------------------------------------------------------at or below
60.500
0
.0000
0
.0000
1
60.500
61.500
61.000
1
.0270
1
.0270
2
61.500
62.500
62.000
2
.0541
3
.0811
3
62.500
63.500
63.000
2
.0541
5
.1351
4
63.500
64.500
64.000
4
.1081
9
.2432
5
64.500
65.500
65.000
8
.2162
17
.4595
6
65.500
66.500
66.000
4
.1081
21
.5676
7
66.500
67.500
67.000
8
.2162
29
.7838
8
67.500
68.500
68.000
5
.1351
34
.9189
9
68.500
69.500
69.000
2
.0541
36
.9730
10
69.500
70.500
70.000
1
.0270
37
1.0000
above 70.500
0
.0000
37
1.0000
-------------------------------------------------------------------------------Mean = 65.811
Standard Deviation = 2.106
Median = 66.0
8
7
Frequency
6
5
4
3
2
1
0
61
62
63
64
65
66
67
68
69
70
height
The histogram for the spot weld shear strength data shows that the data appear to be normally
distributed (the same shape that appears in the stem-leaf-diagram).
20
Frequency
6-39.
10
0
5320 5340 5360 5380 5400 5420 5440 5460 5480 5500
shear strength
6-26
6-40.
Frequency Tabulation for exercise 6-24. Concentration data
-------------------------------------------------------------------------------Lower
Upper
Relative
Cumulative Cum. Rel.
Class
Limit
Limit
Midpoint
Frequency Frequency Frequency
Frequency
-------------------------------------------------------------------------------at or below
29.000
0
.0000
0
.0000
1
29.0000
37.000
33.000
2
.0333
2
.0333
2
37.0000
45.000
41.000
6
.1000
8
.1333
3
45.0000
53.000
49.000
8
.1333
16
.2667
4
53.0000
61.000
57.000
17
.2833
33
.5500
5
61.0000
69.000
65.000
13
.2167
46
.7667
6
69.0000
77.000
73.000
8
.1333
54
.9000
7
77.0000
85.000
81.000
5
.0833
59
.9833
8
85.0000
93.000
89.000
1
.0167
60
1.0000
above
93.0000
0
.0800
60
1.0000
-------------------------------------------------------------------------------Standard Deviation = 12.50
Median = 59.45
Mean = 59.87
Yes, the histogram shows the same shape as the stem-and-leaf display.
6-41.
Yes, the histogram of the distance data shows the same shape as the stem-and-leaf display in
exercise 6-25.
Frequency
20
10
0
220 228 236 244 252 260 268 276 284 292
distance
6-42.
Histogram for the speed data in exercise 6-26. Yes, the histogram of the speed data shows the
same shape as the stem-and-leaf display in exercise 6-26
6-27
Frequency
20
10
0
620
720
770
speed (megahertz)
Yes, the histogram of the wine rating data shows the same shape as the stem-and-leaf display in
exercise 6-27.
7
6
5
Frequency
6-43.
670
4
3
2
1
0
85
90
rating
6-28
95
6-44.
Pareto Chart for
Automobile Defects
81
64.8
80.2
86.4
91.4
96.3
100
100.0
80
72.8
scores
48.6
60
63.0
percent of total
32.4
40
37.0
16.2
0
20
contour
holes/slots lubrication
pits
trim
assembly
dents
deburr
Roughly 63% of defects are described by parts out of contour and parts under trimmed.
Section 6-4
6-45.
Descriptive Statistics
Variable
time
N
8
Variable
time
Mean
2.415
Minimum
1.750
Median
2.440
Maximum
3.150
TrMean
2.415
Q1
1.912
StDev
0.534
Q3
2.973
Time
2.6
2.4
2.2
2.0
1.8
1.6
6-46.
Descriptive Statistics
Variable
PMC
Variable
PMC
N
20
Min
2.000
Mean
4.000
Max
5.200
Median
4.100
Q1
3.150
a) Sample Mean: 4
Sample Variance: 0.867
Sample Standard Deviation: 0.931
6-29
Tr Mean
4.044
Q3
4.800
StDev
0.931
SE Mean
0.208
SE Mean
0.189
6-44.
Pareto Chart for
Automobile Defects
81
64.8
80.2
86.4
91.4
96.3
100
100.0
80
72.8
scores
48.6
60
63.0
percent of total
32.4
40
37.0
16.2
0
20
contour
holes/slots lubrication
pits
trim
assembly
dents
deburr
Roughly 63% of defects are described by parts out of contour and parts under trimmed.
Section 6-4
6-45.
Descriptive Statistics
Variable
time
N
8
Variable
time
Mean
2.415
Minimum
1.750
Median
2.440
Maximum
3.150
TrMean
2.415
Q1
1.912
StDev
0.534
Q3
2.973
Time
2.6
2.4
2.2
2.0
1.8
1.6
6-46.
Descriptive Statistics
Variable
PMC
Variable
PMC
N
20
Min
2.000
Mean
4.000
Max
5.200
Median
4.100
Q1
3.150
a) Sample Mean: 4
Sample Variance: 0.867
Sample Standard Deviation: 0.931
6-29
Tr Mean
4.044
Q3
4.800
StDev
0.931
SE Mean
0.208
SE Mean
0.189
b)
Boxplot of Percentage
5.5
5.0
Percentage
4.5
4.0
3.5
3.0
2.5
2.0
6-47.
Descriptive Statistics
Variable
Temperat
Variable
Temperat
N
9
Min
948.00
Mean
952.44
Max
957.00
Median
953.00
Q1
949.50
Tr Mean
952.44
Q3
955.00
StDev
3.09
SE Mean
1.03
Temperature
955
954
953
952
951
950
949
948
6-48
Descriptive statistics
Variable
drag coefficients
Variable
drag coefficients
N
9
x = 82.00
a) Median: ~
Upper quartile: Q1 =79.50
Lower Quartile: Q3=84.50
Mean
83.11
Minimum
74.00
Median
82.00
Maximum
100.00
b)
6-30
TrMean
83.11
Q1
79.50
StDev
7.11
Q3
84.50
SE Mean
2.37
95
Drag Coef
90
85
80
75
c) Variable
drag coefficients
N
8
Mean
81.00
Variable
drag coefficients
Median
81.50
Minimum
74.00
TrMean
81.00
Maximum
85.00
StDev
3.46
Q1
79.25
SE Mean
1.22
Q3
83.75
Drag Coef1
82.5
80.0
77.5
75.0
Removing the largest observation (100) lowers the mean and median. Removing this outlier
also greatly reduces the variability as seen by the smaller standard deviation and the smaller
difference between the upper and lower quartiles.
6-49.
N
36
Variable
Temp
Mean
65.86
Minimum
31.00
Median
67.50
Maximum
84.00
TrMean
66.66
Q1
58.50
StDev
12.16
SE Mean
2.03
Q3
75.00
a) Median = 67.50
Lower Quartile: Q1=58.50
Upper Quartile: Q3=75.00
Mean
Median
6-31
TrMean
StDev
SE Mean
Temp
35
Variable
Temp
66.86
Minimum
40.00
68.00
Maximum
84.00
67.35
Q1
60.00
10.74
1.82
Q3
75.00
The mean and median have increased and the standard deviation and difference between the upper
and lower quartile has decreased.
c.)Box Plot - The box plot indicates that there is an outlier in the data.
Boxplot of Joint Temp
90
80
Joint Temp
70
60
50
40
30
6-50.
This plot conveys the same basic information as the stem and leaf plot but in a different format. The outliers that were
separated from the main portion of the stem and leaf plot are shown here separated from the whiskers.
Boxplot of Fuel
100
Fuel
95
90
85
6-51
The box plot shows the same basic information as the stem and leaf plot but in a different format. The outliers that
were separated from the main portion of the stem and leaf plot are shown here separated from the whiskers.
6-32
1400
1200
1000
800
600
400
200
0
6-52
The box plot and the stem-leaf-diagram show that the data are very symmetrical about the mean. It also
shows that there are no outliers in the data.
Boxplot of Concentration
90
Concentration
80
70
60
50
40
30
6-53.This plot, as the stem and leaf one, indicates that the data fall mostly in one region and that the measurements toward
the ends of the range are more rare.
6-33
Weld Strength
5450
5400
5350
6-54
The box plot shows that the data are symmetrical about the mean. It also shows that there is an outlier in
the data. These are the same interpretations seen in the stem-leaf-diagram.
Boxplot of Speed
775
750
Speed
725
700
675
650
6-55
We can see that the two distributions seem to be centered at different values.
Boxplot of Female, Male
76
74
72
Data
70
68
66
64
62
60
Female
6-56
Male
The box plot shows that there is a difference between the two formulations. Formulation 2 has a
higher mean cold start ignition time and a larger variability in the values of the start times. The
6-34
first formulation has a lower mean cold start ignition time and is more consistent. Care should be
taken, though since these box plots for formula 1 and formula 2 are made using 8 and 10 data
points respectively. More data should be collected on each formulation to get a better
determination.
Boxplot of Formula 1, Formula 2
3.6
3.2
Data
2.8
2.4
2.0
Formula 1
All distributions are centered at about the same value, but have different variances.
Boxplot of High Dose, Control, Control_1, Control_2
3000
2500
2000
Data
6.57.
Formula 2
1500
1000
500
0
High Dose
Control
Control_1
6-35
Control_2
Section 6-5
6-58.
Stem-leaf-plot of viscosity N = 40
Leaf Unit = 0.10
2
12
16
16
16
16
16
16
16
17
(4)
19
7
42
43
43
44
44
45
45
46
46
47
47
48
48
89
0000112223
5566
2
5999
000001113334
5666689
The stem-leaf-plot shows that there are two different sets of data. One set of data is centered about 43 and
the second set is centered about 48. The time series plot shows that the data starts out at the higher level and
then drops down to the lower viscosity level at point 24. Each plot gives us a different set of information.
If the specifications on the product viscosity are 48.02, then there is a problem with the process
performance after data point 24. An investigation needs to take place to find out why the location of the
process has dropped from around 48.0 to 43.0. The most recent product is not within specification limits.
Time Series Plot of Viscosity
49
48
Viscosity
47
46
45
44
43
42
4
6-59.
12
16
20
Index
24
1|2 represents 12
17|558
18|357
19|00445589
20|1399
21|00238
22|005
23|5678
24|1555899
25|158
6-36
28
32
36
40
Pull-off Force
240
230
220
210
200
190
180
170
4
12
16
20
Index
24
28
32
36
40
In the time series plot there appears to be a downward trend beginning after time 30. The stem and leaf
plot does not reveal this.
16
16
16
16
16
17
17
17
17
17
18
= 50
1
3
5
67
8899
000011111
2233333
44444444444445555
6667
888
1
Time Series Plot of Chem Concentration
18.0
Chem Concentration
6-60.
17.5
17.0
16.5
16.0
1
10
15
20
25
Index
30
35
40
45
50
In the time series plot there appears to be fluctuating. The stem and leaf plot does not reveal this.
6-37
6-61.
0
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
= 100
01224445677777888
001234456667
0113344488
00145567789
011234567788
04579
0223466778
147
2356
024668
13
8
245
128
4
The data appears to decrease between 1790 and 1835, the stem and leaf plot indicates skewed data.
Wolfer sunspot
120
100
80
60
40
20
0
1
20
30
40
50
Index
60
70
80
90
14
Miles Flown
6-62.
10
12
10
6
1
16
24
32
40
48
Index
6-38
56
64
72
80
100
Each year the miles flown peaks during the summer hours. The number of miles flown increased over the
years 1964 to 1970.
6
7
8
9
10
11
12
13
14
15
16
7
246678889
013334677889
01223466899
022334456667888889
012345566
11222345779
1245678
0179
1
2
When grouped together, the yearly cycles in the data are not seen. The data in the stem-leafdiagram appear to be nearly normally distributed.
6-63.
0
0
1
1
1
1
1
2
2
2
2
2
3
3
3
3
3
4
= 105
67
8888
00011
22333333
4444455555555555
6666666777
888888889999
00001111111
222222223333
44455
66667777
89
01
22
45
66
9
1
Number of Earthquakes
40
35
30
25
20
15
10
5
1990
2000
2010
2020
2030
2040 2050
Year
6-39
2060
2070
2080
2090
5
6
7
8
9
10
11
12
13
= 32
00149
012269
3468
00346889
4
178
458
2
1
3
3
4
4
5
5
6
0
0
1
1
1
1
1
2
2
2
2
= 32
2334
66778
00124
5566779
0014
5688
013
N
= 32
6
89
33
4
66777
89999
000011
2233
4445
67
V ariable
P etroleum Imports
Total P etroleum Imports as P erc
P etroleum Imports from P ersian
12000
10000
8000
Data
6-64.
6000
4000
2000
0
75 78 81 84 87 90 93 96 99 02
19 19 19 19 19 19 19 19 19 20
Year
6-40
Section 6-6
6-65.
ML Estimates
95
Mean
74.0044
StDev
0.0043570
90
Percent
80
70
60
50
40
30
20
10
5
1
73.99
74.00
74.01
74.02
Data
The pattern of the data indicates that the sample may not come from a normally distributed population or that the
largest observation is an outlier. Note the slight bending downward of the sample data at both ends of the graph.
6-66.
ML Estimates
95
Mean
14.3589
StDev
18.3769
90
Percent
80
70
60
50
40
30
20
10
5
1
-40
-20
20
40
60
Data
It appears that the data do not come from a normal distribution. Very few of the data points fall on the line.
6-41
6-67.
ML Estimates
95
Mean
43.975
StDev
11.5000
90
Percent
80
70
60
50
40
30
20
10
5
1
0
10
20
30
40
50
60
70
80
Data
The normal probability plot shown below does not seem reasonable for normality.
Probability Plot of Solar intense
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
80
Percent
6-68.
70
60
50
40
30
20
10
5
400
500
600
700
800
900
Solar intense
6-42
1000
1100
1200
810.5
128.3
35
0.888
0.021
6-69.
ML Estimates
95
Mean
65.8611
StDev
11.9888
90
Percent
80
70
60
50
40
30
20
10
5
1
30
40
50
60
70
80
90
100
Data
The data appear to be normally distributed. Although, there are some departures from the line at the ends
of the distribution.
6-70.
ML Estimates
99
Mean
90.5256
StDev
2.88743
Percent
95
90
80
70
60
50
40
30
20
10
5
1
80
90
100
Data
There are a few points outside the confidence limits, indicating that the sample is not perfectly normal. These
deviations seem to be fairly small though.
6-43
6-71.
Mean
1282.78
StDev
539.634
Percent
95
90
80
70
60
50
40
30
20
10
5
1
1000
2000
3000
Data
The data appear to be normally distributed. Although, there are some departures from the line at the ends
of the distribution.
6-72.
Normal Probability Plot for concentration
Data from exercise 6-24
99
ML Estimates
95
Mean
59.8667
StDev
12.3932
90
Percent
80
70
60
50
40
30
20
10
5
1
25
35
45
55
65
75
85
95
Data
The data appear to be normally distributed. Nearly all of the data points fall very close to, or on the line.
6-44
6-73.
Normal Probability Plot for 6-22...6-29
ML Estimates - 95% CI
99
Female
Students
95
Male
Students
90
Percent
80
70
60
50
40
30
20
10
5
1
60
65
70
75
Data
Both populations seem to be normally distributed, moreover, the lines seem to be roughly parallel
indicating that the populations may have the same variance and differ only in the value of their mean.
6-74.
Yes, it is possible to obtain an estimate of the mean from the 50th percentile value of the normal probability
plot. The fiftieth percentile point is the point at which the sample mean should equal the population mean
and 50% of the data would be above the value and 50% below. An estimate of the standard deviation would
be to subtract the 50th percentile from the 84th percentile These values are based on the values from the ztable that could be used to estimate the standard deviation.
Supplemental Exercises
6-75.
Based on the digidot plot and time series plots of these data, in each year the temperature has a
similar distribution. In each year, the temperature will increase until the mid year and then it starts
to decrease.
6-45
6-73.
Normal Probability Plot for 6-22...6-29
ML Estimates - 95% CI
99
Female
Students
95
Male
Students
90
Percent
80
70
60
50
40
30
20
10
5
1
60
65
70
75
Data
Both populations seem to be normally distributed, moreover, the lines seem to be roughly parallel
indicating that the populations may have the same variance and differ only in the value of their mean.
6-74.
Yes, it is possible to obtain an estimate of the mean from the 50th percentile value of the normal probability
plot. The fiftieth percentile point is the point at which the sample mean should equal the population mean
and 50% of the data would be above the value and 50% below. An estimate of the standard deviation would
be to subtract the 50th percentile from the 84th percentile These values are based on the values from the ztable that could be used to estimate the standard deviation.
Supplemental Exercises
6-75.
Based on the digidot plot and time series plots of these data, in each year the temperature has a
similar distribution. In each year, the temperature will increase until the mid year and then it starts
to decrease.
6-45
2000
2001
2002
2003
2004
12.6
13.2
13.8
14.4
Data
15.0
15.6
16
Data
15
14
13
12
1
6
7
Index
Stem-and-leaf of C7
Leaf Unit = 0.10
3
15
23
25
(7)
28
25
16
12
12
12
13
13
14
14
15
15
16
10
11
12
= 60
244
555666677778
12334444
55
2222234
556
112222444
5899
000000111222
6-46
16.2
C7
15
14
13
12
1
12
18
24
30
Index
36
42
48
54
60
Supplemental Exercises
6-76.
s = 1.367
c) A major source of variability might be variability in the reagent material. Furthermore, if the
same setup is used for all measurements it is not expected to affect the variability. However, if
each measurement uses a different setup, then setup differences could also be a major source of
variability.
A low variance is desirable because it indicates consistency from measurement to measurement. This
implies the measurement error has low variability.
6-77.
The unemployment rate drops steadily from 1997 to 2000 and then increases again quite quickly.
It reaches its peak in the middle of 2003 and then begins to drop steadily again.
6-47
6.0
C13
5.5
5.0
4.5
4.0
a)
x
i =1
26
39
52
65
Index
78
6-78.
13
2
i
6
x i = 62,001
i =1
= 10,433
n=6
i
6
i =1
62,001
2
xi
10,433
n
6 = 19.9 2
=
s 2 = i =1
n 1
6 1
s = 19.9 2 = 4.46
2
b)
x
i =1
2
i
= 353
6
xi = 1521
i =1
6-48
n=6
91
104
117
xi
6
1,521
xi2 i =1
353
n
6 = 19.9 2
s 2 = i =1
=
n 1
6 1
s = 19.9 2 = 4.46
Shifting the data from the sample by a constant amount has no effect on the sample variance or standard
deviation.
2
c)
x
i =1
2
i
6
xi = 6200100
i =1
= 1043300
n=6
xi
6
6200100
xi2 i =1
1043300
n
6
s 2 = i =1
=
= 1990 2
n 1
6 1
s = 1990 2 = 44.61
Yes, the rescaling is by a factor of 10. Therefore, s2 and s would be rescaled by multiplying s2 by 102
(resulting in 19902) and s by 10 (44.6).
6-79
a) Sample 1 Range = 4
Sample 2 Range = 4
Yes, the two appear to exhibit the same variability
b) Sample 1 s = 1.604
Sample 2 s = 1.852
No, sample 2 has a larger standard deviation.
c) The sample range is a relatively crude measure of the sample variability as compared to the sample
standard deviation since the standard deviation uses the information from every data point in the sample
whereas the range uses the information contained in only two data points - the minimum and maximum.
6-80
a.) It appears that the data may shift up and then down over the 80 points.
6-49
17
viscosity
16
15
14
13
Index
10
20
30
40
50
60
70
80
b.)It appears that the mean of the second set of 40 data points may be slightly higher than the first set of 40.
c.) Descriptive Statistics: viscosity 1, viscosity 2
Variable
Viscosity1
Viscosity2
N
40
40
Mean
14.875
14.923
Median
14.900
14.850
TrMean
14.875
14.914
StDev
0.948
1.023
SE Mean
0.150
0.162
There is a slight difference in the mean levels and the standard deviations.
6-81.
From the stem-and-leaf diagram, the distribution looks like the uniform distribution. From the
time series plot, there is an increasing trend in energy consumption.
Stem-and-leaf of Energy
Leaf Unit = 100
4
7
9
10
(2)
9
7
5
2
2
2
2
2
2
3
3
3
3
= 21
0011
233
45
7
89
01
22
445
66
6-50
Energy
3200
3000
2800
2600
2400
2200
2000
2
10
12
Index
14
16
18
20
6-82
16
Viscosity
15
14
13
Both sets of data appear to have the same mean although the first half seem to be concentrated a little more tightly.
Two data points appear as outliers in the first half of the data.
6-83
15
sales
10
0
Index
10
20
30
40
50
6-51
60
70
80
90
There appears to be a cyclic variation in the data with the high value of the cycle generally increasing. The
high values are during the winter holiday months.
b) We might draw another cycle, with the peak similar to the last years data (1969) at about 12.7 thousand
bottles.
6-84.
Descriptive Statistics
Variable
temperat
Variable
temperat
N
24
Min
43.000
Mean
48.125
Max
52.000
Median
49.000
Q1
46.000
Tr Mean
48.182
Q3
50.000
StDev
2.692
SE Mean
0.549
52
51
temperatur
50
49
48
47
46
45
44
43
0T|3
0F|4444555
0S|6666777777
0o|8888999
1*|111
1T|22233
1F|45
1S|77
1o|899
6-52
1|2
represents 12
c)
20
springs
15
10
Index
10
20
30
40
The time series plot indicates there was an increase in the average number of nonconforming springs
made during the 40 days. In particular, the increase occurs during the last 10 days.
0
1
2
3
4
= 20
000
0000000
0000
00000
0
3
errors
6-86.
0
Index
10
15
20
The time series plot indicates a slight decrease in the number of errors for strings 16 - 20.
6-53
6-87.
The golf course yardage data appear to be skewed. Also, there is an outlying data point above 7500 yards.
7500
7400
yardage
7300
7200
7100
7000
6900
6800
6-88
Normal Probability Plot for 6-76 First h...6-76 Second
ML Estimates - 95% CI
99
6-76 First
half
95
6-76 Second
half
90
Percent
80
70
60
50
40
30
20
10
5
1
12
13
14
15
16
17
18
Data
Both sets of data appear to be normally distributed and with roughly the same mean value. The difference in slopes for
the two lines indicates that a change in variance might have occurred. This could have been the result of a change in
processing conditions, the quality of the raw material or some other factor.
6-54
6-89
99
ML Estimates
95
Mean
48.125
StDev
2.63490
90
Percent
80
70
60
50
40
30
20
10
5
1
40
45
50
55
Data
There appears to be no evidence that the data are not normally distributed. There are some repeat points in
the data that cause some points to fall off the line.
6-90
Normal Probability Plot for 6-44...6-56
ML Estimates - 95% CI
6-44
99
6-56
95
90
Percent
80
70
60
50
40
30
20
10
5
1
0.5
1.5
2.5
3.5
4.5
Data
Although we do not have sufficient data points to really see a pattern, there seem to be no significant deviations from
normality for either sample. The large difference in slopes indicates that the variances of the populations are very
different.
6-55
6-91.
285
Distance in yards
275
265
255
245
235
225
The plot indicates that most balls will fall somewhere in the 250-275 range. In general, the population is grouped more
toward the high end of the region. This same type of information could have been obtained from the stem and leaf
graph of problem 6-25.
a)
Normal Probability Plot for No. of Cycles
99
ML Estimates
95
90
80
Percent
6-92.
70
60
50
40
30
20
10
5
1
-3000
-2000
-1000
1000
2000
3000
4000
5000
Data
The data do not appear to be normally distributed. There is a curve in the line.
b)
6-56
Mean
1051.88
StDev
1146.43
99
ML Estimates
95
Mean
2.75410
StDev
0.499916
90
Percent
80
70
60
50
40
30
20
10
5
1
1
Data
After the transformation y*=log(y), the normal probability plot shows no evidence that the data are not normally
distributed.
6-93
1100
1000
900
800
700
600
Trial 1
Trial 2
Trial 3
Trial 4
Trial 5
There is a difference in the variability of the measurements in the trials. Trial 1 has the most variability
in the measurements. Trial 3 has a small amount of variability in the main group of measurements, but
there are four outliers. Trial 5 appears to have the least variability without any outliers.
All of the trials except Trial 1 appear to be centered around 850. Trial 1 has a higher mean value
All five trials appear to have measurements that are greater than the true value of 734.5.
The difference in the measurements in Trial 1 may indicate a start-up effect in the data.
There could be some bias in the measurements that is centering the data above the true value.
6-57
6-94.
a) Descriptive Statistics
Variable
N N*
Density
29
0
Variable
Density
Mean
5.4197
Minimum
4.0700
SE Mean
0.0629
Q1 Median
5.2950 5.4600
StDev
0.3389
Variance
0.1148
Q3 Maximum
5.6150
5.8600
Probability Plot of C8
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
5.420
0.3389
29
1.355
<0.005
Percent
80
70
60
50
40
30
20
10
5
4.0
4.5
5.0
5.5
6.0
6.5
C8
c) Due to the very low data point at 4.07, the mean may be lower than it should be. Therefore, the
median would be a better estimate of the density of the earth. The median is not affected by a few
outliers.
9
xi = 559504
i =1
6-95
xi2 = 62572
i =1
n=9
i
9
i =1
559504
2
xi
62572
n
9
s 2 = i =1
=
= 50.61
n 1
9 1
s = 50.61 = 7.11
Subtract 30 and multiply by 10
2
x
i =1
2
i
= 2579200
9
xi = 22848400
i =1
6-58
n=9
8-95
a.)
,2r
1 , 2 r
= P 2 < < 2
2Tr
2Tr
is
2T
r , 2Tr
2 2
2 , 2 r 1 2 , 2 r
8-96
1 =
z 1
1
2
Therefore, 1 1
x2
is
2(489) 2(489)
,
= (28.62,101.98)
34.17 9.59
z 1
1 x22
dx = 1
e dx
2
= ( z 1 ) .
(1 1 ) + (1 2 ) subject to 1 + 2 = .
Therefore, we need to minimize (1 1 ) + (1 + 1 ) .
To minimize L we need to minimize
1 (1 1 ) = 2 e
1
z2
1 (1 + 1 ) = 2 e
1
z2
z2
z2
e
= e 2 . This is solved
by z 1 = z 1 . Consequently, 1 = 1 , 2 1 = and 1 = 2 = 2 .
Upon setting the sum of the two derivatives equal to zero, we obtain
8.97
a) n=1/2+(1.9/.1)(9.4877/4)
n=46
b) (10-.5)/(9.4877/4)=(1+p)/(1-p)
p=0.6004 between 10.19 and 10.41.
8-98
a)
P ( X i ~ ) = 1 / 2
P (allX ~ ) = (1 / 2) n
i
P (allX i ~ ) = (1 / 2) n
8-39
P( A B) = P( A) + P( B) P ( A B)
n
1 1
1
1
= + = 2 =
2 2
2
2
n 1
1
1 P( A B) = P(min( X i ) < ~ < max( X i )) = 1
2
~
b) P (min( X i ) < < max( X i )) = 1
We would expect that 950 of the confidence intervals would include the value of . This is due to
9the definition of a confidence interval.
Let X bet the number of intervals that contain the true mean (). We can use the large sample
approximation to determine the probability that P(930 < X < 970).
Let
p =
950
930
970
= 0.950 p1 =
= 0.930 and p 2 =
= 0.970
1000
1000
1000
p(1 p) 0.950(0.050)
=
n
1000
(
0
.
970
0
.
950
)
(
0
.
930
0
.
950
)
0.950(0.050)
0.950(0.050)
1000
1000
0
.
02
0
.
02
= P Z <
P Z <
= P ( Z < 2.90) P ( Z < 2.90) = 0.9963
0.006892
0.006892
8-40
CHAPTER 8
Section 8-2
8-1
value of z0 which is 2.14. From Table III, (2.14) = P(Z<2.14) = 0.9838 and the
confidence level is 2(0.9838-0.5) = 96.76%.
b) The confidence level for
by the value of z0 which is 2.14. From Table III, (2.49) = P(Z<2.49) = 0.9936 and the
confidence level is is 2(0.9936-0.5) = 98.72%.
c) The confidence level for
by the value of z0 which is 2.14. From Table III, (1.85) = P(Z<1.85) = 0.9678 and the
confidence level is 93.56%.
8-2
8-3
8-4
a) 95% CI for
x z / n x + z / n
1000 1.96(20 / 10 ) 1000 + 1.96(20 / 10 )
987.6 1012.4
n = 25, = 20 x = 1000, z = 1.96
b) .95% CI for ,
x z / n x + z / n
1000 1.96(20 / 25 ) 1000 + 1.96(20 / 25 )
992.2 1007.8
n = 10, = 20 x = 1000, z = 2.58
c) 99% CI for ,
x z / n x + z / n
1000 2.58(20 / 10 ) 1000 + 2.58(20 / 10 )
983.7 1016.3
n = 25, = 20 x = 1000, z = 2.58
d) 99% CI for ,
x z / n x + z / n
1000 2.58(20 / 25 ) 1000 + 2.58(20 / 25 )
989.7 1010.3
e) When n is larger, the CI will get narrower. The higher the confidence level, the wider the CI.
8-1
8-5.
39.2
n=
= 3.84
20
Therefore, n = 4.
b) Find n for the length of the 99% CI to be 40. Za/2 = 2.58
1/2 length = (2.58)(20) / n = 20
51.6 = 20 n
2
51.6
n=
= 6.66
20
Therefore, n = 7.
8-6
8-8
95% Two-sided CI on the breaking strength of yarn: where x = 98, = 2 , n=9 and z0.025 = 1.96
x z 0 .025 / n x + z 0 .025 / n
98 1 . 96 ( 2 ) / 9 98 + 1 . 96 ( 2 ) /
96 . 7 99 . 3
8-9
95% Two-sided CI on the true mean yield: where x = 90.480, = 3 , n=5 and z0.025 = 1.96
x z 0 .025 / n x + z 0 .025 / n
90 . 480 1 . 96 ( 3 ) / 5 90 . 480 + 1 . 96 ( 3 ) /
87 . 85 93 . 11
8-10
99% Two-sided CI on the diameter cable harness holes: where x =1.5045 , = 0.01 , n=10 and
z0.005 = 2.58
x z 0 .005 / n x + z 0 .005 / n
1 . 5045 2 . 58 ( 0 . 01 ) / 10 1 . 5045 + 2 . 58 ( 0 . 01 ) / 10
1 . 4963 1 . 5127
8-2
8-11
n
n
0.001
0.001
74.036 2.58
74.036 + 2.58
15
15
74.0353 74.0367
b) 99% One-sided CI on the true mean piston ring diameter
For = 0.01, z = z0.01 =2.33 and x = 74.036, = 0.001, n=15
x z 0.01
0.001
74.036 2.33
15
74.0354
The lower bound of the one sided confidence interval is less than the lower bound of the twosided confidence. This is because the Type I probability of 99% one sided confidence interval (or
= 0.01) in the left tail (or in the lower bound) is greater than Type I probability of 99% twosided confidence interval (or /2 = 0.005) in the left tail.
8-12
x z 0.025
x + z 0.025
n
n
25
25
1014 1.96
1014 + 1.96
20
20
1003 1025
b) 95% One-sided CI on the true mean piston ring diameter
For = 0.05, z = z0.05 =1.65 and x = 1014, =25 , n=20
x z0.05
25
1014 1.65
20
1005
The lower bound of the one sided confidence interval is lower than the lower bound of the twosided confidence interval even though the level of significance is the same. This is because all of
the Type I probability (or ) is in the left tail (or in the lower bound).
8-3
8-13
n
n
31.62
31.62
3250 1.96
3250 + 1.96
12
12
3232.11 3267.89
x z0.005
x + z0.005
n
n
31.62
31.62
3250 2.58
3250 + 2.58
12
12
3226.4 3273.6
95% Confident that the error of estimating the true mean life of a 75-watt light bulb is less than 5
hours.
For = 0.05, z/2 = z0.025 = 1.96 , and =25 , E=5
2
z
1.96( 25)
n = a/2 =
= 96.04
5
E
Always round up to the next number, therefore n = 97
8-15
Set the width to 6 hours with = 25, z0.025 = 1.96 solve for n.
1/2 width = (1.96)(25) / n = 3
49 = 3 n
2
49
n = = 266.78
3
Therefore, n = 267.
8-16
99% Confident that the error of estimating the true compressive strength is less than 15 psi
For = 0.01, z/2 = z0.005 = 2.58 , and =31.62 , E=15
2
z
2.58(31.62)
n = a/2 =
= 29.6 30
15
E
Therefore, n=30
8-4
8-17
To decrease the length of the CI by one half, the sample size must be increased by 4 times (22).
z / 2 / n = 0.5l
Now, to decrease by half, divide both sides by 2.
( z / 2 / n ) / 2 = (l / 2) / 2
( z / 2 / 2 n ) = l / 4
( z / 2 / 22 n ) = l / 4
Therefore, the sample size must be increased by 22.
8-18
x + z / 2
n
n
z / 2
z
z
1 z / 2
= /2
= /2
=
2n
1.414 n 1.414 n 1.414 n
If n is doubled in Eq 8-7: x z / 2
z / 2
4n
z / 2
2 n
z / 2
2 n
1 z / 2
2 n
x z 0.005
x + z 0.005
n
n
0.5
0.5
13.77 2.57
13.77 + 2.57
11
11
13.383 14.157
b) 95% lower-confidence bound on the mean temperature
For = 0.05, z = z0.05 =1.65 and x = 13.77, = 0.5, n =11
x z 0.05
0.5
13.77 1.65
11
13.521
c) 95% confidence that the error of estimating the mean temperature for wheat grown is
less than 2 degrees Celsius.
For = 0.05, z/2 = z0.025 = 1.96, and = 0.5, E = 2
2
z
1.96(0.5)
n = a/2 =
= 0.2401
2
E
Always round up to the next number, therefore n = 1.
8-5
d) Set the width to 1.5 degrees Celsius with = 0.5, z0.025 = 1.96 solve for n.
1/2 width = (1.96)(0.5) / n = 0.75
0.98 = 0.75 n
2
0.98
n=
= 1.707
0.75
Therefore, n = 2.
Section 8-3
8-20
t 0.025,15 = 2.131
t 0.05,10 = 1.812
t 0.005, 25 = 2.787
t 0.001,30 = 3.385
a)
t 0.025,12 = 2.179
d)
t 0.0005,15 = 4.073
8-22
a)
t 0.05,14 = 1.761
8-23
8-21
t 0.10, 20 = 1.325
b)
t 0.025, 24 = 2.064
c)
t 0.005,13 = 3.012
b)
t 0.01,19 = 2.539
c)
t 0.001, 24 = 3.467
x t 0.025,15
n
n
3645.94
3645.94
60139.7 + 2.131
60139.7 2.131
16
16
58197.33 62082.07
8-24
8-6
d) Set the width to 1.5 degrees Celsius with = 0.5, z0.025 = 1.96 solve for n.
1/2 width = (1.96)(0.5) / n = 0.75
0.98 = 0.75 n
2
0.98
n=
= 1.707
0.75
Therefore, n = 2.
Section 8-3
8-20
t 0.025,15 = 2.131
t 0.05,10 = 1.812
t 0.005, 25 = 2.787
t 0.001,30 = 3.385
a)
t 0.025,12 = 2.179
d)
t 0.0005,15 = 4.073
8-22
a)
t 0.05,14 = 1.761
8-23
8-21
t 0.10, 20 = 1.325
b)
t 0.025, 24 = 2.064
c)
t 0.005,13 = 3.012
b)
t 0.01,19 = 2.539
c)
t 0.001, 24 = 3.467
x t 0.025,15
n
n
3645.94
3645.94
60139.7 + 2.131
60139.7 2.131
16
16
58197.33 62082.07
8-24
8-6
8-25
x = 1.10 s = 0.015 n = 25
95% CI on the mean volume of syrup dispensed
For = 0.05 and n = 25, t/2,n-1 = t0.025,24 = 2.064
s
s
x + t 0.025, 24
x t 0.025, 24
n
n
0.015
0.015
1.10 + 2.064
1.10 2.064
25
25
1.094 1.106
8-26
x t 0.025, 6
n
n
16
315
315 + 2.447
315 2.447
7
7
300.202 329.798
8-27
n
9. 9
118.3 + 2.650
14
125.312
x + t 0.005,13
8-28
x t 0.05, 4
n
n
1.53
1.53
231.67 2.132
231.67 2.132
5
5
230.2 233.1
8-7
By examining the normal probability plot, it appears that the data are normally distributed. There
does not appear to be enough evidence to reject the hypothesis that the frequencies are normally
distributed.
N orm al P rob a b ility P lot for freq u en c ies
M L E s tim a te s - 9 5 % C I
99
M L E s tim a te s
95
M ean
2 3 1 .6 7
S tD e v
1 .3 6 9 4 4
90
Percent
80
70
60
50
40
30
20
10
5
1
226
231
236
D ata
The data appear to be normally distributed based on examination of the normal probability plot
below. Therefore, there is evidence to support that the annual rainfall is normally distributed.
Probability Plot of Rainfall
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
80
Percent
8-29
70
60
50
40
30
20
10
5
200
300
400
500
Rainfall
600
700
x t 0.025,19
n
n
90.34
90.34
485.8 + 2.093
485.8 2.093
20
20
443.520 528.080
8-8
800
485.9
90.30
20
0.288
0.581
8-30
The data appear to be normally distributed based on examination of the normal probability plot
below. Therefore, there is evidence to support that the solar energy is normally distributed.
Probability Plot of Solar
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
Percent
80
70
60
50
40
30
20
10
5
50
55
60
65
Solar
70
75
x t 0.025,15
n
n
4.225
4.225
65.58 + 2.131
65.58 2.131
16
16
63.329 67.831
8-31
x t 0.005,9
n
n
15.7
15.7
317.2 + 3.250
317.2 3.250
10
10
301.06 333.34
8-9
80
65.58
4.225
16
0.386
0.349
8-32
a) The data appear to be normally distributed based on examination of the normal probability plot
below. Therefore, there is evidence to support that the level of polyunsaturated fatty acid is normally
distributed.
Norm al Probability Plot for 8-25
ML Estimates - 95% CI
99
95
90
Percent
80
70
60
50
40
30
20
10
5
1
16
17
18
Data
s
s
x + t 0.005,5
x t 0.005,5
n
n
0.319
0.319
16.98 4.032
16.98 + 4.032
6
6
16.455 17.505
The 99% confidence for the mean polyunsaturated fat is (16.455, 17.505). There is high
confidence that the true mean is in this interval
a) The data appear to be normally distributed based on examination of the normal probability
plot below.
Normal Probability Plot for Strength
ML Estimates - 95% CI
99
ML Estimates
95
90
80
Percent
8-33
70
60
50
40
30
20
10
5
1
2150
2250
2350
Data
8-10
Mean
2259.92
StDev
34.0550
s
s
x + t 0.025,11
x t 0.025,11
n
n
35.6
35.6
2259.9 2.201
2259.9 + 2.201
12
12
2237.3 2282.5
c) 95% lower-confidence bound on mean strength
s
x t 0.05,11
n
35.6
2259.9 1.796
12
2241.4
a) According to the normal probability plot there does not seem to be a severe deviation from
normality for this data. This is due to the fact that the data appears to fall along a straight line.
N orm al P rob ab ility P lot for 8 -2 7
M L E stima te s - 9 5 % C I
99
95
90
80
Percent
8-34
70
60
50
40
30
20
10
5
1
8 .1 5
8 .2 0
8 .2 5
D ata
s
s
x + t 0.025 ,14
x t 0.025,14
n
n
0.025
0.025
8.23 2.145
8.23 + 2.145
15
15
8.216 8.244
8-11
8 .3 0
x + t 0.025,14
0.025
15
8.23 + 1.761
8.241
a) The data appear to be normally distributed based on examination of the normal probability plot
below. Therefore, there is evidence to support that the speed-up of CNN is normally distributed.
Probability Plot of Speed
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
80
Percent
8-35
70
60
50
40
30
20
10
5
3.0
3.5
4.0
4.5
Speed
5.0
5.5
6.0
x t 0.025,12
n
n
0.4328
0.4328
4.313 + 2.179
4.313 2.179
13
13
4.051 4.575
8-12
4.313
0.4328
13
0.233
0.745
8-36
s
x t 0.05, 24
n
0.08
4.05 1.711
25
4.023
There is high confidence that the true mean wall thickness is greater than 4.023 mm.
99
ML Estimates
95
Mean
2.90167
StDev
0.0951169
90
Percent
80
70
60
50
40
30
20
10
5
1
2.6
2.7
2.8
2.9
3.0
3.1
3.2
Data
8-37
a) The data appear to be normally distributed. There is not strong evidence that the percentage of
enrichment deviates from normality.
b) 99% two-sided confidence interval on mean percentage enrichment
For = 0.01 and n = 12, t/2,n-1 = t0.005,11 = 3.106, x = 2.9017
s
s
x + t 0.005,11
x t 0.005,11
n
n
0.0993
0.0993
2.902 3.106
2.902 + 3.106
12
12
2.813 2.991
8-13
s = 0.0993
Section 8-4
8-38
8-39
02.05,10 = 18.31
02.01,12 = 26.22
02.005 , 25 = 46.93
02.95, 20 = 10.85
14(0.008) 2
2
29.14
0.00003075 2
8-40
n = 10 s = 4.8
2 / 2,n 1 = 02.025 ,9 = 19.02 and 12 / 2, n 1 = 02.975 ,9 = 2.70
9( 4.8) 2
9(4.8) 2
2
19.02
2.70
2
10.90 76.80
3.30 < < 8.76
8-41
50(0.37) 2
50(0.37) 2
2
71.42
32.36
0.096 2 0.2115
Taking the square root of the endpoints of this interval we obtain,
0.31 < < 0.46
8-42
n = 17 s = 0.09
2 / 2,n 1 = 02.025 ,16 = 28.85 and 12 / 2, n 1 = 02.975 ,16 = 6.91
16(0.09) 2
16(0.09) 2
2
28.85
6.91
2
0.0045 0.0188
0.067 < < 0.137
8-14
8-43
The data appear to be normally distributed based on examination of the normal probability plot
below. Therefore, there is evidence to support that the mean temperature is normally distributed.
Probability Plot of Mean Temp
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
21.41
0.9463
8
0.352
0.367
Percent
80
70
60
50
40
30
20
10
5
17
18
19
20
21
22
Mean Temp
23
24
25
n = 8 s = 0.9463
2 / 2, n 1 = 02.025 , 7 = 16.01 and 12 / 2, n 1 = 02.975 , 7 = 1.69
7(0.9463) 2
7(0.9463) 2
2
16.01
1.69
2
0.392 3.709
0.626 < < 1.926
8-44
n = 41 s = 15.99
2 / 2,n 1 = 02.025 , 40 = 59.34 and 12 / 2, n 1 = 02.975 , 40 = 24.43
40(15.99) 2
40(15.99) 2
2
59.34
24.43
2
172.35 418.633
13.13 < < 20.46
The data dont appear to be normally distributed based on examination of the normal probability plot
below. Therefore, there is not enough evidence to support that the time of tumor appearance is normally
distributed. So the 95% confidence interval for is invalid.
8-15
Mean
StDev
N
AD
P-Value
95
90
88.78
15.99
41
1.631
<0.005
Percent
80
70
60
50
40
30
20
10
5
8-45
40
60
80
100
TimeOfTumor
120
140
n = 15 s = 0.00831
2 / 2,n 1 = 02.025 ,14 = 26.12 and 12 ,n 1 = 02.95 ,14 = 6.53
14(0.00831) 2
6.53
2
0.000148
0.0122
The data do not appear to be normally distributed based on an examination of the normal probability plot
below. Therefore, the 95% confidence interval for is not valid.
Probability Plot of Gauge Cap
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
Percent
80
70
60
50
40
30
20
10
5
8-46
3.44
3.45
3.46
3.47
3.48
Gauge Cap
8-16
3.49
3.50
3.472
0.008307
15
0.878
0.018
n = 10
9(1.913) 2
2
21.67
1.5199 2
c) 90% lower confidence bound for 2
2
2
For = 0.1 and n = 10, ,n 1 = 0.1, 9 = 14.68
9(1.913) 2
2
14.68
2.2436 2
1.498
d) The lower confidence bound of the 99% two-sided interval is less than the one-sided interval.
The lower confidence bound for 2 is in part (c) is greater because the confidence is lower.
Section 8-5
8-47
a) 95% Confidence Interval on the fraction defective produced with this tool.
p =
13
= 0.04333 n = 300
z / 2 = 1.96
300
p (1 p )
p (1 p )
p p + z / 2
p z / 2
n
n
0.04333(0.95667 )
0.04333(0.95667 )
p 0.04333 + 1.96
300
300
0.02029 p 0.06637
b) 95% upper confidence bound z = z 0.05 = 1.65
0.04333 1.96
p p + z / 2
p (1 p )
n
p 0.04333 + 1.650
0.04333(0.95667)
300
p 0.06273
8-17
n = 10
9(1.913) 2
2
21.67
1.5199 2
c) 90% lower confidence bound for 2
2
2
For = 0.1 and n = 10, ,n 1 = 0.1, 9 = 14.68
9(1.913) 2
2
14.68
2.2436 2
1.498
d) The lower confidence bound of the 99% two-sided interval is less than the one-sided interval.
The lower confidence bound for 2 is in part (c) is greater because the confidence is lower.
Section 8-5
8-47
a) 95% Confidence Interval on the fraction defective produced with this tool.
p =
13
= 0.04333 n = 300
z / 2 = 1.96
300
p (1 p )
p (1 p )
p p + z / 2
p z / 2
n
n
0.04333(0.95667 )
0.04333(0.95667 )
p 0.04333 + 1.96
300
300
0.02029 p 0.06637
b) 95% upper confidence bound z = z 0.05 = 1.65
0.04333 1.96
p p + z / 2
p (1 p )
n
p 0.04333 + 1.650
0.04333(0.95667)
300
p 0.06273
8-17
8-48
a) 95% Confidence Interval on the proportion of such tears that will heal.
n = 37 z / 2 = 1.96
p = 0.676
p (1 p )
p p + z / 2
n
p z / 2
0.676 1.96
p (1 p )
n
0.676(0.324)
0.676(0.324)
p 0.676 + 1.96
37
37
0.5245 p 0.827
b) 95% lower confidence bound on the proportion of such tears that will heal.
p z
0.676 1.64
8-49
p (1 p )
p
n
0.676(0.33)
p
37
0.549 p
a) 95% confidence interval for the proportion of college graduates in Ohio that voted for
George Bush.
p =
412
= 0.536
768
n = 768 z / 2 = 1.96
p z / 2
p (1 p )
p p + z / 2
n
0.536 1.96
p (1 p )
n
0.536(0.464)
0.536(0.464)
p 0.536 + 1.96
768
768
0.501 p 0.571
b) 95% lower confidence bound on the proportion of college graduates in Ohio that voted for
George Bush.
p z
0.536 1.64
8-50
p (1 p )
p
n
0.536(0.464)
p
768
0.506 p
p =
823
= 0.823
1000
n = 1000 z / 2 = 1.96
8-18
p z / 2
0.823 1.96
p (1 p )
p p + z / 2
n
p (1 p )
n
0.823(0.177 )
0.823(0.177 )
p 0.823 + 1.96
1000
1000
0.7993 p 0.8467
b) E = 0.03, = 0.05, z/2 = z0.025 = 1.96 and p = 0.823 as the initial estimate of p,
2
z
1.96
n = / 2 p (1 p ) =
0.823(1 0.823) = 621.79 ,
0.03
E
n 622.
c) E = 0.03, = 0.05, z/2 = z0.025 = 1.96 at least 95% confident
2
z
1.96
n = / 2 (0.25) =
(0.25) = 1067 .11 ,
0.03
E
n 1068.
8-51
12
= 0.4 n = 30 z / 2 = 1.96
30
p (1 p )
p z / 2
p p + z / 2
n
p =
0.4 1.96
p (1 p )
n
0.4(0.6)
0 .4 ( 0 .6 )
p 0.4 + 1.96
30
30
0.225 p 0.575
b) E = 0.02, = 0.05, z/2 = z0.025 = 1.96 and p = 0.4as the initial estimate of p,
2
z
1.96
n = / 2 p (1 p ) =
0.4(1 0.4) = 2304.96 ,
0.02
E
n 2305.
c) E = 0.02, = 0.05, z/2 = z0.025 = 1.96 at least 95% confident
2
z
1.96
n = / 2 (0.25) =
(0.25) = 2401 .
0.02
E
8-52
p =
18
= 0.36
50
n = 50
8-19
z / 2 = 1.96
p z / 2
0.36 1.96
p (1 p )
p p + z / 2
n
p (1 p )
n
0.36(0.64)
0.36(0.64)
p 0.36 + 1.96
50
50
0.227 p 0.493
z
1.96
b) n = / 2 p (1 p ) =
0.36(1 0.36) = 2212.76
0.02
E
n 2213
2
z
1.96
c) n = / 2 p (1 p ) =
0.5(1 0.5) = 2401
0.02
E
8-53
The worst case would be for p = 0.5, thus with E = 0.05 and = 0.01, z/2 = z0.005 = 2.58 we
obtain a sample size of:
2
z
2.58
n = / 2 p(1 p) =
0.5(1 0.5) = 665.64 , n 666
E
0.05
8-54
z
2.58
n = / 2 p(1 p) =
0.5(1 0.5) = 5758.13 , n 5759
E
0.017
Section 8-7
8-55
x t 0.025,15 s 1 +
1
1
x n +1 x + t 0.025,15 s 1 +
n
n
1
1
x n +1 60139.7 + 2.131(3645.94) 1 +
16
16
52131.1 x n +1 68148.3
60139.7 2.131(3645.94) 1 +
The prediction interval is considerably wider than the 95% confidence interval (58,197.3
62,082.07). This is expected because the prediction interval needs to include the variability in the
parameter estimates as well as the variability in a future observation.
8-56
8-20
p z / 2
0.36 1.96
p (1 p )
p p + z / 2
n
p (1 p )
n
0.36(0.64)
0.36(0.64)
p 0.36 + 1.96
50
50
0.227 p 0.493
z
1.96
b) n = / 2 p (1 p ) =
0.36(1 0.36) = 2212.76
0.02
E
n 2213
2
z
1.96
c) n = / 2 p (1 p ) =
0.5(1 0.5) = 2401
0.02
E
8-53
The worst case would be for p = 0.5, thus with E = 0.05 and = 0.01, z/2 = z0.005 = 2.58 we
obtain a sample size of:
2
z
2.58
n = / 2 p(1 p) =
0.5(1 0.5) = 665.64 , n 666
E
0.05
8-54
z
2.58
n = / 2 p(1 p) =
0.5(1 0.5) = 5758.13 , n 5759
E
0.017
Section 8-7
8-55
x t 0.025,15 s 1 +
1
1
x n +1 x + t 0.025,15 s 1 +
n
n
1
1
x n +1 60139.7 + 2.131(3645.94) 1 +
16
16
52131.1 x n +1 68148.3
60139.7 2.131(3645.94) 1 +
The prediction interval is considerably wider than the 95% confidence interval (58,197.3
62,082.07). This is expected because the prediction interval needs to include the variability in the
parameter estimates as well as the variability in a future observation.
8-56
8-20
1
1
x n +1 x + t 0.005,19 s 1 +
n
n
1
1
x n +1 1.25 + 2.861(0.25) 1 +
20
20
0.517 x n +1 1.983
1.25 2.861(0.25) 1 +
The lower bound of the 99% prediction interval is considerably lower than the 99% confidence
interval (1.108 ). This is expected because the prediction interval needs to include the
variability in the parameter estimates as well as the variability in a future observation.
8-57
95% Prediction Interval on the volume of syrup of the next beverage dispensed
x = 1.10 s = 0.015 n = 25 t/2,n-1 = t0.025,24 = 2.064
x t 0.025, 24 s 1 +
1
1
x n +1 x + t 0.025, 24 s 1 +
n
n
1
1
x n +1 1.10 2.064(0.015) 1 +
25
25
1.068 x n +1 1.13
1.10 2.064(0.015) 1 +
1.106
90% prediction interval the value of the natural frequency of the next beam of this type that will
be tested. given x = 231.67, s =1.53 For = 0.10 and n = 5, t/2,n-1 = t0.05,4 = 2.132
x t 0.05, 4 s 1 +
1
1
x n +1 x + t 0.05, 4 s 1 +
n
n
1
1
x n +1 231.67 2.132(1.53) 1 +
5
5
228.1 x n +1 235.2
231.67 2.132(1.53) 1 +
95% Prediction Interval on the volume of syrup of the next beverage dispensed
n = 20 x = 485.8 s = 90.34 t/2,n-1 = t0.025,19 = 2.093
x t 0.025,19 s 1 +
1
1
x n +1 x + t 0.025,19 s 1 +
n
n
1
1
x n +1 485.8 2.093(90.34) 1 +
20
20
292.049 x n +1 679.551
485.8 2.093(90.34) 1 +
8-60
The 95% prediction interval is wider than the 95% confidence interval.
99% prediction interval on the polyunsaturated fat
8-21
1
1
x n +1 x + t 0.005,5 s 1 +
n
n
1
1
x n +1 16.98 + 4.032(0.319) 1 +
6
6
15.59 x n +1 18.37
16.98 4.032(0.319) 1 +
The length of the prediction interval is much longer than the width of the confidence interval
16.455 17.505 .
8-61
x t 0.005,9 s 1 +
1
1
x n +1 x + t 0.005,9 s 1 +
n
n
1
1
x n +1 317.2 3.250(15.7) 1 +
10
10
263.7 x n +1 370.7
317.2 3.250(15.7) 1 +
1
1
x n +1 x + t 0.025,14 s 1 +
n
n
1
1
x n +1 8.23 2.145(0.025) 1 +
15
15
8.17 x n +1 8.29
8.23 2.145(0.025) 1 +
x t 0.05,11 s 1 +
1
1
x n +1 x + t 0.05,11 s 1 +
n
n
1
1
x n +1 2260 + 1.796(35.57) 1 +
12
12
2193.5 x n +1 2326.5
2260 1.796(35.57) 1 +
8-64
8-22
x t 0.05, 24 s 1 +
1
1
x n +1 x + t 0.05, 24 s 1 +
n
n
1
1
x n +1 4.05 1.711(0.08) 1 +
25
25
3.91 x n +1 4.19
4.05 1.711(0.08) 1 +
8-65
90% prediction interval for enrichment data given x = 2.9 s = 0.099 n = 12 for = 0.10
and n = 12, t/2,n-1 = t0.05,11 = 1.796
x t 0.05,12 s 1 +
1
1
x n +1 x + t 0.05,12 s 1 +
n
n
1
1
x n +1 2.9 + 1.796(0.099) 1 +
12
12
2.71 x n +1 3.09
2.9 1.796(0.099) 1 +
x t 0.05,12 s
2.9 1.796(0.099)
1
1
x + t 0.05,12 s
n
n
1
1
2.9 1.796(0.099)
12
12
2.85 2.95
The prediction interval is wider than the CI on the population mean with the same confidence.
The 99% confidence interval is
x t 0.005,12 s
2.9 3.106(0.099)
1
1
x + t 0.005,12 s
n
n
1
1
2.9 + 3.106(0.099)
12
12
2.81 2.99
The prediction interval is even wider than the CI on the population mean with greater confidence.
8-66
x t 0.05, 24 s 1 +
1
x n +1
n
1
x n +1
25
3.91 x n +1
4.05 1.711(0.08) 1 +
The prediction interval bound is much lower than the confidence interval bound of
8-23
4.023 mm
Section 8-7
8-67
95% tolerance interval on the life of the tires that has a 95% CL
given x = 60139.7 s = 3645.94 n = 16 we find k=2.903
x ks, x + ks
99% tolerance interval on the Izod impact strength PVC pipe that has a 90% CL
given x=1.25, s=0.25 and n=20 we find k=3.368
x ks, x + ks
95% tolerance interval on the syrup volume that has 90% confidence level
x = 1.10 s = 0.015 n = 25 and k=2.474
x ks, x + ks
99% tolerance interval on the polyunsaturated fatty acid in this type of margarine that has a
confidence level of 95% x = 16.98 s = 0.319 n=6 and k = 5.775
x ks, x + ks
95% tolerance interval on the rainfall that has a confidence level of 95%
4.023 mm
Section 8-7
8-67
95% tolerance interval on the life of the tires that has a 95% CL
given x = 60139.7 s = 3645.94 n = 16 we find k=2.903
x ks, x + ks
99% tolerance interval on the Izod impact strength PVC pipe that has a 90% CL
given x=1.25, s=0.25 and n=20 we find k=3.368
x ks, x + ks
95% tolerance interval on the syrup volume that has 90% confidence level
x = 1.10 s = 0.015 n = 25 and k=2.474
x ks, x + ks
99% tolerance interval on the polyunsaturated fatty acid in this type of margarine that has a
confidence level of 95% x = 16.98 s = 0.319 n=6 and k = 5.775
x ks, x + ks
95% tolerance interval on the rainfall that has a confidence level of 95%
The 95% tolerance interval is much wider than the 95% confidence interval on the population
mean ( 443.52 528.08 ).
8-72
level
95% tolerance interval on the diameter of the rods in exercise 8-27 that has a 90% confidence
x = 8.23 s = 0.0.25 n=15 and k=2.713
x ks, x + ks
99% tolerance interval on the brightness of television tubes that has a 95% CL
given x = 317.2 s = 15.7 n = 10 we find k=4.433
x ks, x + ks
8-74
90% tolerance interval on the comprehensive strength of concrete that has a 90% CL
given x = 2260 s = 35.57 n = 12 we find k=2.404
x ks, x + ks
x ks, x + ks
8-25
8-76
x ks, x + ks
x ks
4.05 1.702(0.08)
3.91
The lower tolerance bound is of interest if we want to make sure the wall thickness is at least a
certain value so that the bottle will not break.
8-26
Supplemental Exercises
8-77
1 + 2 = . Let = 0.05
Interval for 1 = 2 = / 2 = 0.025
Where
value of z0 which is 1.96. From Table III, we find (1.96) = P(Z<1.96) = 0.975 and the
confidence level is 95%.
Interval for 1 = 0.01, 2 = 0.04
The confidence interval is x 2.33
same since = 0.05 . The symmetric interval does not affect the level of significance; however,
it does affect the length. The symmetric interval is shorter in length.
8-78
= 50
unknown
a) n = 16 x = 52 s = 1.5
52 50
to =
8 / 16
=1
The P-value for t0 = 1, degrees of freedom = 15, is between 0.1 and 0.25. Thus we would
conclude that the results are not very unusual.
b) n = 30
to =
52 50
= 1.37
8 / 30
The P-value for t0 = 1.37, degrees of freedom = 29, is between 0.05 and 0.1. Thus we conclude
that the results are somewhat unusual.
c) n = 100 (with n > 30, the standard normal table can be used for this problem)
zo =
52 50
= 2.5
8 / 100
The P-value for z0 = 2.5, is 0.00621. Thus we conclude that the results are very unusual.
d) For constant values of x and s, increasing only the sample size, we see that the standard error
of X decreases and consequently a sample mean value of 52 when the true mean is 50 is more
unusual for the larger sample sizes.
8-79
= 50, 2 = 5
a) For n = 16 find P( s 7.44) or P( s 2.56)
2
15(7.44)
2
P( S 2 7.44) = P 152
= 0.05 P 15 22.32 0.10
2
5
2
Using Minitab P ( S 7.44) =0.0997
15(2.56)
2
P( S 2 2.56) = P 152
= 0.05 P 15 7.68 0.10
5
2
Using Minitab P ( S 2.56) =0.064
8-27
7.44) or P ( S 2 2.56)
2 29(7.44)
2
P( S 2 7.44) = P 29
2
Using Minitab P ( S 7.44) = 0.044
2 29(2.56)
2
P( S 2 2.56) = P 29
2
Using Minitab P ( S 2.56) = 0.014.
b) For n = 30 find P ( S
70(7.44)
2
P( S 2 7.44) = P 702
= 0.005 P 70 104.16 0.01
5
2
Using Minitab P ( S 7.44) =0.0051
70(2.56)
2
P( S 2 2.56) = P 702
= P( 70 35.84) 0.005
5
2
Using Minitab P ( S 2.56) < 0.001
d) The probabilities get smaller as n increases. As n increases, the sample variance should
approach the population variance; therefore, the likelihood of obtaining a sample variance much
larger than the population variance will decrease.
e) The probabilities get smaller as n increases. As n increases, the sample variance should
approach the population variance; therefore, the likelihood of obtaining a sample variance much
smaller than the population variance will decrease.
8-80
a) The data appear to follow a normal distribution based on the normal probability plot since the
data fall along a straight line.
b) It is important to check for normality of the distribution underlying the sample data since the
confidence intervals to be constructed should have the assumption of normality for the results to
be reliable (especially since the sample size is less than 30 and the central limit theorem does not
apply).
c) No, with 95% confidence, we can not infer that the true mean could be 14.05 since this value is
not contained within the given 95% confidence interval.
d) As with part b, to construct a confidence interval on the variance, the normality assumption
must hold for the results to be reliable.
e) Yes, it is reasonable to infer that the variance could be 0.35 since the 95% confidence interval
on the variance contains this value.
f) i) & ii) No, doctors and children would represent two completely different populations not
represented by the population of Canadian Olympic hockey players. Because neither doctors nor
children were the target of this study or part of the sample taken, the results should not be
extended to these groups.
8-81
a) The probability plot shows that the data appear to be normally distributed. Therefore, there is
no evidence conclude that the comprehensive strength data are normally distributed.
8-28
G
x = 25.12, s = 8.42, n = 9
t 0.01,8 = 2.896
s
x t 0.01,8
n
8.42
25.12 2.896
9
16.99
The lower bound on the 99% confidence interval shows that the mean comprehensive strength is
most likely be greater than 16.99 Megapascals.
c) 98% two-sided confidence interval on the mean
G
x = 25.12, s = 8.42, n = 9
t 0.01,8 = 2.896
s
s
x t 0.01,8
x t 0.01,8
n
n
8.42
8.42
25.12 2.896
25.12 2.896
9
9
16.99 33.25
The bounds on the 98% two-sided confidence interval shows that the mean comprehensive
strength will most likely be greater than 16.99 Megapascals and less than 33.25 Megapascals.
The lower bound of the 99% one sided CI is the same as the lower bound of the 98% two-sided CI
(this is because of the value of )
d) 99% one-sided upper bound on the confidence interval on 2 comprehensive strength
s = 8.42, s 2 = 70.90
02.99 ,8 = 1.65
8(8.42) 2
1.65
2
343.74
The upper bound on the 99% confidence interval on the variance shows that the variance of the
comprehensive strength is most likely less than 343.74 Megapascals2.
e) 98% two-sided confidence interval on 2 of comprehensive strength
8-29
G
x = 23, s = 6.31, n = 9
t 0.01,8 = 2.896
s
s
x t 0.01,8
x + t 0.01,8
n
n
6.31
6.31
23 2.896
23 + 2.896
9
9
16.91 29.09
98% two-sided confidence interval on 2 comprehensive strength
s
s
x t 0.01,8
x t 0.01,8
n
n
8.41
8.41
25 2.896
25 2.896
9
9
16.88 33.12
98% two-sided confidence interval on 2 of comprehensive strength
8-30
a) n
1.96
z
= 0.025 82 =
64 = 39.34 = 40
2.5
2.5
z
2 1.96
b) n = 0.025 6 =
36 = 22.13 = 23
2.5
2.5
As the standard deviation decreases, with all other values held constant, the sample size necessary
to maintain the acceptable level of confidence and the length of the interval, decreases.
8-83
x ks, x + ks
15.33 2.564(0.62 ), 15.33 + 2.564(0.62 )
(13.74,
16.92)
x ks, x + ks
95% prediction interval for the next sample of concrete that will be tested.
given x = 25.12 s = 8.42 n = 9 for = 0.05 and n = 9, t/2,n-1 = t0.025,8 = 2.306
x t 0.025,8 s 1 +
1
1
x n +1 x + t 0.025,8 s 1 +
n
n
1
1
x n +1 25.12 + 2.306(8.42) 1 +
9
9
4.65 x n +1 45.59
25.12 2.306(8.42) 1 +
a) There is no evidence to reject the assumption that the data are normally distributed.
Normal Probability Plot for foam height
ML Estimates - 95% CI
99
ML Estimates
95
90
80
Percent
8-85
70
60
50
40
30
20
10
5
1
178
188
198
208
Data
8-31
218
228
Mean
203.2
StDev
7.11056
G
x = 203.20, s = 7.5, n = 10 t 0.025,9 = 2.262
s
s
x t 0.025,9
x t 0.025,9
n
n
7.50
7.50
203.2 2.262
203.2 + 2.262
10
10
197.84 208.56
c) 95% prediction interval on a future sample
1
1
x t 0.025,9 s 1 +
n
n
x t 0.025,9 s 1 +
1
1
203.2 + 2.262(7.50) 1 +
10
10
185.41 220.99
d) 95% tolerance interval on foam height with 99% confidence k = 4.265
x ks, x + ks
203.2 2.262(7.50) 1 +
e) The 95% CI on the population mean is the narrowest interval. For the CI, 95% of such intervals
contain the population mean. For the prediction interval, 95% of such intervals will cover a future
data value. This interval is quite a bit wider than the CI on the mean. The tolerance interval is the
widest interval of all. For the tolerance interval, 99% of such intervals will include 95% of the true
distribution of foam height.
8-86
a) Normal probability plot for the coefficient of restitution.
There is no evidence to reject the assumption that the data are normally distributed.
99
95
90
Percent
80
70
60
50
40
30
20
10
5
1
0.59
0.60
0.61
0.62
0.63
0.64
0.65
0.66
Data
8-32
x t 0.005,39
0.624 2.7079
c)
n
0.013
x + t 0.005,39
s
n
0.624 + 2.7079
40
0.618 0.630
0.013
40
99% prediction interval on the coefficient of restitution for the next baseball that will be
tested.
x t 0.005,39 s 1 +
1
1
x n +1 x + t 0.005,39 s 1 +
n
n
1
1
x n +1 0.624 + 2.7079(0.013) 1 +
40
40
0.588 x n +1 0.660
0.624 2.7079(0.013) 1 +
d) 99% tolerance interval on the coefficient of restitution with a 95% level of confidence
( x ks, x + ks)
95% Confidence Interval on the proportion of baseballs with a coefficient of restitution that
exceeds 0.635.
8
= 0.2
n = 40
40
p (1 p )
p z
p
n
p =
0.2 1.65
8-88
z = 1.65
0.2(0.8)
p
40
0.0956 p
a) The normal probability shows that the data are mostly follow the straight line, however, there
are some points that deviate from the line near the middle. It is probably safe to assume that the
data are normal.
8-33
99
95
90
Percent
80
70
60
50
40
30
20
10
5
1
0
10
Data
x t 0.025,19
3.265 2.093
c)
n
2.127
x + t 0.025,19
s
n
3.265 + 2.093
20
2.270 4.260
2.127
20
95% prediction interval on the oxygen concentration for the next stream in the system that
will be tested..
x t 0.025,19 s 1 +
1
1
x n +1 x + t 0.025,19 s 1 +
n
n
1
1
x n +1 3.265 + 2.093(2.127) 1 +
20
20
1.297 x n +1 7.827
3.265 2.093(2.127) 1 +
d) 95% tolerance interval on the values of the dissolved oxygen concentration with a 99% level
of confidence
( x ks, x + ks)
8-34
a) There is no evidence to support that the data are not normally distributed. The data points
appear to fall along the normal probability line.
99
ML Estimates
95
Mean
1.529
StDev
0.0556117
90
80
Percent
8-89
70
60
50
40
30
20
10
5
1
1.4
1.5
1.6
1.7
Data
x t 0.005, 29
1.529 2.756
c)
n
0.0566
x + t 0.005, 29
s
n
1.529 + 2.756
30
1.501 1.557
0.0566
30
99% prediction interval on the tar content for the next sample that will be tested..
x t 0.005,19 s 1 +
1
1
x n +1 x + t 0.005,19 s 1 +
n
n
1
1
x n +1 1.529 + 2.756(0.0566) 1 +
30
30
1.370 x n +1 1.688
1.529 2.756(0.0566) 1 +
d) 99% tolerance interval on the values of the tar content with a 95% level of confidence
8-35
( x ks, x + ks)
(1.529 3.350(0.0566), 1.529 + 3.350(0.0566))
(1.339, 1.719)
e) The confidence interval in part (b) is for the population mean and we may interpret this to
imply that 95% of such intervals will cover the true population mean. For the prediction
interval, 95% of such intervals will cover a future observed tar content. For the tolerance
interval, 99% of such intervals will cover 95% of the true distribution
8-90
p z a / 2
0.0067 1.96
p (1 p )
p p + z a / 2
n
p (1 p )
n
0.0067(1 0.0067)
0.0067(1 0.0067)
p 0.0067 + 1.96
1200
1200
0.0021 p 0.0113
b) No, there is not sufficient evidence to support the claim that the fraction of defective units
produced is one percent or less at = 0.05. This is because the upper limit of the control limit is
greater than 0.01.
8-91
p z a / 2
0.005 2.58
p (1 p )
p p + z a / 2
n
p (1 p )
n
0.005(1 0.005)
0.005(1 0.005)
p 0.005 + 2.58
1600
1600
0.0004505 p 0.009549
z
2.58
n = / 2 p (1 p ) =
0.005(1 0.005) = 517.43 , n 518
0.008
E
c) E = 0.008, = 0.01, z/2 = z0.005 = 2.58
2
z
2.58
n = / 2 p (1 p ) =
0.5(1 0.5) = 26001 .56 , n 26002
0.008
E
d) A bound on the true population proportion reduces the required sample size by a substantial
amount. A sample size of 518 is much more reasonable than a sample size of over 26,000.
8-92
117
= 0.242
484
a) 90% confidence interval; z / 2 = 1.645
p =
8-36
p z / 2
p (1 p )
p p + z / 2
n
0.210 p 0.274
p (1 p )
n
With 90% confidence, the true proportion of new engineering graduates who were planning to
continue studying for an advanced degree is between 0.210 and 0.274.
b) 95% confidence interval; z / 2 = 1.96
p z / 2
p (1 p )
p p + z / 2
n
0.204 p 0.280
p (1 p )
n
With 95% confidence, we believe the true proportion of new engineering graduates who were
planning to continue studying for an advanced degree lies between 0.204 and 0.280.
c) Comparison of parts (a) and (b):
The 95% confidence interval is larger than the 90% confidence interval. Higher confidence
always yields larger intervals, all other values held constant.
d) Yes, since both intervals contain the value 0.25, thus there in not enough evidence to determine
that the true proportion is not actually 0.25.
8-93
a) The data appear to follow a normal distribution based on the normal probability plot
since the data fall along a straight line.
b) It is important to check for normality of the distribution underlying the sample data
since the confidence intervals to be constructed should have the assumption of
normality for the results to be reliable (especially since the sample size is less than 30
and the central limit theorem does not apply).
c) 95% confidence interval for the mean
s
s
x t 0.025,10
x + t 0.025,10
n
n
6.33
6.33
22.73 2.228
22.73 + 2.228
11
11
18.478 26.982
d) As with part b, to construct a confidence interval on the variance, the normality
assumption must hold for the results to be reliable.
e) 95% confidence interval for variance
n = 11 s = 6.33
2 / 2 ,n 1 = 02.025 ,10 = 20.48 and 12 / 2 , n 1 = 02.975 ,10 = 3.25
8-37
10(6.33) 2
10(6.33) 2
2
20.48
3.25
2
19.565 123.289
a) The data appear to be normally distributed based on examination of the normal probability plot
below. Therefore, there is evidence to support that the energy intake is normally distributed.
Probability Plot of Energy
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
80
Percent
8-94
70
60
50
40
30
20
10
5
6
Energy
s
s
x t 0.005,9
x + t 0.005,9
n
n
0.5645
0.5645
5.884 3.250
5.884 + 3.250
10
10
5.304 6.464
8-38
5.884
0.5645
10
0.928
0.011
8) The P-value is greater than any acceptable significance level . Therefore, we do not reject the null
hypothesis. There is insufficient evidence to support the claim that the standard deviation is less than
20 microamps.
b) 7) n = 51, s = 20
20 =
50(15.7) 2
= 30.81
400
8) The P-value is less than 0.05, therefore we reject the null hypothesis and conclude that the standard
deviation is significantly less than 20 microamps.
c) Increasing the sample size increases the test statistic 20 and therefore decreases the P-value, providing
more evidence against the null hypothesis.
9-126
9-127
X 0
X 0
X 0 X 0
<
= | = 0 ) + P(
>
| = 0 ) P
/ n
/ n
/ n / n
P( z0 < z ) + P( z0 > z ) = ( z ) + 1 ( z )
P(
= (( )) + (1 (1 )) =
b) = P(z
z when
1 = 0 + d )
or = P( z < Z 0 < z | 1 = 0 + )
= P( z <
x0
< z | 1 = 0 + )
2 /n
= P( z
< Z < z
)
2
/n
2 /n
= ( z
) ( z
)
2
/n
2 /n
9-128
/n
9-65
z0 =
2.076 2
2 / 500
= 1.202
8) Because 1.202 < 1.65, do not reject the null hypothesis and conclude there is insufficient evidence to
indicate that the true mean number of open circuits is greater than 2 at = 0.01
9-129 a) 1) The parameter of interest is the true standard deviation of the golf ball distance .
2) H0: = 10
3) H1: < 10
4) =0.05
5) Because n > 30 we can use the normal distribution
z0 =
S 0
02 /(2n)
13.41 10
10 2 /(200)
= 4.82
8) Since 4.82 > -1.65, do not reject the null hypothesis and conclude there is insufficient evidence to indicate
that the true standard deviation is less than 10 at = 0.05
b) 95% percentile: =
+ 1.645
95% percentile estimator: = X + 1.645S
SE () 2 / n + 1.645 2 2 /(2n)
The statistic S can be used as an estimator for in the standard error formula.
c) 1) The parameter of interest is the true 95th percentile of the golf ball distance .
2) H0: = 285
3) H1: < 285
4) = 0.05
5) Since n > 30 we can use the normal distribution
z0 =
0
SE ()
282.36 285
13.412 / 100 + 1.645 213.412 / 200
= 1.283
8) Because -1.283 > -1.65, do not reject the null hypothesis. There is not sufficient evidence to indicate that
the true is less than 285 at = 0.05
9-130
9-66
02 =
2 X i 0
i =1
2 X i
i =1
6) Reject H0 if
>
2
0
2
a / 2, 2 n
or
02 < 12 a / 2, 2 n
7) compute
2 X i
i =1
02 =
2 X i 0
i =1
2 X i
i =1
8) make conclusions
alternative hypotheses
1) H0 : = 0
H1 : > 0
Reject H0 if
2) H0 : = 0
H1 : < 0
Reject H0 if
02 > a2, 2 n
02 < a2, 2 n
9-67
CHAPTER 9
Section 9-1
9-1
H 0 : = 25, H1 : 25
a)
interest, inequality is in the alternative hypothesis, and the value in the null and alternative
hypotheses matches.
H 0 : > 10, H1 : = 10
H 0 : x = 50, H1 : x 50
b)
c)
H 0 : p = 0.1, H1 : p = 0.3
d)
No, the values in the null and alternative hypotheses do not match and
H 0 : s = 30, H1 : s > 30
e)
No, because the hypothesis is stated in terms of the statistic rather than the
parameter.
9-2
The conclusion does not provide strong evidence that the critical dimension mean equals 100nm.
The conclusion is that we dont have enough evidence to reject the null hypothesis.
9-3
9-4
9-5
X 11.5 12
= P(Z 2)
0.5 / 4
/ n
= 0.02275.
The probability of rejecting the null hypothesis when it is true is 0.02275.
b)
X 11.5 11.25
= P(Z > 1.0)
>
0.5 / 4
/ n
= P
X 11.5 11.5
>
0.5 / 4
/ n
9-1
9-6
X 11.5 12
= P(Z 4) = 0.
/ n 0.5 / 16
a) = P( X 11.5 | = 12) = P
The probability of rejecting the null, when the null is true, is approximately 0 with a sample size
of 16.
X 11.5 11.25
>
0.5 / 16
/ n
X 11.5 11.5
>
0.5 / 16
/ n
X 12 z 0.5 / n
a)
a) = P( X >11.59|=11.5)=P(Z>0.36)=1-0.6406=0.3594
b) = P( X >11.79|=11.5)=P(Z>2.32)=1-0.9898=0.0102
c) Notice that the value of decreases as n increases
9-9
9-10
a)
11.25 12
b)
11.0 12
c)
11.75 12
a) = P(
=
>
2 / 9 2 / 9
2 / 9
2/ 9
9-2
= P
= P(6.75 Z 2.25)
= P(Z 2.25) P(Z 6.75)
= 0.01222 0 = 0.01222
c) = P(98.5
= P(9.75 Z 5.25)
= P(Z 5.25) P(Z 9.75)
=00
= 0.
The probability of accepting the null hypothesis when it is actually false is smaller in part c since the true
mean, = 105, is further from the acceptance region. A larger difference exists.
9-11
Use n = 5, everything else held constant (from the values in exercise 9-6):
a) P( X 98.5) + P( X >101.5)
= P
= P(5.03 Z 1.68)
= P(Z 1.68) P(Z 5.03)
= 0.04648 0
= 0.04648
c) = P(98.5 x 101.5 when = 105)
= P
= P(7.27 Z 3.91)
= P(Z 3.91) P(Z 7.27)
= 0.00005 0
= 0.00005
It is smaller, because it is not likely to accept the product when the true mean is as high as 105.
9-12
X 0 + z / 2
, where =2
n
n
0 z / 2
98.29,101.71
a)
b)
9-3
9-13
c)
97.70,102.30
d)
98.25,101.75
=103-100=3
>0 then = z / 2
n
, where
=2
a) = P(98.69< X <101.31|=103)=P(-6.47<Z<-2.54)=0.0055
b) = P(98.25< X <101.75|=103)=P(-5.31<Z<-1.40)=0.0808
a) As n increases, decreases
9-14
9-15
a) p-value=2(1- ( Z 0 ) )=2(1- (
98 100
) )=2(1- (3) )=2(1-.99865)=0.0027
2/ 9
b) p-value=2(1- ( Z 0 ) )=2(1- (
101 100
) )=2(1- (1.5) )=2(1-.93319)=0.13362
2/ 9
c) p-value=2(1- ( Z 0 ) )=2(1- (
102 100
) )=2(1- (3) )=2(1-.99865)=0.0027
2/ 9
>
20 / 10
20 / 10
= P
20 / 10 20 / 10
= P
= P(Z 0)
= 0.5
c) = P( X 185 when = 195)
20 / 10
20 / 10
= P
= P(Z 1.58)
= 0.05705
9-4
9-16
Using n = 16:
a) = P( X > 185 when = 175)
>
20 / 16
20 / 16
= P
= P(Z > 2)
= 1 P(Z 2) = 1 0.97725 = 0.02275
b) = P( X 185 when = 185)
20 / 16 20 / 16
= P
= P(Z 0) = 0.5.
c) = P( X 185 when = 195)
20 / 16
20 / 16
= P
= P(Z 2) = 0.02275
9-17
20
X 175 + Z
n
a) =0.01, n=10, then z =2.32 and critical value is 189.67
b) =0.05, n=10, then z =1.64 and critical value is 185.93
c) =0.01, n=16, then z = 2.32 and critical value is 186.6
d) =0.05, n=16, then z =1.64 and critical value is 183.2
9-18
a) =0.05, n=10, then the critical value 185.93 (from 9-17 part (b))
= P( X 185.37 when = 185)
X 185
= P
20 / 10
185.93 185
20 / 10
= P(Z 0.147)
= 0.5584
b) =0.05, n=16, then the critical value 183.2 (from 9-17(d)), then
= P( X 183.2 when = 185)
20 / 16
20 / 16
= P
c)
= P(Z -0.36)
= 0.3594
as n increases, decreases
9-5
9-19
9-20
X 0
/ n
180 175
a) X =180 then Z 0 =
= 0.79
20 / 10
p-value=1- (0.79) = 1 0.7852 = 0.2148
190 175
b) X =190 then Z 0 =
= 2.37
20 / 10
p-value=1- (2.37) = 1 0.991106 = 0.008894
170 175
c) X =170 then Z 0 =
= 0.79
20 / 10
p-value=1- (0.79) = 1 0.214764 = 0.785236
p-value=1- ( Z 0 ) ) where Z 0 =
a) = P(
= P(2.83 Z 0.566)
= P(Z 0.566) P(Z 2.83)
= 0.71566 0.00233
= 0.71333
1 = 0.2867
9-21
Using n = 16:
a) = P( X 4.85 | = 5) + P( X > 5.15 | = 5)
X 5
4.85 5 X 5
5.15 5
+ P
>
0.25 / 16 0.25 / 16 0.25 / 16 0.25 / 16
= P
9-6
4.85 5.1
= P
0.25 / 16
X 5.1
5.15 5.1
0.25 / 16 0.25 / 16
= 0.25, 0 = 5
a) =0.01, n=8 then
a= 0 + z / 2 /
b= 0 z / 2 /
n =5-2.57*.25/ 8 =4.77
b= 0 z / 2 /
n =5-2.57*.25/ 16 =4.8393
9-23
a= 0 + z / 2 /
b= 0 z / 2 /
n =5-1.96*.25/ 16 =4.8775
p-value=2(1 - ( Z 0 ) ) where
x 0
/ n
5.2 5
= 2.26
.25 / 8
p-value=2(1- ( 2.26)) = 2(1 0.988089) = 0.0238
a) x =5.2 then z 0
z0 =
x =4.7 then z 0 =
4.7 5
= 3.39
.25 / 8
p-value=2(1- (3.39)) = 2(1 0.99965) = 0.0007
5.1 5
c) x =5.1 then z 0 =
= 1.1313
.25 / 8
p-value=2(1- (1.1313)) = 2(1 0.870762) = 0.2585
b)
9-7
9-24
a) = P(4.845< X <5.155|=5.05)=P(-2.59<Z<1.33)=0.9034
b) = P(4.8775< X <5.1225|=5.05)=P(-2.76<Z<1.16)=0.8741
c) As the n increases, decreases
9-25
Accept Region:
Reject Region:
p2 = 8/15 = 0.533
0.267 p 0.533
p < 0.267 or p > 0.533
0 .267 0 .4
0 .533 0 .4
+P Z >
=P Z <
0 .4 ( 0 .6 )
0 .4 ( 0 .6 )
15
15
b) When p = 0.2,
0.267 0.2
0.533 0.2
= P(0.267 p 0.533) = P
Z
0.2(0.8)
0.2(0.8)
15
15
= P(0.65 Z 3.22)
= P( Z 3.22) P( Z 0.65)
= 0.99936 0.74215
= 0.2572
9-26
0.1
Reject region: p
Use the normal approximation for parts a), b) and c):
a) When p =0.3 =
0.1 0.3
P( p < 0.1) = P Z
0.3(0.7)
10
9-8
= P ( Z 1.38)
= 0.08379
0.1 0.2
> 0.1) = P Z >
b) When p = 0.2 = P ( p
0.2(0.8)
10
= P ( Z > 0.79)
= 1 P ( Z < 0.79)
= 0.78524
9-27
The problem statement implies H0: p = 0.6, H1: p > 0.6 and defines an acceptance region as
400
= 0.80
500
p > 0.80
0.80 0.60
0.6(0.4)
500
= P(Z>9.13)=1-P(Z 9.13) 0
b) = P( p 0.8 when p=0.75) = P(Z 2.58)=0.99506
9-28
x
185
= P Z
= P Z
20 / 10
20 / 10
185
P Z
=
20 / 10
178
181
184
187
190
193
196
199
P(Z 1.11) =
P(Z 0.63) =
P(Z 0.16) =
P(Z 0.32) =
P(Z 0.79) =
P(Z 1.26) =
P(Z 1.74) =
P(Z 2.21) =
0.8665
0.7357
0.5636
0.3745
0.2148
0.1038
0.0409
0.0136
0.1335
0.2643
0.4364
0.6255
0.7852
0.8962
0.9591
0.9864
9-9
0.6
0.4
0.2
0
175
180
185
190
195
200
b)
Power Function Curve
1
0.8
0.6
0.4
0.2
0
175
180
185
190
Section 9-2
H 0 : = 10, H 1 : > 10
b) H 0 : = 7, H 1 : 7
c) H 0 : = 5, H 1 : < 5
9-29
a)
9-30
a) =0.01, then a=
9-31
a) =0.01, then a=
z 2.33
b) =0.05, then a= z 1.64
9-10
195
200
0.6
0.4
0.2
0
175
180
185
190
195
200
b)
Power Function Curve
1
0.8
0.6
0.4
0.2
0
175
180
185
190
Section 9-2
H 0 : = 10, H 1 : > 10
b) H 0 : = 7, H 1 : 7
c) H 0 : = 5, H 1 : < 5
9-29
a)
9-30
a) =0.01, then a=
9-31
a) =0.01, then a=
z 2.33
b) =0.05, then a= z 1.64
9-10
195
200
c) =0.1, then a=
z 1.29
9-32
z1 -2.33
b) =0.05, then a= z1 -1.64
c) =0.1, then a= z1 -1.29
9-33
a) =0.01, then a=
0.02
b) p-value=1- ( Z 0 ) =1- (1.84) 0.97
c) p-value=1- ( Z 0 ) =1- (0.4) 0.34
9-35
a) p-value= ( Z 0 ) = ( 2.05)
9-36
0.98
b) p-value= ( Z 0 ) = (1.84) 0.03
c) p-value= ( Z 0 ) = (0.4) 0.65
z0 =
98 100
2/ 9
= 3.0
8) Since -3.0 < 1.65 do not reject H0 and conclude the water temperature is not significantly different
greater than 100 at = 0.05.
b) P-value = 1 ( 3.0) = 1 0.00135 = 0.99865
c)
100 104
= z 0.05 +
2/ 9
= (1.65 + 6)
= (-4.35)
0
9-37
9-11
7) x = 2.78, = 0.9
z0 =
2.78 3
0.9 / 15
= 0.95
8) Since 0.95 > -1.96, do not reject the null hypothesis and conclude there is not sufficient evidence
to support the claim the mean crankshaft wear is not equal to 3 at = 0.05.
b)
= z 0.025 +
3 3.25
3 3.25
z 0.025 +
0.9 / 15
0.9 / 15
c)
9-38
n=
(z
+ z ) 2
2
/2
(z 0.025 + z 0.10 )2 2
(3.75 3) 2
z0 =
154.2 155
= 1.69
1.5 / 10
8) Since 1.69 > -2.58, do not reject the null hypothesis and conclude there is not sufficient evidence
to support the claim the mean melting point is not equal to 155 F at = 0.01.
b) P-value = 2*P(Z <- 1.69) =2* 0.045514 = 0.091028
= z0.005
c)
n
n
z0.005
(155 150) 10
(155 150) 10
2.58
= 2.58
1.5
1
.
5
= (-7.96)- (-13.12) = 0 0 = 0
d)
n=
(z
+ z ) 2
2
/2
(z 0.005 + z 0.10 )2 2
(150 155) 2
n 2.
9-39
a.) 1) The parameter of interest is the true mean battery life in hours, .
2) H0 : = 40
3) H1 : > 40
4) = 0.05
x
5) z0 =
/ n
9-12
z0 =
40.5 40
1.25 / 10
= 1.26
8) Since 1.26 < 1.65 do not reject H0 and conclude the battery life is not significantly different greater than
40 at = 0.05.
b) P-value = 1 (1.26) = 1 0.8962 = 0.1038
40 42
= z 0.05 +
1.25 / 10
c)
= (1.65 + 5.06)
= (-3.41)
0.000325
d)
n=
(z
+ z ) 2
2
(z 0.05 + z 0.10 )2 2
(40 44) 2
40 . 5 + 1 . 65 (1 . 25 ) / 10
39 . 85
The lower bound of the 90 % confidence interval must be greater than 40 to verify that the true mean exceeds
40 hours.
9-40
a)
1) The parameter of interest is the true mean tensile strength, .
2) H0 : = 3500
3) H1 : 3500
4) = 0.01
5) z0 =
x
/ n
z0 =
3450 3500
= 2.89
60 / 12
8) Since -2.89 < 2.58, reject the null hypothesis and conclude the true mean tensile strength
is significantly different from 3500 at = 0.01.
b) Smallest level of significance =
P-value = 2[1 (2.89) ]= 2[1 .998074] = 0.004
The smallest level of significance at which we are willing to reject the null hypothesis is 0.004.
c)
9-13
= z0.005
n
n
z0.005
(3470 3500) 12
(3470 3500) 12
= 2.58
2.58
60
60
x z0.005
x + z0.005
n
n
60
60
3450 2.58
3450 + 2.58
12
12
3405.313 3494.687
With 99% confidence, we believe the true mean tensile strength is between 3405.313 psi and
3494.687 psi. We can test the hypotheses that the true mean tensile strength is not equal to 3500
by noting that the value is not within the confidence interval. Hence we reject the null hypothesis.
9-41
x
/ n
z0 =
102.2 100
4/ 8
= 1.56
8) Since 1.56> 1.65, do not reject the null hypothesis and conclude the there is insufficient
evidence to conclude that the true speed strength is less than 100 at = 0.05.
b)
c)
= 1 z 0.05
(95 100) 8
= 1-(-1.65 - 3.54) = 1-(1.89) = 0.02938
d) n =
e)
(z
+ z ) 2
2
(z 0.05 + z 0.15 )2 2
(95 100) 2
x + z 0.05
9-14
102.2 + 1.65
104.53
Because 100 is included in the CI then we dont have enough confidence to reject the null
hypothesis.
9-42
x
/ n
z0 =
1.4975 1.50
0.01 / 25
= 1.25
8) Since 2.58 < -1.25 < 2.58, do not reject the null hypothesis and conclude the true mean
hole diameter is not significantly different from 1.5 in. at = 0.01.
b) p-value=2(1- ( Z 0 ) )=2(1- (1.25) ) 0.21
c)
n
n
= z 0.005
z 0.005
(1.495 1.5) 25
(1.495 1.5) 25
2.58
= 2.58
0.01
0.01
( z / 2 + z ) 2 2
( z 0.005 + z 0.10 ) 2 2
(1.495 1.50) 2
n 60.
e) For = 0.01, z/2 = z0.005 = 2.58
x z0.005
x + z0.005
n
n
0.01
0.01
1.4975 + 2.58
1.4975 2.58
25
25
1.4923 1.5027
The confidence interval constructed contains the value 1.5, thus the true mean hole diameter
could possibly be 1.5 in. using a 99% level of confidence. Since a two-sided 99% confidence
9-15
interval is equivalent to a two-sided hypothesis test at = 0.01, the conclusions necessarily must
be consistent.
9-43
x
/ n
z0 =
4.05 4
= 1.77
0.2 / 50
8) Since 1.77>1.65, reject the null hypothesis and conclude that there is sufficient evidence
to conclude that the true average battery life exceeds 4 hours at = 0.05.
b) p-value=1- ( Z 0 ) =1- (1.77)
c)
= z0.05
d) n =
(z
0.04
(4.5 4) 50
= (1.65 17.68) = (-16.03) = 0
0.2
Power = 1- = 1-0 = 1
+ z ) 2
2
(z 0.05 + z 0.1 )2 2
(4.5 4) 2
n2
e) x z0.05
0.2
4.05 1.65
50
4.003
Since the lower limit of the CI is just slightly above 4, we conclude that average life is greater
than 4 hours at =0.05.
Section 9-3
9-44
9-45
9-46
9-16
interval is equivalent to a two-sided hypothesis test at = 0.01, the conclusions necessarily must
be consistent.
9-43
x
/ n
z0 =
4.05 4
= 1.77
0.2 / 50
8) Since 1.77>1.65, reject the null hypothesis and conclude that there is sufficient evidence
to conclude that the true average battery life exceeds 4 hours at = 0.05.
b) p-value=1- ( Z 0 ) =1- (1.77)
c)
= z0.05
d) n =
(z
0.04
(4.5 4) 50
= (1.65 17.68) = (-16.03) = 0
0.2
Power = 1- = 1-0 = 1
+ z ) 2
2
(z 0.05 + z 0.1 )2 2
(4.5 4) 2
n2
e) x z0.05
0.2
4.05 1.65
50
4.003
Since the lower limit of the CI is just slightly above 4, we conclude that average life is greater
than 4 hours at =0.05.
Section 9-3
9-44
9-45
9-46
9-16
9-47
0.025 p 0.05
b) 1 0.05 p 1 0.025 then 0.95 p 0.975
c) 0.25 p 0.4
9-48
a)
9-49
a)
9-50
2) H0 : = 22.5
3) H1 : 22.5
4) = 0.05
5) t0
x
s/ n
t0 =
22.496 22.5
0.378 / 5
= 0.00237
8) Since 0.00237 >- 2.776, we cannot reject the null hypothesis. There is not sufficient evidence to
conclude that the true mean interior temperature is not equal to 22.5 C at = 0.05.
2*0.4 <P-value < 2* 0.5 ; 0.8 < P-value <1.0
b.) The points on the normal probability plot fall along the line. Therefore, there is no evidence to
conclude that the interior temperature data is not normally distributed.
Normal Probability Plot for temp
ML Estimates - 95% CI
99
ML Estimates
95
90
Percent
80
70
60
50
40
30
20
10
5
1
21.5
22.5
Data
9-17
23.5
Mean
22.496
StDev
0.338384
c.) d =
| 0 | | 22.75 22.5 |
=
=
= 0.66
0.378
Using the OC curve, Chart VII e) for = 0.05, d = 0.66, and n = 5, we get 0.8 and
power of 10.8 = 0.2.
d) d =
| 0 | | 22.75 22.5 |
=
=
= 0.66
0.378
Using the OC curve, Chart VII e) for = 0.05, d = 0.66, and 0.1 (Power=0.9),
n = 40 .
e) 95% two sided confidence interval
s
s
x + t 0.025, 4
x t 0.025, 4
n
n
0.378
0.378
22.496 2.776
22.496 + 2.776
5
5
22.027 22.965
We cannot conclude that the mean interior temperature is not equal to 22.5 since the value is included
inside the confidence interval.
9-51
x
s/ n
t0 =
98.264 98.6
= 3.48
0.4821 / 25
8) Since 3.48 > 2.064, reject the null hypothesis and conclude that the there is sufficient evidence to
conclude that the true mean female body temperature is not equal to 98.6 F at = 0.05.
P-value = 2* 0.001 = 0.002
b)
9-18
99
95
90
Percent
80
70
60
50
40
30
20
10
5
1
97
98
99
D ata
c) d =
| 0 | | 98 98.6 |
=
=
= 1.24
0.4821
Using the OC curve, Chart VII e) for = 0.05, d = 1.24, and n = 25, we get 0 and
power of 10 1.
d) d =
| 0 | | 98.2 98.6 |
=
=
= 0.83
0.4821
Using the OC curve, Chart VII e) for = 0.05, d = 0.83, and 0.1 (Power=0.9),
n = 20 .
s
s
x t0.025, 24
x + t0.025, 24
n
n
0.4821
0.4821
98.264 2.064
98.264 + 2.064
25
25
98.065 98.463
We can conclude that the mean female body temperature is not equal to 98.6 since the value is not included
inside the confidence interval.
9-52 a) 1) The parameter of interest is the true mean rainfall, .
2) H0 : = 25
3) H1 : > 25
4) = 0.01
5) t0 =
s/ n
26.04 25
4.78 / 20
= 0.97
8) Since 0.97 < 2.539, do not reject the null hypothesis and conclude there is insufficient evidence to indicate
that the true mean rainfall is greater than 25 acre-feet at = 0.01. The 0.10 < P-value < 0.25.
b) the data on the normal probability plot fall along the straight line. Therefore there is evidence
9-19
99
ML Estimates
95
Mean
26.035
StDev
4.66361
90
Percent
80
70
60
50
40
30
20
10
5
1
20
30
40
Data
c) d =
| 0 | | 27 25 |
=
=
= 0.42
4.78
Using the OC curve, Chart VII h) for = 0.01, d = 0.42, and n = 20, we get 0.7 and
power of 10.7 = 0.3.
d) d =
| 0 | | 27.5 25 |
=
=
= 0.52
4.78
Using the OC curve, Chart VII h) for = 0.05, d = 0.42, and 0.1 (Power=0.9),
n = 75 .
e) 99% lower confidence bound on the mean diameter
s
x t0.01,19
n
4.78
26.04 2.539
20
23.326
Since the lower limit of the CI is less than 25, we conclude that there is insufficient evidence to indicate that
the true mean rainfall is greater than 25 acre-feet at = 0.01.
9-53
a)
1) The parameter of interest is the true mean sodium content, .
2) H0 : = 130
3) H1 : 130
4) = 0.05
5) t0 =
x
s/ n
9-20
t0 =
129.753 130
= 1.456
0.929 / 30
8) Since 1.456 < 2.064, do not reject the null hypothesis and conclude that the there is not
sufficient evidence that the true mean sodium content is different from 130mg at = 0.05.
From table V the t0 value is found between the values of 0.05 and 0.1 with 29 degrees of freedom,
so 2*0.05<P-value < 2* 0.1 Therefore, 0.1< P-value < 0.2.
b) The assumption of normality appears to be reasonable.
Normal Probability Plot for 9-33
ML Estimates - 95% CI
99
95
90
80
Percent
70
60
50
40
30
20
10
5
1
127
128
129
130
131
132
Data
c) d =
| 0 | | 130.5 130 |
=
=
= 0.538
0.929
Using the OC curve, Chart VII e) for = 0.05, d = 0.53, and n = 30, we get 0.2 and power
of 10.20 = 0.80
d) d =
| 0 | | 130.1 130 |
=
=
= 0.11
0.929
Using the OC curve, Chart VII e) for = 0.05, d = 0.11, and 0.25 (Power=0.75), n = 100 .
e) 95% two sided confidence interval
s
s
x t 0.025, 29
x + t 0.025, 29
n
n
0.929
0.929
129.753 + 2.045
129.753 2.045
30
30
129.406 130.100
9-21
There is no evidence that the mean differs from 130 because that value is inside the confidence
interval.
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying distribution
is normal.
1) The parameter of interest is the true mean coefficient of restitution, .
2) H0 : = 0.635
3) H1 : > 0.635
4) = 0.05
5) t0 =
s/ n
0.624 0.635
0.013 / 40
= 5.35
8) Since 5.25 < 1.685, do not reject the null hypothesis and conclude that there is not
sufficient evidence to indicate that the true mean coefficient of restitution is greater than
0.635 at = 0.05.
The area to right of -5.35 under the t distribution is greater than 0.9995 from table V.
Minitab gives P-value = 1.
b) From the normal probability plot, the normality assumption seems reasonable:
Probability Plot of Baseball Coeff of Restitution
Normal
99
95
90
80
Percent
9-54
70
60
50
40
30
20
10
5
c) d =
0.59
0.60
0.61
0.62
0.63
0.64
Baseball Coeff of Restitution
0.65
0.66
| 0 | | 0.64 0.635 |
=
=
= 0.38
0.013
Using the OC curve, Chart VII g) for = 0.05, d = 0.38, and n = 40, we get 0.25 and power
of 10.25 = 0.75.
d) d =
| 0 | | 0.638 0.635 |
=
=
= 0.23
0.013
Using the OC curve, Chart VII g) for = 0.05, d = 0.23, and 0.25 (Power=0.75),
n = 40 .
9-22
e)
s
= 0.6205
n
Since 0.635 > 0.6205, then we fail to reject the null hypothesis.
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying distribution
is normal.
1) The parameter of interest is the true mean oxygen concentration, .
2) H0 : = 4
3) H1 : 4
4) = 0.01
5) t0 =
s/ n
3.265 4
2.127 / 20
= 1.55
8) Because -2.861<-1.55 do not reject the null hypothesis and conclude that there is
insufficient evidence to indicate that the true mean oxygen differs from 4 at = 0.01.
P-Value: 2*0.05<P-value<2*0.10 therefore 0.10< P-value<0.20
b) From the normal probability plot, the normality assumption seems reasonable:
Probability Plot of O2 concentration
Normal
99
95
90
80
Percent
9-55
70
60
50
40
30
20
10
5
c.) d =
0.0
2.5
5.0
O2 concentration
7.5
| 0 | | 3 4 |
=
=
= 0.47
2.127
Using the OC curve, Chart VII f) for = 0.01, d = 0.47, and n = 20, we get 0.70 and
power of 10.70 = 0.30.
d) d =
| 0 | | 2.5 4 |
=
=
= 0.71
2.127
Using the OC curve, Chart VII f) for = 0.01, d = 0.71, and 0.10 (Power=0.90), n = 40 .
e)
9-23
s
s
x t / 2,n1
x + t / 2,n1
= 1.9 4.62
n
n
Because 4 is within the confidence interval, we fail to reject the null hypothesis.
9-56
a)
1) The parameter of interest is the true mean sodium content, .
2) H0 : = 300
3) H1 : > 300
4) = 0.05
5) t0 =
x
s/ n
t0 =
315 300
= 2.48
16 / 7
8) Since 2.48>1.943, reject the null hypothesis and conclude that there is sufficient evidence that
the leg strength exceeds 300 watts at = 0.05.
The p-value is between .01 and .025
b) d =
| 0 | | 305 300 |
=
=
= 0.3125
16
s
= 303.2
n
s/ n
n = 16 x = 60,139.7 s = 3645.94
60139.7 60000
t0 =
3645.94 / 16
= 0.15
8) Since 0.15 < 1.753., do not reject the null hypothesis and conclude there is insufficient evidence to indicate
that the true mean tire life is greater than 60,000 kilometers at = 0.05. The P-value > 0.40.
b.) d =
| 0 | | 61000 60000 |
=
=
= 0.27
3645.94
Using the OC curve, Chart VII g) for = 0.05, d = 0.27, and 0.1 (Power=0.9),
9-24
n = 4.
Yes, the sample size of 16 was sufficient.
9-58
In order to use t statistics in hypothesis testing, we need to assume that the underlying distribution is normal.
1) The parameter of interest is the true mean impact strength, .
2) H0 : = 1.0
3) H1 : > 1.0
4) = 0.05
5) t0 =
s/ n
1.25 1.0
= 4.47
0.25 / 20
8) Since 4.47 > 1.729, reject the null hypothesis and conclude there is sufficient evidence to indicate that the
true mean impact strength is greater than 1.0 ft-lb/in at = 0.05. The P-value < 0.0005
9-59
In order to use t statistic in hypothesis testing, we need to assume that the underlying distribution is normal.
1) The parameter of interest is the true mean current, .
2) H0 : = 300
3) H1 : > 300
4) = 0.05
5) t0 =
s/ n
t 0.05,9 = 1.833
n = 10 x = 317.2 s = 15.7
317.2 300
t0 =
15.7 / 10
= 3.46
8) Since 3.46 > 1.833, reject the null hypothesis and conclude there is sufficient evidence to indicate that the
true mean current is greater than 300 microamps at = 0.05. The 0.0025 <P-value < 0.005
9-60
a)
1) The parameter of interest is the true mean height of female engineering students, .
2) H0 : = 65
3) H1 : > 65
4) = 0.05
5) t0 =
s/ n
65.811 65
2.11 / 37
= 2.34
8) Since 2.34 > 1.68, reject the null hypothesis and conclude that there is sufficient evidence to
indicate that the true mean height of female engineering students is not equal to 65 at = 0.05.
P-value: 0.01<P-value<0.025.
b.) From the normal probability plot, the normality assumption seems reasonable:
9-25
95
90
Percent
80
70
60
50
40
30
20
10
5
c)
60
d=
62
62 65
2.11
64
66
Female heights
68
70
72
= 1.42 , n=37 so, from the OC Chart VII g) for = 0.05, we find that 0.
d.) d =
64 65
2.11
*
n = 30 .
9-61
= 0.47 so, from the OC Chart VII g) for = 0.05, and 0.2 (Power=0.8).
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying
distribution is normal.
1) The parameter of interest is the true mean distance, .
2) H0 : = 280
3) H1 : > 280
4) = 0.05
5) t0 =
s/ n
260.3 280
13.41 / 100
= 14.69
8) Since 14.69 < 1.6604, do not reject the null hypothesis and conclude that there is
insufficient evidence to indicate that the true mean distance is greater than 280 at = 0.05.
From table V the t0 value in absolute value is greater than the value corresponding to
0.0005. Therefore, the P-value is greater than 0.9995.
b) From the normal probability plot, the normality assumption seems reasonable:
9-26
Percent
95
90
80
70
60
50
40
30
20
10
5
1
0.1
c) d =
220
230
240
250
260
270
280
Distance for golf balls
290
300
310
| 0 | | 290 280 |
=
=
= 0.75
13.41
Using the OC curve, Chart VII g) for = 0.05, d = 0.75, and n = 100, 0 and power of 10 =
1.
d) d =
| 0 | | 290 280 |
=
=
= 0.75
13.41
Using the OC curve, Chart VII g) for = 0.05, d = 0.75, and 0.20 (Power=0.80), n = 15 .
9-62
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying
distribution is normal.
1) The parameter of interest is the true mean concentration of suspended solids, .
2) H0 : = 55
3) H1 : 55
4) = 0.05
5) t0 =
s/ n
59.87 55
12.50 / 60
= 3.018
8) Since 3.018 > 2.000, reject the null hypothesis and conclude that there is sufficient evidence to
indicate that the true mean concentration of suspended solids is not equal to 55 at = 0.05.
From table V the t0 value is between the values of 0.001 and 0.0025 with 59 degrees of freedom.
Therefore 2*0.001<P-value < 2* 0.0025 and 0.002< P-value<0.005. Minitab gives a P-value of
0.0038.
b) From the normal probability plot, the normality assumption seems reasonable:
9-27
Percent
90
80
70
60
50
40
30
20
10
5
1
0.1
d) d =
20
50 55
12.50
30
40
50
60
70
Concentration of solids
80
90
100
= 0.4 , n=60 so, from the OC Chart VII e) for = 0.05, d= 0.4 and n=60 we
d=
50 55
12.50
= 0.4
that power=0.90, n = 75 .
9-28
Section 9-4
9-63
a) =0.01, n=20, from table V we find the following critical values 6.84 and 38.58
b) =0.05, n=12, from table V we find the following critical values 3.82 and 21.92
c) =0.10, n=15, from table V we find the following critical values 6.57 and 23.68
9-64
2 ,n1 = 36.19
2 ,n1 = 19.68
2 ,n1 = 21.06
12 ,n1 = 7.63
12 ,n1 = 4.57
12 ,n1 = 7.79
9-65
9-66
9-67
9-68
a) 0.1<P-value<0.5
b) 0.1<P-value<0.5
c) 0.99<P-value<0.995
9-69
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we need
to assume that the underlying distribution is normal.
1) The parameter of interest is the true standard deviation of performance time . However,
the answer can be found by performing a hypothesis test on 2.
2) H0 : 2 = .752
3) H1 : 2 >.752
4) = 0.05
5) 20 =
( n 1)s2
2
6) Reject H0 if
7) n = 17, s = 0.09
20 =
(n 1) s 2
16(0.09) 2
= 0.23
.75 2
8) Because 0.23 < 26.30 do not reject H0 and conclude there is insufficient evidence to indicate
the true variance of performance time content exceeds 0.752 at = 0.05.
P-value: Because 20 =0.23 the P-value>0.995
b) The 95% one sided confidence interval given below, includes the value 0.75. Therefore, we are
not be able to conclude that the standard deviation is greater than 0.75.
9-29
16(.09) 2
2
26.3
0.07
9-70
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we need
to assume that the underlying distribution is normal.
1) The parameter of interest is the true measurement standard deviation . However, the
answer can be found by performing a hypothesis test on 2.
2) H0 : 2 = .012
3) H1 : 2 .012
4) = 0.05
5) 20 =
( n 1)s2
2
02.975,14 = 5.63 or
20 > 2 ,2 ,n 1 where
02.025,14 = 26.12
7) n = 15, s = 0.0083
20 =
(n 1) s 2
14(.0083) 2
= 9.6446
.012
14(.0083) 2
14(.0083) 2
2
26.12
5.63
2
0.00607 0.013
9-71
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we need
to assume that the underlying distribution is normal.
1) The parameter of interest is the true standard deviation of titanium percentage, . However,
the answer can be found by performing a hypothesis test on 2.
2) H0 : 2 = (0.25)2
3) H1 : 2 (0.25)2
4) = 0.05
5) 20 =
( n 1)s2
2
7) n = 51, s = 0.37
20 =
( n 1)s2
50(0.37) 2
(0.25) 2
= 109.52
8) Since 109.52 > 71.42 we reject H0 and conclude there is sufficient evidence to indicate the
true standard deviation of titanium percentage is significantly different from 0.25 at = 0.01.
9-30
50(0.37) 2
50(0.37) 2
2
71.42
32.36
0.096 2 0.2115
Taking the square root of the endpoints of this interval we obtain,
0.31 < < 0.46
Since 0.25 falls below the lower confidence bound we would conclude that the population
standard deviation is not equal to 0.25.
9-72
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we need to
assume that the underlying distribution is normal.
1) The parameter of interest is the true standard deviation of Izod impact strength, . However, the
answer can be found by performing a hypothesis test on 2.
2) H0 : 2 = (0.10)2
3) H1 : 2 (0.10)2
4) = 0.01
5) 20 =
( n 1)s2
2
02.995,19 = 6.84 27 or
= 38.58
7) n = 20, s = 0.25
20 =
(n 1) s 2
19(0.25) 2
= 118.75
(0.10) 2
8) Since 118.75 > 38.58 reject H0 and conclude there is sufficient evidence to indicate the true
standard deviation of Izod impact strength is significantly different from 0.10 at = 0.01.
b.) P-value: The P-value<0.005
c.) 99% confidence interval for :
First find the confidence interval for 2 :
For = 0.01 and n = 20, 2 / 2 , n 1 =
02.995,19 = 6.84
and 12 / 2 ,n 1 =
02.005,19 = 38.58
19(0.25) 2
19(0.25) 2
2
38.58
6.84
2
0.03078 0.1736
Since 0.01 falls below the lower confidence bound we would conclude that the population standard deviation
is not equal to 0.01.
9-73
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we need
to assume that the underlying distribution is normal.
1) The parameter of interest is the standard deviation of tire life, . However, the answer can
be found by performing a hypothesis test on 2.
2) H0 : 2 = 40002
3) H1 : 2 <40002
4) = 0.05
9-31
5)
20
(n 1) s 2
6) Reject H0 if
02 < 12 ,n1
where
02.95,15 = 7.26
7) n = 16, s2 = (3645.94)2
20 =
(n 1) s 2
15(3645.94) 2
= 12.46
4000 2
8) Since 12.46 > 7.26, fail to reject H0 and conclude there is not sufficient evidence to indicate
the true standard deviation of tire life is less than 4000 km at = 0.05.
P-value = P(2 <12.46) for 15 degrees of freedom 0.5<1-P-value < 0.9
Then 0.1<P-value<0.5
b) The 95% one sided confidence interval below, includes the value 4000, therefore, we could not
be able to conclude that the variance was not equal to 40002.
15(3645.94) 2
7.26
5240
9-74
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we need to
assume that the underlying distribution is normal.
1) The parameter of interest is the true standard deviation of the diameter,. However, the answer can
be found by performing a hypothesis test on 2.
2) H0 : 2 = 0.0001
3) H1 : 2 > 0.0001
4) = 0.01
5) 20 =
( n 1) s2
2
( n 1)s2
b)
0.015
Using the chart in the Appendix, with =
= 1.5 and n = 15 we find = 0.50.
0.01
c)
0.0125
=
= 1.25
0
0.01
9-75
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we need to
assume that the underlying distribution is normal.
1) The parameter of interest is the true variance of sugar content, 2. However, the
answer can be found by performing a hypothesis test on 2.
9-32
2) H0 : 2 = 18
3) H1 : 2 18
4) = 0.05
5) 20 =
( n 1)s2
2
02.975,9 = 2.70 or
20 > 2 ,2 ,n 1 where
02.025,9 = 19.02
7) n = 10, s = 4.8
20 =
(n 1) s 2
9(4.8) 2
= 11.52
18
8) Since 11.52 < 19.02 do not reject H0 and conclude there is insufficient evidence to indicate the true
variance of sugar content is significantly different from 18 at = 0.01.
P-value: The 20 is between 0.10 and 0.50. Therefore, 0.2<P-value<1
=2
40
= 1.49 and = 0.10, n = 30.
18
9-33
Section 9-5
9-76
a)
1) The parameter of interest is the true fraction of satisfied customers.
2) H0 : p = 0.9
3) H1 : p 0.9
4) = 0.05
5) z 0 =
x np0
np0 (1 p 0 )
or z 0 =
p p0
p 0 (1 p 0 )
n
same conclusion
6) Reject H0 if z0 < z/2 where z/2 = z0.025 = 1.96 or z0 > z/2 where z/2 = z0.025
= 1.96
850
= 0.85
1000
x np0
850 1000(0.9)
=
= 5.27
1000(0.9)(0.1)
np0 (1 p0 )
=
7) x = 850 n = 1000 p
z0 =
8) Because -5.27<-1.96 reject the null hypothesis and conclude the true fraction of
satisfied customers is significantly different from 0.9 at = 0.05.
The P-value: 2(1-(5.27)) 2(1-1) 0
b)
The 95% confidence interval for the fraction of surveyed customers is:
p z / 2
.85 1.96
p (1 p )
p p + z / 2
n
p (1 p )
n
0.85(0.15)
0.85(0.15)
p .85 + 1.96
1000
1000
0.827 p 0.87
Because 0.9 is not included in the confidence interval, we reject the null hypothesis at
= 0.05.
9-77
a)
1) The parameter of interest is the true fraction of rejected parts
2) H0 : p = 0.03
3) H1 : p < 0.03
4) = 0.05
5) z0 =
x np0
or z 0 =
np0 (1 p0 )
p p0
; Either approach will yield the same
p0 (1 p0 )
n
conclusion
6) Reject H0 if z0 < z where z = z0.05 = 1.65
=
7) x = 10 n = 500 p
10
= 0.02
500
x np0
10 500(0.03)
=
= 1.31
z0 =
500(0.03)(0.97)
np0 (1 p0 )
9-34
8) Because 1.31 > 1.65, do not reject the null hypothesis. There is not enough
evidence to conclude that the true fraction of rejected parts is significantly less than 0.03
at = 0.05.
P-value = (-1.31) = 0.095
b)
The upper one-sided 95% confidence interval for the fraction of rejected parts is:
p p z
p (1 p )
n
p .02 + 1.65
0.02(0.98)
500
p 0.0303
Because 0.03<0.0303 the we fail to reject the null hypothesis
9-78
a)
1) The parameter of interest is the true fraction defective integrated circuits
2) H0 : p = 0.05
3) H1 : p 0.05
4) = 0.05
5)
z0 =
x np0
np0 (1 p 0 )
or
z0 =
p p0
p 0 (1 p 0 )
n
conclusion
6) Reject H0 if z0 < z/2 where z/2 = z0.025 = 1.96 or z0 > z/2 where z/2 = z0.025 = 1.96
7) x = 13 n = 300 p =
z0 =
13
= 0.043
300
x np 0
np0 (1 p 0 )
13 300(0.05)
300(0.05)(0.95)
= 0.53
8) Since 0.53 > 1.65, do not reject null hypothesis and conclude the true fraction of defective
integrated circuits is not significantly less than 0.05, at = 0.05.
The P-value: 2(1-(0.53))=2(1-0.70194)=0.59612
b) The 95% confidence interval is:
p z / 2
p (1 p )
p p + z / 2
n
p (1 p )
n
0.043(0.957)
0.043(0.957)
p 043 + 1.96
300
300
0.02004 p 0.065
Since 0.02004 p = 0.05 0.065 , then we fail to reject the null hypothesis.
.043 1.96
9-79
a)
1) The parameter of interest is the true success rate
9-35
2) H0 : p = 0.78
3) H1 : p > 0.78
4) The value for is not given. We assume = 0.05.
5)
z0 =
x np 0
np 0 (1 p 0 )
or
p p0
z0 =
p 0 (1 p 0 )
n
conclusion
6) Reject H0 if z0 > z where z = z0.05 = 1.65
289
0.83
350
x np0
289 350(0.78)
=
= 2.06
np0 (1 p0 )
350(0.78)(0.22)
=
7) x = 289 n = 350 p
z0 =
8) Because 2.06 > 1.65, reject the null hypothesis and conclude the true success rate is
significantly greater than 0.78, at = 0.05.
P-value=1-0.9803=0.0197
b) The 95% lower confidence interval:
p (1 p )
p
n
p z
.83 1.65
0.83(0.17)
p
350
0.7969 p
a)
1) The parameter of interest is the true percentage of polished lenses that contain surface defects,
p.
2) H0 : p = 0.02
3) H1 : p < 0.02
4) = 0.05
5)
z0 =
x np0
or z0 =
np0 (1 p0 )
p p0
; Either approach will yield the
p0 (1 p0 )
n
same conclusion
6) Reject H0 if z0 < z where z = z0.05 = 1.65
=
7) x = 6 n = 250 p
6
= 0.024
250
p p0
=
z0 =
p0 (1 p0 )
n
0.024 0.02
= 0.452
0.02(1 0.02)
250
8) Since 0.452 > 1.65 do not reject the null hypothesis and conclude the machine cannot be
qualified at the 0.05 level of significance.
9-36
p p + z
p (1 p )
n
p .024 + 1.65
0.024(0.976)
250
p 0.0264
Since
9-81
a)
1) The parameter of interest is the true percentage of football helmets that contain flaws, p.
2) H0 : p = 0.1
3) H1 : p > 0.1
4) = 0.01
5)
z0 =
x np0
or z0 =
np0 (1 p0 )
p p0
; Either approach will yield the same
p0 (1 p0 )
n
conclusion
6) Reject H0 if z0 > z where z = z0.01 = 2.33
=
7) x = 16 n = 200 p
16
= 0.08
200
p p 0
=
z0 =
p 0 (1 p 0 )
n
0.08 0.10
0.10(1 0.10)
200
= 0.94
8) Because -0.94 < 2.33 do not reject the null hypothesis. There is not enough evidence that the
proportion of football helmets with flaws exceeds 10%.
P-value = 1- (-0.94) =0.8264
b) The 99% lower confidence interval:
p (1 p )
p
n
p z
.08 2.33
Because 0.035
9-82
0.08(0.92)
p
200
0.035 p
a) 1) The parameter of interest is the true proportion of engineering students planning graduate studies
2) H0 : p = 0.50
3) H1 : p 0.50
4) = 0.05
9-37
x np 0
z0 =
5)
np 0 (1 p 0 )
or
z0 =
p p0
p 0 (1 p 0 )
n
6) Reject H0 if z0 < z/2 where z/2 = z0.025 = 1.96 or z0 > z/2 where z/2 = z0.025 = 1.96
117
= 0.2423
484
x np 0
117 484(0.5)
=
= 11.36
np 0 (1 p0 )
484(0.5)(0.5)
7) x = 117 n = 484
z0 =
p =
8) Since 11.36 > 1.65, reject the null hypothesis and conclude the true proportion of engineering students
planning graduate studies is significantly different from 0.5, at = 0.05.
P-value =2[1 (11.36)] 0
b.)
p =
117
= 0.242
484
p z / 2
0.242 1.96
p (1 p )
p p + z / 2
n
p (1 p )
n
0.242(0.758)
0.242(0.758)
p 0.242 1.96
484
484
0.204 p 0.280
Since the 95% confidence interval does not contain the value 0.5, then conclude that the true proportion of
engineering students planning graduate studies is significantly different from 0.5.
9-83
1) The parameter of interest is the true proportion of batteries that fail before 48 hours, p.
2) H0 : p = 0.002
3) H1 : p < 0.002
4) = 0.01
5)
z0 =
x np0
np0 (1 p0 )
or
z0 =
p p0
p0 (1 p0 )
n
p =
15
= 0.003
5000
p p0
=
z0 =
p 0 (1 p 0 )
n
0.003 0.002
0.002(1 0.998)
5000
= 1.58
8) Since 1.58 > -2.33 do not reject the null hypothesis and conclude the proportion of proportion
of cell phone batteries that fail is not less than 0.2% at =0.01.
9-38
9-84.
The problem statement implies that H0: p = 0.6, H1: p > 0.6 and defines an acceptance region as
315
and rejection region as p > 0.63
p
= 0.63
500
a) The probability of a type 1 error is
.
0.63 0.6
500
9-85
a) The parameter of interest is the true proportion of engine crankshaft bearings exhibiting surface roughness.
2) H0 : p = 0.10
3) H1 : p > 0.10
4) = 0.05
5)
z0 =
x np0
np0 (1 p 0 )
or
z0 =
p p 0
p 0 (1 p 0 )
n
z0 =
10
= 0.118
85
x np 0
10 85(0.10)
=
= 0.54
np0 (1 p 0 )
85(0.10)(0.90)
p =
8) Because 0.54 < 1.65, do not reject the null hypothesis. There is not enough evidence to conclude that
the true proportion of crankshaft bearings exhibiting surface roughness exceeds 0.10, at = 0.05.
P-value = 1 (0.54) = 0.295
b) p= 0.15, p0=0.10, n=85, and z/2=1.96
p0 p + z / 2 p0 (1 p0 ) / n p0 p z / 2 p0 (1 p0 ) / n
p
(
1
p
)
/
n
p
(
1
p
)
/
n
0.10 0.15+1.96 0.10(1 0.10) / 85 0.10 0.151.96 0.10(1 0.10) / 85
0
.
15
(
1
0
.
15
)
/
85
0
.
15
(
1
0
.
15
)
/
85
z / 2 p0 (1 p0 ) z p(1 p)
n =
p p0
0
.
15
0
.
10
2
= (10.85) = 117.63 118
9-39
Section 9-7
9-86
e x
where = [0(24) + 1(30) + 2(31) + 3(11) + 4( 4)] / 100 = 1.41
P ( X = x) =
x!
Value
Observed Frequency
Expected Frequency
0
24
30.12
1
30
36.14
2
31
21.69
3
11
8.67
4
4
2.60
Since value 4 has an expected frequency less than 3, combine this category with the previous category:
Value
Observed Frequency
Expected Frequency
0
24
30.12
1
30
36.14
2
31
21.69
3-4
15
11.67
(Oi Ei )2
i =1
Ei
02 =
6) Reject H0 if
7)
2o
>
20.05,3
= 7.81
2
2
2
2
(
24 30.12) (30 36.14) (31 21.69) (1511.67)
=
+
+
+
= 7.23
2
0
30.12
36.14
21.69
11.67
8) Since 7.23 < 7.81 do not reject H0. We are unable to reject the null hypothesis that the distribution of
X is Poisson.
b) The P-value is between 0.05 and 0.1 using Table IV. From Minitab the P-value = 0.0649.
9-40
9-87
e x
where = [1(1) + 2(11) + + 7(10) + 8(9)] / 75 = 4.907
P ( X = x) =
x!
Estimated mean = 4.907
Value
Observed Frequency
Expected Frequency
1
1
2.7214
2
11
6.6770
3
8
10.9213
4
13
13.3977
5
11
13.1485
6
12
10.7533
7
10
7.5381
8
9
4.6237
Since the first category has an expected frequency less than 3, combine it with the next category:
Value
Observed Frequency
Expected Frequency
1-2
12
9.3984
3
8
10.9213
4
13
13.3977
5
11
13.1485
6
12
10.7533
7
10
7.5381
8
9
4.6237
6) Reject H0 if
2o
>
20.01,5
( O i E i )2
i =1
Ei
= 15.09
7)
20 =
(12 9.3984) 2 +
( 9 4.6237) 2
= 6.955
9.3984
4.6237
8) Since 6.955 < 15.09 do not reject H0. We are unable to reject the null hypothesis that the distribution
of the number of flaws is Poisson.
9-41
9-88
Value
Rel. Freq
Observed
(Days)
Expected
(Days)
5
0.067
2
6
0.067
2
8
0.100
3
9
0.133
4
10
0.200
6
11
0.133
4
12
0.133
4
13
0.067
2
14
0.033
1
15
0.067
2
1.0626
1.7942
3.2884
3.7016
3.7501
3.4538
2.9159
2.2724
1.6444
1.1106
Since there are several cells with expected frequencies less than 3, the revised table would be:
Value
Observed
(Days)
Expected
(Days)
5-8
7
9
4
10
6
11
4
12-15
9
6.1452
3.7016
3.7501
3.4538
7.9433
( O i E i )2
i =1
Ei
02
7)
(7 6 . 1452 )2 + (4 3 . 7016 )2 + (6 3 . 7501 )2 + (4 3 . 4538 )2 + (9 7 . 9433 )2 = 1.72
=
6 . 1452
3 . 7016
3 . 7501
3 . 4538
7 . 9433
8) Since 1.72 < 7.81 do not reject H0. We are unable to reject the null hypothesis that the distribution
for the number of calls is Poisson.
b) The P-value is between 0.9 and 0.5 using Table IV. P-value = 0.6325 (found using Minitab)
9-42
9-89
Use the binomial distribution to get the expected frequencies with the mean = np = 6(0.25) = 1.5
Value
Observed
Expected
0
4
8.8989
1
21
17.7979
2
10
14.8315
3
13
6.5918
4
2
1.6479
The expected frequency for value 4 is less than 3. Combine this cell with value 3:
Value
Observed
Expected
0
4
8.8989
1
21
17.7979
2
10
14.8315
3-4
15
8.2397
6) Reject H0 if
2o
>
20.05,3
( O i E i )2
i =1
Ei
= 7.81
7)
20 =
( 4 8.8989) 2 +
8.8989
(15 8.2397) 2
8.2397
= 10.39
8) Since 10.39 > 7.81 reject H0. We can conclude that the distribution is not binomial with n = 6 and p =
0.25 at = 0.05.
b) P-value = 0.0155 (from Minitab)
9-90
p sample =
0
39
38.1426
1
23
26.1571
2
12
8.5952
3
1
1.8010
Since value 3 has an expected frequency less than 3, combine this category with that of value 2:
Value
Observed
Expected
0
39
38.1426
1
23
26.1571
2-3
13
10.3962
9-43
20 =
6) Reject H0 if
2o
>
20.05,1
20 =
( O i E i )2
i =1
Ei
= 384
.
.
( 39 381426
) 2 + (23 26.1571)2 + (13 10.3962) 2
= 1.053
38.1426
26.1571
10.39
8) Since 1.053 < 3.84 do not reject H0. We are unable to reject the null hypothesis that the distribution
of the number of underfilled cartons is binomial at = 0.05.
7)
b) The P-value is between 0.5 and 0.1 using Table IV. From Minitab the P-value = 0.3048.
9-91
6) Reject H0 if
2o
>
20.05,24
( O i E i )2
i =1
Ei
= 36.42
8) Since 769.57 >>> 36.42, reject H0. We can conclude that the distribution is not Poisson at = 0.05.
b) P-value = 0 (found using Minitab)
9-92
P ( X = x) =
e x
where = [6(1) + 7(1) + + 39(1) + 41(1)] / 105 = 19.514
x!
Observed
Frequency
Expected
Frequency
0.03
0.08
0.18
10
0.78
11
1.38
12
2.24
13
3.36
14
4.69
15
11
6.10
16
7.43
17
8.53
18
9.25
19
9.50
20
9.27
9-44
21
8.62
22
7.64
23
6.48
24
5.27
25
4.12
26
3.09
27
2.23
28
1.56
29
1.05
30
0.68
31
0.43
32
0.26
34
0.09
35
36
39
41
1
2
1
1
0.05
0.03
0.00
0.00
After combining categories with frequencies less than 3, we get the following table:
Number of
Earthquake
s
Observed
Frequenc
y
6-7-8-10-1112
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27-28-2930-31-3234-35-3639-41
Expected
Frequenc
y
13.00
6
5
11
7
3
8
4
4
7
8
4
3
2
4
4.68
3.36
4.69
6.10
7.43
8.53
9.25
9.50
9.27
8.62
7.64
6.48
5.27
4.12
3.09
16
6.38
9-45
a)
1) The variable of interest is the form of the distribution for the number of earthquakes per year
of magnitude 7.0 and greater..
2) H0: The form of the distribution is Poisson
3) H1: The form of the distribution is not Poisson
4) = 0.05
5) The test statistic is
20 =
6) Reject H0 if
7)
>
2
o
2
0.05 ,14
( O i E i )2
i =1
Ei
= 23.68
2
(
13 4.86)
=
2
0
4.86
2
(
16 6.38)
+
= 48.89
6.38
8) Since 48.89 > 23.68 reject H0. The form of the distribution of the number of earthquakes is
not Poisson.
b) P-value < 0.005
Section 9-8
9-93
=
2
0
(O
ij
Eij )
Eij
i =1 j =1
= 12.592
2
The calculated test statistic is 0 = 11.65
2
7.
8. 20 >/ 20.05,6 , do not reject H0 and conclude that the data provide insufficient evidence to claim that
machine breakdown and shift are dependent at = 0.05.
P-value = 0.070 (using Minitab)
9-94
=
2
0
i =1 j =1
(O
ij
Eij )
Eij
= 6.637
2
7. The calculated test statistic is 0 = 0.033
2
8.
02 >/ 02.01,1
, do not reject H0 and conclude that the evidence is not sufficient to claim that surgical-
9-46
a)
1) The variable of interest is the form of the distribution for the number of earthquakes per year
of magnitude 7.0 and greater..
2) H0: The form of the distribution is Poisson
3) H1: The form of the distribution is not Poisson
4) = 0.05
5) The test statistic is
20 =
6) Reject H0 if
7)
>
2
o
2
0.05 ,14
( O i E i )2
i =1
Ei
= 23.68
2
(
13 4.86)
=
2
0
4.86
2
(
16 6.38)
+
= 48.89
6.38
8) Since 48.89 > 23.68 reject H0. The form of the distribution of the number of earthquakes is
not Poisson.
b) P-value < 0.005
Section 9-8
9-93
=
2
0
(O
ij
Eij )
Eij
i =1 j =1
= 12.592
2
The calculated test statistic is 0 = 11.65
2
7.
8. 20 >/ 20.05,6 , do not reject H0 and conclude that the data provide insufficient evidence to claim that
machine breakdown and shift are dependent at = 0.05.
P-value = 0.070 (using Minitab)
9-94
=
2
0
i =1 j =1
(O
ij
Eij )
Eij
= 6.637
2
7. The calculated test statistic is 0 = 0.033
2
8.
02 >/ 02.01,1
, do not reject H0 and conclude that the evidence is not sufficient to claim that surgical-
9-46
9-95.
=
2
0
(O
ij
Eij )
Eij
i =1 j =1
= 21.665
2
7. The calculated test statistic is 0 = 25.55
8.
02 > 02.01,9
.01, 9
Therefore, reject H0 and conclude that the grades are not independent at = 0.01.
P-value = 0.002
9-96
=
2
0
i =1 j =1
(O
Eij )
ij
Eij
= 9.488
2
7. The calculated test statistic is 0 = 2.46
8.
02 >/ 02.05, 4
0.05 , 4
, do not reject H0 and conclude that the evidence is not sufficient to claim that the data are
=
2
0
i =1 j =1
(O
ij
Eij )
Eij
= 11.344
2
7. The calculated test statistic is 0 = 10.71
2
8.
02 >/ 02.01,3
, do not reject H0 and conclude that the evidence is not sufficient to claim that the type of
9-47
=
2
0
(O
ij
Eij )
Eij
i =1 j =1
= 7.815
2
7. The calculated test statistic is .0 = 26.97 .
2
8.
02 >>> 02.05,3 , reject H0 and conclude that the opinions on the change are not independent of class
standing. P-value 0
9-99
a)
1. The variable of interest is successes.
2. H0: successes are independent of size of stone.
3. H1: successes are not independent of size of stone.
4. = 0.05
5. The test statistic is:
r
=
2
0
i =1 j =1
(O
ij
Eij )
Eij
>
2
0
2
0.05,1 ,
2
0
reject H0 and conclude that there is enough evidence to claim that number of
=0.01
85 86
= (2.33 0.31) = (2.02) = 0.9783
16 / 25
85 86
= (2.33 0.63) = (1.70) = 0.9554
n=100 =
z 0.01 +
16 / 100
85 86
= (2.33 1.25) = (1.08) = 0.8599
n=400 =
z 0.01 +
16 / 400
85 86
= (2.33 3.13) = (0.80) = 0.2119
n=2500 =
z 0.01 +
16 / 2500
a.) n=25
= z 0.01 +
9-48
=
2
0
(O
ij
Eij )
Eij
i =1 j =1
= 7.815
2
7. The calculated test statistic is .0 = 26.97 .
2
8.
02 >>> 02.05,3 , reject H0 and conclude that the opinions on the change are not independent of class
standing. P-value 0
9-99
a)
1. The variable of interest is successes.
2. H0: successes are independent of size of stone.
3. H1: successes are not independent of size of stone.
4. = 0.05
5. The test statistic is:
r
=
2
0
i =1 j =1
(O
ij
Eij )
Eij
>
2
0
2
0.05,1 ,
2
0
reject H0 and conclude that there is enough evidence to claim that number of
=0.01
85 86
= (2.33 0.31) = (2.02) = 0.9783
16 / 25
85 86
= (2.33 0.63) = (1.70) = 0.9554
n=100 =
z 0.01 +
16 / 100
85 86
= (2.33 1.25) = (1.08) = 0.8599
n=400 =
z 0.01 +
16 / 400
85 86
= (2.33 3.13) = (0.80) = 0.2119
n=2500 =
z 0.01 +
16 / 2500
a.) n=25
= z 0.01 +
9-48
z0 =
b) n=25
n=100
z0 =
n=400
z0 =
n=2500
86 85
= 0.31
16 / 25
86 85
16 / 100
86 85
P-value:
9-101
a.
Sample Size, n
50
p(1 p)
n
Sampling Distribution
Normal
b.
80
Normal
c.
100
Normal
Sample Mean
p
Sample Variance
p(1 p)
50
p(1 p)
80
p(1 p)
100
d) As the sample size increases, the variance of the sampling distribution decreases.
9-102
a.
n
50
b.
100
c.
500
1000
d.
Test statistic
z0 =
0.095 0.10
= 0.12
0.10(1 0.10) / 50
0.095 0.10
z0 =
= 0.15
0.10(1 0.10) / 100
0.095 0.10
z0 =
= 0.37
0.10(1 0.10) / 500
0.095 0.10
z0 =
= 0.53
0.10(1 0.10) / 1000
a) n = 20:
5 20
= (0.163) = 0.564
= 1.96
12
9-49
P-value
0.4522
conclusion
Do not reject H0
0.4404
Do not reject H0
0.3557
Do not reject H0
0.2981
Do not reject H0
5 50
12
b) n = 50:
c) n = 100:
5 100
12
d) , which is the probability of a Type II error, decreases as the sample size increases because the variance
of the sample mean decreases. Consequently, the probability of observing a sample mean in the
acceptance region centered about the incorrect value of 200 ml/h decreases with larger n.
9-104
a) n = 20:
5 20
= (0.362) = 0.6406
= 1.96
14
b) n = 50:
5 50
= (0.565) = 1 (0.565) = 1 0.7123 = 0.2877
= 1.96
14
c) n = 100:
5 100
= (1.611) = 1 (1.611) = 1 0.9463 = 0.0537
= 1.96
14
d) The probability of a Type II error increases with an increase in the standard deviation.
9-105
= 8, = 204 200 = 4,
4 20
= ( 0.28) = 1 (0.28) = 1 0.61026 = 0.38974
a) n = 20: = 196
.
8
4 50
= ( 2.58) = 1 (2.58) = 1 0.99506 = 0.00494
.
b) n = 50: = 196
8
4 100
= ( 3.04) = 1 (3.04) = 1 0.99882 = 0.00118
c) n = 100: = 196
.
8
a) =0.05
n=100
0.5 0.6
= (1.65 2.0) = (0.35) = 0.3632
0.5(0.5) / 100
n=150
= z 0.05 +
= z 0.05 +
0.5 0.6
9-50
0.5 0.6
= (1.65 3.46) = (1.81) = 0.03515
0.5(0.5) / 300
n=300
= z 0.05 +
b) =0.01
0.5 0.6
= (2.33 2.0) = (0.33) = 0.6293
0.5(0.5) / 100
n=100
= z 0.01 +
0.5 0.6
= (2.33 2.45) = (0.12) = 0.4522
0
.
5
(
0
.
5
)
/
100
n=150
= z 0.01 +
0.5 0.6
= (2.33 3.46) = (1.13) = 0.1292
0.5(0.5) / 300
n=300
= z 0.01 +
Decreasing the value of decreases the power of the test for the different sample sizes.
c) =0.05
n=100
0.5 0.8
= (1.65 6.0) = (4.35) 0.0
0.5(0.5) / 100
Power = 1 = 1 0 1
= z 0.05 +
The true value of p has a large effect on the power. The further p is away from p0 the larger the power of the
test.
d)
z / 2 p0 (1 p0 ) z p(1 p)
n=
p p0
0.6 0.5
z / 2 p0 (1 p0 ) z p(1 p)
n =
p p0
0.8 0.5
The true value of p has a large effect on the sample size. The further p is away from p0 the smaller the
sample size that is required.
9-51
9-107
a) Rejecting a null hypothesis provides a stronger conclusion than failing to reject a null hypothesis.
Therefore, place what we are trying to demonstrate in the alternative hypothesis.
Assume that the data follow a normal distribution.
b) 1) the parameter of interest is the mean weld strength, .
2) H0 : = 150
3) H1 : > 150
4) Not given
5) The test statistic is:
t0 =
x 0
s/ n
t0 =
P-value =
153.7 150
11.3 / 20
= 1.46
8) There is some modest evidence to support the claim that the weld strength exceeds 150 psi.
If we used = 0.01 or 0.05, we would not reject the null hypothesis, thus the claim would not be
supported. If we used = 0.10, we would reject the null in favor of the alternative and conclude the
weld strength exceeds 150 psi.
9-108
a)
d=
| 0 | | 73 75 |
=
=
=2
Using the OC curve for = 0.05, d = 2, and n = 10, we get 0.0 and
power of 10.0 1.
d=
| 0 | | 72 75 |
=
=
=3
Using the OC curve for = 0.05, d = 3, and n = 10, we get 0.0 and
power of 10.0 1.
b) d =
| 0 | | 73 75 |
=
=
=2
Using the OC curve, Chart VII e) for = 0.05, d = 2, and 0.1 (Power=0.9),
n* = 5 .
d=
Therefore,
n=
n* + 1 5 + 1
=
=3
2
2
| 0 | | 72 75 |
=
=
=3
Using the OC curve, Chart VII e) for = 0.05, d = 3, and 0.1 (Power=0.9),
n* = 3 .
Therefore,
n=
n* + 1 3 + 1
=
=2
2
2
9-52
c) = 2 .
d=
| 0 | | 73 75 |
=
=
=1
2
Using the OC curve for = 0.05, d = 1, and n = 10, we get 0.10 and
power of 10.10 0.90.
d=
| 0 | | 72 75 |
=
=
= 1.5
2
Using the OC curve for = 0.05, d = 1.5, and n = 10, we get 0.04 and
power of 10.04 0.96.
d=
| 0 | | 73 75 |
=
=
=1
Using the OC curve, Chart VII e) for = 0.05, d = 1, and 0.1 (Power=0.9),
n * = 10 . Therefore, n =
d=
n * + 1 10 + 1
=
= 5.5 n 6
2
2
| 0 | | 72 75 |
=
=
= 1.5
Using the OC curve, Chart VII e) for = 0.05, d = 3, and 0.1 (Power=0.9),
n * = 7 . Therefore, n =
n* + 1 7 + 1
=
=4
2
2
Increasing the standard deviation lowers the power of the test and increases the sample size
required to obtain a certain power.
9-109
20 =
( n 1)s2
2
6) 20.99 ,7 = 124
. reject H0 if 20 < 1.24
7) s = 0.00001 and = 0.01
02 =
7(0.00001) 2
= 1.75
(0.00002) 2
1.75 > 1.24, do not reject the null hypothesis; that is, there is insufficient evidence to conclude the
standard deviation is at most 0.00002 mm.
b) Although the sample standard deviation is less than the hypothesized value of 0.00002, it is not
significantly less (when = 0.01) than 0.00002. The value of 0.00001 could have occurred as a result of
sampling variation.
9-110
9-53
4) not given
20 =
( n 1)s2
2
6) will be determined based on the P-value
7) s = 0.004 and n = 10
02 =
9(0.004) 2
= 0.000009
(4) 2
P value 0.
P-value = P < 0.00009 ;
The P-value is approximately 0, therefore we reject the null hypothesis and conclude that the standard
deviation of the concentration is less than 4 grams per liter.
2
9-111
Create a table for the number of nonconforming coil springs (value) and the observed number of times the
number appeared. One possible table is:
Value
Obs
0
0
1
0
2
0
3
1
4
4
5
3
6
4
7
6
8
4
9
3
10
0
11
3
12
3
13
2
14
1
15
1
16
0
17
2
sample mean =
p sample
Value
Observed
Expected
0
0
0.00165
1
0
0.01889
2
0
0.10608
3
1
0.38911
4
4
1.04816
5
3
2.21073
6
4
3.80118
7
6
5.47765
8
4
6.74985
9
3
7.22141
10
0
6.78777
11
3
5.65869
12
3
4.21619
13
2
2.82541
14
1
1.71190
15
1
0.94191
16
0
0.47237
17
2
0.21659
18
1
0.09103
19
2
0.03515
Since several of the expected values are less than 3, some cells must be combined resulting in the
following table:
Value
0-5
6
7
8
Observed
8
4
6
4
9-54
Expected
3.77462
3.80118
5.47765
6.74985
18
1
19
2
9
10
11
12
13
3
0
3
3
9
7.22141
6.78777
5.65869
4.21619
6.29436
( O i E i )2
i =1
Ei
9-112
Create a table for the number of errors in a string of 1000 bits (value) and the observed number of times the
number appeared. One possible table is:
Value
0
1
2
3
4
5
Obs
3
7
4
5
1
0
The value of p must be estimated. Let the estimate be denoted by psample
sample mean =
p sample
Value
Observed
Expected
0
3
3.64839
1
7
6.21282
2
4
5.28460
3
5
2.99371
4
1
1.27067
5
0
0.43103
Since several of the expected values are less than 3, some cells must be combined resulting in the
following table:
Value
0
1
2
3
Observed
3
7
4
6
Expected
3.64839
6.21282
5.28460
4.69541
The degrees of freedom are k p 1 = 4 1 1 = 2
a) 1) The variable of interest is the form of the distribution for the number of errors in a string of 1000 bits.
2) H0: The form of the distribution is binomial
3) H1: The form of the distribution is not binomial
4) = 0.05
5) The test statistic is
k
(Oi Ei )2
i =1
Ei
=
2
0
9-55
02 =
(3 3.64839)2 +
3.64839
(6 4.69541)2
4.69541
= 0.88971
8) Since 0.88971 < 9.49 do not reject H0. We are unable to reject the null hypothesis that the distribution
of the number of errors is binomial at = 0.05.
b) P-value = 0.6409 (found using Minitab)
9-113
Divide the real line under a standard normal distribution into eight intervals with equal
probability. These intervals are [0,0.32), [0.32, 0.675), [0.675, 1.15), [1.15, ) and their negative
counterparts. The probability for each interval is p = 1/8 = 0.125 so the expected cell frequencies
are E = np = (100) (0.125) = 12.5. The table of ranges and their corresponding frequencies is
completed as follows.
Interval Obs. Frequency. Exp. Frequency.
x 5332.5
1
12.5
5332.5< x 5357.5
4
12.5
5357.5< x 5382.5
7
12.5
5382.5< x 5407.5
24
12.5
5407.5< x 5432.5
30
12.5
5432.5< x 5457.5
20
12.5
5457.5< x 5482.5
15
12.5
x 5482.5
5
12.5
(15 - 12.5)
12 . 5
( 5 12 . 5 ) 2
= 63 . 36
12 . 5
9-114
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying distribution
is normal.
1) The parameter of interest is the true mean concentration of suspended solids, .
2) H0 : = 50
3) H1 : < 50
4) = 0.05
5) Since n>>30 we can use the normal distribution
z0 =
s/ n
59.87 50
12.50 / 60
= 6.12
8) Since 6.12>-1.65, do not reject the null hypothesis and conclude there is insufficient evidence to indicate that
the true mean concentration of suspended solids is less than 50 ppm at = 0.05.
9-56
c) We can divide the real line under a standard normal distribution into eight intervals with equal
probability. These intervals are [0,.32), [0.32, 0.675), [0.675, 1.15), [1.15, ) and their negative
counterparts. The probability for each interval is p = 1/8 = .125 so the expected cell frequencies
are E = np = (60) (0.125) = 7.5. The table of ranges and their corresponding frequencies is
completed as follows.
Interval Obs. Frequency. Exp. Frequency.
x 45.50 9
7.5
45.50< x 51.43 5
7.5
51.43< x 55.87 7
7.5
55.87< x 59.87 11
7.5
59.87< x 63.87 4
7.5
63.87< x 68.31 9
7.5
68.31< x 74.24 8
7.5
x 74.24
6
7.5
The test statistic is:
2o =
(9 7.5) 2 (5 7.5) 2
+
+
7.5
7.5
(8 7.5) 2 (6 7.5) 2
+
= 5.06
7.5
7.5
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying distribution
is normal.
1) The parameter of interest is the true mean overall distance for this brand of golf ball, .
2) H0 : = 270
3) H1 : < 270
4) = 0.05
5) Since n>>30 we can use the normal distribution
z0 =
s/ n
260.30 270.0
13.41 / 100
= 7.23
8) Since 7.23<-1.65, reject the null hypothesis and conclude there is sufficient evidence to indicate that the
true mean distance is less than 270 yard at = 0.05.
b) The P-value 0
c) We can divide the real line under a standard normal distribution into eight intervals with equal
probability. These intervals are [0,.32), [0.32, 0.675), [0.675, 1.15), [1.15, ) and their
negative counterparts. The probability for each interval is p = 1/8 = .125 so the expected cell
frequencies are E = np = (100) (0.125) = 12.5. The table of ranges and their corresponding
frequencies is completed as follows.
Interval Obs. Frequency. Exp. Frequency.
x 244.88
16
12.5
9-57
244.88< x 251.25
251.25< x 256.01
256.01< x 260.30
260.30< x 264.59
264.59< x 269.35
269.35< x 275.72
x 275.72
6
17
9
13
8
19
12
12.5
12.5
12.5
12.5
12.5
12.5
12.5
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying distribution
is normal.
1) The parameter of interest is the true mean coefficient of restitution, .
2) H0 : = 0.635
3) H1 : > 0.635
4) = 0.01
5) Since n>30 we can use the normal distribution
z0 =
s/ n
0.624 0.635
0.0131 / 40
= 5.31
8) Since 5.31< 2.33, do not reject the null hypothesis and conclude there is insufficient evidence to indicate
that the true mean coefficient of restitution is greater than 0.635 at = 0.01.
b) The P-value
(5.31) 1
c.) If the lower bound of the CI was above the value 0.635 then we could conclude that the mean coefficient
of restitution was greater than 0.635.
9-117
a) In order to use t statistics in hypothesis testing, we need to assume that the underlying distribution
is normal. Use the t-test to test the hypothesis that the true mean is 2.5 mg/L.
1) State the parameter of interest: The parameter of interest is the true mean dissolved oxygen level, .
2) State the null hypothesis H0 : = 2.5
3) State the alternative hypothesis H1 : 2.5
4) Give the significance level = 0.05
5) Give the statistic
t0 =
s/ n
9-58
x
s/ n
s/ n
3.265 2.5
2.127 / 20
= 1.608
8) Since 1.608 < 2.093, do not reject the null hypotheses and conclude that the true mean is not significantly
different from 2.5 mg/L
c) The value of 1.608 is found between the columns of 0.05 and 0.1 of table V. Therefore the P-value is
between 0.1 and 0.2. Minitab provides a value of 0.124
d) The confidence interval found in exercise 8-81 b. agrees with the hypothesis test above. The value of 2.5 is
within the 95% confidence limits. The confidence interval shows that the interval is quite wide due to the large
sample standard deviation value.
x t 0.025,19
3.265 2.093
n
2.127
x + t 0.025,19
3.265 + 2.093
20
2.270 4.260
2.127
20
9-118 a)
1) The parameter of interest is the true mean sugar concentration, .
2) H0 : = 11.5
3) H1 : 11.5
4) = 0.05
5) t0 =
x
s/ n
t0 =
11.47 11.5
0.022 / 20
= 6.10
8) Since 6.10 > 2.093, reject the null hypothesis and conclude that there is sufficient evidence that
the true mean sugar concentration is different from 11.5 at = 0.05.
From table V the t0 value in absolute value is greater than the value corresponding to 0.0005 with
19 degrees of freedom. Therefore 2*0.0005 = 0.001>P-value
9-59
b) d =
| 0 | | 11.4 11.5 |
=
=
= 4.54
0.022
Using the OC curve, Chart VII e) for = 0.05, d = 4.54, and n = 20 we find
0 and the Power 1.
c) d =
| 0 | | 11.45 11.5 |
=
=
= 2.27
0.022
Using the OC curve, Chart VII e) for = 0.05, d = 2.27, and 1->0.9 (<0.1),
We find that n should be at least 5.
d) 95% two sided confidence interval
s
s
x t 0.025,19
x + t 0.025,19
n
n
0.022
0.022
11.47 2.093
11.47 + 2.093
20
20
11.46 11.48
We can conclude that the mean sugar concentration content is not equal to 11.5 because that value
is not inside the confidence interval.
e) The normality plot below indicates that the normality assumption is reasonable.
Probability Plot of Sugar Concentration
Normal
99
95
90
Percent
80
70
60
50
40
30
20
10
5
11.40
9-119
11.42
11.44
11.46
11.48
Sugar Concentration
11.50
a)
1) The parameter of interest is the true mean body weight, .
2) H0 : = 300
3) H1 : 300
4) = 0.05
5) t0 =
x
s/ n
9-60
11.52
t0 =
325.5 300
198.79 / 27
= 0.667
8) Since 0.667< 2.056, we fail to reject the null hypothesis and conclude that there is not sufficient
evidence that the true mean body weight is different from 300 at = 0.05.
b) From table V the t0 value is found between the values of 0.25 and 0.4 with 26 degrees of
freedom, so 0.5 < P-value < 0.8 The smallest level of significance at which we would willing to
reject the null hypothesis is the P-value
c) 95% two sided confidence interval
s
s
x t 0.025,19
x + t 0.025,19
n
n
198.79
198.79
325.5 2.056
325.5 2.056
27
27
246.84 404.16
There is not enough evidence to conclude that the mean body weight differs from 300 because 300
is within the 95% confidence interval.
9-120
a)
1) The parameter of interest is the true mean percent protein, .
2) H0 : = 80
3) H1 : > 80
4) = 0.05
5) t0 =
s/ n
80.68 80
= 0.37
7.38 / 16
8) Since 0.37 < 1.753, do not reject the null hypothesis and conclude that there is not
sufficient evidence to indicate that the true mean percent protein is greater than 80 at =
0.05.
b) From the normal probability plot, the normality assumption seems reasonable:
9-61
95
90
Percent
80
70
60
50
40
30
20
10
5
60
70
80
percent protein
90
100
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we
need to assume that the underlying distribution is normal.
1) The parameter of interest is the true variance of tissue assay, 2.
2) H0 : 2 = 0.6
3) H1 : 2 0.6
4) = 0.01
5) 20 =
( n 1)s2
2
02.005,11 = 26.76
7) n = 12, s = 0.758
20 =
(n 1) s 2
11(0.758) 2
= 10.53
0.6
8) Since 2.6< 10.53 <26.76 we fail to reject H0 and conclude there is not sufficient evidence to
indicate the true variance of tissue assay is significantly different from 0.6 at = 0.01.
b) P-value: 0.1<p-value/2<0.5 then 0.2<p-value<1
c) 99% confidence interval for , first find the confidence interval for 2
For = 0.05 and n = 12,
11(0.758) 2
11(0.758) 2
2
26.76
2.60
0.236 2 2.43
Because 0.6 falls within the 99% confidence bounds we would conclude that we dont have
enough evidence to conclude that the population variance is different from 0.6
9-122
9-62
a) In order to use the 2 statistic in hypothesis testing and confidence interval construction, we
need to assume that the underlying distribution is normal.
1) The parameter of interest is the true variance of the ratio between the numbers of
symmetrical and total synapses, 2.
2) H0 : 2 = 0.02
3) H1 : 2 0.02
4) = 0.05
5) 20 =
( n 1)s2
2
02.025,30 = 46.98
7) n = 31, s = 0.198
20 =
(n 1) s 2
30(0.198) 2
= 58.81
0.02
8) Because 58.81 > 46.98 reject H0 and conclude that the true variance of the ratio between the
numbers of symmetrical and total synapses is significantly different from 0.02 at = 0.05.
b) P-value/2<0.005 so that P-value<0.01
9-123
a)
1) The parameter of interest is the true mean of cut-on wave length, .
2) H0 : = 6.5
3) H1 : 6.5
4) A value for is not given. We assume that = 0.05.
5) t0 =
x
s/ n
t0 =
6.55 6.5
= 0.47
0.35 / 11
8) Since 0.47< 2.228, we fail to reject the null hypothesis and conclude that there is not
sufficient
evidence that the true mean of cut-on wave length is different from 6.5 at = 0.05.
b) From table V the t0 value is found between the values of 0.25 and 0.4 with 10 degrees of
freedom, so 0.5<P-value<0.8
c)
d=
| 0 | | 6.25 6.5 |
=
=
= 0.71
0.35
Using the OC curve, Chart VII e) for = 0.05, d = 0.71, and 1- > 0.95 (<0.05),
We find that n should be at least 30.
d)
d=
| 0 | | 6.95 6.5 |
=
=
= 1.28
0.35
Using the OC curve, Chart VII e) for = 0.05, n=11, d = 1.28, we find 0.1
9-63
9-124
( n 1)s2
2
02.995,5 = 0.41 reject H0 if 02 < 0.41 or 02.005,5 = 16.75 reject H0 if 02 > 16.75
7) n = 6, s = 0.319
5(0.319) 2
=
= 0.509
12
2
0
2
0.995, 50
20 =
( n 1)s2
2
or
7) n = 51, s = 0.319
02 =
50(0.319) 2
= 5.09
12
20 =
( n 1)s2
2
6) Since no critical value is given, we will calculate the p-value
7) n = 10, s = 15.7
20 =
9(15.7) 2
= 5546
.
400
01
. < P value < 0.5
9-64
8) The P-value is greater than any acceptable significance level . Therefore, we do not reject the null
hypothesis. There is insufficient evidence to support the claim that the standard deviation is less than
20 microamps.
b) 7) n = 51, s = 20
20 =
50(15.7) 2
= 30.81
400
8) The P-value is less than 0.05, therefore we reject the null hypothesis and conclude that the standard
deviation is significantly less than 20 microamps.
c) Increasing the sample size increases the test statistic 20 and therefore decreases the P-value, providing
more evidence against the null hypothesis.
9-126
9-127
X 0
X 0
X 0 X 0
<
= | = 0 ) + P(
>
| = 0 ) P
/ n
/ n
/ n / n
P( z0 < z ) + P( z0 > z ) = ( z ) + 1 ( z )
P(
= (( )) + (1 (1 )) =
b) = P(z
z when
1 = 0 + d )
or = P( z < Z 0 < z | 1 = 0 + )
= P( z <
x0
< z | 1 = 0 + )
2 /n
= P( z
< Z < z
)
2
/n
2 /n
= ( z
) ( z
)
2
/n
2 /n
9-128
/n
9-65
Mind-Expanding Exercises
11-91
The correlation coefficient for the n pairs of data (xi, zi) will not be near unity. It will be near zero. The data for the
pairs (xi, zi) where
z i = y i2
y i = xi
y i = xi
correlation coefficient.
11-92
a)
1 =
S xY
S xx
0 = Y 1 x
2
Cov(Y , SxY ) Cov(Yi , Yi ( xi x )) ( xi x )
Cov(Y , 1 ) =
=
=
= 0.
S xx
nSxx
nSxx
Cov(1, 1 ) = V (1 ) =
S xx
2
x 2
Cov( 0 , 1 ) =
S xx
b) The requested result is shown in part a).
11-93
a)
(Y
MS E =
1 xi ) 2
2
i
n2
n2
E (ei ) = E (Yi ) E ( 0 ) E ( 1 ) xi = 0
V (ei ) = 2 [1 ( 1n +
E ( MS E ) =
)]
Therefore,
E (e ) = V (e )
2
i
n2
=
=
( xi x ) 2
S xx
n2
[1 ( n1 +
( xi x ) 2
S xx
)]
n2
2 [n 1 1]
n2
=2
E ( MS R ) = E ( 12 S xx ) = S xx V ( 1 ) + [ E ( 1 )]2
= S xx
+ 12 = 2 + 12 S xx
S xx
2
11-94
1 =
S x1Y
S x1 x1
11-59
Therefore,
E Yi ( x1i x1 )
=
E ( 1 ) = i =1
S x1 x1
(
i =1
+ 1 x1i + 2 x2 i )( x1i x1 )
S x1 x1
1S x x + 2 x ( x x1 )
2i
1 1
1i
i =1
S x1 x1
= 1 +
2Sx x
1 2
S x1 x1
V ( 1 ) =
S xx
2
11-95
. To minimize
V ( 1 ), S xx
S xx = ( xi x ) 2 , Sxx is
i =1
maximized by choosing approximately half of the observations at each end of the range of x.
From a practical perspective, this allocation assumes the linear model between Y and x holds throughout the range of x
and observing Y at only two x values prohibits verifying the linearity assumption. It is often preferable to obtain some
observations at intermediate values of x.
n
11-96
w (y
i =1
1 xi ) 2
x
)
=
2
wi ( yi 0 1 xi )
i i 0 1i
0 i =1
i =1
n
n
2
w
(
y
x
)
=
2
wi ( yi 0 1 xi )xi
i i 0 1i
1 i =1
i =1
0 wi + 1 wi xi = wi yi
0 wi xi + 1 wi xi 2 = wi xi yi
and these equations are solved as follows
1=
( w x y )( w ) w y
( w )( w x ) ( w x )
i
0 =
11-97
y = y + r
sy
sx
i
2
w y w x
w w
i
2
i
(x x)
11-60
= y+
S xy
(y
(x
S xx S yy
= y+
y) 2 ( x x )
x)2
S xy
(x x)
S xx
= y + 1 x 1 x = 0 + 1 x
11-98
a)
n
n
2
(
y
x
)
=
2
( yi 0 1 xi )xi
i 0 1i
1 i =1
i =1
0 xi + 1 xi 2 = xi yi
Therefore,
1 =
x y x
x
i
0
2
b) V ( 1 )
c)
x (y )
x
i
xi (Yi 0) xi 2 2
2
=
=V
=
xi 2 [ xi 2 ]2 xi 2
1 t / 2,n 1
2
2
xi
( xi x ) 2
freedom is slightly smaller than the corresponding t value based on n-2 degrees of freedom.
11-61
CHAPTER 11
Section 11-2
11-1
a)
y i = 0 + 1 xi + i
S xx = 157.42 4314 = 25.348571
2
14
14
SS R = 1S xy = 2.3298017(59.057143)
= 137.59
SS E = S yy SS R
= 159.71429 137.59143
= 22.123
SS E
22.123
2 = MS E =
=
= 1.8436
n2
12
b) y = 0 + 1 x
y = 48.012962 2.3298017(4.3) = 37.99
c) y = 48.012962 2.3298017(3.7) = 39.39
d) e = y y = 46.1 39.39 = 6.71
11-2
a)
y i = 0 + 1 xi + i
2
S xx = 143215.8 1478
= 33991.6
20
)(12.75 )
S xy = 1083.67 (147820
= 141.445
1 =
S xy
S xx
141.445
= 0.00416
33991.6
11-1
0.8
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0
-50
50
100
x
y = 0.32999 + 0.00416(85) = 0.6836
c) y = 0.32999 + 0.00416(90) = 0.7044
d) = 0.00416
b)
11-3
Coef
-5.558
12.652
SE Coef
9.159
1.243
R-Sq = 78.7%
T
-0.61
10.18
P
0.549
0.000
R-Sq(adj) = 78.0%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
28
29
SS
3378.5
913.7
4292.2
MS
3378.5
32.6
F
103.53
y i = 0 + 1 xi + i
(219.55) 2
= 21.106
30
(219.55)(2611)
S xy = 19375.21
= 267.037
30
S xy 267.037
1 =
=
= 12.652
S xx
21.106
S xx = 1627.847
2611
219.55
(12.652)
= 5.56
30
30
y = 5.56 + 12.652 x
SS E
913.7
2 = MS E =
=
= 32.6
n2
28
0 =
11-2
P
0.000
120
5.71252
78.7%
78.0%
Rating Pts
110
100
90
80
70
60
5
7
Yds per Att
1 = 12.652
1
d)
10 = 0.79
12.652
e) y = 5.56 + 12.652(7.21) = 85.66
e = y y
c)
a)
Regression Analysis - Linear model: Y = a+bX
Dependent variable: SalePrice
Independent variable: Taxes
-------------------------------------------------------------------------------Standard
T
Prob.
Parameter
Estimate
Error
Value
Level
Intercept
13.3202
2.57172
5.17948
.00003
Slope
3.32437
0.390276
8.518
.00000
-------------------------------------------------------------------------------Analysis of Variance
Source
Sum of Squares
Df Mean Square
F-Ratio Prob. Level
Model
636.15569
1
636.15569
72.5563
.00000
Residual
192.89056
22
8.76775
-------------------------------------------------------------------------------Total (Corr.)
829.04625
23
Correlation Coefficient = 0.875976
R-squared = 76.73 percent
Stnd. Error of Est. = 2.96104
2 = 8.76775
If the calculations were to be done by hand, use Equations (11-7) and (11-8).
y = 13.3202 + 3.32437 x
b) y = 13.3202 + 3.32437(7.5) = 38.253
c)
11-3
d) All the points would lie along a 45 degree line. That is, the regression model would estimate the values exactly. At
this point, the graph of observed vs. predicted indicates that the simple linear regression model provides a reasonable
fit to the data.
Plot of Observed values versus predicted
50
Predicted
45
40
35
30
25
25
30
35
40
45
Observed
11-4
50
11-5
a)
Regression Analysis - Linear model: Y = a+bX
Dependent variable: Usage
Independent variable: Temperature
-------------------------------------------------------------------------------Standard
T
Prob.
Parameter
Estimate
Error
Value
Level
Intercept
-6.3355
1.66765
-3.79906
.00349
Slope
9.20836
0.0337744
272.643
.00000
-------------------------------------------------------------------------------Analysis of Variance
Source
Sum of Squares
Df Mean Square
F-Ratio Prob. Level
Model
280583.12
1
280583.12
74334.4
.00000
Residual
37.746089
10
3.774609
-------------------------------------------------------------------------------Total (Corr.)
280620.87
11
Correlation Coefficient = 0.999933
R-squared = 99.99 percent
Stnd. Error of Est. = 1.94284
2 = 3.7746
If the calculations were to be done by hand, use Equations (11-7) and (11-8).
y = 6.3355 + 9.20836 x
b) y = 6.3355 + 9.20836(55) = 500.124
d)
a)
The regression equation is
MPG = 39.2 - 0.0402 Engine Replacement
Predictor
Constant
Engine Replacement
S = 3.74332
Coef
39.156
-0.040216
R-Sq = 59.1%
SE Coef
2.006
0.007671
T
19.52
-5.24
P
0.000
0.000
R-Sq(adj) = 57.0%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
19
20
SS
385.18
266.24
651.41
MS
385.18
14.01
F
27.49
2 = 14.01
y = 39.2 0.0402 x
b) y = 39.2 0.0402(150) = 33.17
c) y = 34.2956
e = y y = 41.3 34.2956 = 7.0044
11-5
P
0.000
11-7
a)
60
50
40
850
950
1050
Predictor
Coef
StDev
T
P
Constant
-16.509
9.843
-1.68
0.122
x
0.06936
0.01045
6.64
0.000
S = 2.706
R-Sq = 80.0%
R-Sq(adj) = 78.2%
Analysis of Variance
Source
DF
SS
MS
F
P
Regression
1
322.50
322.50
44.03
0.000
Error
11
80.57
7.32
Total
12
403.08
2 = 7.3212
y = 16.5093 + 0.0693554 x
e = y y = 1.39592
b) y = 46.6041
c) y = 16.5093 + 0.0693554(950) = 49.38
a)
9
8
7
6
5
11-8
4
3
2
1
0
60
70
80
90
100
Yes, a linear regression would seem appropriate, but one or two points might be outliers.
Predictor
Constant
x
S = 1.318
Coef
SE Coef
T
P
-10.132
1.995
-5.08
0.000
0.17429
0.02383
7.31
0.000
R-Sq = 74.8%
R-Sq(adj) = 73.4%
Analysis of Variance
Source
DF
Regression
1
Residual Error
18
Total
19
b)
2 = 1.737
c)
y = 4.68265
and
SS
92.934
31.266
124.200
MS
92.934
1.737
y = 10.132 + 0.17429 x
at x = 85
11-6
F
53.50
P
0.000
11-9
a)
250
200
150
100
10
20
30
40
b)
c)
d)
a)
40
30
11-10
20
10
0
0.0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
Yes, a simple linear regression model seems appropriate for these data.
Predictor
Constant
x
S = 3.716
Coef
StDev
T
P
0.470
1.936
0.24
0.811
20.567
2.142
9.60
0.000
R-Sq = 85.2%
R-Sq(adj) = 84.3%
Analysis of Variance
Source
Regression
Error
Total
DF
1
16
17
SS
1273.5
220.9
1494.5
MS
1273.5
13.8
F
92.22
11-7
P
0.000
2 = 13.81
y = 0.470467 + 20.5673 x
c)
d)
y = 10.1371 e = 1.6629
a)
2600
2100
1600
0
10
15
20
25
Yes, a simple linear regression (straight-line) model seems plausible for this situation.
Predictor
Constant
x
S = 99.05
Coef
2625.39
-36.962
StDev
45.35
2.967
R-Sq = 89.6%
Analysis of Variance
Source
DF
SS
Regression
1
1522819
Error
18
176602
Total
19
1699421
b)
c)
T
57.90
-12.46
P
0.000
0.000
R-Sq(adj) = 89.0%
MS
1522819
9811
F
155.21
P
0.000
2 = 9811.2
y = 2625.39 36.962 x
y = 2625.39 36.962(20) = 1886.15
D
d) If there were no error, the values would all lie along the 45 line. The plot indicates age is reasonable
regressor variable.
2600
2500
2400
2300
FITS1
11-11
b)
2200
2100
2000
1900
1800
1700
1600
2100
2600
11-8
11-12
a)
The regression equation is
Porosity = 55.6 - 0.0342 Temperature
Predictor
Constant
Temperature
S = 8.79376
Coef
55.63
-0.03416
SE Coef
32.11
0.02569
R-Sq = 26.1%
T
1.73
-1.33
P
0.144
0.241
R-Sq(adj) = 11.3%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
5
6
SS
136.68
386.65
523.33
MS
136.68
77.33
F
1.77
P
0.241
y = 55.63 0.03416 x
2 = 77.33
b)
c)
y = 4.39 e = 7.012
d)
Scatterplot of Porosity vs Temperature
30
25
Porosity
20
15
10
5
0
1100
1200
1300
Temperature
1400
1500
The simple linear regression model doesnt seem appropriate because the scatter plot doesnt show a linear
relationship between the data.
11-13
Coef
0.6578
0.17806
S = 0.287281
SE Coef
0.1657
0.01400
R-Sq = 94.7%
Analysis of Variance
Source
DF
SS
Regression
1 13.344
T
3.97
12.72
P
0.003
0.000
R-Sq(adj) = 94.1%
MS
13.344
F
161.69
11-9
P
0.000
Residual Error
Total
9
10
y = 0.658 + 0.178 x
2 = 0.083
b)
0.743
14.087
0.083
c) 0.178(3) = 0.534
d)
e)
Scatterplot of y-hat vs y
4.5
4.0
3.5
y-hat
3.0
2.5
2.0
1.5
1.0
0.5
1.0
1.5
2.0
2.5
3.0
3.5
4.0
All the points would lie along the 45 degree line y = y . That is, the regression model would estimate the
values exactly. At this point, the graph of observed vs. predicted indicates that the simple linear regression
model provides a reasonable fit to the data.
11-14
a)
The regression equation is
Deflection = 32.0 - 0.277 Stress level
Predictor
Constant
Stress level
Coef
32.049
-0.27712
SE Coef
2.885
0.04361
S = 1.05743
R-Sq = 85.2%
T
11.11
-6.35
P
0.000
0.000
R-Sq(adj) = 83.1%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
7
8
SS
45.154
7.827
52.981
MS
45.154
1.118
F
40.38
2 = 1.118
11-10
P
0.000
Stress level
70
65
60
55
50
11
12
13
b) y = 32.05 0.277(65)
c) (-0.277)(5) = -1.385
14
15
Deflection
16
17
18
19
= 14.045
1
= 3.61
0.277
e) y = 32.05 0.277(68) = 13.214 e = y y = 11.640 13.214 = 1.574
d)
11-15
Scatterplot of y vs x
2.7
2.6
2.5
2.4
y
2.3
2.2
2.1
2.0
1.9
1.8
0
10
15
20
25
30
35
Its possible to fit this data with straight-line model, but its not a good fit. There is a curvature shown on
the scatter plot.
a)
The regression equation is
y = 2.02 + 0.0287 x
Predictor
Constant
x
Coef
2.01977
0.028718
S = 0.159159
SE Coef
0.05313
0.003966
R-Sq = 67.7%
T
38.02
7.24
P
0.000
0.000
R-Sq(adj) = 66.4%
Analysis of Variance
Source
DF
SS
MS
11-11
Regression
Residual Error
Total
1
25
26
1.3280
0.6333
1.9613
1.3280
0.0253
52.42
0.000
y = 2.02 + 0.0287 x
2 = 0.0253
b) y = 2.02 + 0.0287(11) = 2.3357
c)
Scatterplot of y vs y-hat
2.7
2.6
2.5
2.4
y
2.3
2.2
2.1
2.0
1.9
1.8
2.0
2.1
2.2
2.3
2.4
2.5
y-hat
2.6
2.7
2.8
2.9
If the relationship between length and age was deterministic, the points would fall on the 45 degree
line y = y . Because the points in this plot vary substantially from this line, it appears that age is not a
reasonable choice for the regressor variable in this model.
11-16
11-17
a)
11-18
11-19
0 + 1 x = ( y 1 x ) + 1 x = y
a) The slopes of both regression models will be the same, but the intercept will be shifted.
b) y = 2132.41 36.9618 x
0 = 2625.39
1 = 36.9618
vs.
0 = 2132.41
1 = 36.9618
11-12
(y
2 ( y i xi ) ( xi ) = 2[ y i xi + xi ] = 0
2
Therefore,
b)
yx
x
i i
2
i
y = 21.031461x .
45
40
35
30
chloride
11-20
25
20
15
10
0
0
0.2
0.4
0.6
0.8
1.2
watershed
11-13
1.4
1.6
1.8
Section 11-4
11-21
f0 =
MS R
SS R / 1
=
MS E SS E /( n 2)
SS R = 1S xy = 2.3298017(59.057143)
= 137.59
SS E = S yy SS R
= 159.71429 137.59143
= 22.123
f0 =
137.59
= 74.63
22.123 / 12
8) Since 74.63 > 4.75 reject H 0 and conclude that compressive strength is significant in predicting intrinsic
permeability of concrete at = 0.05. We can therefore conclude that the model specifies a useful linear
relationship between these two variables.
P value 0.000002
b)
2 = MS E =
1
x2
+
=
n S xx
c) se ( ) =
0
11-22
SS E
22.123
=
= 1.8436
n2
12
and se ( 1 ) =
2
S xx
1 . 8436
= 0 .2696
25 . 3486
1
3 . 0714 2
1 . 8436
+
= 0 . 9043
14 25 . 3486
f0 =
MS R
SS R / 1
=
MS E
SS E /( n 2 )
1 xy
= 0.5886
SS E = S yy SS R
2
.75 ) 0.5886
= (8.86 1220
= 0.143275
f0 =
0.5886
= 73.95
0.143275 / 18
8) Since 73.95 > 4.414, reject H 0 and conclude the model specifies a useful relationship at = 0.05.
P value 0.000001
11-14
b)
2
.00796
se( 1 ) =
=
= 4.8391x10 4
S xx
33991.6
1
1
73.9 2
x
=
.
00796
+
se( 0 ) = 2 +
= 0.04091
20 33991.6
n S xx
11-23
a)
The regression equation is
Rating Pts = - 5.558 + 12.65 Yds per Att
S = 5.71252
R-Sq = 78.7%
R-Sq(adj) = 78.0%
Analysis of Variance
Source
Regression
Error
Total
DF
1
28
29
SS
3378.53
913.72
4292.25
MS
3378.53
32.63
F
103.53
P
0.000
Refer to ANOVA
H 0 : 1 = 0
H 1 : 1 0
= 0.01
Because p-value = 0.000 < = 0.01, reject H0. We can conclude that there is a linear relationship between
these two variables.
2
b) = 32.63
se( 1 ) =
2
S xx
32.63
= 1.243
21.106
1 x 2
1 7.318 2
=
32
.
63
se( 0 ) = 2 +
+
= 9.159
30 21.106
n S xx
c) 1)The parameter of interest is the regressor variable coefficient, 1.
2) H 0 :1 = 10
3) H 1:1 10
4) = 0.01
5) The test statistic is t 0 =
1 1,0
se( 1 )
t0 =
12.652 10
= 2.134
1.243
8) Because 2.134 < 2.763, fail to reject H 0 . There is not enough evidence to conclude that the slope
differs from 10 at = 0.01.
11-24
11-15
3) H 1:1 0
4) = 0.05, using t-test
5) The test statistic is t 0 =
1
se( 1 )
MSR
SS R / 1
=
MS E SSE / ( n 2)
c) se (
)=
1
)=
se (
0
=
S xx
8 . 7675
= . 39027
57 . 5631
1
x
2 +
=
n
S
xx
1
6 . 4049 2
8 . 7675
+
= 2 . 5717
24 57 . 5631
H 0 : 0 = 0
H1 : 0 0
se( 0 )
t0 =
13.3201
= 5.179
2.5717
8) Since 5.179 > 2.074 reject H 0 and conclude the intercept is not zero at = 0.05.
11-16
11-25
f0 =
MS R
SS R / 1
=
MS E SS E /(n 2)
f0 =
280583.12 / 1
= 74334.4
37.746089 / 10
8) Since 74334.4 > 10.049, reject H 0 and conclude the model is useful = 0.01. P-value < 0.000001
b) se( 1 ) = 0.0337744, se( 0 ) = 1.66765
c) 1) The parameter of interest is the regressor variable coefficient, 1.
2) H 0 :1 = 10
3) H 1:1 10
4) = 0.01
5) The test statistic is
t0 =
1 1,0
se( 1 )
t0 =
9.21 10
= 23.37
0.0338
8) Since 23.37 < 3.17 reject H 0 and conclude the slope is not 10 at = 0.01. P-value 0.
d) H0: 0 = 0
H1: 0 0
t0 =
6.3355 0
= 3.8
1.66765
P-value < 0.005. Reject H0 and conclude that the intercept should be included in the model.
11-26
H 0 : 1 = 0
H 1 : 1 0
= 0.01
f 0 = 4.53158
f 0.01,1,18 = 8.285
f 0 >/ f ,1,18
Therefore, do not reject H0. P-value = 0.04734. Insufficient evidence to conclude that the model is a useful
relationship.
b)
) = 0.0166281
se(
1
se( ) = 2.61396
0
11-17
c)
H 0 : 1 = 0.05
H 1 : 1 < 0.05
= 0.01
0.0354 ( 0.05 )
= 0.87803
0.0166281
t .01,18 = 2.552
t0 =
t 0 </ t ,18
Therefore, do not reject H0. P-value = 0.804251. Insufficient evidence to conclude that 1 is < -0.05.
d)
H 0 : 0 = 0 H1 : 0 0
= 0.01
t 0 = 12.8291
t .005,18 = 2.878
t 0 > t / 2,18
Therefore, reject H0. P-value 0
11-27
H1 : 1 0
= 0.05
f 0 = 44.0279
f .05 ,1,11 = 4.84
f 0 > f ,1,11
) = 0.0104524
se(
1
se( ) = 9.84346
0
c)
H 0 : 0 = 0
H1 : 0 0
= 0.05
t0 = 1.67718
t.025 ,11 = 2.201
| t0 |</ t / 2 ,11
Therefore, do not rejectH0. P-value = 0.12166.
11-18
11-28
H 0 : 1 = 0
H 1 : 1 0
= 0.05
f 0 = 53.50
f .05,1,18 = 4.414
f 0 > f ,1,18
) = 0.0256613
se(
1
se( ) = 2.13526
0
c)
H 0 : 0 = 0
H1 : 0 0
= 0.05
t0 = - 5.079
t.025,18 = 2.101
| t0 |> t / 2,18
Therefore, reject H0. P-value = 0.000078.
11-29
H 0 : 1 = 0
H 1 : 1 0
= 0.01
f 0 = 155.2
f .01,1,18 = 8.285
f 0 > f ,1,18
) = 45.3468
se(
1
se( ) = 2.96681
0
c)
H 0 : 1 = 30
H 1 : 1 30
= 0.01
36 .9618 ( 30 )
= 2.3466
2.96681
t .005 ,18 = 2.878
t0 =
| t 0 |>/ t / 2 ,18
Therefore, do not reject H0. P-value = 0.0153(2) = 0.0306.
d)
H 0 : 0 = 0
H1 : 0 0
= 0.01
11-19
t0 = 57.8957
t.005 ,18 = 2.878
t 0 > t / 2,18 , therefore, reject H0. P-value < 0.00001.
e) H0 : 0 = 2500
H1 : 0 > 2500
= 0.01
2625.39 2500
= 2.7651
45.3468
t.01,18 = 2.552
t0 =
H 0 : 1 = 0
H 1 : 1 0
= 0.01
f 0 = 92.224
f .01,1,16 = 8.531
f 0 > f ,1,16
Therefore, reject H 0 .
b) P-value < 0.00001
c)
) = 2.14169
se(
1
se( ) = 1.93591
0
d)
H 0 : 0 = 0
H1 : 0 0
= 0.01
t0 = 0.243
t.005 ,16 = 2.921
t0 >/ t / 2 ,16
Therefore, do not reject H0. There is not sufficient evidence to conclude that the intercept differs from zero. Based
on this test result, the intercept could be removed from the model.
11-20
11-31
H 0 : 1 = 0
H 1 : 1 0
= 0.01
Because the P-value = 0.000 < = 0.01, reject H0. There is evidence of a linear relationship between these
two variables.
b) = 0.083
The standard errors for the parameters can be obtained from the computer output or calculated as follows.
2
se( 1 ) =
=
S xx
0.083
= 0.014
420.91
1 10.09 2
x2
2 1
se( 0 ) = +
= 0.083 +
= 0.1657
11 420.91
n S xx
c)
1) The parameter of interest is the intercept 0.
2) H 0 : 0 = 0
3) H 1 : 0 0
4)
= 0.01
0
se( 0 )
6) Reject H0 if t 0 < t / 2,n 2 where t 0.005,9 = 3.250 or t 0 > t / 2, n 2 where t 0.005,9 = 3.250
7) Using the results from Exercise 11-13
t0 =
0.6578
= 3.97
0.1657
8) Because t0 = 3.97 > 3.250 reject H0 and conclude the intercept is not zero at = 0.01.
11-32
H 0 : 1 = 0
H 1 : 1 0
= 0.01
Because the P-value = 0.000 < = 0.01, reject H0. There is evidence of a linear relationship between these
two variables.
b) Yes
2
c) = 1.118
2
1.118
se( 1 ) =
=
= 0.0436
S xx
588
11-21
1 65.67 2
x2
2 1
=
+
=
1.118 +
se( 0 )
= 2.885
588
n
S
9
xx
11-33
a)
H 0 : 1 = 0
H 1 : 1 0
= 0.05
Because the P-value = 0.310 > = 0.05, fail to reject H0. There is not sufficient evidence of a linear
relationship between these two variables.
The regression equation is
BMI = 13.8 + 0.256 Age
Predictor
Constant
Age
Coef
13.820
0.2558
S = 5.53982
SE Coef
9.141
0.2340
R-Sq = 14.6%
Analysis of Variance
Source
DF
SS
Regression
1
36.68
Residual Error
7 214.83
Total
8 251.51
T
1.51
1.09
P
0.174
0.310
R-Sq(adj) = 2.4%
MS
36.68
30.69
F
1.20
P
0.310
2
b) = 30.69, se( 1 ) = 0.2340, se( 0 ) = 9.141 from the computer output
c) se( 0 ) =
11-34
t0 =
1
n
2 +
1 38.256 2
x2
=
30
.
69
+
= 9.141
9
560
.
342
S xx
2 / S xx
= b . Therefore, t0 =
11-35
d=
| 10 (12.5) |
5.5 29 / 21.106
b
1 , S xx = a 2 S xx , x = ax , 0 = b0 , and
a
b1 / a
(b ) 2 / a 2 S xx
= t0 .
= 0.3878
Assume = 0.05, from Chart VI and interpolating between the curves for n = 20 and n = 30,
11-36
a)
2
i
= 21.031461, = 3.611768,
H 0 : 0 = 0
and
2
i
= 14.7073 .
11-22
0.70 .
.025 ,12
Y | x t .025 ,12 2 ( 1n +
0
( x0 x ) 2
S xx
( 2 .5 3 .0714 ) 2
25 .3486
y 0 t .025 ,12 2 (1 + n1 +
( x0 x ) 2
S xx
( 2.5 3.0714 ) 2
25.348571
a) 1
(t 0.005,18 )se( 1 )
0.0041612 (2.878)(0.000484)
0.0027682 1 0.0055542
b) 0 t 0.005,18 se( 0 )
0.3299892 ( 2.878)(0.04095)
0.212250 0 0.447728
c) 99% confidence interval on when x 0 = 85D F .
11-23
Y | x = 0 .683689
0
Y | x t.005 ,18 2 ( 1n +
0
( x0 x ) 2
S xx
)
( 85 73 .9 ) 2
33991 .6
0 .683689 0 .0594607
0 .6242283 Y | x 0 0 .7431497
x0 = 90D F .
y 0 = 0.7044949
2
y 0 t.005,18 2 (1 + 1n + ( x0Sxxx ) )
2
73.9 )
0.7044949 2.878 0.00796(1 + 201 + ( 9033991
.6 )
0.7044949 0.263567
0.420122 y 0 0.947256
11-39
1 t / 2,n 2 se( 1 )
12.652 t 0.025,5 (1.243)
12.652 2.048(1.243)
10.106 15.198
1
0 t / 2,n 2 se( 0 )
5.558 t 0.025,5 (9.159)
5.558 2.048(9.159)
24.316 13.200
0
c) 95% confidence interval for the mean rating when the average yards per attempt is 8.0
t 0.025, 28 2 +
n
(x0 x )2
S xx
1 (8 7.318)2
95.656 2.048 32.6 +
21.106
30
95.656 2.048(1.344)
92.903 98.409
d) 95% prediction interval on x0 = 8.0
11-24
y t 0.025, 28
1 ( x 0 x )2
1 + +
S xx
n
2
1 (8 7.318)
95.656 2.048 32.61 +
+
21.106
30
95.656 2.048(5.866)
83.642 y 0 107.67
Note for Problems 11-40 through 11-41. These computer printouts were obtained from Statgraphics. For Minitab users, the
standard errors are obtained from the Regression subroutine.
11-40
a) 2.51479 1 4.13395.
b) 7.98547 0 18.6549.
c)
5 6.40492 )
38.253 (2.074) 8.76775( 241 + ( 7.57
)
.563139
38.253 1.5353
36.7177 Y | x0 39.7883
1
24
( 7.5 6.40492 ) 2
)
57.563139
38.253 6.3302
31.9228 y0 44.5832
11-41
a) 9.10130 1 9.31543
b) 11.6219 0 1.04911
1 +
c) 500.124 (2.228) 3.774609( 12
(5546.5) 2
)
3308.9994
500.124 14037586
.
498.72024 Y|x 50152776
.
0
1 +
d) 500.124 (2.228) 3.774609(1 + 12
(5546.5) 2
)
3308.9994
500.124 4.5505644
495.57344 y0 504.67456
It is wider because the prediction interval includes errors for both the fitted model and for a future observation.
11-25
11-42
a) 0 . 07034 1 0 . 00045
b) 28 . 0417 0 39 . 027
c) 28 . 225 ( 2 . 101 ) 13 . 39232 ( 201 +
( 150 149 . 3 ) 2
48436 . 256
28 .225 1 .7194236
26 .5406 y | x 0 29 .9794
d) 28 . 225 ( 2 . 101 ) 13 . 39232 (1 +
28 . 225 7 . 87863
20 . 3814 y 0 36 . 1386
11-43
1
20
( 150 149 . 3 ) 2
48436 . 256
a) 0 . 03689 1 0 . 10183
b) 47 .0877 0 14 .0691
( 910 939 ) 2
67045 . 97
46 .6041 2 .514401
44 .0897 y | x 0 49 .1185 )
d) 46 . 6041 ( 3 . 106 ) 7 . 324951 (1 +
46 . 6041 8 . 779266
1
13
( 910 939 ) 2
67045 . 97
37 . 8298 y 0 55 . 3784
11-44
a) 0 . 11756 1 0 . 22541
b) 14 . 3002 0 5 . 32598
c) 4.76301 ( 2.101) 1.982231 ( 201 +
( 85 82 .3 ) 2
3010 .2111
4.76301 0.6772655
4.0857 y | x0 5.4403
d) 4 .76301 ( 2 .101 ) 1 .982231 (1 +
1
20
( 85 82 . 3 ) 2
3010 .2111
4 .76301 3 .0345765
1 .7284 y 0 7 .7976
11-45
a)
201.552 1 266.590
b) 4.67015 0
2.34696
( 30 24 . 5 ) 2
1651 . 4214
(1 0 . 806111 ) 2
3 .01062
a) 14 . 3107 1 26 . 8239
b) 5 . 18501 0 6 . 12594
c) 21 .038 ( 2 .921 ) 13 .8092 ( 181 +
21 . 038 2 . 8314277
18 .2066 y | x 23 .8694
0
(1 0.806111 ) 2
3.01062
21 .038 11 .217861
9 .8201 y 0 32 .2559
11-26
11-47
a) 43 .1964 1 30 .7272
b) 2530 .09 0 2720 .68
c) 1886 .154 ( 2 .101 ) 9811 .21( 201 +
( 20 13 . 3375 ) 2
1114 .6618
1
20
( 20 13 . 3375 ) 2
1114 . 6618
11-48
1 t / 2,n 2 se( 1 )
0.034 t 0.025,5 (0.026)
0.034 2.571(0.026)
0.101 0.033
1
0 t / 2,n 2 se( 0 )
55.63 t 0.025,5 (32.11)
55.63 2.571(32.11)
26.89 138.15
0
(
x0 x )
2 1
t0.025,5 +
n
S
xx
1 (1500 1242.86 )2
7
117142
.
8
4.63 2.571(7.396)
14.39 4.63
d) 95% prediction interval when x 0 = 1500
11-27
1 ( x 0 x )2
1 + +
S xx
n
y t 0.025,5
1 (1500 1242.86 )2
4.63 2.571 77.331 + +
117142.8
7
4.63 2.571(11.49)
24.91 y 0 34.17
Its wider because it depends on both the error associated with the fitted model as well as that of the future
observation.
11-49
1 t / 2,n 2 se( 1 )
0.178 t 0.005,9 (0.014)
0.178 3.250(0.014)
0.1325 0.2235
1
0 t / 2,n 2 se( 0 )
0.6578 t 0.005,9 (0.1657)
0.6578 3.250(0.1657)
0.119 1.196
0
when x0 = 8
y|x t0.025,9
0
1 ( x0 x )2
+
S xx
n
2
1 (8 10.09)2
11
420
.
91
11-28
Section 11-7
11-50
S
2 25.35
= 0.8617
R 2 = 12 XX = ( 2.330 )
159.71
S YY
The model accounts for 86.17% of the variability in the data.
11-51
a) R = 0.787
The model accounts for 78.7% of the variability in the data.
2
b) There is no major departure from the normality assumption in the following graph.
Normal Probability Plot of the Residuals
(response is Rating Pts)
99
95
90
Percent
80
70
60
50
40
30
20
10
5
-15
-10
-5
0
Residual
10
Residual
-5
-10
60
11-52
70
80
90
Fitted Value
100
110
4.9176
5.0208
4.5429
4.5573
5.0597
29.6681073
30.0111824
28.4224654
28.4703363
30.1405004
11-29
-3.76810726
-0.51118237
-0.52246536
-2.57033630
-0.24050041
29.9
30.9
28.9
35.9
31.5
31.0
30.9
30.0
36.9
41.9
40.5
43.9
37.5
37.9
44.5
37.9
38.9
36.9
45.8
3.8910
5.8980
5.6039
5.8282
5.3003
6.2712
5.9592
5.0500
8.2464
6.6969
7.7841
9.0384
5.9894
7.5422
8.7951
6.0831
8.3607
8.1400
9.1416
26.2553078
32.9273208
31.9496232
32.6952797
30.9403441
34.1679762
33.1307723
30.1082540
40.7342742
35.5831610
39.1974174
43.3671762
33.2311683
38.3932520
42.5583567
33.5426619
41.1142499
40.3805611
43.7102513
3.64469225
-2.02732082
-3.04962324
3.20472030
0.55965587
-3.16797616
-2.23077234
-0.10825401
-3.83427422
6.31683901
1.30258260
0.53282376
4.26883165
-0.49325200
1.94164328
4.35733807
-2.21424985
-3.48056112
2.08974865
b) Assumption of normality does not seem to be violated since the data appear to fall along a straight line.
Normal Probability Plot
99.9
99
cumulative percent
95
80
50
20
5
1
0.1
-4
-2
Residuals
c) There are no serious departures from the assumption of constant variance. This is evident by the random pattern of
the residuals.
Plot of Residuals versus Taxes
Residuals
Residuals
-2
-2
-4
-4
26
29
32
35
38
41
44
3.8
Predicted Values
d)
11-53
4.8
5.8
6.8
7.8
8.8
Taxes
R 2 76.73% ;
11-30
9.8
95
80
50
20
5
1
0.1
-2.6
-0.6
1.4
3.4
5.4
Residuals
c) There might be lower variance at the middle settings of x. However, this data does not indicate a serious departure
from the assumptions.
Plot of Residuals versus Temperature
5.4
5.4
3.4
3.4
Residuals
Residuals
1.4
1.4
-0.6
-0.6
-2.6
-2.6
180
280
380
480
580
21
680
31
51
61
71
81
a) R = 20.1121%
b) These plots might indicate the presence of outliers, but no real problem with assumptions.
2
Residuals Versus x
ResidualsVersusthe Fitted Values
(response is y)
(responseis y)
10
10
Residual
11-54
41
Temperature
Predicted Values
l
a
u 0
id
se
R
-10
-10
100
200
300
23
24
25
26
27
FittedValue
11-31
28
29
30
31
Residual
10
-10
-2
-1
Normal Score
a) R = 0.879397
b) No departures from constant variance are noted.
2
Residuals Versus x
(response is y)
(response is y)
30
30
20
20
10
10
Residual
Residual
11-55
-10
-10
-20
-20
-30
-30
0
10
20
30
40
80
130
Residual
10
0
-10
-20
-30
-1.5
-1.0
-0.5
0.0
0.5
1.0
1.5
Normal Score
a) R 2 = 71.27%
b) No major departure from normality assumptions.
Normal Probability Plot of the Residuals
(response is y)
3
Residual
11-56
180
Fitted Value
-1
-2
-2
-1
Normal Score
11-32
230
Residuals Versus x
(response is y)
(response is y)
Residual
Residual
-1
-1
-2
-2
60
70
80
90
100
a) R 2 = 85 . 22 %
b) Assumptions appear reasonable, but there is a suggestion that variability increases slightly with
11-57
Residuals Versus x
y .
(response is y)
(response is y)
Residual
Residual
Fitted Value
-5
-5
0.0
0.2
0.4
0.6
0.8
1.0
1.2
1.4
1.6
1.8
10
20
30
40
Fitted Value
c) Normality assumption may be questionable. There is some bending away from a line in the tails of the normal
probability plot.
Normal Probability Plot of the Residuals
(response is y)
Residual
-5
-2
-1
Normal Score
11-58
a)
The regression equation is
Compressive Strength = - 2150 + 185 Density
Predictor
Constant
Density
Coef
-2149.6
184.55
S = 339.219
SE Coef
332.5
11.79
R-Sq = 86.0%
T
-6.46
15.66
P
0.000
0.000
R-Sq(adj) = 85.6%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
40
41
SS
28209679
4602769
32812448
MS
28209679
115069
11-33
F
245.15
P
0.000
d) R =
2
SS R
SS
28209679
= 1 E =
= 0.8597 = 85.97%
SS T
SS T 32812448
95
90
Percent
80
70
60
50
40
30
20
10
5
-1000
-500
0
Residual
500
1000
Residual
500
-500
-1000
1000
2000
3000
Fitted Value
4000
5000
11-34
Normal Score
-1
-2
-200
-100
100
Residual
c) Rn2ew
model
= 0.9573
Larger, because the model is better able to account for the variability in the data with these two outlying data points
removed.
d) ol2 d
2
new
model
model
= 9811 . 21
= 4022 . 93
Yes, reduced more than 50%, because the two removed points accounted for a large amount of the error.
a)
60
11-60
50
40
850
950
1050
b)
y = 0.677559 + 0.0521753 x
H0 : 1 = 0
H 1 : 1 0
= 0.05
f 0 = 7.9384
f.05,1,12 = 4.75
f 0 > f ,1,12
Reject Ho.
2 = 25.23842
2
d) orig = 7.502
c)
The new estimate is larger because the new point added additional variance that was not accounted for by the model.
e)
11-35
Residual
10
-10
-2
-1
Normal Score
g) Constant variance assumption appears valid except for the added point.
Residuals Versus the Fitted Values
Residuals Versus x
(response is y)
(response is y)
10
Residual
Residual
10
-10
-10
850
950
45
1050
50
11-61
Yes, when the residuals are standardized the unusual residuals are easier to identify.
1.0723949 -0.7005724 -0.1382218 0.6488043 -2.3530824 -2.1686819 0.4684192
0.4242137 0.0982116 0.5945823 0.2051244 -0.8806396 0.7921072 0.7242900
-0.4737526 0.9432146 0.1088859 0.3749259 1.0372187 -0.7774419
11-62
[+
1
n
= 2 1 ( 1n +
55
Fitted Value
] 2
)]
( xi x )
S xx
( xi x ) 2
S xx
[+
[+
1
n
2 1
n
( xi x ) 2
S xx
( xi x )
S xx
]
]
a) Because ei is divided by an estimate of its standard error (when is estimated by 2 ), ri has approximately unit
variance.
b) No, the term in brackets in the denominator is necessary.
c) If xi is near x and n is reasonably large, ri is approximately equal to the standardized residual.
d) If xi is far from x , the standard error of ei is small. Consequently, extreme points are better fit by least squares
regression than points near the middle range of x. Because the studentized residual at any point has variance of
approximately one, the studentized residuals can be used to compare the fit of points to the regression line over the
range of x.
2
11-63
Using
E
R 2 = 1 SS
S yy ,
F0 =
S yy SS E
SS E
n2
Also,
11-36
S yy SS E
2
SS E = ( yi 0 1 xi ) 2
= ( yi y 1 ( xi x )) 2
= ( yi y ) + 1
= ( yi y ) 2
S yy SS E = 1
Therefore, F0 =
( x x ) 2
( x x )
2
( yi y )(xi x )
( xi x ) 2
12
2 / S xx
= t 02
Because the square of a t random variable with n-2 degrees of freedom is an F random variable with 1 and n-2 degrees
of freedom, the usually t-test that compares
| t0 |
to t / 2 , n 2 is equivalent to comparing
f ,1,n2 = t 2 / 2 ,n 2 .
a) f = 0 . 9 ( 23 ) = 207 . Reject H 0 : 1 = 0 .
0
1 0 .9
2
b) Because f .05 ,1, 23 = 4 . 28 , H 0 is rejected if 23 R > 4 . 28 .
1 R2
That is, H0 is rejected if
23 R 2 > 4 .28 (1 R 2 )
27 .28 R 2 > 4 .28
R 2 > 0 .157
11-37
f 0 = t02
to
Section 11-8
11-64
a) H 0 : = 0
= 0.05
H1 : 0
t0 =
0.8 20 2
1 0.64
= 5.657
t 0.025,18 = 2.101
| t 0 |> t 0.025,18
Reject H0. P-value = (<0.0005)(2) = <0.001
b) H 0 : = 0 .5
= 0.05
H 1 : 0 .5
.025
. . 0 . 5534 0 . 9177 .
where z.025 = 196
Because = 0 and = 0.5 are not in the interval, so reject H0.
11-65
a) H 0 : = 0
= 0.05
H1 : > 0
t0 =
0.75 20 2
1 0.752
= 4.81
t 0.05,18 = 1.734
t 0 > t 0.05,18
Reject H0. P-value < 0.0005
b) H 0 : = 0 .5
H 1 : > 0 .5
= 0.05
z 0.05
17
) where z .05
= 1.64
2 . 26
Because = 0 and = 0.5 are not in the interval, reject the nul hypotheses from parts (a) and (b).
11-66
n = 25 r = 0.83
a) H 0 : = 0
H 1 : 0 = 0.05
11-38
t0 =
n2
1 r 2
0 . 83 23
1 ( 0 .83 ) 2
= 7 . 137
t .025 , 23 = 2 . 069
t 0 > t / 2 , 23
Reject H0 . P-value = 0.
b) tanh(arctanh 0.83 -
z.025
22
) tanh(arctanh 0.83 +
z.025
22
. . 0 . 6471 0 . 9226 .
where z.025 = 196
a)
H 0 : = 0. 8
H 1 : 0 .8
= 0.05
n = 50 r = 0.62
a) H 0 : = 0
= 0.01
H1 : 0
t0 =
n2
1 r 2
0 . 62
48
1 ( 0 . 62 ) 2
= 5 . 475
t .005 , 48 = 2 . 682
t 0 > t 0 .005 , 48
Reject H0. P-value 0
.005
a) r = 0.933203
b)
H0 : = 0
H1 : 0
t0 =
n2
1 r 2
= 0.05
0 . 933203
15
1 ( 0 . 8709 )
= 10 . 06
H 0 : 1 = 0
H 1 : 1 0
= 0.05
f 0 = 101 . 16
f .05 ,1 ,15 = 4 . 543
f 0 >> f ,1 ,15
Reject H0. Conclude that the model is significant at = 0.05. This test and the one in part b) are identical.
d) No problems with model assumptions are noted.
11-39
Residuals Versus x
(response is y)
Residual
Residual
(response is y)
-1
-1
-2
-2
10
20
30
40
50
15
Fitted Value
Residual
2
1
-1
-2
-2
-1
Normal Score
a) y = 0.0280411 + 0.990987 x
Scatterplot of Stat vs OR
100
90
Stat
11-69
80
70
60
65
70
75
80
85
90
OR
b) H 0 : 1 = 0
H 1 : 1 0
f 0 = 79 . 838
= 0.05
f .05 ,1 ,18 = 4 . 41
f 0 >> f ,1 ,18
Reject H0 .
c) r = 0 . 816 = 0 . 903
d) H 0 : = 0
H1 : 0
= 0.05
11-40
95
100
25
t0 =
R n2
1 R
t .025 ,18 = 2 . 101
2
0 . 90334 18
= 8 . 9345
1 0 . 816
t 0 > t / 2 ,18
Reject H0.
e) H 0 : = 0 . 5
= 0.05
H 1 : 0 .5
z 0 = 3 . 879
z . 025 = 1 . 96
z 0 > z / 2
Reject H0.
f) tanh(arcta nh 0.90334 -
z.025
17
) tanh(arcta nh 0.90334 +
z.025
17
) where
z0.025 = 1.96 .
0 . 7677 0 . 9615 .
11-70
a) y = 69 . 1044 + 0 . 419415 x
b) H 0 : 1 = 0
H 1 : 1 0
= 0.05
f 0 = 35 .744
f .05 ,1, 24 = 4 .260
f 0 > f ,1, 24
Reject H0.
c)
r = 0.77349
d) H 0 : = 0
H1 : 0
t0 =
0 .77349
= 0.05
24
1 0 .5983
= 5 .9787
t .025 , 24 = 2 .064
t 0 > t / 2 , 24
Reject H0.
e) H 0 : = 0 . 6
= 0.05
H 1 : 0 .6
z 0 = (arctanh 0 .77349 arctanh 0 .6 )( 23 ) 1 / 2 = 1 .6105
z .025 = 1 .96
z 0 >/ z / 2
Do not reject H0 .
f) tanh(arcta nh 0.77349 - z ) tanh(arcta nh 0.77349 +
23
0 . 5513 0 . 8932 .
. 025
11-71
z .025
23
. .
) where z.025 = 196
a)
11-41
Current WithoutElect(mA)
50
40
30
20
10
0
3
4
Supply Voltage
Coef
5.503
6.7342
SE Coef
3.104
0.7999
R-Sq = 89.9%
T
1.77
8.42
P
0.114
0.000
R-Sq(adj) = 88.6%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
8
9
SS
1493.7
168.6
1662.3
MS
1493.7
21.1
F
70.88
P
0.000
y = 5.50 + 6.73x
Yes, because the P-value 0, the regression model is significant at = 0.05.
b) r =
0.899 = 0.948
c)
H0 : = 0
H1 : 0
t0 =
r n2
t 0.025,8
1 r2
= 2.306
0.948 10 2
1 0.948 2
= 8.425
11-42
z
z
1.96
19.6
tanh arctan h 0.948 +
0.7898 0.9879
a)
Scatterplot of 2001 vs 2000
16
15
2001
11-72
14
13
12
12
13
14
2000
15
16
Coef
-0.0144
1.01127
S = 0.110372
SE Coef
0.3315
0.02321
R-Sq = 99.5%
T
-0.04
43.56
P
0.966
0.000
R-Sq(adj) = 99.4%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
10
11
SS
23.117
0.122
23.239
MS
23.117
0.012
F
1897.63
P
0.000
y = 0.014 + 1.011x
Yes, because the P-value 0, the regression model is significant at = 0.05.
b) r =
0.995 = 0.9975
c)
11-43
H 0 : = 0.9
H 1 : 0.9
z 0 = (arctan h R arctan h 0 )(n 3)
1/ 2
1/ 2
z 0 = 5.6084
z / 2 = z 0.025 = 1.96
| z 0 |> z 0.025
Reject H0. P-value = (1-1)(2) = 0.000.
d)
z
z
1.96
19.6
tanh arctan h 0.9975
tanh arctan h 0.9975 +
12 3
12 3
0.9908 0.9993
11-73
a) r= 0.787 =0.887
b)
H0 : = 0
H1 : 0
t0 =
r n2
0.887 30 2
1 r2
1 0.887 2
t.025,28 = 2.048
= 10.164
t0 > t.025, 28
Reject H0. The P-value < 0.0005
c)
1.96
19.6
tanh arctan h 0.887
0.7741 0.9452
d)
11-44
H 0 : = 0.7
H 1 : 0.7
z 0 = (arctan h R arctan h 0 )(n 3)
1/ 2
1/ 2
z 0 = 2.808
z / 2 = z 0.025 = 1.96
| z 0 |> z 0.025
Reject H0. The P-value = 2(0.0025) = 0.005
11-74
Scatterplot of y vs x
4
3
2
1
y
0
-1
-2
-3
-4
-5
-5
-4
-3
-2
-1
r = 0. The correlation coefficient does not detect the relationship between x and y because the relationship
is not linear as shown on the graph above.
Section 11-9
11-76
a)
Yes,
b)
No
c)
Yes,
d)
Yes,
ln y = ln 0 + 1 ln x + ln
ln y = ln 0 + x ln 1 + ln
1
1
= 0 + 1 +
y
x
800
11-75
700
600
500
400
300
200
100
0
280
330
11-45
380
H 0 : = 0.7
H 1 : 0.7
z 0 = (arctan h R arctan h 0 )(n 3)
1/ 2
1/ 2
z 0 = 2.808
z / 2 = z 0.025 = 1.96
| z 0 |> z 0.025
Reject H0. The P-value = 2(0.0025) = 0.005
11-74
Scatterplot of y vs x
4
3
2
1
y
0
-1
-2
-3
-4
-5
-5
-4
-3
-2
-1
r = 0. The correlation coefficient does not detect the relationship between x and y because the relationship
is not linear as shown on the graph above.
Section 11-9
11-76
a)
Yes,
b)
No
c)
Yes,
d)
Yes,
ln y = ln 0 + 1 ln x + ln
ln y = ln 0 + x ln 1 + ln
1
1
= 0 + 1 +
y
x
800
11-75
700
600
500
400
300
200
100
0
280
330
11-45
380
15
10
0
0
500
1000
1500
2000
2500
3000
3500
b) y = - 1956.3 + 6.686 x
c)
Source
Regression
Residual Error
Total
DF
SS
491662
124403
616065
1
9
10
MS
491662
13823
F
35.57
d)
There is a curve in the residuals.
Residuals Versus the Fitted Values
(response is Vapor Pr)
200
Residual
100
-100
-200
-100
100
200
300
400
500
600
Fitted Value
7
6
Ln(VP)
5
4
3
2
1
0 .0 0 2 7
0 .0 03 2
0 .0 0 3 7
1/T
ln y = 20.6 - 5201(1/x)
11-46
P
0.000
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
9
10
SS
28.511
0.004
28.515
MS
28.511
0.000
F
66715.47
P
0.000
0.02
Residual
0.01
0.00
-0.01
-0.02
-0.03
1
Fitted Value
There is still curvature in the data, but now the plot is convex instead of concave.
a)
15
10
11-77
0
0
500
1000
1500
2000
2500
3000
b)
c)
y = 0.8819 + 0.00385 x
H 0 : 1 = 0
H1 : 1 0
= 0.05
f 0 = 122.03
f 0 > f 0.05,1, 48
Reject H0. Conclude that regression model is significant at = 0.05
d) No, it seems the variance is not constant, there is a funnel shape.
11-47
3500
Residual
0
-1
-2
-3
-4
-5
0
10
Fitted Value
e)
11-48
Supplemental Exercises
n
11-78
a)
( y i y i ) =
i =1
yi
i =1
and
i =1
= n 0 + 1
Then,
(n 0 + 1
x ) y
i
i =1
= n 0 + 1
= n 0 + 1
x (
i
i =1
n
+ 1 xi )
i =1
xi n 0 1
i =1
b)
=0
i =1
( y i yi )xi = yi xi yi xi
i =1
i =1
and
i =1
i =1
i =1
i =1
yi xi = 0 xi + 1 xi
i =1
i =1
i =1
i =1
i =1
i =1
i =1
0 xi + 1 xi 2 (0 + 1xi ) xi =
0 xi + 1 xi 2 0 xi 1 xi 2 = 0
c)
1
n
=y
i =1
y = ( 0 +1 x)
1 n
1
y i = n ( 0 + 1 xi )
n i =1
1
= (n 0 + 1 xi )
n
1
= (n( y 1 x ) + 1 xi )
n
1
= (ny n1 x + 1 xi )
n
= y x + x
1
=y
11-49
a)
Plot of y vs x
2.2
1.9
1.6
y
11-79
1.3
0.7
1.1
1.3
1.5
1.7
1.9
2.1
2) H 0 :1 = 0
3) H 1:1 0
4) = 0.05
5) The test statistic is
f0 =
SS R / k
SS E /(n p )
f0 =
1.96613 / 1
= 255263.9
0.0000623515 / 8
8) Because 2552639 > 5.32 reject H0 and conclude that the regression model is significant at = 0.05.
P-value < 0.000001
d)
1.53667 1 1.55085
e) 2) H 0 : 0 = 0
11-50
3) H 1: 0 0
4) = 0.05
5) The test statistic is t 0 =
0
se( 0 )
a) y = 93.34 + 15.64 x
b)
H 0 : 1 = 0
H1 : 1 0
f 0 = 12.872
= 0.05
f .05,1,14 = 4.60
f 0 > f 0.05,1,14
Reject H0. Conclude that 1 0 at = 0.05.
c) (7.961 1 23.322)
d) (74.758 0 111.923)
e) y = 93.34 + 15.64(2.5) = 132.44
.325 )
132.44 2.145 136.27 161 + ( 2.57.2017
132.44 6.468
125.97 Y |x0 =2.5 138.91
11-81
11-82
y = 1.2232 + 0.5075 x
where y = 1 /
The residual plots indicate a possible outlier.
2
y = 4.5755 + 2.2047 x , r = 0.992, R = 98.40%
The model appears to be an excellent fit. The R is large and both regression coefficients are significant. No, the
existence of a strong correlation does not imply a cause and effect relationship.
11-83
y = 0.7916 x
Even though y should be zero when x is zero, because the regressor variable does not usually assume values near zero, a
model with an intercept fits this data better. Without an intercept, the residuals plots are not satisfactory.
a)
110
100
90
80
days
11-84
70
60
50
40
30
16
17
index
11-51
18
y = 193 + 15.296 x
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
14
15
SS
1492.6
7926.8
9419.4
MS
1492.6
566.2
F
2.64
P
0.127
Do not reject Ho. We do not have evidence of a relationship. Therefore, there is not sufficient evidence to conclude that
the seasonal meteorological index (x) is a reliable predictor of the number of days that the ozone level exceeds 0.20
ppm (y).
95% CI on 1
c)
1 t / 2 , n 2 se ( 1 )
15 . 296 t .025 ,12 ( 9 . 421 )
15 . 296 2 . 145 ( 9 . 421 )
4 . 912 1 35 . 504
d) The normality plot of the residuals is satisfactory. However, the plot of residuals versus run order exhibits a strong
downward trend. This could indicate that there is another variable should be included in the model and it is one that
changes with time.
40
30
20
Residual
Normal Score
10
0
-10
-1
-20
-30
-2
-40
-30
-20
-10
10
20
30
-40
40
Residual
10
Observation Order
a)
0.7
0.6
0.5
11-85
0.4
0.3
0.2
0.3
0.4
0.5
0.6
0.7
b)
y = 0.6714 2964x
11-52
0.8
0.9
1.0
12
14
16
c) Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
6
7
SS
0.03691
0.13498
0.17189
MS
0.03691
0.02250
F
1.64
P
0.248
R2 = 21.47%
Because the P-value > 0.05, reject the null hypothesis and conclude that the model is significant.
d) There appears to be curvature in the data. There is a dip in the middle of the normal probability plot and the plot of
the residuals versus the fitted values shows curvature.
1.5
0.2
0.1
0.5
Residual
Normal Score
1.0
0.0
0.0
-0.5
-0.1
-1.0
-0.2
-1.5
-0.2
-0.1
0.0
0.1
0.4
0.2
0.6
a)
940
11-86
0.5
Fitted Value
Residual
930
920
920
930
940
b) y = 33.3 + 0.9636 x
c)Predictor
Constant
Therm
Coef
66.0
0.9299
S = 5.435
R-Sq = 71.2%
SE Coef
194.2
0.2090
T
0.34
4.45
P
0.743
0.002
R-Sq(adj) = 67.6%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
8
9
SS
584.62
236.28
820.90
MS
584.62
29.53
F
19.79
P
0.002
Reject the null hypothesis and conclude that the model is significant. Here 77.3% of the variability is explained by the
model.
d)
H 0 : 1 = 1
H1 : 1 1
= 0.05
11-53
t0 =
1 1 0.9299 1
=
= 0.3354
0.2090
se( 1 )
> t a / 2,n 2 , we cannot reject Ho and we conclude that there is not enough evidence to reject the claim
that the devices produce different temperature measurements. Therefore, we assume the devices produce equivalent
measurements.
e) The residual plots to not reveal any major problems.
Normal Probability Plot of the Residuals
(response is IR)
Normal Score
-1
-5
Residual
Residual
-5
920
930
940
Fitted Value
a)
8
7
6
5
11-87
4
3
2
1
2
11-54
5
= 0.699 + 1.66 x
b) y
c)
Source
DF
Regression
1
Residual Error
8
Total
9
SS
28.044
9.860
37.904
MS
28.044
1.233
F
22.75
P
0.001
Reject the null hypothesis and conclude that the model is significant.
d)
x0 = 4.25
y|x = 4.257
0
1 (4.25 4.75) 2
10
4.257 2.306(0.3717)
3.399 y| xo 5.114
e) The normal probability plot of the residuals appears linear, but there are some large residuals in the lower fitted
values. There may be some problems with the model.
Residual
Normal Score
-1
-1
-2
-2
-1
11-88
a)
The regression equation is
No. Of Atoms (x 10E9) = - 0.300 + 0.0221 power(mW)
Predictor
Constant
power(mW)
Coef
-0.29989
0.0221217
S = 0.0337423
SE Coef
0.02279
0.0006580
R-Sq = 98.9%
T
-13.16
33.62
P
0.000
0.000
R-Sq(adj) = 98.8%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
13
14
Fitted Value
Residual
SS
1.2870
0.0148
1.3018
MS
1.2870
0.0011
F
1130.43
11-55
P
0.000
0.8
0.0337423
98.9%
98.8%
0.6
0.4
0.2
0.0
10
20
30
power(mW)
40
50
0.989 = 0.994
H0 : = 0
H1 : 0
t0 =
r n2
0.994 15 2
1 r2
t 0.025,13 = 2.160
1 .994 2
= 32.766
1 t / 2,n 2 se( 1 )
0.022 t 0.025,13 (0.00066)
0.022 2.160(0.00066)
0.0206 0.0234
1
11-89
a)
11-56
price
3000
2500
2000
1500
1000
0.30
0.35
0.40
carat
0.45
0.50
The relationship between carat and price is not linear. Yes, there is one outlier, observation number 33.
b) The person obtained a very good pricehigh carat diamond at low price.
c) All the data
The regression equation is
price = - 1696 + 9349 carat
Predictor
Constant
carat
Coef
-1696.2
9349.4
S = 331.921
SE Coef
298.3
794.1
R-Sq = 78.5%
T
-5.69
11.77
P
0.000
0.000
R-Sq(adj) = 77.9%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
38
39
SS
15270545
4186512
19457057
MS
15270545
110171
F
138.61
P
0.000
/ 2,n 2
Coef
-1841.2
9809.2
S = 296.218
SE Coef
269.9
722.5
R-Sq = 83.3%
T
-6.82
13.58
P
0.000
0.000
R-Sq(adj) = 82.8%
Analysis of Variance
Source
Regression
DF
1
SS
16173949
MS
16173949
11-57
F
184.33
P
0.000
Residual Error
Total
37
38
3246568
19420517
87745
/ 2,n 2
11-90
The regression equation is
Population = 3549143 + 651828 Count
Predictor
Constant
Count
S = 183802
Coef
3549143
651828
SE Coef
131986
262844
R-Sq = 33.9%
T
26.89
2.48
P
0.000
0.029
R-Sq(adj) = 28.4%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
1
12
13
SS
2.07763E+11
4.05398E+11
6.13161E+11
MS
2.07763E+11
33783126799
F
6.15
P
0.029
y = 3549143 + 651828 x
Yes, the regression is significant at = 0.05. Care needs to be taken in making cause and effect statements
based on a regression analysis. In this case, it is surely not the case that an increase in the stork count is
causing the population to increase, in fact, the opposite is most likely the case. However, unless a designed
experiment is performed, cause and effect statements should not be made on regression analysis alone. The
existence of a strong correlation does not imply a cause and effect relationship.
11-58
SS R = R 2 ( S yy ) = 0.94(0.50) = 0.47
SS E = S yy SS R = 0.5 0.47 = 0.03
H 0 : 1 = 2 = ... = 6 = 0
H 1 : at least one j 0
= 0.05
f0 =
SS R / k
0.47 / 6
=
= 18.28
SS E / n p 0.03 / 7
p=6
R 2 = 0.92
c)
SS E
0.04
=
= 0.005
8
n p
0.03
MS E ( full ) =
= 0.004
7
MS E (reduced ) =
No, the MSE is larger for the reduced model, although not by much. Generally, if adding a variable to a model
reduces the MSE it is an indication that the variable may be useful in explaining the response variable. Here the
decrease in MSE is not very great because the added variable had no real explanatory power.
Mind-Expanding Exercises
12-98.
Because
R2 =
SS E
SS R / k
, F0 =
and this is the usual F-test for significance
S yy
SS E /(n k 1)
0.90 / 4
F0 =
= 33.75 and the critical value is f.05,4,15 = 3.06 . Because
(1 0.9) /( 20 4 1)
SS R
S yy
of regression. Then,
and
1 R2 =
12-73
12-99.
R2 / 4
3.06
(1 R 2 ) / 15
R2
0.816
(1 R 2 )
Then, R 2 0.449 results in a significant regression.
12-100. Because
+ hi ,i +1 h j ,i +1 + ... + hi , j (1 h j , j ) + hi , j +1 h j , j +1 + ... + hi , n h j , n ) 2
The expression in parentheses is recognized as the ijth element of (I-H)(I-H') = I-H. Therefore, Cov(ei , e j ) = hij 2 .
12-102.
1 2
p
Z
= 2 X ' y + 2 ( X ' X ) + 2 T '
Z
Z
= 0,
= 2(T c ) . Set
Z
= 0.
Then we get
12-74
b) This solution would be appropriate in situations where you have tested the hypothesis that T = c and concluded
that this hypothesis cannot be rejected.
12-104. a) For the piecewise linear function to be continuous at x =
show that
+ ( x x* )
1
0
y=
0 + 2 ( x x* )
x x*
x > x
Let
0,
z=
1,
x x
x > x
x1 = x x
x2 = (x x )z
0 = 0
1 = 1
2 = 2 1
Then, y = 0 * + 1 * x1 + 2 * x2 .
b) Should refer to exercise 12-80. If there is a discontinuity at
+ x
1
0
y=
0 + 1 x
0,
z=
Let
1,
x x*
x > x
x x
x > x
0 = 0
1 = 1
2 = 0 0
3 = 1 1
x1 = x
x 2 = xz
c) Should refer to exercise 12-80. One could estimate x* as a parameter in the model. A simple approach is to refit the
model in Exercise 12-80 with different choices for x* and to select the value for x* that minimizes the residual sum of
squares.
12-75
CHAPTER 12
Sections 12-1
12-1.
223
553
10
X y = 43550.8
104736.8
b)
c)
12-2.
171.055
a)
= ( X X ) 1 X y
1.9122
= 0.0931
0.2532
b)
12-3.
a)
x2 = 2
x2 = 8
Model 1
y = 100 + 2 x1 + 8
Model 2
y = 95 + 15
. x1 + 3( 2) + 4 x1
y = 108 + 2 x1
y = 101 + 5.5x1
y = 100 + 2 x1 + 4(8)
y = 95 + 15
. x 1 + 3(8) + 16x 1
y = 132 + 2 x1
y = 119 + 17.5x1
MODEL 1
y = 132 + 2
160
140
120
100
80
60
40
20
0
y = 108 + 2
12-1
10
MODEL 2
y = 119 + 17.5 x
350
300
250
200
150
100
50
0
y = 101 + 5.5 x
10
15
The interaction term in model 2 affects the slope of the regression equation. That is, it modifies the
amount of change per unit of x1 on y .
b) x 2 = 5
y = 100 + 2 x1 + 4(5)
y = 120 + 2 x1
x2 = 2
x2 = 8
y = 95 + 15
. x1 + 3(5) + 2 x1 (5)
y = 101 + 5.5x1
y = 119 + 17.5x1
y = 110 + 115
. x1
Change per
unit of x1
11.5
5.5
17.5
Yes, result does depend on the value of x2, because x2 interacts with x1.
12-4.
Predictor
Constant
X1
X2
X3
X4
S = 11.79
Coef
-123.1
0.7573
7.519
2.483
-0.4811
SE Coef
157.3
0.2791
4.010
1.809
0.5552
R-Sq = 85.2%
T
-0.78
2.71
1.87
1.37
-0.87
P
0.459
0.030
0.103
0.212
0.415
R-Sq(adj) = 76.8%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
b)
c)
d)
DF
4
7
11
SS
5600.5
972.5
6572.9
MS
1400.1
138.9
F
10.08
P
0.005
2 = 139.00
se( 0 ) = 157.3 , se( 1 ) = 0.2791 , se( 2 ) = 4.010 , se( 3 ) = 1.809 , and se( 4 ) = 0.5552
y = 123.1 + 0.7573(75) + 7.519(24) + 2.483(90) 0.4811(98) = 290.476
12-2
12-5.
Coef
49.90
-0.01045
-0.00120
-0.0032364
0.292
-3.855
0.1897
S = 2.22830
SE Coef
19.67
0.02338
0.01631
0.0009459
1.765
1.329
0.2730
R-Sq = 89.3%
T
2.54
-0.45
-0.07
-3.42
0.17
-2.90
0.69
P
0.024
0.662
0.942
0.004
0.871
0.012
0.498
R-Sq(adj) = 84.8%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
DF
6
14
20
SS
581.898
69.514
651.412
MS
96.983
4.965
F
19.53
P
0.000
2 = 4.965
se( 0 ) = 19.67 , se( 1 ) = 0.02338 , se( 2 ) = 0.01631 , se( 3 ) = 0.0009459 ,
b)
12-6.
0.102 x4 - 2.16 x5
T
P
1.03
0.320
-0.11
0.912
3.83
0.002
-1.91
0.077
-0.90
0.382
R-Sq(adj) = 57.8%
MS
5.5780
0.7792
F
7.16
P
0.002
a)
b)
c)
2 = .7792
se( 0 ) = 7.226 , se( 2 ) = .2633 , se( 3 ) = .1359 , se( 4 ) = .05339 and se( 5 ) = 2.395
d)
12-7.
12-3
y = 8.996
Predictor
Coef
SE Coef
T
P
Constant
47.82
49.94
0.96
0.353
x1
-9.604
3.723
-2.58
0.020
x2
0.4152
0.2261
1.84
0.085
x3
18.294
1.323
13.82
0.000
--------------------------------------------------------------------------------
a)
b)
c)
d)
12-8.
2 = 12.3
se( 0 ) = 49.94 , se( 1 ) = 3.723 , se( 2 ) = 0.2261 , and se( 3 ) = 1.323
y = 47.82 9.604(14.5) + 0.44152(220) + 18.29(5) = 91.38
Predictor
Constant
temp
soaktime
soakpct
difftime
diffpct
Coef
-0.03023
0.00002856
0.0023182
-0.003029
0.008476
-0.002363
S = 0.002296
SE Coef
0.06178
0.00003437
0.0001737
0.005844
0.001218
0.008078
R-Sq = 96.8%
T
-0.49
0.83
13.35
-0.52
6.96
-0.29
P
0.629
0.414
0.000
0.609
0.000
0.772
R-Sq(adj) = 96.2%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
DF
5
26
31
SS
0.00418939
0.00013708
0.00432647
MS
0.00083788
0.00000527
F
158.92
P
0.000
b)
c)
d)
12-9.
2 = 5.27 x10 6
The standard errors are listed under the StDev column above.
Coef
-440.39
19.147
68.080
SE Coef
94.20
3.460
5.241
R-Sq = 84.3%
T
-4.68
5.53
12.99
P
0.000
0.000
0.000
R-Sq(adj) = 83.5%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
37
39
SS
11076473
2055837
13132310
MS
5538237
55563
12-4
F
99.67
P
0.000
x1 = mAmps x2 = ExposureTime
b)
= 55563
se( 0 ) = 94.20 , se( 1 ) = 3.460 , and se( 2 ) = 5.241
c)
Coef
0.4875
-0.000767
-0.02274
-0.04150
13.240
S = 0.236010
SE Coef
0.4272
0.003508
0.04747
0.08408
1.679
R-Sq = 81.2%
T
1.14
-0.22
-0.48
-0.49
7.89
P
0.271
0.830
0.638
0.628
0.000
R-Sq(adj) = 76.5%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
DF
4
16
20
SS
3.84906
0.89121
4.74028
MS
0.96227
0.05570
F
17.28
P
0.000
2 = 0.05570
se( 0 ) = 0.4272 , se( 1 ) = 0.003508 , se( 2 ) = 0.04747 , se( 3 ) = 0.08408 , and
b)
se( 4 ) = 1.679
c) y = 0.4875 0.000767(30) 0.02274(5) 0.04150(5) + 13.240(0.135) = 1.9307
12-11. The regression equation is
density = - 0.110 + 0.407 dielectric constant + 2.11 loss factor
Predictor
Constant
dielectric constant
loss factor
S = 0.00883422
Coef
-0.1105
0.4072
2.108
SE Coef
0.2501
0.1682
5.834
R-Sq = 99.7%
T
-0.44
2.42
0.36
P
0.670
0.042
0.727
R-Sq(adj) = 99.7%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
8
10
SS
0.23563
0.00062
0.23626
MS
0.11782
0.00008
12-5
F
1509.64
P
0.000
b)
x1 = DielectricConst x2 = LossFactor
= 0.00008
se( 0 ) = 0.2501 , se( 1 ) = 0.1682 , and se( 2 ) = 5.834
2
Coef
-171.26
7.029
12.696
S = 3.07827
SE Coef
28.40
1.539
1.539
R-Sq = 93.7%
T
-6.03
4.57
8.25
P
0.001
0.004
0.000
R-Sq(adj) = 91.6%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
6
8
SS
842.37
56.85
899.22
MS
421.18
9.48
F
44.45
P
0.000
2 = 9.48
se( 0 ) = 28.40 , se( 1 ) = 1.539 , and se( 2 ) = 1.539
c)
Coef
238.56
0.3339
-2.7167
SE Coef
45.23
0.6763
0.6887
R-Sq = 75.6%
T
5.27
0.49
-3.94
P
0.002
0.639
0.008
R-Sq(adj) = 67.4%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
6
8
SS
24518
7928
32446
MS
12259
1321
F
9.28
b)
x1 = % Brightness x2 = %Contrast
= 1321
2
12-6
P
0.015
Coef
-39.92
0.7156
1.2953
-0.1521
S = 3.24336
SE Coef
11.90
0.1349
0.3680
0.1563
T
-3.36
5.31
3.52
-0.97
R-Sq = 91.4%
P
0.004
0.000
0.003
0.344
R-Sq(adj) = 89.8%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
3
17
20
SS
1890.41
178.83
2069.24
MS
630.14
10.52
F
59.90
P
0.000
a)
b)
c)
2 = 10.52
se( 0 ) = 11.90 , se( 1 ) = 0.1349 , se( 2 ) = 0.3680 , and se( 3 ) = 0.1563
d)
Coef
3.220
1.2243
4.4231
-4.0914
S = 1.92094
SE Coef
6.519
0.1080
0.2799
0.4953
T
0.49
11.33
15.80
-8.26
R-Sq = 97.8%
P
0.626
0.000
0.000
0.000
R-Sq(adj) = 97.5%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
DF
3
26
29
SS
4196.3
95.9
4292.2
MS
1398.8
3.7
F
379.07
P
0.000
c)
2 = 3.7
se( 0 ) = 6.519 , se( 3 ) = 0.1080 , se( 7 ) = 0.2799 , and se( 10 ) = 0.4953
d)
b)
12-16.
Predictor
Coef
SE Coef
12-7
Constant
GF
GA
ADV
PPGF
PCTG
PEN
BMI
AVG
SHT
PPGA
PKPCT
SHGF
SHGA
FG
-29.46
0.12356
-0.15953
0.1236
-0.764
3.100
0.2402
0.0036
-18.86
-0.03494
0.0526
0.0505
0.5833
0.3733
0.1578
S = 3.58133
93.98
0.05319
0.04458
0.2023
1.269
4.406
0.3453
0.1151
28.31
0.03419
0.2137
0.7221
0.3935
0.2712
0.2227
R-Sq = 90.1%
-0.31
2.32
-3.58
0.61
-0.60
0.70
0.70
0.03
-0.67
-1.02
0.25
0.07
1.48
1.38
0.71
0.758
0.035
0.003
0.550
0.556
0.492
0.497
0.976
0.515
0.323
0.809
0.945
0.159
0.189
0.489
R-Sq(adj) = 80.9%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
14
15
29
SS
1754.28
192.39
1946.67
MS
125.31
12.83
F
9.77
P
0.000
2 = 12.83
The standard errors of the coefficients are listed under the SE Coef column above.
12-17.
Predictor
Constant
Xl
X2
S = 12.35
Coef
SE Coef
T
P
383.80
36.22
10.60
0.002
-3.6381
0.5665
-6.42
0.008
-0.11168
0.04338
-2.57
0.082
R-Sq = 98.5%
R-Sq(adj) = 97.5%
Analysis of Variance
Source
DF
Regression
2
Residual Error
3
Total
5
SS
29787
458
30245
MS
14894
153
F
97.59
P
0.002
a)
c)
d) Predictor
Coef
SE Coef
T
P
Constant
484.0
101.3
4.78
0.041
Xl
-7.656
3.846
-1.99
0.185
X2
-0.2221
0.1129
-1.97
0.188
X1*X2
0.004087
0.003871
1.06
0.402
S = 12.12
R-Sq = 99.0%
R-Sq(adj) = 97.6%
Analysis of Variance
Source
DF
SS
MS
F
Regression
3
29951.4
9983.8
67.92
Residual Error
2
294.0
147.0
Total
5
30245.3
P
0.015
12-8
f)
12-18.
a)
f
= 2 [ yi 0' 1 ( xi1 x1 ) 2 ( xi 2 x2 )]
0'
f
= 2 [ yi 0' 1 ( xi1 x1 ) 2 ( xi 2 x2 )]( xi1 x1 )
1
f
= 2 [ yi 0' 1 ( xi1 x1 ) 2 ( xi 2 x2 )]( xi 2 x2 )
2
n 0' = yi
c) Substituting y i y for
yi
0' = 0 .
Sections 12-2
12-19.
a) n = 10, k = 2,
p = 3,
= 0.05
H 0 : 1 = 2 = ... = k = 0
H 1 : j 0 for at least one j
(1916) 2
= 4490
10
y i 1916
S yy = 371595.6
1916
' X ' y = [171.055 3.713 1.126] 43550.8 = 371511.9
104736.8
1916 2
SS R = 371511.9
= 4406.3
10
SS E = S yy SS R = 4490 4406.3 = 83.7
f0 =
SS R
k
SS E
n p
4406.3 / 2
= 184.25
83.7 / 7
f 0.05, 2, 7 = 4.74
f 0 > f 0.05, 2,7
12-9
f)
12-18.
a)
f
= 2 [ yi 0' 1 ( xi1 x1 ) 2 ( xi 2 x2 )]
0'
f
= 2 [ yi 0' 1 ( xi1 x1 ) 2 ( xi 2 x2 )]( xi1 x1 )
1
f
= 2 [ yi 0' 1 ( xi1 x1 ) 2 ( xi 2 x2 )]( xi 2 x2 )
2
n 0' = yi
c) Substituting y i y for
yi
0' = 0 .
Sections 12-2
12-19.
a) n = 10, k = 2,
p = 3,
= 0.05
H 0 : 1 = 2 = ... = k = 0
H 1 : j 0 for at least one j
(1916) 2
= 4490
10
y i 1916
S yy = 371595.6
1916
' X ' y = [171.055 3.713 1.126] 43550.8 = 371511.9
104736.8
1916 2
SS R = 371511.9
= 4406.3
10
SS E = S yy SS R = 4490 4406.3 = 83.7
f0 =
SS R
k
SS E
n p
4406.3 / 2
= 184.25
83.7 / 7
f 0.05, 2, 7 = 4.74
f 0 > f 0.05, 2,7
12-9
Reject H0 and conclude that the regression model is significant at = 0.05. P-value = 0.000
b)
2 = MS E =
SS E
= 11.957
n p
H 0 : 1 = 0
H 1 : 1 0
1
t0 =
se( 1 )
2 = 0
2 0
t0 =
3.713
= 16.21
0.229
t / 2, 7 = t 0.025, 7 = 2.365
=
2
se( 2 )
1.126
= 11.04
0.10199
S yy = 742.00
a)
H 0 : 1 = 2 = 0
H 1 : j 0 for at least one j
= 0.01
n
SS R = ' X ' y
( yi ) 2
i =1
220
220 2
= ( 1.9122 0.0931 0.2532 ) 36768
9965 10
= 5525.5548 4840
= 685.55
SS E = S yy SS R
= 742 685.55
= 56.45
SS R
685.55 / 2
f 0 = SSk =
= 42.51
E
56.45 / 7
n p
f 0.01, 2, 7 = 9.55
f 0 > f 0.01, 2,7
Reject H0 and conclude that the regression model is significant at = 0.01. P-value = 0.000121
2 = MS E =
SS E
56.45
=
= 8.0643
n p
7
b)
H 0 : 1 = 0
H 1 : 1 0
1
t0 =
se( )
1
0.0931
= 3.67
0.0254
t 0.005,7 = 3.499
=
| t 0 |> t 0.005,7
Reject H0 and conclude that
P-value =
c)
12-21.
a)
b)
2 (1 P ( t < t 0 ))
1 : t0 = 4.82
2 : t0 = 8.21
3 : t0 = 0.98
H 0 : 3 = 0
H1 : 3 0
= 0.05
t0 = 0.98
t 0.025, 22 = 2.074
| t 0 |>/ t 0.025, 23
Do not reject
12-22. a)
H 0 : 1 = 2 = 3 = 4 = 0
H1 at least one j 0
= 0.05
f 0 = 10.08
f 0.05, 4, 7 = 4.12
f 0 > f 0.05, 4, 7
Reject H 0 P-value = 0.005
b) = 0.05
H 0 : 1 = 0 2 = 0
H 1 : 1 0
2 0
t0 = 1.87
t0 = 2.71
t / 2, n p = t0.025 , 7 = 2.365
3 = 0
3 0
t0 = 1.37
4 = 0
4 0
t0 = 0.87
12-23.
a)
H 0 : 1 = 2 = 3 = 4 = 5 = 6 = 0
H1: at least one 0
f 0 = 19.53
f ,6,14 = f 0.05, 6,14 = 2.848
f 0 > f 0.05, 6,14
Reject H0 and conclude regression model is significant at = 0.05
b) The t-test statistics for 1 through 6 are -0.45, -0.07, -3.42, 0.17, -2.90, 0.69. Because t0.025,14=-2.14,
the regressors that contribute to the model at =0.05 are etw and axle.
12-24.
a)
H0 : j = 0
for all j
H0 :2 = 0
H1:2 0
t 0 = 0. 113
3 = 0
3 0
4 = 0
4 0
t 0 = 3. 83
| t 0 | >/ t / 2 ,14
| t 0 | > t / 2 ,14
Do not reject H0
Reject H0
All variables do not contribute to the model.
12-25.
P-value = 0.0023
5 = 0
5 0
t 0 = 1. 91
| t 0 | >/ t / 2 ,14
t0 = 0. 9
| t 0 | >/ t / 2 ,14
Do not reject H0
Do not reject H0
a) H 0 : 1 = 2 = 3 = 0
H1: j 0
f 0 = 828.31
f .05,3,16 = 3.34
f 0 > f 0.05,3,16
Reject H0 and conclude regression is significant at = 0.05
b) 2 = 12.2856
= 0.05
t / 2 , n p = t.025,16 = 2. 12
H 0 : 1 = 0
H1:1 0
t 0 = 2.58
12-26.
2 = 0
3 = 0
3 0
t 0 = 13.82
2 0
t 0 = 1.84
| t 0 |> t 0.025,16
| t 0 |>/ t 0.025,16
| t0 |> t0.025,16
Reject H0
Do not reject H0
Reject H0
Coef
0.0011
0.008581
-0.0208
0.0097
SE Coef
0.9067
0.007083
0.1018
0.1798
12-12
T
0.00
1.21
-0.20
0.05
P
0.999
0.242
0.841
0.958
S = 0.506197
R-Sq = 8.1%
R-Sq(adj) = 0.0%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
DF
3
17
20
SS
0.3843
4.3560
4.7403
MS
0.1281
0.2562
F
0.50
P
0.687
H 0 : 1 = 2 = 3 = 0
H1 : j 0
= 0.05
f 0 = 0.50
f 0.05,3,17 = 3.197
f 0 < f 0.05,3,17
Do not reject H0. There is insufficient evidence to conclude that the model is significant at =0.05. The Pvalue =0.687.
b) H 0 : 1 = 0
H 1 : 1 0
= 0.05
t0 =
1
0.008581
=
= 1.21
se( 1 ) 0.007083
t0.025,17 = 2.11
| t0 |< t / 2,17 . Fail to reject H 0 , there is not enough evidence to conclude that 1 is significant in the
model at = 0.05.
H0 : 2 = 0
H1 : 2 0
= 0.05
t0 =
2
0.0208
=
= 0.2
0.1018
se( 2 )
t0.025,17 = 2.11
| t0 |< t / 2,17 . Fail to reject H 0 , there is not enough evidence to conclude that 2 is significant in the
model at = 0.05.
H 0 : 3 = 0
H1 : 3 0
= 0.05
t0 =
3
0.0097
=
= 0.05
se( 3 ) 0.1798
12-13
t0.025,17 = 2.11
| t0 |< t / 2,17 . Fail to reject H 0 , there is not enough evidence to conclude that 3 is significant in the
model at = 0.05.
12-27. a)
H 0 : 1 = 2 = 0
H 0 : for at least one j 0
= 0.05
f 0 = 99.67
f 0.05, 2,37 = 3.252
f 0 > f 0.05, 2,37
The regression equation is
rads = - 440 + 19.1 mAmps + 68.1 exposure time
Predictor
Constant
mAmps
exposure time
S = 235.718
Coef
-440.39
19.147
68.080
SE Coef
94.20
3.460
5.241
R-Sq = 84.3%
T
-4.68
5.53
12.99
P
0.000
0.000
0.000
R-Sq(adj) = 83.5%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
37
39
SS
11076473
2055837
13132310
MS
5538237
55563
F
99.67
P
0.000
Reject H0 and conclude regression model is significant at = 0.05 P-value < 0.000001
b)
2 = MS E = 55563
se( 1 ) = 2 c jj = 3.460
H 0 : 1 = 0
H 1 : 1 0
= 0.05
t0 =
1
se( 1 )
19.147
= 5.539
3.460
t 0.025, 40 3 = t 0.025,37 = 2.0262
=
| t 0 |> t / 2,37 ,
Reject
H0
se( 2 ) = 2 c jj = 5.241
H0 : 2 = 0
12-14
H1 : 2 0
= 0.05
t0 =
2
se( 2 )
68.080
= 12.99
5.241
t 0.025, 40 3 = t 0.025,37 = 2.0262
=
| t 0 |> t / 2,37 ,
Reject
H0
conclude that
12-28.
The regression equation is
y = - 171 + 7.03 x1 + 12.7 x2
Predictor
Constant
x1
x2
Coef
-171.26
7.029
12.696
S = 3.07827
SE Coef
28.40
1.539
1.539
R-Sq = 93.7%
T
-6.03
4.57
8.25
P
0.001
0.004
0.000
R-Sq(adj) = 91.6%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
DF
2
6
8
SS
842.37
56.85
899.22
MS
421.18
9.48
F
44.45
P
0.000
H 0 : 1 = 2 = 0
H 1 : for at least one j 0
= 0.05
f0 =
SS R
k
SS E
n p
842.37 / 2
= 44.45
56.85 / 6
f 0.05, 2, 6 = 5.14
f 0 > f 0.05, 2,6
Reject H0 and conclude regression model is significant at = 0.05 P-value 0
b)
2 = MS E = 9.48
se( 1 ) = 2 c jj = 1.539
H 0 : 1 = 0
H 1 : 1 0
= 0.05
12-15
t0 =
1
se( 1 )
7.03
= 4.568
1.539
t 0.025,93 = t 0.025, 6 = 2.447
=
se( 2 ) = 2 c jj = 1.539
H0 : 2 = 0
H1 : 2 0
= 0.05
t0 =
2
se( 2 )
12.7
= 8.252
1.539
t 0.025,93 = t 0.025, 6 = 2.447
=
| t 0 |> t / 2, 6 , Reject H 0
conclude that
c) With a smaller sample size, the difference in the estimate from the hypothesized value needs to be
greater to be significant.
12-29.
Useful range (ng) = 239 + 0.334 Brightness (%) - 2.72 Contrast (%)
Predictor
Constant
Brightness (%)
Contrast (%)
S = 36.3493
Coef
238.56
0.3339
-2.7167
SE Coef
45.23
0.6763
0.6887
R-Sq = 75.6%
T
5.27
0.49
-3.94
P
0.002
0.639
0.008
R-Sq(adj) = 67.4%
Analysis of Variance
Source
Regression
Residual Error
Total
a)
DF
2
6
8
SS
24518
7928
32446
MS
12259
1321
F
9.28
H 0 : 1 = 2 = 0
H 1 : for at least one j 0
= 0.05
12-16
P
0.015
f0 =
SS R
k
SS E
n p
24518 / 2
= 9.28
7928 / 6
2 = MS E = 1321
se( 1 ) = 2 c jj = 0.6763
H 0 : 1 = 0
H 1 : 1 0
= 0.05
t0 =
1
se( 1 )
0.3339
= 0.49
0.6763
t 0.025,93 = t 0.025, 6 = 2.447
=
| t 0 |< t / 2, 6 , Fail to reject H 0 , there is no enough evidence to conclude that 1 is significant in the model at =
0.05
se( 2 ) = 2 c jj = 0.6887
H0 : 2 = 0
H1 : 2 0
= 0.05
t0 =
2
se( 2 )
2.7167
= 3.94
0.6887
t 0.025,93 = t 0.025, 6 = 2.447
=
| t 0 |> t / 2, 6 ,
Reject
H0
conclude that
12-30.
The regression equation is
Stack Loss(y) = - 39.9 + 0.716 X1 + 1.30 X2 - 0.152 X3
Predictor
Constant
X1
X2
X3
Coef
-39.92
0.7156
1.2953
-0.1521
S = 3.24336
SE Coef
11.90
0.1349
0.3680
0.1563
R-Sq = 91.4%
T
-3.36
5.31
3.52
-0.97
P
0.004
0.000
0.003
0.344
R-Sq(adj) = 89.8%
12-17
Analysis of Variance
Source
Regression
Residual Error
Total
DF
3
17
20
SS
1890.41
178.83
2069.24
MS
630.14
10.52
F
59.90
P
0.000
a)
H 0 : 1 = 2 = 3 = 0
H 1 : j 0 for at least one j
= 0.05
f0 =
SS R
k
SS E
n p
189.41 / 3
= 59.90
178.83 / 17
f 0.05,3,17 = 3.20
f 0 > f 0.05,3,17
Reject H0 and conclude that the regression model is significant at = 0.05 P-value < 0.000001
b)
2 = MS E = 10.52
se( 1 ) = 2 c jj = 0.1349
H 0 : 1 = 0
H 1 : 1 0
= 0.05
t0 =
1
se( 1 )
0.7156
= 5.31
0.1349
t 0.025, 214 = t 0.025,17 = 2.110
=
| t 0 |> t / 2,17 . Reject H 0 and conclude that 1 is significant in the model at = 0.05.
se( 2 ) = 2 c jj = 0.3680
H0 : 2 = 0
H1 : 2 0
= 0.05
t0 =
2
se( 2 )
1.2953
= 3.52
0.3680
t 0.025, 214 = t 0.025,17 = 2.110
=
12-18
se( 3 ) = 2 c jj = 0.1563
H 0 : 3 = 0
H1 : 3 0
= 0.05
t0 =
3
se( 3 )
0.1521
= 0.97
0.1563
t 0.025, 214 = t 0.025,17 = 2.110
=
| t 0 |< t / 2,17 . Fail to reject H 0 , there is not enough evidence to conclude that 3
0.05.
12-31.
Rating Pts = 3.22 + 1.22 Pct Comp + 4.42 Pct TD - 4.09 Pct Int
Predictor
Constant
Pct Comp
Pct TD
Pct Int
Coef
3.220
1.2243
4.4231
-4.0914
S = 1.92094
SE Coef
6.519
0.1080
0.2799
0.4953
R-Sq = 97.8%
T
0.49
11.33
15.80
-8.26
P
0.626
0.000
0.000
0.000
R-Sq(adj) = 97.5%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
3
26
29
SS
4196.3
95.9
4292.2
MS
1398.8
3.7
F
379.07
P
0.000
H 0 : 3 = 7 = 10 = 0
H 1 : j 0 for at least one j
a)
= 0.05
f 0 = 379.07
f.05,3, 26 = 2.98
f 0 >> f 0.05,3, 26
Reject H0 and conclude that the regression model is significant at = 0.05 P-value
b) All at = 0.05
t.025,26 = 2.056
H 0 : 3 = 0
H 0 : 7 = 0
H 0 : 10 = 0
H1:3 0
H1 : 7 0
H 1 : 10 0
t0 = 11.33
t 0 = 15.80
t 0 = 8.26
| t 0 |> t / 2, 26
| t 0 |> t / 2, 26
| t 0 |> t / 2, 26
Reject H0
Reject H0
Reject H0
12-19
c)
The regression equation is
Rating Pts = - 33.6 + 2.18 Pct Comp - 4.54 Pct Int
Predictor
Constant
Pct Comp
Pct Int
Coef
-33.60
2.1842
-4.540
S = 6.13862
SE Coef
19.46
0.2855
1.580
R-Sq = 76.3%
T
-1.73
7.65
-2.87
P
0.096
0.000
0.008
R-Sq(adj) = 74.5%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
27
29
SS
3274.8
1017.4
4292.2
MS
1637.4
37.7
F
43.45
P
0.000
f0 =
2558.9 / 1 921.5
=
= 249.1
MS E
3.7
f 0.05,1, 26 = 4.26
f 0 >> f 0.05,1, 24
Reject H0 and conclude that the TD percentage regressor is significant at = 0.05
P-value 0
Note that f0 = 691.6 = t-statistic2 = 15.802 = 249.6 (except for rounding error)
12-32.
a) H 0 : j = 0
H1: j 0
for all j
for at least one j
f 0 = 158.9902
f .05,5, 26 = 2.59
f 0 > f ,5, 26
Reject H0 and conclude regression is significant at = 0.05
P-value < 0.000001
b) = 0.05
t / 2 , n p = t.025, 26 = 2. 056
H 0 : 1 = 0
2 = 0
3 = 0
4 = 0
5 = 0
H 1 : 1 0
t 0 = 0.83
2 0
3 0
4 0
t0 = 12.25
t0 = 0.52
t 0 = 6.96
5 0
t 0 = 0.29
Do not Reject H0
Reject H0
Do not reject H0
Reject H0
Do not reject H0
c)
d) H 0 : j = 0
H1: j 0
for all j
for at least one j
f 0 = 308.455
f.05, 2, 29 = 3.33
f 0 > f 0.05, 2, 29
Reject H0 and conclude regression is significant at = 0.05
12-20
= 0.05
t / 2 ,n p = t .025 , 29 = 2.045
H 0 : 1 = 0
H 1 : 1 0
t 0 = 18.31
2 = 0
2 0
t 0 = 6 . 37
| t 0 | > t / 2,29
| t 0 | > t / 2 , 29
Reject H0 for each regressor variable and conclude that both variables are significant at = 0.05
e) part ( d ) = 6.7 E 6 . Part c) is smaller, suggesting a better model.
12-33.
y = 0 + 1 x1 + 2 x2 ,
a) H 0 : 1 = 2 = 0
H1 at least one j 0
f 0 = 97.59
f 0.05, 2,3 = 9.55
f 0 > f 0.05, 2,3
Reject H0 P-value = 0.002
b)
H 0 : 1 = 0
t 0 = 6.42
t / 2,3 = t 0.025,3 = 3.182
H0 : 2 = 0
H1 : 2 0
t 0 = 2.57
t / 2,3 = t 0.025,3 = 3.182
| t 0 |> t 0.025,3
| t 0 |>/ t 0.025,3
H1 : 1 0
c)
SS R ( 2 | 1 , 0 ) = 1012
H0 : 2 = 0
H1 : 2 0
= 0.05
f 0 = 6.629
d) H 0 : 1 = 2 = 12 = 0
H1 at least one
j 0
= 0.05
f 0 = 7.714
f ,3, 2 = f 0.05,3, 2 = 19.16
f 0 >/ f 0.05,3, 2
Do not reject H0
12-21
e) H
: 12 = 0
H 1 : 12 0
= 0.05
SSR 163 .9
=
= 1 .11
MS E
147
2 = 147.0
2 (no interaction term) = 153.0
MS E ( 2 )
12-34.
a)
H0 : j = 0
for all j
H1 : j 0
f 0 = 9.77
f .05,14,15 = 2.424
f 0 > f 0.05,14,15
Reject H0 and conclude that the regression model is significant at = 0.05
b) H 0 : j = 0
H1 : j 0
t.025,15 = 2.131
GF :
t 0 = 2.32
Reject H0
GA :
t 0 = 3.58
Reject H0
ADV :
t 0 = 0.61
Do not reject H0
PPGF :
t 0 = 0.60
t 0 = 0.70
Do not reject H0
PCTG :
PEN :
t 0 = 0.70
Do not reject H0
BMI :
t 0 = 0.03
Do not reject H0
AVG :
t 0 = 0.67
Do not reject H0
SHT :
t 0 = 1.02
Do not reject H0
PPGA :
t 0 = 0.25
Do not reject H0
PKPCT :
t 0 = 0.07
Do not reject H0
SHGF :
t 0 = 1.48
Do not reject H0
SHGA :
t 0 = 1.38
Do not reject H0
FG :
t 0 = 0.71
Do not reject H0
Do not reject H0
12-22
H 0 : 1 = 0
H 1 : 1 0
t 0 = 5.70
4 = 0
4 0
t 0 = 0.86
Reject H0
Do not reject H0
Based on the t-test, power play goals for (PPGF) is not a logical choice to add to the model that already contains the
number of goals for.
Sections 12-3 and 12-4
12-35.
a) 0 t / 2, n p
2 c
00
1 t / 2,n p 2 c
11
2 t / 2,n p 2 c
22
x 2 = 43
y 0 = 189.471
X 0' ( X ' X ) 1 X 0 = 0.305065
189.471 (2.365) 550.7875(0.305065)
158.815 Y | x0 220.127
c) = 0.05
x1 = 18
x2 = 43
y 0 = 189.471
12-23
H 0 : 1 = 0
H 1 : 1 0
t 0 = 5.70
4 = 0
4 0
t 0 = 0.86
Reject H0
Do not reject H0
Based on the t-test, power play goals for (PPGF) is not a logical choice to add to the model that already contains the
number of goals for.
Sections 12-3 and 12-4
12-35.
a) 0 t / 2, n p
2 c
00
1 t / 2,n p 2 c
11
2 t / 2,n p 2 c
22
x 2 = 43
y 0 = 189.471
X 0' ( X ' X ) 1 X 0 = 0.305065
189.471 (2.365) 550.7875(0.305065)
158.815 Y | x0 220.127
c) = 0.05
x1 = 18
x2 = 43
y 0 = 189.471
12-23
1.9122 (2.365)(10.055)
1.9122 23.78
25.6922 0 21.8678
1 t / 2,n p 2 c
11
2 t / 2,n p 2 c
22
x 2 = 50
y 0 = 29.37
X 0' ( X ' X ) 1 X 0 = 0.211088
x1 = 200
x 2 = 50
y 0 = 29.37
X 0' ( X ' X ) 1 X 0 = 0.211088
y t / 2,n p 2 (1 + X 0' ( X ' X ) 1 X 0 )
29.37 (2.365) 85.694(1.211088)
29.37 24.093
5.277 y 0 53.463
12-37.
a)
20.477 1 1.269
12-24
0.245 2 1.076
14.428 3 22.159
se( y 0 ) = 4.721
b) Y | x = 91.372
0
91.372 2.921(4.721)
77.582 y 0 105.162
c) Y | x = 91.372
se( Y | x0 ) = 3.163
0
91.372 (2.921)(3.163)
82.133 Y | x0 100.611
12.38.
t .005,16 = 2.921
a) 95 % CI on coefficients
0.0973 1 1.4172
1.9646 2 17.0026
1.7953 3 6.7613
1.7941 4 0.8319
b) Y | x = 290.44
0
t.025, 7 = 2.365
se( Y | x0 ) = 7.61
Y | x t / 2,n p se( Y | x )
0
290.44 (2.365)(7.61)
272.44 Y | x0 308.44
c)
y 0 t / 2, n p 2 (1 + x 0 ( X X) 1 x 0 )
290.44 2.365(14.038)
257.25 y0 323.64
12-39.
6.9467 1 0.3295
0.3651 2 0.1417
b) 45.8276 1 30.5156
1.3426 2 0.8984
a)
0.03433 12 0.04251
These part b. intervals are much wider.
Yes, the addition of this term increased the standard error of the regression coefficient estimators.
12-40.
2 0.535
0.229 3 0.812
0.216 4 0.013
7.2968 5 2.9756
a) 0.595
b) Y | x = 8. 99568
0
se ( Y | x0 ) = 0. 472445
12-25
t.025,14 = 2. 145
Y | x t / 2, n p se( Y | x )
0
8.99568 (2.145)(0.472445)
7.982 Y | x0 10.009
se( y 0 ) = 100121
.
c) y0 = 8. 99568
8.99568 2.145(1.00121)
6.8481 y0 11.143
12-41. a)
1 t / 2,n p 2 c11
2 t / 2,n p 2 c
22
t 0.005,37 = 2.7154
se( Y | x0 ) = 54.6
Y | x t / 2,n p se( Y | x )
0
85.1 ( 2.7154)(54.6)
233.4 Y | x0 63.2
se( y 0 ) = 241.95
85.1 2.7154(241.95)
742.09 y 0 571.89
c) y 0 = 85.1
12-42.
The regression equation is
ARSNAILS = 0.001 + 0.00858 AGE - 0.021 DRINKUSE + 0.010 COOKUSE
Predictor
Constant
AGE
DRINKUSE
COOKUSE
Coef
0.0011
0.008581
-0.0208
0.0097
S = 0.506197
SE Coef
0.9067
0.007083
0.1018
0.1798
R-Sq = 8.1%
T
0.00
1.21
-0.20
0.05
P
0.999
0.242
0.841
0.958
R-Sq(adj) = 0.0%
Analysis of Variance
12-26
Source
Regression
Residual Error
Total
DF
3
17
20
SS
0.3843
4.3560
4.7403
MS
0.1281
0.2562
F
0.50
P
0.687
2.617 0 2.634
0.012 1 0.0291
0.316 2 0.274
0.511 3 0.531
b)
Y | x0 = 0.214
se( Y | x0 ) = 0.216
t 0.005,16 = 2.898
Y |x t / 2,n p se( Y |x )
0
0.214 ( 2.898)(0.216)
0.412 Y |x0 0.840
se( y 0 ) = 0.5503
0.214 2.898(0.5503)
1.381 y 0 1.809
c) y0 = 0.214
0.0943 1 0.7201
8.743 2 12.959
b)
Y | x0 = 0.8787
t 0.005,16 = 1.860
se( Y | x0 ) = 0.00926
Y | x t / 2,n p se( Y | x )
0
0.8787 (1.860)(0.00926)
0.86148 Y | x0 0.89592
se( y 0 ) = 0.0134
0.8787 1.86(0.0134)
0.85490 y0 0.90250
c) y 0 = 0.8787
12-44.
The regression equation is
12-27
Coef
-171.26
7.029
12.696
S = 3.07827
SE Coef
28.40
1.539
1.539
R-Sq = 93.7%
T
-6.03
4.57
8.25
P
0.001
0.004
0.000
R-Sq(adj) = 91.6%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
6
8
SS
842.37
56.85
899.22
MS
421.18
9.48
F
44.45
P
0.000
a)
1 t / 2,n p 2 c11
2 t / 2,n p 2 c
12.7 t .025, 6
22
se( 2 )
12.7 (2.447)(1.539)
12.7 3.766
8.934 1 16.466
b)
New
Obs
1
Fit
140.82
SE Fit
6.65
Y |x = 140.82
0
95% CI
(124.54, 157.11)
t 0.025, 6 = 2.447
Y |x t / 2,n p se( Y |x )
0
95% PI
(122.88, 158.77)XX
140.82 ( 2.447)(6.65)
124.55 Y |x0 157.09
se( y 0 ) = 7.33
140.82 2.447(7.33)
122.88 y 0 158.76
c) y 0 = 140.82
12-28
Coef
238.56
0.3339
-2.7167
SE Coef
45.23
0.6763
0.6887
R-Sq = 75.6%
T
5.27
0.49
-3.94
P
0.002
0.639
0.008
R-Sq(adj) = 67.4%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
2
6
8
SS
24518
7928
32446
MS
12259
1321
F
9.28
P
0.015
a) t 0.005, 6 = 3.707
2.173 1 2.841
5.270 2 0.164
b)
Predicted Values for New Observations
New
Obs
1
Fit
44.6
SE Fit
21.9
99% CI
(-36.7, 125.8)
99% PI
(-112.8, 202.0)
Contrast
(%)
80.0
Brightness (%)
70.0
Y | x = 44.6
0
t 0.005, 6 = 3.707
se( Y | x0 ) = 21.9
Y |x t / 2,n p se( Y |x )
0
44.6 (3.707)(21.9)
36.7 Y |x0 125.8
se( y 0 ) = 42.44
c) y 0 = 44.6
44.6 3.707(42.44)
112.8 y 0 202.0
Fit
187.3
SE Fit
21.6
99% CI
(107.4, 267.2)
99% PI
(30.7, 344.0)
12-29
CI: 107.4
PI:
Contrast
(%)
25.0
Brightness (%)
50.0
Y | x 267.2
0
30.7 y0 344.0
These intervals are wider because the regressors are set at extreme values in the x space and the standard
errors are greater.
12-46. a) t 0.025,17 = 2.110
0.431 1 1.00
0.519 2 2.072
0.482 3 0.178
b) Y | x = 36.023
0
t 0.025,17 = 2.110
se( Y | x0 ) = 1.803
Y |x t / 2,n p se( Y | x )
0
36.023 (2.110)(1.803)
32.219 Y | x0 39.827
se( y 0 ) = 3.698
36.023 2.110(3.698)
28.194 y 0 43.852
c) y 0 = 36.023
12-47.
CI: 21.030
Y |x 34.559
PI: 18.173
y0 37.417
Y |x = 27.795
0
a) t0.025, 26 = 2.056
10.183 0 16.623
1.002 3 1.446
3.848 7 4.999
5.110 10 3.073
b)
Y |x t.025,24 se( Y |x )
0
82.079 (2.056)(0.3877)
82.079 0.7971
81.282 Y |x0 82.876
12-48.
a)
0.000042 1 0.000099
12-30
0.00196 2 0.00268
0.01504 3 0.00898
0.00597 4 0.010979
0.01897 5 0.01424
b) Y | x = 0.022466
se( Y | x0 ) = 0.000595 t.025, 26 = 2.056
0
0.0220086 (2.056)(0.000595)
0.0212 Y | x0 0.0237
c)
Y |x = 0.0171
se( Y |x ) = 0.000548 t .025, 29 = 2.045
0.0171 (2.045)(0.000548)
0.0159 Y | x0 0.0183
0
a)
t.005,14=2.977
8.658 0 108.458
0.08 2 0.059
0.05 3 0.047
0.006 7 0
4.962 8 5.546
7.811 9 0.101
b)
Y |x
1.102 10 0.523
= 29.71
se( Y |x0 ) = 1.395
Y |x t.005,14 se( Y |x )
0
29.71 ( 2.977)(1.395)
25.557 Y |x0 33.863
c)
d) The intervals in part c) are narrower. All of the regressors used in part c) are significant, but not all of
those used in part a) are significant. The model used in part c) is preferable.
12-50.
0.0102 1 0.2369
b) y = 19.329 + 0.25956 x1
a)
12-31
c)
0.1735 1 0.3456
d) The simple linear regression model has the shorter interval. Obviously there are extraneous variables in
the model from part a), but there are two good predictors in that model, only one of which is included in
the model of part b). Still the shorter interval is an initial indicator that the original model with all variables
might be improved.
Section 12-5
12-51.
a)
The regression equation is
mpg = 49.9 - 0.0104 cid - 0.0012 rhp - 0.00324 etw + 0.29 cmp - 3.86 axle
+ 0.190 n/v
Predictor
Constant
cid
rhp
etw
cmp
axle
n/v
Coef
49.90
-0.01045
-0.00120
-0.0032364
0.292
-3.855
0.1897
S = 2.22830
SE Coef
19.67
0.02338
0.01631
0.0009459
1.765
1.329
0.2730
R-Sq = 89.3%
T
2.54
-0.45
-0.07
-3.42
0.17
-2.90
0.69
P
0.024
0.662
0.942
0.004
0.871
0.012
0.498
R-Sq(adj) = 84.8%
Score
-1
-2
-3
-2
-1
0
Standardized Residual
12-32
c)
0.1735 1 0.3456
d) The simple linear regression model has the shorter interval. Obviously there are extraneous variables in
the model from part a), but there are two good predictors in that model, only one of which is included in
the model of part b). Still the shorter interval is an initial indicator that the original model with all variables
might be improved.
Section 12-5
12-51.
a)
The regression equation is
mpg = 49.9 - 0.0104 cid - 0.0012 rhp - 0.00324 etw + 0.29 cmp - 3.86 axle
+ 0.190 n/v
Predictor
Constant
cid
rhp
etw
cmp
axle
n/v
Coef
49.90
-0.01045
-0.00120
-0.0032364
0.292
-3.855
0.1897
S = 2.22830
SE Coef
19.67
0.02338
0.01631
0.0009459
1.765
1.329
0.2730
R-Sq = 89.3%
T
2.54
-0.45
-0.07
-3.42
0.17
-2.90
0.69
P
0.024
0.662
0.942
0.004
0.871
0.012
0.498
R-Sq(adj) = 84.8%
Score
-1
-2
-3
-2
-1
0
Standardized Residual
12-32
Standardized Residual
-1
-2
-3
20
25
30
Fitted Value
35
40
Standardized Residual
-1
-2
-3
100
200
300
cid
400
Standardized Residual
-1
-2
-3
3000
3500
4000
4500
etw
5000
5500
6000
9.75
10.00
Standardized Residual
-1
-2
-3
8.50
8.75
9.00
9.25
cmp
9.50
12-33
500
Standardized Residual
-1
-2
-3
2.50
2.75
3.00
3.25
axle
3.50
3.75
4.00
4.25
Standardized Residual
-1
-2
-3
20
25
30
n/v
35
40
d)
95
90
80
Percent
12-52.
70
60
50
40
30
20
10
5
-20
-10
0
Residual
12-34
10
20
Residuals Versus x1
Residuals Versus x2
(response is y)
10
10
5
Residual
Residual
(response is y)
-5
-5
-10
-10
-15
-15
20
30
40
50
60
70
80
90
21
22
23
24
Residuals Versus x3
Residuals Versus x4
(response is y)
(response is y)
10
10
26
-5
-5
-10
-10
-15
-15
85
86
87
88
89
90
91
92
93
94
90
95
100
x3
105
110
x4
a) R = 97.8 %
b) Assumption of normality appears adequate.
2
95
90
Percent
80
70
60
50
40
30
20
10
5
-4
-3
-2
-1
0
1
Residual
x7 (percentage of TDs)
1
Residual
c)
Residual
12-53.
25
x2
Residual
Residual
x1
0
-1
0
-1
-2
-2
-3
-3
-4
-4
-5
-5
70
80
90
100
Fitted Value
110
120
55.0
12-35
57.5
60.0
62.5
Pct Comp
65.0
67.5
70.0
1
Residual
Residual
0
-1
0
-1
-2
-2
-3
-3
-4
-4
-5
-5
2
6
Pct TD
10
1.0
1.5
2.0
2.5
3.0
Pct Int
a) R2= 0.969
b) Normality is acceptable
Normal Probability Plot of the Residuals
(response is PITCH)
Residual
0.005
0.000
-0.005
-2
-1
Normal Score
c) Plot is acceptable.
Residuals Versus the Fitted Values
(response is PITCH)
0.005
Residual
12-54.
0.000
-0.005
0.01
0.02
0.03
0.04
Fitted Value
d)
12-36
0.05
0.06
0.07
3.5
4.0
4.5
0.0191 0.0003 0.0026 0.0009 0.0293 0.1112 0.1014 0.0131 0.0076 0.0004 0.0109
0.0000 0.0140 0.0039 0.0002 0.0003 0.0079 0.0022 4.5975 0.0033 0.0058 0.1412
0.0161 0.0268 0.0609 0.0016 0.0029 0.3391 0.3918 0.0134 0.0088 0.5063
The 19th observation is influential
2
12-55. a) R = 84.3 %
b) Assumption of normality appears adequate.
Normal Probability Plot of the Residuals
(response is rads)
99
95
90
Percent
80
70
60
50
40
30
20
10
5
-500
-250
0
Residual
250
500
750
c) There are funnel shapes in the graphs, so the assumption of constant variance is violated. The model is
inadequate.
Residuals Versus the Fitted Values
(response is rads)
750
Residual
500
250
-250
-500
-500
500
Fitted Value
1000
1500
500
500
250
250
Residual
Residual
(response is rads)
750
-250
-250
-500
-500
0
10
exposure time
15
20
12-37
10
15
20
25
mAmps
30
35
40
0.032728
0.029489
0.023724
0.014663
0.077299
0.3436
0.008489
0.007592
0.001985
0.002068
0.021386
0.105059
0.000643
0.000375
0.0002
0.000209
0.006095
0.005442
0.0043
0.002564
0.015557
0.077846
0.07828
0.070853
0.020725
0.021539
0.177299
0.731526
No, none of the observations has a Cooks distance greater than 1.
0.008279
0.006018
0.000926
0.002467
0.0014
0.057512
12-56. a) R 2 = 8.1 %
b) Assumption of normality appears is not adequate.
Normal Probability Plot of the Residuals
(response is ARSNAILS)
99
95
90
Percent
80
70
60
50
40
30
20
10
5
-1.0
-0.5
0.0
0.5
1.0
1.5
Residual
c) The graphs indicate non-constant variance. Therefore, the model is not adequate.
Residuals V er sus the Fitted V alues
(response is ARSNAILS)
4
Standardized Residual
-1
0.0
0.1
0.2
0.3
0.4
Fitted Value
0.5
0.6
0.7
Residual
1.0
0.5
0.0
-0.5
0
10
20
30
40
50
60
70
80
AGE
12-38
90
0.008611
0.003612
0.000823
0.013062
0.001459
0.036157
(response is ARSNAILS)
1.5
1.5
1.0
1.0
Residual
Residual
(response is ARSNAILS)
0.5
0.5
0.0
0.0
-0.5
-0.5
1
3
DRINKUSE
2.0
2.5
3.0
3.5
COOKUSE
4.0
4.5
5.0
95
90
Percent
80
70
60
50
40
30
20
10
5
-0.02
-0.01
0.00
Residual
0.01
0.02
Residual
0.005
0.000
-0.005
-0.010
0.7
0.8
0.9
1.0
Fitted Value
12-39
1.1
1.2
0.010
0.010
0.005
0.005
Residual
Residual
(response is density)
0.000
0.000
-0.005
-0.005
-0.010
-0.010
2.0
2.2
2.4
2.6
dielectric constant
2.8
3.0
0.015
0.020
0.025
0.058551
0.485253
12-58. a) R 2 = 93.7 %
b) The normal assumption appears inadequate
Normal Probability Plot of the Residuals
(response is y)
99
95
90
Percent
80
70
60
50
40
30
20
10
5
-7.5
-5.0
-2.5
0.0
Residual
Standardized Residual
1.5
1.0
0.5
0.0
-0.5
-1.0
60
70
80
Fitted Value
90
100
12-40
2.5
0.030
loss factor
5.0
0.035
0.040
0.045
0.077203
Residuals Versus x1
Residuals Versus x2
(response is y)
1
Residual
Residual
(response is y)
0
-1
0
-1
-2
-2
-3
-3
-4
-4
13.0
13.5
14.0
x1
14.5
15.0
11.0
11.5
12.0
x2
12.5
13.0
95
90
Percent
80
70
60
50
40
30
20
10
5
-80
-60
-40
-20
0
Residual
20
40
60
80
c) Assumption of constant variance is a possible concern. One point is a concern as a possible outlier.
12-41
Residual
40
20
0
-20
-40
50
75
100
125
Fitted Value
150
175
80
80
60
60
40
40
Residual
Residual
200
20
20
-20
-20
-40
-40
20
30
40
50
Contrast (%)
60
70
50
80
60
70
80
Brightness (%)
0.000075
12-60. a) R 2 = 91.4 %
b) Assumption of normality appears adequate.
Normal Probability Plot of the Residuals
(response is Stack Loss(y))
99
95
90
Percent
80
70
60
50
40
30
20
10
5
-8
-6
-4
-2
0
Residual
12-42
90
100
Residuals Versus X1
5.0
5.0
2.5
2.5
Residual
Residual
0.0
-2.5
0.0
-2.5
-5.0
-5.0
-7.5
-7.5
5
10
15
20
25
Fitted Value
30
35
50
40
55
60
Residuals Versus X2
70
75
80
Residuals Versus X3
5.0
5.0
2.5
2.5
Residual
Residual
65
X1
0.0
-2.5
-5.0
0.0
-2.5
-5.0
-7.5
-7.5
16
18
20
22
X2
24
26
28
70
75
80
85
90
95
X3
0.004048
0.035866
0.065473
0.019565
0.065066
0.001122
12-61.
a) R 2 = 0.985
b) R 2 = 0.99
R2 increases with addition of interaction term. No, adding additional regressor will always increase r2
12-62.
a) R = 0.955 . Yes, the R2 using these two regressors is nearly as large as the R2 from the model with
five regressors.
b) Normality is acceptable, but there is some indication of outliers.
2
(response is PITCH)
(response is PITCH)
0.005
Residual
Residual
0.005
0.000
-0.005
-0.005
0.01
0.02
0.03
0.04
0.05
0.06
0.07
Fitted Value
c)
0.000
-2
-1
Normal Score
0.0202 0.0008 0.0021 0.0003 0.0050 0.0000 0.0506 0.0175 0.0015 0.0003 0.0087
12-43
0.0001 0.0072 0.0126 0.0004 0.0021 0.0051 0.0007 0.0282 0.0072 0.0004 0.1566
0.0267 0.0006 0.0189 0.0179 0.0055 0.1141 0.1520 0.0001 0.0759 2.3550
The last observation is very influential
a) There is some indication of nonconstant variance since the residuals appear to fan out with increasing
values of y.
Residual Plot for y
8
5
Residuals
12-63.
-1
-4
-7
0
40
80
120
160
200
240
Predicted
b)
Source
Model
Error
Total (Corr.)
Sum of Squares
30531.5
193.725
30725.2
DF
3
16
19
R-squared = 0.993695
R-squared (Adj. for d.f.) = 0.992513
Mean Square
10177.2
12.1078
F-Ratio
840.546
P-value
.0000
R 2 = 0.9937 or 99.37 %;
y = 6.22489 016647
.
x1 0.000228x 2 + 0.157312 x 3
d)
12-44
Residuals
0.05
-0.05
-0.1
-0.15
4
4.3
4.6
4.9
5.2
5.5
Predicted
Residuals
0.05
-0.05
-0.1
-0.15
3.3
5.3
7.3
9.3
11.3
x3
12-45
Residuals
0.05
0.02
-0.01
-0.04
-0.07
3.8
4.1
4.4
4.7
5.3
Predicted
Using 1/x3
Coef
-19.329
0.25956
S = 5.43788
SE Coef
8.943
0.04220
R-Sq = 57.5%
T
-2.16
6.15
P
0.039
0.000
R-Sq(adj) = 55.9%
b) R-Sq = 57.5
c) Model appears adequate.
Residuals V er sus the Fitted V alues
(response is W)
2
Standardized Residual
12-64.
-1
-2
25
30
35
40
Fitted Value
45
50
d) No, the residuals do not seem to be related to PPGF. Since there is no pattern evident in the plot, it does not
12-46
Residuals V er sus P P GF
(response is W)
2
Standardized Residual
-1
-2
40
12-65.
50
60
PPGF
70
80
a) p = k + 1 = 2 + 1 = 3
Average size = p/n = 3/25 = 0.12
b) Leverage point criteria:
hii > 2( p / n)
h18,18 = 0.2929
Points 17 and 18 are leverage points.
Sections 12-6
12-66.
f 0 = 17.2045
f .05, 2,5 = 5.79
H1:11 0
= 0.05
t 0 = 2.45
t ,n p = t .025,83 = t .025,5 = 2.571
| t 0 |>/ 2.571
Do not reject H0 and conclude insufficient evidence to support value of quadratic term in model at = 0.05
d) One residual is an outlier
12-47
Residuals V er sus P P GF
(response is W)
2
Standardized Residual
-1
-2
40
12-65.
50
60
PPGF
70
80
a) p = k + 1 = 2 + 1 = 3
Average size = p/n = 3/25 = 0.12
b) Leverage point criteria:
hii > 2( p / n)
h18,18 = 0.2929
Points 17 and 18 are leverage points.
Sections 12-6
12-66.
f 0 = 17.2045
f .05, 2,5 = 5.79
H1:11 0
= 0.05
t 0 = 2.45
t ,n p = t .025,83 = t .025,5 = 2.571
| t 0 |>/ 2.571
Do not reject H0 and conclude insufficient evidence to support value of quadratic term in model at = 0.05
d) One residual is an outlier
12-47
Normal Score
1.0
0.5
0.0
-0.5
-1.0
-1.5
-50
-40
-30
-20
-10
10
20
30
40
50
Residual
Residual
20
10
0
-10
-20
-30
-40
-50
500
600
700
Fitted Value
12-67.
12-48
800
Normal Score
-1
-0.004
-0.002
0.000
0.002
0.004
0.006
Residual
Residuals Versus x
(response is y)
0.006
Residual
0.004
0.002
0.000
-0.002
-0.004
0.0
0.5
1.0
1.5
2.0
2.5
Residual
0.004
0.002
0.000
-0.002
-0.004
-8
-7
-6
-5
-4
-3
Fitted Value
12-68.
12-49
-2
-1
H1 : j 0
= 0.05
f 0 = 1044.99
f .05, 2,9 = 4.26
H 0 : 11 = 0
H1 : 11 0
t 0 = 2.97
= 0.05
t.025,9 = 2.262
| t 0 |> t 0.025,9
Reject H0 and conclude that 11 is significant at = 0.05
d) Observation number 9 is an extreme outlier.
Normal Probability Plot of the Residuals
(response is y)
2
Normal Score
-1
-2
-5
-4
-3
-2
-1
75
85
Residual
Residual
-1
-2
-3
-4
-5
25
35
45
55
65
Fitted Value
12-50
e)
Do not reject H0 and conclude that cubic term is not significant at = 0.05
12-69.
a) Predictor
Constant
xl
x2
x3
x1x2
x1x3
x2x3
x1^2
x2^2
x3^3
Coef
-1.769
0.4208
0.2225
-0.12800
-0.01988
0.009151
0.002576
-0.01932
-0.00745
0.000824
S = 0.06092
SE Coef
1.287
0.2942
0.1307
0.07025
0.01204
0.007621
0.007039
0.01680
0.01205
0.001441
R-Sq = 91.7%
T
-1.37
1.43
1.70
-1.82
-1.65
1.20
0.37
-1.15
-0.62
0.57
P
0.188
0.172
0.108
0.087
0.118
0.247
0.719
0.267
0.545
0.575
R-Sq(adj) = 87.0%
Analysis of Variance
Source
Regression
Residual Error
Total
DF
9
16
25
SS
0.655671
0.059386
0.715057
MS
0.072852
0.003712
F
19.63
P
0.000
1 = 2 = 3 =
H1 at least one j 0
b) H0 all
23 = 0
f 0 = 19.628
f .05,9,16 = 2.54
f 0 > f 0.05,9,16
Reject H0 and conclude that the model is significant at = 0.05
c) Model is acceptable.
12-51
Normal Score
-1
-2
-0.05
0.00
0.05
0.10
Residual
Residual
0.10
0.05
0.00
-0.05
0.0
0.1
0.2
0.3
0.4
0.5
Fitted Value
d)
H 0 : 11 = 22 = 33 = 12 = 13 = 23 = 0
H1 at least one 0
f0 =
0.0359
SS R ( 11 , 22 , 33 , 12 , 13 , 23 | 1 , 2 , 3 , 0 ) / r
6
=
= 1.612
MS E
0.003712
f .05,6,16 = 2.74
f 0 >/ f .05, 6,16
Do not reject H0
SS R ( 11 22 33 12 13 23 | 1 2 3 0 )
= SS R ( 11 22 33 12 13 23 1 2 3 | 0 ) SS R ( 1 2 3 | 0 )
Reduced Model:
= 0.65567068 0.619763
= 0.0359
y = 0 + 1 x1 + 2 x 2 + 3 x 3
12-70.
a) Create an indicator variable for sex (e.g. 0 for male, 1 for female) and include this variable in the model.
b)
12-52
Coef
-0.2139
-0.0081
0.0276
0.007937
0.1675
S = 0.514000
SE Coef
0.9708
0.1050
0.1844
0.007251
0.2398
R-Sq = 10.8%
T
-0.22
-0.08
0.15
1.09
0.70
P
0.828
0.940
0.883
0.290
0.495
R-Sq(adj) = 0.0%
There are three possible transmission types: L4, L5 and M6. So, two indicator variables could be used
where x3=1 if trns=L5, 0 otherwise and x4=1 if trns=M6, 0 otherwise.
b)
H 0 : 4 = 0 is 0.02 which is significant for values of > 0.02. Thus, it appears that whether or not the
transmission is manual affects mpg, but there is not a significant difference between the types of automatic
transmission.
12-72.
y = 0 + 1 x1 + 2 x 2 + 12 x12
y = 11.503 + 0.153x1 6.094 x 2 0.031x12
0 for tool type 302
where x 2 =
1 for tool type 416
Test of different slopes:
H 0 : 12 = 0
H1 : 12 0
t 0 = 1.79
= 0.05
t .025,16 = 2.12
| t 0 |>/ t 0.025,16
Do not reject H0. Conclude that data is insufficient to claim that (2) regression models are needed.
Test of different intercepts and slopes using extra sums of squares:
H 0 : 2 = 12 = 0
SS ( 2 , 12 | 0 ) = SS ( 1 , 2 , 12 | 0 ) SS ( 1 | 0 )
= 1013.35995 130.60910
= 882.7508
12-53
f0 =
SS ( 2 , 12 | 0 ) / 2 882.7508 / 2
=
= 1087.40
MS E
0.4059
Reject H0.
The default settings for F-to-enter and F-to-remove for Minitab were used. Different settings can change the models
generated by the method.
a) The min MSE equation is: x1, x2, x 3
C p = 3.8 MS E = 134 .6
x2,
y = 0.20052 + 0.467864 x1
The min MS E equation is the same as the min Cp.
b) Same as equation in part (a).
c) Same as equation in part (a).
d) Same as equation in part (a).
e) All methods give the same equation with either min Cp or min MSE.
12-76.
The default settings for F-to-enter and F-to-remove for Minitab were used. Different settings can change the models
generated by the method.
a) The min MSE equation is: x1, x3, x 4
C p = 2.6 MS E = 0.6644
x4
12-54
e) The minimum MSE and forward models all are the same. Stepwise and backward regressions generate the minimum
Cp model. The minimum Cp model has fewer regressors and it might be preferred, but MSE has increased.
y = 253.06 2.5453 x 2
The min MS E equation is the same as the min Cp.
b) Same as equation in part (a).
c) Same as equation in part (a).
d) Same as equation in part (a).
e) All methods give the same equation with either min Cp or min MSE.
12-78.
The default settings for F-to-enter and F-to-remove for Minitab were used. Different settings can change the models
generated by the method.
a)
Vars
1
1
2
2
3
3
4
4
5
5
6
6
7
R-Sq
63.5
57.5
86.3
75.1
87.2
86.9
87.7
87.7
88.2
88.1
88.6
88.4
89.2
R-Sq(adj)
62.1
55.9
85.3
73.3
85.7
85.4
85.8
85.7
85.7
85.6
85.7
85.4
85.8
Mallows
C-p
29.5
38.6
-3.3
13.8
-2.5
-2.2
-1.4
-1.3
-0.1
0.1
1.3
1.5
2.4
S
5.0407
5.4379
3.1391
4.2359
3.1017
3.1287
3.0886
3.1010
3.0958
3.1122
3.1028
3.1276
3.0888
12-55
P
A P
G G D G
F A V F
X
X
X X
X
X
X X
X X
X
X X
X
X X
X X
X
X X
X
X X
X
X X
X X
P
P
P K
C P B A S P P
T E M V H G C
G N I G T A T
S
H
G
F
S
H
G F
A G
X
X
X
X
X
X
X
X
X
X
X
X
X X
X
X
X
X
X
X
7
8
8
9
9
10
10
11
11
12
12
13
13
14
89.2
89.6
89.4
89.8
89.6
89.9
89.8
90.0
89.9
90.1
90.1
90.1
90.1
90.1
85.8
85.6
85.3
85.3
84.9
84.5
84.5
84.0
83.7
83.1
83.1
82.1
82.1
80.9
2.4
3.8
4.1
5.4
5.8
7.4
7.4
9.1
9.4
11.0
11.1
13.0
13.0
15.0
3.0926
3.1068
3.1386
3.1453
3.1793
3.2245
3.2259
3.2823
3.3097
3.3647
3.3709
3.4677
3.4682
3.5813
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X
X X
X X
X
X X
X X
X
X
X X X
X
X X
X X X
X X X X
X X
X
X X X
X X X
X
X X
X X X
X
X X
X X X
X
X X X X X X X
X
X X X
X X X
X
X X
X X X X
X
X X X X X X X
X X X X X
X X X
X X X X X X X X X
e) The model with minimum Cp is a good choice. Its MSE is not much larger than the minimum MSE
model and it is a simple model with few regressors.
12-81.
The default settings for F-to-enter and F-to-remove for Minitab were used. Different settings can change the models
generated by the method.
Vars
1
1
2
2
3
3
4
4
5
5
6
6
7
7
8
8
9
9
10
R-Sq
78.7
76.6
91.9
90.0
97.8
96.6
100.0
99.5
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
R-Sq(adj)
78.0
75.8
91.3
89.3
97.5
96.2
100.0
99.4
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
100.0
Mallows C-p
35767.9
39318.4
13596.8
16755.2
3736.4
5635.3
8.3
879.8
8.9
9.0
5.4
5.5
5.6
6.1
7.2
7.2
9.0
9.1
11.0
S
5.7125
5.9891
3.5886
3.9830
1.9209
2.3568
0.17007
0.95852
0.16921
0.16942
0.15408
0.15463
0.15072
0.15258
0.15261
0.15285
0.15583
0.15597
0.15977
12-56
C C
A o o Y
t m m d
t p p s
X
X
X
X
X
X
X
X
X
X
X
X X
X X
X
X
X
X
X X
X
X
X
X
X
X X
X
X X
p
e
r
P
P
c
c
t
t L
A
o I I
t T T n n n
t D D g t t
X
X
X
X
X
X
X
X
X
X
X
X
X X
X
X
X
X
X
X
X
X X X
X
X X X
X
X X X
X X
X X X
X X
X X X X X X
X X X X X X
X X X X X X
X X X X X X
X X X X X X
a) Note that R-squared in the models shown is rounded by Minitab because it is near 100%. Because MSE is not zero,
R-squared is not exactly 100%.
Minimum Cp=5.4:
y = 3.2992 + 0.82871x pctcomp + 0.00056 x yds + 3.828 x yds / att 0.08458 xtd +
y = 0.038 + 0.00850 x1
The min MS E model is the same as the min Cp model
b) The full model that contains all 3 variables
Vars
1
1
2
2
3
3
4
4
5
5
6
6
R-Sq
66.0
59.1
81.6
78.1
88.8
88.8
90.3
89.9
90.7
90.7
91.0
91.0
R-Sq(adj)
64.2
57.0
79.5
75.7
86.8
86.8
87.8
87.3
87.6
87.6
87.2
87.1
Mallows
C-p
26.5
35.3
8.6
13.0
1.3
1.4
1.5
2.0
2.9
2.9
4.5
4.5
S
3.4137
3.7433
2.5813
2.8151
2.0718
2.0755
1.9917
2.0302
2.0057
2.0064
2.0442
2.0487
c r e c
i h t m
d p w p
X
X
X
X
X
X
X
X
X
X
X
X
X X
X
X
X
X
X
X X
12-57
a
x n
c
l / h c o
e v c o 2
X
X
X X
X
X
X X
X
X
X
X
X
X
X
X X X
X
X
X
7
7
8
8
9
91.3
91.2
91.4
91.4
91.4
86.6
86.4
85.6
85.6
84.4
6.2
6.3
8.0
8.1
10.0
2.0927
2.1039
2.1651
2.1654
2.2562
X X
X
X X
X X X
X X X
X
X X X X X
X
X X X
X X X X X
X
X X X
X X X X X
X
X
X
X
a)
The minimum Cp (1.3) model is:
y = 49.5 0.017547 x cid 0.0034252 x etw + 1.29 x cmp 3.184 x axle 0.0096 x c 02
b)
e) The minimum Cp model is preferred because it has a very low MSE as well (4.29)
f) Only one indicator variable is used for transmission to distinguish the automatic from manual types and two
indicator variables are used for drv:
xtrans=0 for automatic (L4, L5) and 1 for manual (M6) and
xdrv1=0 if drv=4 or R and 1 if drv =F; xdrv2=0 if drv = 4 or F and 1 if drv=R.
The minimum Cp (4.0) model is the same as the minimum MSE (2.267) model:
y = 39.12 0.0044 xetw + 0.271xn / v 4.5 xtrns + 3.2 xdrv1 + 1.7 xdrv 2
Forward selection:
y = 0 + 1 x'+ 11 ( x' ) 2
y = 759.395 7.607 x'0.331( x' ) 2
y = 759.395 7.607( x 297.125) 0.331( x 297.125) 2
y = 26202.14 + 189.09 x 0.331x 2
a) y = 759.395 90.783x'47.166( x' ) 2 , where x' =
x x
Sx
285 297.125
= 1.016
11.9336
y = 759.395 90.783(1.106) 47.166(1.106)2 = 802.943 psi
b) At x = 285
x' =
2
c) y = 759.395 90.783 x 297.125 47.166 x 297.125
119336
.
119336
.
12-58
e)
where
y' =
y y
Sy
and
x' =
xx
Sx
The default settings for F-to-enter and F-to-remove for Minitab were used. Different settings can change the models
generated by the method.
a)
C p = 4.66 MS E = 0.004
c) The forward selection model in part (a) is more parsimonious with a lower Cp and equivalent MSE. Therefore, we
prefer the model in part (a).
12-86.
R2 =
SS R
SS
= 1 E
S yy
S yy
SS E = MS E (n p)
= 100(30 10)
= 2000
0. 92 = 1 2000
S
yy
25000 = Syy
SS R = S yy SS E
= 25000 2000 = 23000
23000
SS
MS R = R =
= 2555.56
k
9
f0 =
MS R 2555.56
=
= 25.56
100
MS E
f .05,9, 20 = 2.39
f 0 > f ,9, 20
Reject H0 and conclude at least one
b) k = 4 p = 5
j is significant at = 0.05.
SS E = 2200
12-59
SS E
2200
=
= 88
n p 30 5
Yes, MSE is reduced with new model (k = 4).
MS E =
c) C = SS E ( p ) n + 2 p
p
2
12-87.
Cp =
2200
30 + 2(5) = 2
100
SS E = 300
SS E
300
=
= 14.29
n p 25 4
SS E
Cp =
n + 2p
MS E ( full )
MS E =
300
25 + 2(4)
10
= 5 + 8 = 13
YES, Cp > p
b) p = 5 SSE = 275
MS E =
SS E
275
=
= 11
n p 30 5
Cp =
275
25 + 2(5) = 12.5
10
12-60
Supplemental Exercises
a) y = 4203 0.231x 3 +21.485 x 4 + 1.6887 x 5
b) H 0 : 3 = 4 = 5 = 0
H1: j 0 for at least one j
= 0.01
f 0 = 1651.25
f .01,3,36 = 4.38
t.005, 36 = 2. 72
H 0 : 3 = 0
H1: 3 0
t 0 = 2.06
| t 0 | >/ t / 2 , 36
Do not reject H0
d) R2 = 0. 993
H0 :4 = 0
H1:4 0
H 0 :5 = 0
H1: 5 0
t 0 = 22.91
| t 0 | > t / 2 , 36
t 0 = 3.00
| t 0 | > t / 2 , 36
Reject H0
Adj. R 2 = 0. 9925
Reject H0
Normal Score
-1
-2
-100
100
Residual
f) Plot is satisfactory.
Residuals Versus the Fitted Values
(response is y)
100
Residual
12-88.
-100
3000
4000
5000
Fitted Value
12-61
Residual
100
-100
29000
29500
30000
x3
f 0 = 1321.39
f .01,3,36 = 4.38
f 0 >> f ,3,36
t.005, 36 = 2. 72
H0:3
H1: 3
=0
0
t 0 = 1.45
H0:4 = 0
H1:4 0
H 0 :5 = 0
| t 0 | > t / 2 , 36
| t 0 | >/ t / 2 , 36
H1:5 0
t 0 = 19.95
| t 0 | >/ t / 2 , 36
t 0 = 2.53
Reject H0
Do not reject H0
c) Curvature is evident in the residuals plots from this model.
Do not reject H0
95
90
80
Percent
12-89.
70
60
50
40
30
20
10
5
-0.03
-0.02
-0.01
0.00
Residual
0.01
12-62
0.02
0.03
(response is y*)
(response is y*)
0.03
0.02
0.02
0.01
0.01
Residual
Residual
0.03
0.00
-0.01
-0.01
-0.02
-0.02
8.0
12-90.
0.00
8.1
8.2
8.3
Fitted Value
8.4
10.26
8.5
10.27
10.28
10.29
x3*
10.30
10.31
10.32
H1 : at least one j 0
= 0.01
f 0 = 4.81
f .01,5,19 = 4.17
f 0 > f ,5,19
Reject H0.
b) = 0.05
P-value = 0.005
t.025,19 = 2.093
H 0 :3 = 0
H 0 : 1 = 0
H1: 1 0
t 0 = 2.47
| t 0 | > t / 2 , 19
H1:2 0
H1:3 0
H0 :4 = 0
H1:4 0
t 0 = 2.74
| t 0 | > t / 2 ,19
t 0 = 2.42
| t 0 | > t / 2 ,19
| t 0 | > t / 2 ,19
| t 0 | >/ t / 2 ,19
Reject H0
Reject H0
Reject H0
Reject H0
Do not reject H0
H0 :2 = 0
t 0 = 2.80
H0 :5 = 0
H1: 5 0
t0 = 0.26
H 0 : 1 = 2 = 3 = 4 = 0
for at least one j
H1 : j 0
= 0.05
f 0 = 6.28
f .05, 4, 20 = 2.87
f 0 > f , 4, 20
Reject H0.
= 0.05
H 0 : 1 = 0
H1:1 0
t0 = 2.53
| t 0 | > t / 2 , 20
t. 025, 20 = 2. 086
H 0 :2 = 0
H1:2 0
t 0 = 2.89
| t 0 | > t / 2 , 20
H0 :3 = 0
H 0 :4 = 0
t 0 = 2.49
t 0 = 3.05
| t 0 | > t / 2 , 20
H1:3 0
| t 0 | > t / 2 , 20
Reject H0
H1:4 0
Reject H0
Reject H0
Reject H 0
th
d) The addition of the 5 regressor results in a loss of one degree of freedom in the denominator and the reduction in
SSE is not enough to compensate for this loss.
12-63
H1 : j 0
= 0.05
f 0 = 19.53
f .05, 4,19 = 2.90
f 0 > f 0.05, 4,19
Reject H0.
t 0 = 3.96
| t0 |> t0.025,19
t.025,19 = 2. 093
H 0 :2 = 0
H1:2 0
t 0 = 6.43
| t0 |> t0.025,19
t 0 = 3.64
| t0 |> t0.025,19
| t0 |> t0.025,19
Reject H0
Reject H0
Reject H0
Reject H0
= 0.05
H 0 : 1 = 0
H1:1 0
H0 :3 = 0
H0 :4 = 0
H1:3 0
H1:4 0
t0 = 3.39
Residuals Versus x1
(response is y)
(response is y)
2
Residual
Residual
-1
-1
-2
-2
0
8
x1
10
12
14
16
10
20
25
Residuals Versus x4
Residuals Versus x3
(response is y)
(response is y)
2
Residual
2
Residual
15
x2
-1
-1
-2
-2
2
10
12
8
x4
x3
12-64
10
12
12-91.
H1 : at least one j 0
= 0.05
f 0 = 21.79
f .05, 4, 20 = 2.87
f 0 > f 0.05, 4, 20
Reject H0 and conclude that regression is significant at = 0.05.
= 0.05
t .025, 20 = 2.086
H0 :1 = 0
H1:1 0
H 0 :2 = 0
H1:2 0
H0 :3 = 0
H0 :4 = 0
H1: 3 0
H1:4 0
t 0 = 5.76
t 0 = 5.96
t 0 = 2.90
| t 0 |> t 0.025,20
| t 0 |> t 0.025,20
| t 0 |> t 0.025,20
t 0 = 4.99
| t 0 |> t 0.025,20
Reject H0
Reject H0
Reject H0
Reject H0
c) The residual plots are more satisfactory than those in Exercise 12-90.
Residuals Versus the Fitted Values
(response is y*)
1.0
0.5
Residual
0.0
-0.5
-1.0
-1.5
-2.0
-2
-1
0
Fitted Value
1.0
1.0
0.5
0.5
0.0
0.0
Residual
Residual
(response is y*)
-0.5
-0.5
-1.0
-1.0
-1.5
-1.5
-2.0
-2.0
0.2
0.3
0.4
0.5
0.6
0.7
0.2
x1*
0.3
0.4
0.5
x2*
12-65
0.6
0.7
(response is y*)
1.0
1.0
0.5
0.5
0.0
0.0
Residual
Residual
(response is y*)
-0.5
-0.5
-1.0
-1.0
-1.5
-1.5
-2.0
-2.0
0.3
0.5
x3*
0.6
1.5
0.7
2.0
H1 : at least one j 0
= 0.05
f 0 = 300.11
f .05, 2, 7 = 4.74
f 0 >> f 0.05, 2, 7
Reject H0.
c) H 0 :11 = 0
H1:11 0
= 0.05
SS ( | ) / r 2.4276 / 1
F0 = R 11 1
=
0.04413
MS E
f 0 = 55.01
f .05,1, 7 = 5.59
f 0 >> f ,1, 7
Reject H0.
d) There is some indication of non-constant variance.
Residuals Versus the Fitted Values
(response is y)
0.3
0.2
0.1
Residual
12-92.
0.4
0.0
-0.1
-0.2
-0.3
11
12
13
14
Fitted Value
12-66
15
16
2.5
x4*
3.0
3.5
Normal Score
-1
-0.3
-0.2
-0.1
0.0
0.1
0.2
0.3
Residual
12-93.
MS E = 0.013156
C p = 4.0
VIF ( 5 ) = 29.1
*
Yes, VIFs for x3 and x5 exceed 10
12-94.
MS E ( p) = 0.41642
Min C p = 5.0
b) Syy = 1230. 5
H 0 : 1 = 2 = 0
H1 : at least one j 0
= 0.05
12-67
f 0 = 55.37
f .05, 2,12 = 3.89
f 0 > f 0.05, 2,12
Reject H0 and conclude that the regression model is significant at = 0.05
c) R 2 = SS R = 1110.2 = 0.9022 or 90.22%
1230.5
S yy
d) k = 3 p = 4
SSE ' = 117. 20
SS E ' 117.2
MS E ' =
=
= 10.65
n p
11
No, MSE increased with the addition of x3 because the reduction in SSE was not enough to compensate for the loss in
one degree of freedom in the error sum of squares. This is why MSE can be used as a model selection criterion.
e) SS R = S yy SS E = 1230.5 117.20 = 1113.30
SS R ( 3 | 2 , 1 , 0 ) = SS R ( 3 2 1 | 0 ) SS R ( 2 , 1 | 0 )
= 1113 .30 1110 .20
= 3 .1
H0 : 3 = 0
H1 : 3 0
= 0.05
f0 =
SS R ( 3 | 2 , 1 , 0 ) / r
3.1 / 1
=
= 0.291
117.2 / 11
SS E ' / n p
f .05,1,11 = 4.84
f 0 >/ f 0.05,1,11
Do not reject H0.
a)
The model assumptions are not violated as shown in the graphs below.
Where X1 = AVG, X2 = R, X3 = H, X4 = 2B, X5 = 3B, X6 = HR, X7 = RBI, X8 = BB, X9 = SO, X10 = SB
X11 = GIDP, X12 = LOB and X13 = OBP
The model assumptions appear reasonable.
Nor mal P r obability P lot of the Residuals
(response is W)
2
Score
12-96.
-1
-2
-3
-2
-1
0
1
Standardized Residual
12-68
(response is W)
(response is W)
2
Standardized Residual
Standardized Residual
-1
-2
-1
-2
65
70
75
80
85
Fitted Value
90
95
100
20
40
60
80
Standardized Residual
-1
-2
800
900
1000
1100
1200
1300
SO_B
Standardized Residual
-1
-2
400
450
500
550
600
650
BB_B
Standardized Residual
-1
-2
100
120
140
160
180
HR_B
100
SB
200
220
240
260
b)
12-69
120
140
160
Based on the graphs below, the model assumptions are not violated
Nor mal P r obability P lot of the Residuals
(response is W)
2
Score
-1
-2
-3
-2
-1
0
1
Standardized Residual
Standardized Residual
-1
-2
60
70
80
Fitted Value
90
100
Standardized Residual
-1
-2
25
30
35
40
SV
45
50
55
Standardized Residual
-1
-2
120
140
160
180
200
220
HR_P
12-70
Standardized Residual
-1
-2
350
400
450
500
BB_P
550
600
650
Standardized Residual
-1
-2
900
1000
1100
SO_P
1200
1300
0.010207 x so _ b + 0.07897 xlob 870.2 xobp 134.79 xera + 0.81681xer 0.06698 xhr _ p
0.032314 xbb _ p + 0.008755 x so _ p
Every variable in the above model is significant at =0.10. If is decreased to 0.05, SO_P is no longer
significant. The residual plots do not show any violations of the model assumptions (only a few plots of
residuals vs. the predictors are shown)
12-71
Score
-1
-2
-3
-2
-1
0
1
Standardized Residual
(response is W)
(response is W)
2
Standardized Residual
Standardized Residual
-1
-2
60
70
80
Fitted Value
90
100
(response is W)
Standardized Residual
-1
-2
0.250
0.255
0.260
0.265
0.270
AVG_B
0.275
0.280
0.285
Standardized Residual
-1
-1
-2
3.5
R2 =
4.0
4.5
ERA
800
900
1000
1100
SO_B
a)
-2
50
12-97.
5.0
5.5
SS R
S yy
12-72
1200
1300
SS R = R 2 ( S yy ) = 0.94(0.50) = 0.47
SS E = S yy SS R = 0.5 0.47 = 0.03
H 0 : 1 = 2 = ... = 6 = 0
H 1 : at least one j 0
= 0.05
f0 =
SS R / k
0.47 / 6
=
= 18.28
SS E / n p 0.03 / 7
p=6
R 2 = 0.92
c)
SS E
0.04
=
= 0.005
8
n p
0.03
MS E ( full ) =
= 0.004
7
MS E (reduced ) =
No, the MSE is larger for the reduced model, although not by much. Generally, if adding a variable to a model
reduces the MSE it is an indication that the variable may be useful in explaining the response variable. Here the
decrease in MSE is not very great because the added variable had no real explanatory power.
Mind-Expanding Exercises
12-98.
Because
R2 =
SS E
SS R / k
, F0 =
and this is the usual F-test for significance
S yy
SS E /(n k 1)
0.90 / 4
F0 =
= 33.75 and the critical value is f.05,4,15 = 3.06 . Because
(1 0.9) /( 20 4 1)
SS R
S yy
of regression. Then,
and
1 R2 =
12-73
i =1 j =1
MS E =
13-47.
(
j =1
ij
( y
yi ) 2
ij
and
a (n 1)
y ij = + a i + ij .
Then
y ij y i = ij i.
and
i. )
is recognized to be the sample variance of the independent random variables
n 1
i1 , i 2 ,, in .
n
2
( ij i. )
j =1
=2
E=
n 1
Therefore,
and
i =1
E ( MS E ) =
= 2.
The development would not change if the random effects model had been specified because
y ij y i = ij i.
13-48.
n ( y i. y.. ) 2
F=
i =1
MSE
F=
Now,
n
2
is too large.
( y1. y 2. ) 2
MSE
( y1. y 2. ) 2
and MSE = s 2p .
MSE n2
Consequently, F = t 2 . Also, the distribution of the square of a t random variable with a(n - 1) degrees of
freedom is an F distribution with 1 and a(n - 1) degrees of freedom. Therefore, if the tabulated t value for a
two-sided t-test of size is t 0 , then the tabulated F value for the F test above is t 2 . Therefore, t > t 0
0
13-49. MS E
i =1
j =1
( yij yi. ) 2
2(n 1)
and
j =1
( yij yi. ) 2
n 1
in the t-test.
13-50.
i =1
i =1
V (Yi. ) = ni i2 .
Then,
i =1
i =1
V ( ciYi. ) = 2 ci2 ni
13-41
13-51.
If b, c, and d are the coefficients of three orthogonal contrasts, it can be shown that
( bi yi. ) 2
i =1
a
( ci yi. ) 2
i =1
a
2
i
i =1
i =1
a
2
i
i =1
i =1
we have
( yi. ) 2
i =1
= yi2.
2
i
yi2 y..2
n N
by n,
( d i yi. ) 2
i =1
obtained from a geometrical argument. The square of the distance of any point in four-dimensional space from the zero
point can be expressed as the sum of the squared distance along four orthogonal axes. Let one of the axes be the 45
degree line and let the point be ( y1. , y2. , y3. , y4. ). The three orthogonal contrasts are the other three axes. The square of
a
yi2. and this equals the sum of the squared distances along each of the
i =1
four axes.
13-52.
Because
2 =
n ( i ) 2
i =1
a
D2
( i ) 2 .
2
i =1
2
( 1 x) + ( 2 x) 2 is minimized for x equal
a 2
(1 -
1 + 2
2
) 2 + (2 -
1 + 2
2
) 2 (1 ) 2 + (2 )2 (i ) 2
i =1
2
2
2
a
Then, 1 2 + 2 1 = D + D = D ( ) 2 .
i
13-53.
MS E =
i =1
( yij yi. ) 2
j =1
a(n 1)
i =1
i =1
si2
where
( n 1) S i2
s i2 =
j =1
( y ij y i. ) 2
n 1
. Because s2 is the
i
sample variance of
y i1 , y i 2 ,..., y in ,
freedom. Then,
a ( n 1) MS E has a
2
P( 12 , a ( n 1)
2
a(n 1) MS E
2 , a ( n 1) ) = 1
2
a(n 1) MS
a(n 1) MS E
E
= P
2
2
12 , a ( n 1)
, a ( n 1)
2
2
13-42
13-54.
( N a ) MS E
V (Yi. ) = 2 +
2
n
(a 1) MS T
n( 2 + n )
2
(a 1) MS T
n 2 + 2
degrees of freedom. Using the independence of MST and MSE, we conclude that
MS T MS E
/ 2
2
2
n +
has an F( a 1) ,( N a ) distribution.
Therefore,
P ( f 1 , a 1, N a
2
MS T
2
f ,a 1, N a ) = 1
2
MS E n 2 + 2
1 1
2 1 1
MS T
MS T
1
= P f
1 2 f
n 1 2 , a 1, N a MS E
n 2 , a 1, N a MS E
by an algebraic solution for
13-55.
a) As in Exercise 13-54,
and P( L
MS T
2
MS E n 2 + 2
2
U) .
2
has an F( a 1) ,( N a ) distribution.
and
2
U)
2
2
1
1
= P( 2 )
U
L
1 = P( L
1
1
+ 1 2 + 1 + 1)
U
L
2
= P(
2
L
U
)
= P(
2
2
L + 1 + U + 1
13-43
b)
2
U)
2
2 +1
= P ( L + 1 2 U + 1)
2 + 2
= P ( L + 1 2 U + 1)
2
1
1
)
= P(
2
2
U + 1 +
L +1
1 = P( L
Therefore,
13-56. MST
i =1
1
1
,
)
U +1 L +1
2 + 2
ni ( yi. y.. )2
and for any random variable X, E ( X 2 ) = V ( X ) + [ E ( X )]2 .
a 1
Then,
a
E ( MST ) =
i =1
a 1
Now, Y1. Y.. = ( )Y11 + ... + ( n1 N1 )Y1n N1 Y21 ... N1 Y2 n ... N1 Ya1 ... N1 Yan
1
2
a
1
1
n1
1
N
and
N n1 2
2
V (Y1. Y.. ) = ( n11 N1 ) 2 n1 +
= ( n11 N1 )
2
N
n
n
E (Y1. Y.. ) = ( n11 N1 )n1 1 2 2 ... a a = 1from the constraint
N
N
Then,
a
E ( MST ) =
a
=2 +
Because
13-57.
i =1
i =1
ni {( n1i N1 ) 2 + i2 }
a 1
i =1
a 1
ni i2
a 1
E ( MS E ) = 2 , this does suggest that the null hypothesis is as given in the exercise.
t ,a ( n 1)
2
t 2,a ( n1)
2
As in Exercise 13-48, t , a ( n 1)
2
n=
[(1 nNi ) 2 + ni i2 ]
2 MS E
n
2 MS E
n
= A2
2 MS E F ,1, a ( n 1)
A2
13-44
=A
b) Because n determines one of the degrees of freedom of the tabulated F value on the right-side of the equation in part (a),
some approximation is needed. Because the value for a 95% confidence interval based on a normal distribution is 1.96, we
approximate
t ,a ( n 1)
by 2 and we approximate
by 4.
Then,
n=
2(4)(4)
= 8 . With n = 8, a(n - 1) = 35 and F0.05,135
= 4.12 .
,
4
The value 4.12 can be used for F in the equation for n and a new value can be computed for n as
n=
2(4)(4.12)
= 8.24 8
4
Because the solution for n did not change, we can use n = 8. If needed, another iteration could be used to
refine the value of n.
13-45
CHAPTER 13
Section 13-2
13-1.
F
14.76
P
0.000
Reject H0 and conclude that cotton percentage affects mean breaking strength.
STRENGTH
25
15
5
15
20
25
30
35
COTTON
b) Tensile strength seems to increase up to 30% cotton and declines at 35% cotton.
Mean Strength
20
18
16
14
12
10
15
Level
15
20
25
30
35
N
5
5
5
5
5
Pooled StDev =
20
Mean
9.800
15.400
17.600
21.600
10.800
2.839
25
Cotton Percentage
StDev
3.347
3.130
2.074
2.608
2.864
30
35
c) The normal probability plot and the residual plots show that the model assumptions are reasonable.
13-1
(response is STRENGTH)
(response is STRENGTH)
6
6
4
3
Residual
Residual
1
0
2
1
0
-1
-1
-2
-2
-3
-3
-4
-4
10
15
20
-2
-1
Fitted Value
Normal Score
Residual
3
2
1
0
-1
-2
-3
-4
15
25
35
COTTON
F
3.59
P
0.053
Do not reject H0. There is no evidence that flow rate affects etch uniformity.
UNIFORMIT
125
160
200
FLOW
Normal Score
13-2
-1
-2
-1
Residual
13-2
(response is UNIFORMI)
(response is UNIFORMI)
Residual
Residual
-1
-1
120
130
140
150
160
170
180
190
3.5
200
4.0
13-3.
4.5
Fitted Value
FLOW
F
12.73
P
0.000
(response is STRENGTH)
200
100
100
Residual
200
-100
-100
-200
-200
1
2650
-1
-2
-200
-100
2750
2850
2950
Fitted Value
TECHNIQU
Normal Score
Residual
(response is STRENGTH)
100
200
Residual
13-3
3050
3150
13-4
F
4.01
P
0.046
Reject H0
b)There is some indication of greater variability in circuit two. There is some curvature in the normal
probability plot.
Normal Score
-1
-2
-10
10
Residual
(response is RESPONSE)
(response is RESPONSE)
10
Residual
Residual
10
-10
-10
1
18
CIRCUITT
23
Fitted Value
y3 t 0.025,12
18.4 2.179
13-5.
MSE
MSE
i y3 + t 0.025,12
n
n
32.6
32.6
1 18.4 + 2.179
5
5
12.84 1 23.96
13-4
28
b) There is some indication of that the variability of the response may be increasing as the mean response
increases. There appears to be an outlier on the normal probability plot.
Residuals Versus COATINGT
(response is CONDUCTI)
5
5
Residual
Residual
-5
-5
-10
-10
130
COATINGT
135
140
145
Fitted Value
Normal Score
-1
-2
-10
-5
Residual
MSE
MSE
i y1 + t 0.015,15
n
n
y1 t 0.025,15
16.2
16.2
1 145.00 + 2.131
4
4
140.71 1 149.29
145.00 2.131
99% confidence interval on the difference between the means of coating types 1 and 4.
y1 y 4 t 0.005,15
(145.00 129.25) 2.947
2MSE
2MSE
1 4 y1 y4 + t 0.005,15
n
n
2(16.2)
2(16.2)
1 4 (145.00 129.25) 2.947
4
4
7.36 1 4 24.14
13-5
13-6
F
30.85
P
0.000
Reject H0
b) P-value 0
c)
Residuals Versus ORIFICE
Residual
0
-1
0
-1
-2
-2
-3
-3
-4
-4
-5
-5
0.5
1.0
1.5
2.0
60
70
ORIFICE
80
Fitted Value
Normal Score
Residual
(response is RADON)
-1
-2
-5
-4
-3
-2
-1
Residual
y5 t 0.025,18
65 2.101
MSE
MSE
i y5 + t 0.015,18
n
n
7.35
7.35
1 65 + 2.101
4
4
62.15 1 67.84
13-6
13-7.
F
1.87
P
0.214
Do not reject H0
b) P-value = 0.214
c) The residual plot suggests nonconstant variance. The normal probability plot looks acceptable.
Residuals Versus RODDING
(response is STRENGTH)
(response is STRENGTH)
100
Residual
Residual
100
-100
-100
10
15
20
25
1500
1550
RODDING
1600
Fitted Value
Normal Score
-1
-2
-100
100
Residual
13-8
13-7
(response is TEMPERAT)
(response is TEMPERAT)
Residual
Residual
-1
-1
-2
-2
1
12
13
PREPMETH
14
15
Fitted Value
Normal Score
-1
-2
-2
-1
Residual
MSE
MSE
i y1 + t 0.015,16
n
n
y1 t 0.025,16
14.8 2.120
13-9.
0.497
0.497
3 14.8 + 2.120
5
5
14.13 1 15.47
F
8.30
P
0.002
Reject H0
b) P-value = 0.002
c) The residual plots indicate that the constant variance assumption is reasonable. The normal probability
plot has some curvature in the tails but appears reasonable.
13-8
Residual
10
-10
75
85
95
Fitted Value
(response is STRENGTH)
(response is STRENGTH)
2
10
Normal Score
Residual
1
0
-10
-1
-2
AIRVOIDS
-10
10
Residual
d) 95% Confidence interval on the mean of retained strength where there is a high level of air voids
y3 t 0.025,21
75.5 2.080
MSE
MSE
i y3 + t 0.015,21
n
n
74.1
74.1
3 75.5 + 2.080
8
8
69.17 1 81.83
e) 95% confidence interval on the difference between the means of retained strength at the high level and the
low levels of air voids.
y1 y3 t 0.025,21
(92.875 75.5) 2.080
2MSE
2MSE
1 3 y1 y3 + t 0.025, 21
n
n
2(74.1)
2(74.1)
1 4 (92.875 75.5) 2.080
8
8
8.42 1 4 26.33
13-9
13-10. a)
ANOVA
Source
Factor
Error
Total
DF
5
30
35
SS
2.5858
0.8217
3.4075
MS
0.5172
0.0274
F
18.88
P
0.000
Data
8.8
8.6
8.4
8.2
8.0
-0.5
0.5
Yes, the box plot and ANOVA show that there is a difference in the cross-linker level.
b) Anova table in part (a) showed the p-value = 0.000 < = 0.05. Therefore there is at least one
level of cross-linker is different. The variability due to random error is SSE = 0.8217
c) Domain spacing seems to increase up to the 0.5 cross-linker level and declines once crosslinker level reaches 1.
Scatterplot of Mean domain spacing vs Cross-linker level
8.9
8.8
8.7
8.6
8.5
8.4
8.3
8.2
8.1
8.0
-1.0
-0.5
0.0
Cross-linker level
13-10
0.5
1.0
d) The normal probability plot and the residual plots show that the model assumptions are
reasonable.
99
0.2
Residual
Percent
90
50
10
1
-0.4
-0.2
0.0
Residual
0.2
0.0
-0.2
-0.4
0.4
8.0
8.2
8.4
8.6
Fitted Value
8.8
6.0
Residual
Frequency
0.2
4.5
3.0
1.5
0.0
-0.3
-0.2
-0.1
0.0
Residual
0.1
0.2
0.0
-0.2
-0.4
0.3
10
15
20
25
Observation Order
13-11. a) No, the diet does not affect the protein content of cows milk.
Comparative boxplots
Boxplot of C 3 by C 4
4.5
C3
4.0
3.5
3.0
2.5
Barley
Barley+lupins
C4
lupins
ANOVA
Source
C4
Error
Total
DF
2
76
78
SS
0.235
12.364
12.599
S = 0.4033
MS
0.118
0.163
F
0.72
P
0.489
R-Sq = 1.87%
R-Sq(adj) = 0.00%
13-11
30
35
c) The Barley diet has the highest average protein content and lupins the lowest.
Scatter plot of mean pr otein vs diet type
3.900
mean protein
3.875
3.850
3.825
3.800
3.775
3.750
Barley
Barley+Lupins
Diet type
Lupins
Score
1
0
-1
-2
-3
-1.0
-0.5
0.0
Residual
0.5
1.0
1.5
Residual
0.5
0.0
-0.5
-1.0
3.750
3.775
3.800
3.825
Fitted Value
3.850
3.875
3.900
RESI1
0.5
0.0
-0.5
-1.0
Barley
Barley+Lupins
Diet
Lupins
13-12
13-12. a) From the analysis of variance shown below, F0.05, 3,8 = 4.07 > F0 = 3.43 there is no
difference in the spoilage percentage when using different AO solutions.
ANOVA
Source
AO solutions
Error
Total
DF
3
8
11
SS
3364
2617
5981
MS
1121
327
F
3.43
P
0.073
b) From the table above, the P-value = 0.073 and the variability due to random error is SSE =
2617.
c) A 400ppm AO solution should be used because it produces the lowest average spoilage
percentage.
Scatterplot of Avg. Spoilage vs AO solution
70
Avg. Spoilage
60
50
40
30
20
0
100
200
AO solution
300
400
d) The normal probability plot and the residual plots show that the model assumptions are
reasonable.
99
20
Residual
Percent
90
50
10
1
-20
-40
-20
0
Residual
20
40
30
40
50
Fitted Value
60
70
1
0
13-13
20
Residual
Frequency
-20
-20
-10
0
10
Residual
20
30
13-13
F
2.62
4 5 6 7 8 9
Observation Order
P
0.083
10 11 12
Total
21
Do not reject H0
0.4582
b) P-value = 0.083
c) Residuals are acceptable
Normal Probability Plot of the Residuals
(response is density)
2
Normal Score
-1
-2
-0.2
-0.1
0.0
0.1
0.2
Residual
0.2
0.2
0.1
0.1
Residual
Residual
(response is DENSITY)
0.0
-0.1
0.0
-0.1
-0.2
-0.2
100
110
120
130
140
150
160
170
180
21.55
TEMPERAT
13-14
21.65
Fitted Value
13-15.
13-14
21.75
13-16.
Significant differences between levels 15 and 20, 15 and 25, 15 and 30, 20 and 30, 20 and 35, 25 and 30, 25
and 35, and 30 and 35.
Fisher's pairwise comparisons
Family error rate = 0.0251
Individual error rate = 0.0100
Critical value = 3.055
Intervals for (column level mean) - (row level mean)
1
2
-18.426
3.626
-8.626
13.426
-1.226
20.826
13-18
1.40
-1.777
6.277
Significant differences between levels 0.37 and all other levels; 0.51
and 1.02, 1.40, and 1.99; 0.71 and 1.40 and 1.99; 1.02 and 1.40 and 1.99
13-15
13-19
2
3
4
5
-8.642
8.142
5.108
21.892
7.358
24.142
-8.642
8.142
5.358
22.142
7.608
24.392
-8.392
8.392
-6.142
10.642
-22.142
-5.358
-24.392
-7.608
13-21.
13-22.
a)
LSD = t 0.025, 25
2MS E
2 0.0274
= 2.042
= 0.1952
b
6
13-16
level
-0.75
-1
0
0.5
1
Lower
-0.5451
-0.7451
-0.1118
0.0215
-0.1451
Cross-linker
level
-0.75
-1
0
0.5
1
Center
-0.3500
-0.5500
0.0833
0.2167
0.0500
Upper
-0.1549
-0.3549
0.2785
0.4118
0.2451
---------+---------+---------+---------+
(---*---)
(---*---)
(---*---)
(---*---)
(---*---)
---------+---------+---------+---------+
-0.50
0.00
0.50
1.00
Lower
-0.3951
0.2382
0.3715
0.2049
Center
-0.2000
0.4333
0.5667
0.4000
Upper
-0.0049
0.6285
0.7618
0.5951
---------+---------+---------+---------+
(---*---)
(---*---)
(---*---)
(---*---)
---------+---------+---------+---------+
-0.50
0.00
0.50
1.00
Lower
0.4382
0.5715
0.4049
Center
0.6333
0.7667
0.6000
Upper
0.8285
0.9618
0.7951
---------+---------+---------+---------+
(---*---)
(---*---)
(---*---)
---------+---------+---------+---------+
-0.50
0.00
0.50
1.00
Lower
-0.0618
-0.2285
Center
0.1333
-0.0333
Upper
0.3285
0.1618
---------+---------+---------+---------+
(---*---)
(---*---)
---------+---------+---------+---------+
-0.50
0.00
0.50
1.00
Lower
Center
Upper
13-17
-0.3618
Cross-linker
level
1
-0.1667
0.0285
---------+---------+---------+---------+
(---*---)
---------+---------+---------+---------+
-0.50
0.00
0.50
1.00
Cross-linker levels -0.5, 0, 0.5 and 1 do not differ. Cross-linker levels -0.75 and -1 do not differ
from one other but both are significantly different to the others.
b) The mean values are
8.0667, 8.2667, 8.6167, 8.7, 8.8333, 8.6667
MS E
0.0274
=
= 0.0676
b
6
X =
1
0.8
0.6
0.4
0.2
0
8
8.2
8.4
8.6
8.8
With a scaled normal distribution over this plot, the conclusions are similar to those from the LSD
method.
13-23.
a) There is no significant difference in protein content between the three diet types.
Fisher 99% Individual Confidence Intervals
All Pairwise Comparisons among Levels of C4
Simultaneous confidence level = 97.33%
C4 = Barley subtracted from:
C4
Barley+lupins
lupins
Lower
-0.3207
-0.4218
Center
-0.0249
-0.1260
Upper
0.2709
0.1698
C4
Barley+lupins
lupins
-------+---------+---------+---------+-(-----------*-----------)
(-----------*-----------)
-------+---------+---------+---------+--0.25
0.00
0.25
0.50
Lower
-0.3911
Center
-0.1011
Upper
0.1889
-------+---------+---------+---------+-(-----------*-----------)
-------+---------+---------+---------+--
13-18
-0.25
0.00
0.25
0.50
b) The mean values are: 3.886, 3.8611, 3.76 (barley, b+l, lupins)
From the ANOVA the estimate of can be obtained
Source DF
SS
MS
F
P
C4
2
0.235 0.118 0.72 0.489
Error
76 12.364 0.163
Total
78 12.599
S = 0.4033
R-Sq = 1.87%
R-Sq(adj) = 0.00%
The minimum sample size could be used to calculate the standard error of
a sample mean
X =
MS E
=
b
0.163
= 0.081
25
The graph would not show any differences between the diets.
13-24.
= 55 , 1 = -5, 2 = 5, 3 = -5, 4 = 5.
n(100)
2 =
= n,
a 1 = 3
4(25)
Various choices for n yield:
n
4
5
2
4
5
a (n 1) = 4(n 1)
2
2.24
a(n-1)
12
16
Power=1-
0.80
0.90
Therefore, n = 5 is needed.
13-25
2
7.32
10.98
2.7
3.13
a(n-1)
5
10
Power=1-
0.55
0.95
Therefore, n = 3 is needed.
Section 13-3
13-26
a)
Analysis of Variance for UNIFORMITY
Source
DF
SS
MS
WAFERPOS
3
16.220
5.407
Error
8
5.217
0.652
Total
11
21.437
F
8.29
P
0.008
Reject H0, and conclude that there are significant differences among wafer positions.
2 =
d) Greater variability at wafer position 1. There is some slight curvature in the normal probability plot.
13-19
-0.25
0.00
0.25
0.50
b) The mean values are: 3.886, 3.8611, 3.76 (barley, b+l, lupins)
From the ANOVA the estimate of can be obtained
Source DF
SS
MS
F
P
C4
2
0.235 0.118 0.72 0.489
Error
76 12.364 0.163
Total
78 12.599
S = 0.4033
R-Sq = 1.87%
R-Sq(adj) = 0.00%
The minimum sample size could be used to calculate the standard error of
a sample mean
X =
MS E
=
b
0.163
= 0.081
25
The graph would not show any differences between the diets.
13-24.
= 55 , 1 = -5, 2 = 5, 3 = -5, 4 = 5.
n(100)
2 =
= n,
a 1 = 3
4(25)
Various choices for n yield:
n
4
5
2
4
5
a (n 1) = 4(n 1)
2
2.24
a(n-1)
12
16
Power=1-
0.80
0.90
Therefore, n = 5 is needed.
13-25
2
7.32
10.98
2.7
3.13
a(n-1)
5
10
Power=1-
0.55
0.95
Therefore, n = 3 is needed.
Section 13-3
13-26
a)
Analysis of Variance for UNIFORMITY
Source
DF
SS
MS
WAFERPOS
3
16.220
5.407
Error
8
5.217
0.652
Total
11
21.437
F
8.29
P
0.008
Reject H0, and conclude that there are significant differences among wafer positions.
2 =
d) Greater variability at wafer position 1. There is some slight curvature in the normal probability plot.
13-19
(response is UNIFORMI)
(response is UNIFORMI)
Residual
Residual
-1
-1
WAFERPOS
Fitted Value
Normal Score
-1
-2
-1
Residual
13-27
F
5.77
P
0.003
2 =
(response is OUTPUT)
0.2
0.2
0.1
0.1
Residual
Residual
(response is OUTPUT)
0.0
0.0
-0.1
-0.1
-0.2
-0.2
1
3.8
3.9
4.0
Fitted Value
LOOM
13-20
4.1
Normal Score
-1
-2
-0.2
-0.1
0.0
0.1
0.2
Residual
13-28. a) Yes, the different batches of raw material significantly affect mean yield at = 0.01 because pvalue is small.
Source
Batch
Error
Total
DF
5
24
29
SS
56358
58830
115188
MS
11272
2451
F
4.60
P
0.004
2 =
d) The normal probability plot and the residual plots show that the model assumptions are
reasonable.
13-21
99
100
50
Residual
Percent
90
50
0
-50
10
1
-100
-50
0
Residual
50
-100
100
100
50
13-29.
-75
-50
-25
0
25
Residual
50
75
1600
0
-50
2
0
1550
Fitted Value
Residual
Frequency
1500
-100
100
8 10 12 14 16 18 20 22 24 26 28 30
Observation Order
P
0.538
Do not reject H0, there is no significant difference among the chemical types.
18.0 24.0
= 1 . 2
5
b)
2 =
c)
2 = 24.0
set equal to 0
d) Variability is smaller in chemical 4. There is some curvature in the normal probability plot.
Residuals Versus CHEMICAL
(response is BRIGHTNE)
(response is BRIGHTNE)
10
Residual
Residual
10
-5
-5
78
79
80
Fitted Value
CHEMICAL
13-22
81
82
Normal Score
-1
-2
-5
10
Residual
13-30
a)
2
2
total
= position
+ 2 = 2.237
b)
2
position
= 0.709
2
total
13-31.
a) Instead of testing the hypothesis that the individual treatment effects are zero, we are testing
whether there is variability in protein content between all diets.
H 0 : 2 = 0
H 1 : 2 0
b) The statistical model is
i = 1,2,..., a
y = + i + ij
j = 1,2,..., n
i ~ N (0, 2 ) and i ~ N (0, 2 )
c)
The last TWO observations were omitted from two diets to generate equal sample sizes with n = 25.
DF
2
72
74
S = 0.405122
SS
0.2689
11.8169
12.0858
MS
0.1345
0.1641
R-Sq = 2.23%
F
0.82
P
0.445
R-Sq(adj) = 0.00%
2 = MS E = 0.1641
MS tr MS E 0.1345 0.1641
2 =
=
= 0.001184
n
25
13-23
Section 13-4
13-32.
a) Analysis of variance for Glucose
Source
Time
Min
Error
Total
DF
1
1
5
7
SS
36.13
128.00
3108.75
3272.88
MS
36.125
128.000
621.750
F
0.06
0.21
P
0.819
0.669
99
20
Residual
Percent
90
50
10
1
-50
-25
0
Residual
25
105
Fitted Value
110
20
Residual
Frequency
100
13-33.
-20
-40
50
1
0
-40
-30
-20
-10
0
Residual
10
0
-20
-40
20
13-24
F
8.92
4.39
3
4
5
6
Observation Order
P
0.000
0.007
1.15
1.05
1.05
SHAPE
SHAPE
1.15
0.95
0.95
0.85
0.85
0.75
0.75
11.73
14.37
16.59
20.43
23.46
28.74
b)
NOZZLE
VELOCITY
(response is SHAPE)
(response is SHAPE)
0.1
Residual
Residual
0.1
0.0
-0.1
0.0
-0.1
10
20
30
VELOCITY
NOZZLE
13-25
Residual
0.1
0.0
-0.1
0.7
0.8
0.9
1.0
Fitted Value
Normal Score
-1
-2
-0.1
0.0
0.1
Residual
13-34
MS
0.12833
0.27500
0.00889
F
14.44
30.94
P
0.001
0.000
Reject H0, and conclude that there are significant differences in hardness measurements between the tips.
b)
Fisher's pairwise comparisons
Family error rate = 0.184
Individual error rate = 0.0500
Critical value = 2.179
Intervals for (column level mean) - (row level mean)
1
2
3
2
-0.4481
0.3981
3
-0.2981
-0.2731
0.5481
0.5731
4
-0.7231
-0.6981
-0.8481
0.1231
0.1481
-0.0019
Significant difference between tip types 3 and 4
13-26
0.15
0.15
0.10
0.10
0.05
0.05
Residual
Residual
(response is HARDNESS)
0.00
-0.05
0.00
-0.05
-0.10
-0.10
1
SPECIMEN
0.15
Residual
0.10
0.05
0.00
-0.05
-0.10
9.2
9.3
9.4
9.5
9.6
9.7
9.8
9.9
10.0
10.1
10.2
Fitted Value
Normal Score
-1
-2
-0.10
-0.05
0.00
0.05
0.10
0.15
Residual
13-35.
TIPTYPE
F
3.00
30.32
13-27
P
0.125
0.001
0.02
0.02
0.01
0.01
0.00
0.00
Residual
Residual
(response is ARSENIC)
-0.01
-0.02
-0.02
-0.03
-0.03
1
SUBJECT
(response is ARSENIC)
0.02
Residual
0.01
0.00
-0.01
-0.02
-0.03
0.05
0.10
0.15
Fitted Value
-1
-2
-0.03
-0.02
-0.01
TEST
Normal Score
-0.01
0.00
0.01
0.02
Residual
13-28
13-36.
Normal Score
-1
-2
-500
500
1000
Residual
1000
1000
500
500
Residual
Residual
(response is PROTOPE)
-500
-500
LOT
10
STORAGE
13-29
20
Residual
500
-500
500
1000
1500
Fitted Value
13-37.
A version of the electronic data file has the reading for length 4 and width 5 as 2. It should be 20.
a) Analysis of Variance for LEAKAGE
Source
DF
SS
MS
LENGTH
3
72.66
24.22
WIDTH
4
90.52
22.63
Error
12
180.83
15.07
Total
19
344.01
F
1.61
1.50
P
0.240
0.263
Do not reject H0, mean leakage voltage does not depend on the channel length.
b) One unusual observation in width 5, length 4. There are some problems with the normal probability plot,
including the unusual observation.
Residuals Versus WIDTH
(response is LEAKAGE)
(response is LEAKAGE)
Residual
10
Residual
10
WIDTH
10
Fitted Value
(response is LEAKAGE)
2
10
Normal Score
Residual
-1
LENGTH
-2
0
Residual
13-30
10
F
6.16
3.86
P
0.009
0.031
Reject H0. And conclude that the mean leakage voltage does depend on channel length. By removing the
data point that was erroneous, the analysis results in a conclusion. The erroneous data point that was an
obvious outlier had a strong effect the results of the experiment.
Supplemental Exercises
a) Analysis of Variance for SURFACE ROUGNESS
Analysis of Variance for y
Source
DF
SS
MS
F
Material
3
0.2402
0.0801
4.96
Error
11
0.1775
0.0161
Total
14
0.4177
P
0.020
Reject H0
b) One observation is an outlier.
0.2
Residual
13-38
0.1
0.0
-0.1
-0.2
1
Material
13-31
F
6.16
3.86
P
0.009
0.031
Reject H0. And conclude that the mean leakage voltage does depend on channel length. By removing the
data point that was erroneous, the analysis results in a conclusion. The erroneous data point that was an
obvious outlier had a strong effect the results of the experiment.
Supplemental Exercises
a) Analysis of Variance for SURFACE ROUGNESS
Analysis of Variance for y
Source
DF
SS
MS
F
Material
3
0.2402
0.0801
4.96
Error
11
0.1775
0.0161
Total
14
0.4177
P
0.020
Reject H0
b) One observation is an outlier.
0.2
Residual
13-38
0.1
0.0
-0.1
-0.2
1
Material
13-31
Normal Score
-1
-2
-0.2
-0.1
0.0
0.1
0.2
0.3
Residual
c) There appears to be a problem with constant variance. This may be due to the outlier in the data.
Residuals Versus the Fitted Values
(response is Surf Rou)
0.3
Residual
0.2
0.1
0.0
-0.1
-0.2
0.1
0.2
0.3
0.4
0.5
Fitted Value
d) 95% confidence interval on the difference in the means of EC10 and EC1
y1 y4 t0.025,,11
(0.490 0.130) 2.201
13-39.
MSE MSE
MSE MSE
+
1 3 y1 y4 + t0.025,,11
+
n1
n2
n1
n2
(0.0161) (0.0161)
(0.0161) (0.0161)
+
1 4 (0.490 0.130) + 2.201
+
4
2
4
2
0.118 1 4 0.602
F
76.09
P
0.000
Reject H0, the type of alloy has a significant effect on mean contact resistance.
13-32
MSE
MSE
i y3 + t 0.005,27
n
n
y3 t 0.005,27
71.9
71.9
3 140.4 + 2.771
10
10
132.97 1 147.83
140.4 2.771
(response is RESISTAN))
(response is RESISTAN)
30
20
20
10
10
Residual
30
-10
-10
-20
-20
1
100
ALLOY
(response is RESISTAN)
-1
-2
-20
110
120
Fitted Value
Normal Score
Residual
d) Variability of the residuals increases with the response. The normal probability plot has some
curvature in the tails, indicating a problem with the normality assumption. A transformation of the
response should be conducted.
-10
10
20
30
Residual
13-33
130
140
13-40
F
1.68
P
0.211
c) There is some curvature in the normal probability plot. There appears to be some differences in the
variability for the different methods. The variability for method one is larger than the variability for method
3.
Normal Probability Plot of the Residuals
(response is score)
2
Normal Score
-1
-2
-5
-4
-3
-2
-1
Residual
Residual
0
-1
-2
-3
-4
-5
1
METHOD
d.)
13-41.
2 =
F
7.84
P
0.007
b) P-value = 0.007
c) Fisher's pairwise comparisons
Family error rate = 0.116
Individual error rate = 0.0500
Critical value = 2.179
Intervals for (column level mean) - (row level mean)
70
75
13-34
75
-16.7
72.7
35.3
124.7
80
7.3
96.7
There are significant differences in the mean volume for temperature levels 70 and 80; and 75 and 80. The
highest temperature results in the smallest mean volume.
d) There are some relatively small differences in the variability at the different levels of temperature. The
variability decreases with the fitted values. There is an unusual observation on the normal probability plot.
Normal Score
-1
-2
-50
50
(response is VOLUME)
(response is VOLUME)
50
50
Residual
Residual
Residual
-50
-50
70
75
80
1170
TEMPERAT
13-42.
1180
1190
1200
1210
1220
1230
1240
Fitted Value
13-35
1250
2.0401
2.4434
2.8234
2.5168
1.8368
(response is WEIGHTGA)
(response is WEIGHTGA)
0.5
Residual
Residual
0.5
0.0
-0.5
0.0
-0.5
100
150
200
50
60
MEANWEIG
70
80
AIRTEMP
Residual
0.5
0.0
-0.5
0
Fitted Value
Normal Score
-1
-2
-0.5
0.0
0.5
Residual
13-43.
13-36
F
6.23
4.27
P
0.000
0.002
90
100
(response is PCTERROR)
(response is PCTERROR)
500
500
Residual
1000
Residual
1000
-500
-500
1
PROJECT
ALGORITH
1000
Residual
500
-500
0
500
1000
Fitted Value
Normal Score
-1
-2
-500
500
1000
Residual
c) The best choice is algorithm 5 because it has the smallest mean and a low variability.
13-44. a) The normal probability plot shows that the normality assumption is not reasonable.
13-37
Mean
StDev
N
AD
P-Value
95
90
293.8
309.4
30
2.380
<0.005
Percent
80
70
60
50
40
30
20
10
5
-500
500
Obs
1000
1500
b) The normal probability plot shows that the normality assumption is reasonable.
Probability Plot of log(obs)
Normal - 95% CI
99
Mean
StDev
N
AD
P-Value
95
90
2.209
0.5060
30
0.687
0.066
Percent
80
70
60
50
40
30
20
10
5
Source
Treatments
Error
Total
0.5
DF
2
27
29
1.0
1.5
SS
1.188
6.237
7.425
2.0
2.5
log(obs)
MS
0.594
0.231
3.0
F
2.57
3.5
4.0
P
0.095
There is evidence to support the claim that the treatment means differ at = 0.1 for the
transformed data since the P-value = 0.095.
c) The normal probability plot and the residual plots show that the model assumptions are
reasonable.
13-38
90
0.5
Residual
Percent
50
10
1
-1.0
-0.5
0.0
Residual
0.5
0.0
-0.5
-1.0
1.0
1.0
0.5
4
2
0
13-45.
-0.8
a)
-0.4
0.0
Residual
0.4
2.2
2.3
Fitted Value
-0.5
-1.0
0.8
8 10 12 14 16 18 20 22 24 26 28 30
Observation Order
2 =0.284, =0.5333
Numerator degrees of freedom = a 1 = 4 = 1
b)
50
3.56
a(n-1)
Power = 1-
1.89
245
0.05
0.95
6.25
4.6875
2.5
0.0
=1.6,
2.4
Residual
Frequency
2.1
a(n-1)
Power = 1-
2.5
2.165
12
8
0.05
0.25
0.95
0.75
13-39
13-40
Mind-Expanding Exercises
15-43
Under the null hypothesis each rank has probability of 0.5 of being either positive or negative. Define the
random variable Xi as
1
Xi =
0
rank is positive
rank is negative
Then,
n
R + = xi i and E ( R + ) = E ( xi )i =
i =1
i =1
where V ( R ) = V ( xi )i
+
n(n + 1)
1
i=
2
4
since
E ( xi ) =
1
2
by independence
1 1 1
1
1
+ 0 2 [ E ( X i )] 2 = =
2 4 4
2
2
n(n + 1)(2n + 1)
Then, V ( R + ) =
24
V ( X i ) = 12
15-44
a)
p = P( X i +5 > X i ) =
4
There appears to be an upward trend.
5
and p = 0.5.
c) V = 4
1. The parameter of interest is number of values of i for which
X i +5 > X i
2. H 0 : there is no trend
3 H 1 : there is an upward trend
4. = 0.05
5. The test statistic is the observed number of values where X i +5 > X i or V = 4.
6. We reject H0 if the P-value corresponding to V = 4 is less than or equal to = 0.05.
7. Using the binomial distribution with n = 5 and p = 0.5, P-value = P( V 4| p = 0.5) = 0.1875
8. Conclusion: do not reject H0. There is not enough evidence to suspect an upward trend in the wheel
opening dimensions at =0.05.
15-45
15-19
CHAPTER 15
Section 15-2
15-1
a)
1. The parameter of interest is median of pH.
2. H 0 : ~ = 7.0
3 H : ~ 7.0
1
4. =0.05
5. The test statistic is the observed number of plus differences or r+ = 8.
6. We reject H0 if the P-value corresponding to r+ = 8 is less than or equal to = 0.05.
7. Using the binomial distribution with n = 10 and p = 0.5, P-value = 2P(R+ 8 | p = 0.5) = 0.1
8. Conclusion: we cannot reject H0. There is not enough evidence to reject the manufacturers claim that the
median of the pH is 7.0
b)
1. The parameter of interest is median of pH.
2. H 0 : ~ = 7.0
3 H : ~ 7.0
1
4. =0.05
5. The test statistic is
z0 =
r * 0.5n
0.5 n
r * 0.5n 8 0.5(10)
z0 =
=
= 1.90
0.5 n
0.5 10
8. Conclusion: we cannot reject H0. There is not enough evidence to reject the manufacturers claim that
the median of the pH is 7.0
P-value = 2[1 - P(|Z0| < 1.90)] = 2(0.0287) = 0.0574
15-2
a)
1. The parameter of interest is median titanium content.
2. H 0 : ~ = 8.5
3 H : ~ 8.5
1
4. =0.05
5. The test statistic is the observed number of plus differences or r+ = 7.
6. We reject H0 if the P-value corresponding to r+ = 7 is less than or equal to =0.05.
7. Using the binomial distribution with n=10 and p=0.5, P-value = 2P(R*7|p=0.5)=0.359
8. Conclusion: we cannot reject H0. There is not enough evidence to reject the manufacturers claim that
the median of the titanium content is 8.5.
b)
1. Parameter of interest is the median titanium content
2. H 0 : ~ = 8.5
3. H : ~ 8.5
1
4. =0.05
5. Test statistic is
z0 =
r + 0.5n
0.5 n
0.5 20
8. Conclusion, cannot reject H0. There is not enough evidence to conclude that the median titanium content
differs from 8.5. The P-value = 2*P( |Z| > 1.34) = 0.1802.
15-1
15-3
a)
1. Parameter of interest is the median impurity level.
2. H 0 : ~ = 2.5
3. H : ~ < 2.5
1
4. =0.05
5. The test statistic is the observed number of plus differences or r+ = 2.
6. We reject H0 if the P-value corresponding to r+ = 2 is less than or equal to = 0.05.
7. Using the binomial distribution with n = 22 and p = 0.5, P-value = P(R+ 2 | p = 0.5) = 0.0002
8. Conclusion, reject H0. The data supports the claim that the median is impurity level is less than 2.5.
b)
1. Parameter of interest is the median impurity level
2. H 0 : ~ = 2.5
3. H : ~ < 2.5
1
4. =0.05
5. Test statistic is
z0 =
r + 0.5n
0.5 n
0.5 22
8. Conclusion, reject H0 and conclude that the median impurity level is less than 2.5. The P-value = P(Z<3.84) = 0.000062
15-4
a)
1. Parameter of interest is the median margarine fat content
2. H 0 : ~ = 17.0
3. H : ~ 17.0
1
4. =0.05
5. The test statistic is the observed number of plus differences or r+ = 3.
6. We reject H0 if the P-value corresponding to r+ = 7 is less than or equal to =0.05.
7. Using the binomial distribution with n=6 and p=0.5, P-value = 2*P(R*3|p=0.5,n=6)=1.
8. Conclusion, cannot reject H0. There is not enough evidence to conclude that the median fat content differs
from 17.0.
b)
1. Parameter of interest is the median margarine fat content
2. H 0 : ~ = 17.0
3. H : ~ 17.0
1
4. =0.05
+
0. 5 n
0.5 6
8. Conclusion, cannot reject H0. The P-value = 1-(0) = 1 0.5 = 0.5. There is not enough evidence to
conclude that the median fat content differs from 17.0.
15-2
15-5
a)
1. Parameter of interest is the median compressive strength
2. H 0 : ~ = 2250
3. H : ~ > 2250
1
4. = 0.05
5. The test statistic is the observed number of plus differences or r+ = 7.
6. We reject H0 if the P-value corresponding to r+ = 7 is less than or equal to =0.05.
7. Using the binomial distribution with n = 12 and p = 0.5, P-value = P( R+ 7 | p = 0.5) = 0.3872
8. Conclusion, cannot reject H0. There is not enough evidence to conclude that the median compressive
strength is greater than 2250.
b)
1. Parameter of interest is the median compressive strength
2. H 0 : ~ = 2250
3. H : ~ > 2250
1
4. =0.05
+
0. 5 n
0.5 12
8. Conclusion, cannot reject H0. The P-value = 1-(0.58) = 1 - 0.7190 = 0.281. There is not enough
evidence to conclude that the median compressive strength is greater than 2250.
15-6
a)
1. Parameters of interest are the median caliper measurements
2. H 0 : ~D = 0
3. H : ~ 0
1
4. =0.05
5. The test statistic is r = min( r+ , r-).
6. Because =0.05 and n = 12, Appendix A, Table VIII gives the critical value of
r0*.05 = 2.
We reject
H0 in favor of H1 if r 2.
7. There are four ties that are ignored so that r+ = 6 and r - = 2 and r = min(6,2) = 2
8. Conclusion, reject H0. There is a significant difference in the median measurements of the two calipers at
= 0.05.
b)
1. Parameters of interest are the median caliper measurements
2. H 0 : ~D = 0
3. H : ~ 0
1
4. =0.05
5. Test statistic
z0 =
r + 0.5n
0.5 n
0.5 8
8. Conclusion, do not reject H0. There is not a significant difference in the median measurements of the two
calipers at = 0.05.
The P-value = 2[1-P( |Z0| < 1.41) = 0.159.
15-7
a)
15-3
1. Parameters of interest are the median hardness readings for the two tips
2. H 0 : ~D = 0
3. H : ~ 0
1
4. =0.05
5. The test statistic is r = min( r+, r-).
6. Because = 0.05 and n = 8, Appendix A, Table VIII gives the critical value of
r0*.05 = 0.
We reject
H0 in favor of H1 if r 0.
7. r+ = 6 and r - = 2 and r = min(6,2) = 2
8. Conclusion, cannot reject H0. There is not a significant difference in the tips.
b) P-value = 2P(R 6 | p = 0.5) = 0.289. The P-value is greater than 0.05, therefore we cannot reject H0
and conclude that there is no significant difference between median hardness readings for the two tips. This
result agrees with the result found in Exercise 15-9.
15-8
a)
1. Parameters of interest are the median drying times for the two formulations
2. H 0 : ~D = 0
3. H : ~ 0
1
4. =0.01
5. The test statistic is r = min( r+, r-).
6. Because = 0.01 and n = 20, Appendix A, Table VIII gives the critical value of
r0*.01 = 3.
We reject
H0 in favor of H1 if r 3.
7. There are two ties that are ignored so that r+ = 3 and r - = 15 and r = min(3,15) = 3
8. Conclusion, reject H0. There is a significant difference in the drying times of the two formulations at =
0.01.
b)
1. Parameters of interest are the median drying times of the two formulations
2. H 0 : ~D = 0
3. H : ~ 0
1
4. =0.01
5. Test statistic
z0 =
r + 0.5n
0.5 n
0.5 18
8. Conclusion, reject H0 and conclude that there is a significant difference between the drying times for the
two formulations at = 0.01.
The P-value = 2[1-P( |Z0| < 2.83) = 0.005.
c) P-value = 2P(R 3 | p = 0.5) = 0.0075. The exact P-value computed here agrees with the normal
approximation in Exercise 15-11 in the sense that both calculations would lead to the rejection of H0.
3.5
15-9
a)
f ( x) = e
~ =4.5)
c) = P(Type II error)=P(R* > 1|
15-4
e x = 0.5 .
3.5
~ =4.5) = 0.963
n =10, p = 0.4167, = P(Type II error)=P(R > 1|
*
15-10
a)
1. Parameters of interest are the median blood cholesterol levels
2. H 0 : ~D = 0
3. H : ~ 0
1
4. =0.05
5. The test statistic r = min( r+ , r-).
6. Because = 0.05 and n = 15, Appendix A, Table VIII gives the critical value of
r0*.05 = 3.
We reject
H0 in favor of H1 if r 3.
7. r+ = 14 and r - = 1 and r = min(14,1) = 1
8. Conclusion, reject H0. There a significant difference in the blood cholesterol levels at = 0.05.
b)
1. Parameter of interest is the difference in blood cholesterol levels
2. H 0 : ~D = 0
3. H : ~ 0
1
4. =0.05
5. Test statistic
z0 =
r + 0.5n
0.5n
0.5 15
8. Conclusion, reject H0. There is a difference between the blood cholesterol levels.
The P-value = 2*P (|Z| > 3.36) = 0.0008.
15-11
X
= 1.96 | = 1 = P( Z < 1.20) = 0.115
/ n
= P
c) The sign test that rejects if R 1 has = 0.011 based on the binomial distribution.
d)
0.1587 and n = 10 when = 1. Then = 0.487. The value of is greater for the sign test than for the normal
test because the Z-test was designed for the normal distribution.
15-5
Section 15-3
Note to Instructors: When an observation equals the hypothesized mean the observation is dropped and not
considered in the analysis.
15-12
a)
1. Parameter of interest is the mean titanium content
2 . H 0 : = 8 .5
3 . H 1 : 8 .5
4. =0.05
5. The test statistic is w = min( w + , w -).
6. We reject H0 if w
w0*.05 = 52, because = 0.05 and n = 20, Appendix A, Table IX gives the critical
value.
7. w+ = 80.5 and w - = 109.5 and w = min(80.5,109.5) = 80.5
8. Conclusion, because 80.5 > 52, we cannot reject H0. The mean titanium content is not significantly
different from 8.5 at = 0.05.
b)
1. Parameter of interest is the mean titanium content
2 . H 0 : = 8 .5
3 . H 1 : 8 .5
4. =0.05
W + n ( n + 1) / 4
n ( n + 1)( 2 n + 1) / 24
6. We reject H0 if the |Z0| > Z0.025 = 1.96, at = 0.05
7. w+ = 80.5 and w - = 109.5 and
5. The test statistic is Z =
0
Z0 =
80 . 5 19 ( 20 ) / 4
19 ( 20 )( 39 ) / 24
= 0 . 58
8. Conclusion, because 0.58 < 1.96, we cannot reject H0. The mean titanium content is not significantly
different from 8.5 at =0.05.
15-13
2. H 0 : = 7. 0
3 . H 1 : 7 .0
4. = 0.05
5. The test statistic is w = min( w + , w -).
6. We reject H0 if w
w0*.05 = 8, because = 0.05 and n = 10, Appendix A, Table IX gives the critical
value.
7. w+ = 50.5 and w - = 4.5 and w = min(50.5, 4.5) = 4.5
8. Conclusion, because 4.5 < 8, we reject H0 and conclude that the mean pH is not equal to 7.
15-14
2. H 0 : D = 0
3. H 1 : D 0
4. =0.05
5. The test statistic is w = min( w + , w -).
6. We reject H0 if w
w0*.05 = 3 because = 0.05 and n = 8 Appendix A, Table IX gives the critical value.
15-6
15-15
2. H 0 : = 2. 5
3. H 1 : < 2. 5
4. = 0.05
5. The test statistic is w+
6. We reject H0 if w+
value.
7. w+ = 5 and w - = 205
8. Conclusion, because 5 < 60, we reject H0 and conclude that the mean impurity level is less than 2.5 ppm.
15-16
2. H 0 : D = 0
3. H 1 : D 0
4. =0.01
5. The test statistic is w = min( w + , w -).
6. We reject H0 if w
value.
7. w+ = 141.5 and w - = 29.5 and w = min(141.5, 29.5) = 29.5
8. Conclusion, not enough evidence to reject H0.
15-17
1. Parameter of interest is the difference in mean hardness readings for the two tips
2. H 0 : D = 0
3. H 1 : D 0
4. = 0.05
5. The test statistic is w = min( w + , w -).
6 We reject H0 if w
w0*.05 = 3, because = 0.05 and n = 8 Appendix A, Table IX gives the critical value.
a)
1. The parameters of interest are the mean current (note: set circuit 1 equal to sample 2 so that Table X can
be used. Therefore 1=mean of circuit 2 and 2=mean of circuit 1)
2. H 0 : 1 = 2
3. H 1 : 1 > 2
4. = 0.025
5. The test statistic is w = ( n1 + n 2 )( n1 + n 2 + 1) w
2
1
2
6. We reject H0 if w2
w0*.025 = 51, because =0.025 and n1=8 and n2=9, Appendix A, Table X gives the
critical value.
7. w1 = 78and w2 = 75 and because 75 is less than 51, do not reject H0
8. Conclusion, do not reject H0. There is not enough evidence to conclude that the mean of circuit 1 exceeds
the mean of circuit 2.
b)
1. The parameters of interest are the mean image brightness of the two tubes
2. H 0 : 1 = 2
3. H 1 : 1 > 2
4. = 0.025
15-7
15-15
2. H 0 : = 2. 5
3. H 1 : < 2. 5
4. = 0.05
5. The test statistic is w+
6. We reject H0 if w+
value.
7. w+ = 5 and w - = 205
8. Conclusion, because 5 < 60, we reject H0 and conclude that the mean impurity level is less than 2.5 ppm.
15-16
2. H 0 : D = 0
3. H 1 : D 0
4. =0.01
5. The test statistic is w = min( w + , w -).
6. We reject H0 if w
value.
7. w+ = 141.5 and w - = 29.5 and w = min(141.5, 29.5) = 29.5
8. Conclusion, not enough evidence to reject H0.
15-17
1. Parameter of interest is the difference in mean hardness readings for the two tips
2. H 0 : D = 0
3. H 1 : D 0
4. = 0.05
5. The test statistic is w = min( w + , w -).
6 We reject H0 if w
w0*.05 = 3, because = 0.05 and n = 8 Appendix A, Table IX gives the critical value.
a)
1. The parameters of interest are the mean current (note: set circuit 1 equal to sample 2 so that Table X can
be used. Therefore 1=mean of circuit 2 and 2=mean of circuit 1)
2. H 0 : 1 = 2
3. H 1 : 1 > 2
4. = 0.025
5. The test statistic is w = ( n1 + n 2 )( n1 + n 2 + 1) w
2
1
2
6. We reject H0 if w2
w0*.025 = 51, because =0.025 and n1=8 and n2=9, Appendix A, Table X gives the
critical value.
7. w1 = 78and w2 = 75 and because 75 is less than 51, do not reject H0
8. Conclusion, do not reject H0. There is not enough evidence to conclude that the mean of circuit 1 exceeds
the mean of circuit 2.
b)
1. The parameters of interest are the mean image brightness of the two tubes
2. H 0 : 1 = 2
3. H 1 : 1 > 2
4. = 0.025
15-7
=72 and
w2
=108
78 72
= 0 . 58
10 . 39
Because Z0 < 1.96, cannot reject H0
8. Conclusion, fail to reject H0. There is not a significant difference in the heat gain for the heating units at
= 0.05. P-value =2[1 - P( Z < 0.58 )] = 0.5619
z0 =
15-19
2. H 0 : 1 = 2
3. H 1 : 1 2
4. =0.01
5. The test statistic is w = ( n1 + n 2 )( n1 + n 2 + 1) w
2
1
2
6. We reject H0 if w
w0*.01 = 23, because =0.01 and n1=6 and n2=6, Appendix A, Table X gives the
critical value.
7. w1 = 40 and w2 = 38 and because 40 and 38 are greater than 23, we cannot reject H0
8. Conclusion, do not reject H0. There is no significant difference in the flight delays at = 0.01.
15-20
a)
1. The parameters of interest are the mean heat gains for heating units
2. H 0 : 1 = 2
3. H 1 : 1 2
4. = 0.05
5. The test statistic is w = ( n1 + n 2 )( n1 + n 2 + 1) w
2
1
2
6. We reject H0 if w
w0*.01 = 78, because = 0.01 and n1 = 10 and n2 = 10, Appendix A, Table X gives the
critical value.
7. w1 = 77 and w2 = 133 and because 77 is less than 78, we can reject H0
8. Conclusion, reject H0 and conclude that there is a significant difference in the heating units at = 0.05.
b)
1. The parameters of interest are the mean heat gain for heating units
2. H 0 : 1 = 2
3. H 1 : 1 2
4. =0.05
=105 and
w2
=175
77 105
= 2 . 12
13 . 23
Because |Z0 | > 1.96, reject H0
8. Conclusion, reject H0 and conclude that there is a difference in the heat gain for the heating units at
=0.05. P-value =2[1 - P( Z < 2.19 )] = 0.034
z0 =
15-21
a)
1. The parameters of interest are the mean etch rates
2. H 0 : 1 = 2
3. H 1 : 1 2
4. = 0.05
15-8
w0*.05 = 78, because = 0.05 and n1 = 10 and n2 = 10, Appendix A, Table X gives
2. H 0 : 1 = 2
3. H 1 : 1 2
4. = 0.05
=232.5 and
w2
=581.25
258 232 . 5
= 1 . 06
24 . 11
Because |Z0| < 1.96, do not reject H0
8. Conclusion, do not reject H0. There is not a significant difference in the pipe deflection temperatures at
= 0.05. P-value =2[1 - P( Z < 1.06 )] = 0.2891
z0 =
15-22
a)
1. The parameters of interest are the mean temperatures
2. H 0 : 1 = 2
3. H 1 : 1 2
4. = 0.05
5. The test statistic is w = ( n1 + n 2 )( n1 + n 2 + 1) w
2
1
2
6. We reject H0 if w
w0*.05 = 185, because = 0.05 and n1 = 15 and n2 = 15, Appendix A, Table X gives
2. H 0 : 1 = 2
3. H 1 : 1 2
4. =0.05
=105 and
w2
=175
73 105
= 2 . 42
13 . 23
Because |Z0| > 1.96, reject H0
8. Conclusion, reject H0 and conclude that there is a significant difference between the mean etch rates. Pvalue = 0.0155
z0 =
15-9
15-23
a)
Distance
244
258
260
263
263
265
267
268
270
271
271
271
273
275
275
279
281
283
286
287
Rank
1
2
3
4.5
4.5
6
7
8
9
11
11
11
13
14.5
14.5
16
17
18
19
20
The sum of the ranks for group 1 is w1=135.5 and for group 2, w2=74.5
Since w2 is less than w0.05 = 78 , we reject the null hypothesis that both groups have
the same mean.
b) When the sample sizes are equal it does not matter which group we select for w1
10(10 + 10 + 1)
= 105
2
10 * 10(10 + 10 + 1)
W21 =
= 175
12
135.5 105
Z0 =
= 2.31
175
W =
1
Because z0 > z0.025 = 1.96, we reject H0 and conclude that the sample means are different
for the two groups.
When z0 = 2.31 the P-value = 0.021
Section 15-5
15-24
N
5
5
5
15
Median
550.0
553.0
528.0
Ave Rank
8.7
10.7
4.6
8.0
15-10
Z
0.43
1.65
-2.08
15-23
a)
Distance
244
258
260
263
263
265
267
268
270
271
271
271
273
275
275
279
281
283
286
287
Rank
1
2
3
4.5
4.5
6
7
8
9
11
11
11
13
14.5
14.5
16
17
18
19
20
The sum of the ranks for group 1 is w1=135.5 and for group 2, w2=74.5
Since w2 is less than w0.05 = 78 , we reject the null hypothesis that both groups have
the same mean.
b) When the sample sizes are equal it does not matter which group we select for w1
10(10 + 10 + 1)
= 105
2
10 * 10(10 + 10 + 1)
W21 =
= 175
12
135.5 105
Z0 =
= 2.31
175
W =
1
Because z0 > z0.025 = 1.96, we reject H0 and conclude that the sample means are different
for the two groups.
When z0 = 2.31 the P-value = 0.021
Section 15-5
15-24
N
5
5
5
15
Median
550.0
553.0
528.0
Ave Rank
8.7
10.7
4.6
8.0
15-10
Z
0.43
1.65
-2.08
H = 4.83
H = 4.84
DF = 2
DF = 2
P = 0.089
P = 0.089 (adjusted for ties)
Do not reject H0. The conditioning method does not have an effect on the breaking strength at = 0.05.
b) The P-value is about 0.089 (use the chi-square distribution)
15-25
15-26
Z
-1.27
0.83
2.31
-1.88
Do not reject H0. There is not a significant difference between the manufacturers at = 0.01.
b) The P-value is about 0.039 (use the chi-square distribution)
15-27 a) Following is the data and ranks for the conductivity test.
Type
Coating Type 1
141
143
146
150
10
12.5
15
19
Coating Type 2
137
143
149
152
12.5
18
20
9
Coating Type 3
127
132
133
134
1.5
5.5
7
8
Coating Type 4
127
129
129
132
3.5
3.5
5.5
1.5
Coating Type 5
142
144
147
148
11
14
16
17
Since there is a fairly large number of ties, we use Equation 15-12 as the test statistic.
1
20(21) 2
[2868
]
20 1
4
= 34.89474
s2 =
h=
1
20(212 )
2694
.
125
= 14.02
34.89474
4
15-11
r_i.
56.5
59.5
22
14
58
b) An approximate P-value can be obtained from the approximate chi-squared distribution for H.
Here a = 5 and each n = 4 so that the guidelines for the adequacy of the approximation are not
quite satisfied. The guideline specifies each n > 4.
But if we apply the chi-squared result we use a 1 = 4 degrees of freedom and determine
0.005 < P-value < 0.01
From Minitab we obtain the following results that agree with the calculated values.
Kruskal-Wallis Test on C1
C2
N Median Ave Rank
1
4
144.5
14.1
2
4
146.0
14.9
3
4
132.5
5.5
4
4
129.0
3.5
5
4
145.5
14.5
Overall 20
10.5
H = 13.98
H = 14.02
Since
h>
DF = 4
DF = 4
2
0.05,4
P = 0.007
P = 0.007
Z
1.37
1.65
-1.89
-2.65
1.51
= 9.49 , we would reject the null hypothesis and conclude that the treatments
N
6
6
6
18
DF = 2
DF = 2
Median
3.100
4.400
3.800
Ave Rank
5.8
13.0
9.7
9.5
Z
-2.06
1.97
0.09
P = 0.067
P = 0.066 (adjusted for ties)
Do not reject H0. There is not a significant difference between the flow rates on the uniformity at = 0.05.
b) P-value = 0.066 (use the chi-square distribution)
15-29
a) D A A D D A A A D D D A
The total number of runs is r = 6.
The observed value of 6 is below the mean of 7. For a sample of size n = 12 with equal samples
from each group, the number of runs can range from 2 to 12. Therefore, the P-value can be
computed as the probability of 6 or fewer or 8 or greater runs. This will be greater usual levels of
significance so that the null hypothesis is not rejected.
The normal approximation can be used for a simpler calculation
2n1 n 2 + N 2(6)(6) + 12
=
=7
N
12
2n n (2n n N ) 2 6 6 (2(6)(6) 12)
R2 = 1 2 2 1 2
=
= 2.727
N ( N 1)
12 2 (12 1)
r R
67
z=
=
= 0.6055
R
2.727
At = 0.05 reject H0 when z 1.96 . Because z = 0.6055 < 1.96, fail to reject H0. There is
R =
15-12
b) The sum of the ranks for group 1 is w1=40 and for group 2, w2=38
Since both w1 and w2 are both greater than w0.05 =26, there is insufficient evidence to reject the
null hypothesis that both groups have the same mean.
In part (a) we test whether the two samples come from the same distribution. In part (b) we test
whether their means are equal.
15-30
a) The sample numbers of the ranked data are:
12112112221121122
The number of runs is r =10.
2n1n2 + N 2(9)(8) + 17
= 9.47
=
17
N
2n n ( 2n n N ) 2 9 8 ( 2(9)(8) 17)
=
= 3.955
R2 = 1 2 2 1 2
17 2 (17 1)
N ( N 1)
r R 10 9.47
=
= 0.2665
z=
R
3.955
At = 0.05 reject H0 when | z | 1.96 . Because z = 0.2665 < 1.96, we fail to reject H0. There is
R =
b)
The sum of the ranks for group 2 (which has the smaller sample size) is 78. Since w0.05=51, we
fail to reject H0. There is insufficient evidence to conclude that the groups have the same mean.
Yes, the hypotheses differ. The one from part (a) test whether the samples came from the same
distribution. Here we test whether the populations have the same mean.
15-31
The sample numbers of the ranked data are: 222111112221211112211222; thus, r = 9
15-13
2n1n2 + N 2(12)(12) + 24
= 13
=
24
N
2n n ( 2n n N ) 2 12 12 ( 2(12)(12) 24)
=
= 5.739
R2 = 1 2 2 1 2
24 2 ( 24 1)
N ( N 1)
9 13
r R
=
= 1.67
z=
R
5.739
R =
At = 0.05, fail to reject H0 since z>z0=-1.96. There is not enough evidence to conclude
that the distributions differ.
15-32
Let 0 be the data below 2000 psi and 1 be the data above 2000 psi
1 0 1 1 1 0 0 1 1 1 1 1 0 1 0 1 1 1 1 0 The total number of runs is r = 10.
2n1 n2 + N 2(6)(14) + 20
=
= 9.4
N
20
2n n (2n n N ) 2 6 14 (2(6)(14) 20)
R2 = 1 2 2 1 2
= 3.272
=
N ( N 1)
20 2 (20 1)
r R 10 9.4
z=
=
= 0.3317
R
3.272
R =
15-14
Supplemental Exercises
15-33
a)
1. Parameter of interest is median surface finish
2. H 0 : ~ = 10.0
3 H : ~ 10.0
1
4. = 0.05
5. The test statistic is the observed number of plus differences or r+ = 5.
6. We reject H0 if the P-value corresponding to r+ = 5 is less than or equal to = 0.05.
7. Using the binomial distribution with n = 10 and p = 0.5, P-value = 2P( R* 5| p = 0.5) =1.0
8. Conclusion, we cannot reject H0. We cannot reject the claim that the median is 10 in.
b)
1. Parameter of interest is median of surface finish
2. H 0 : ~ = 10.0
3 H : ~ 10.0
1
4. =0.05
+
0. 5 n
7. Computation: z = 5 0.5(10) = 0
0
0.5 10
8. Conclusion, cannot reject H0. There is not sufficient evidence that the median surface finish differs from
10 microinches.
The P-value = 2[1-(0)] = 1
15-34
a)
1. The parameter of interest is the median fluoride emissions.
2. H 0 : ~ = 6
3. H : ~ < 6
1
4. =0.05
5. The test statistic is the observed number of plus differences or r+ = 4.
6. We reject H0 if the P-value corresponding to r+ = 4 is less than or equal to = 0.05.
7. Using the binomial distribution with n = 15 and p = 0.5, P-value = P(R+ 4 | p = 0.5) = 0.1334
8. Conclusion, do not reject H0. The data does not support the claim that the median fluoride impurity level
is less than 6.
Using Minitab (Sign Rank Test)
Sign test of median = 6.000 versus <
N Below Equal Above
y
15
9
2
4
Do not reject H0
b)
1. Parameter of interest is median fluoride impurity level
2. H 0 : ~ = 6.0
3 H : ~ < 6.0
1
4. =0.05
+
0. 5 n
0.5 13
15-15
6.000
P
0.1334
Median
4.000
8. Conclusion, do not reject H0. There is not enough evidence to conclude that the median fluoride impurity
level is less than 6.0 ppm. The P-value = 2[1-(1.39)] = 0.1645.
15-35
2. H 0 : D = 0
3. H 1 : D 0
4. =0.05
5. The test statistic is w = min( w +, w -).
6. We reject H0 if w
w0*.05 = 8 because = 0.05 and n = 10 Appendix A, Table IX gives the critical value.
2. H 0 : ~D = 0
3. H : ~ 0
1
4. = 0.05
5. The test statistic r = min(r+, r-).
6. Because = 0.05 and n = 7, Appendix A, Table VIII gives the critical value of
r0*.05 = 0. We reject
H0 in favor of H1 if r 10.
7. r+ = 1 and r - = 6 and r = min(1,6) = 1
8. Conclusion, cannot reject H0. There is not a significant difference in the impurity levels.
15-37
a)
1. The parameters of interest are the mean mileages for a Volkswagen and Mercedes.
2. H 0 : 1 = 2
3. H 1 : 1 2
4. = 0.01
5. The test statistic is w = ( n1 + n 2 )( n1 + n 2 + 1) w
2
1
2
6. We reject H0 if w
w0*.01 = 71, because = 0.01 and n1 = 10 and n2 = 10, Appendix A, Table X gives the
critical value.
7. w1 = 55 and w2 = 155 and because 55 is less than 71, we reject H0
8. Conclusion, reject H0 and conclude that there is a difference in the mean mileages of the two different
vehicles.
b)
1. The parameters of interest are the mean mileages for a Volkswagen and Mercedes
2. H 0 : 1 = 2
3. H 1 : 1 2
4. = 0.01
=105 and
w2
=175
55 105
= 3 . 78
13 . 23
Because |Z0| > 2.58, reject H0
8. Conclusion, reject H0 and conclude that there is a significant difference in the mean mileages at = 0.01.
P-value =2[1 - P(Z < 3.78 )] = 0.00016
z0 =
15-16
15-38
a)
1. The parameters of interest are the in mean fill volumes
2. H 0 : 1 = 2
3. H 1 : 1 2
4. =0.05
5. The test statistic w = ( n1 + n 2 )( n1 + n 2 + 1) w
2
1
2
6. We reject if H0 w
w0*.05 = 78, because = 0.05 and n1 = 10 and n2 = 10, Appendix A, Table X gives
2. H 0 : 1 = 2
3. H 1 : 1 2
4. =0.05
= 105 and
w2
= 175
58 105
= 3 . 55
13 . 23
Because |Z0| > 1.96, reject H0
8. Conclusion, reject H0 and conclude that there is a significant difference in the mean mileages at = 0.05.
P-value =2[1 - P(Z < 3.55 )] 0.000385
z0 =
15-39
15-40
Z
-4.31
-0.09
4.40
P = 0.000
P = 0.000 (adjusted for ties)
N
9
9
9
9
36
Median
1106.0
672.2
646.2
377.6
H = 11.61
DF = 3
Ave Rank
24.8
19.7
20.9
8.7
18.5
Z
2.06
0.38
0.79
-3.23
P = 0.009
Reject H0 at = 0.05
15-41
15-17
Z
2.69
-0.06
-2.63
H = 9.57
DF = 2
Reject H0 at = 0.05
15-42
Kruskal-Wallis Test on C4
C5
1
2
3
Overall
N
10
10
10
30
Median
68.55
379.20
149.75
Ave Rank
13.2
20.4
12.9
15.5
Z
-1.01
2.16
-1.14
H = 4.65 DF = 2 P = 0.098
Since P=0.098<0.1, we would reject the null hypothesis and conclude that the treatments
differ.
15-18
, ARL=3.63.
Individual V alue
30
1
U C L=25.26
20
_
X=15.65
10
LC L=6.04
1
14
21
15
M oving Range
28
35
O bser vation
42
49
56
63
U C L=11.81
10
__
M R=3.61
LC L=0
0
1
14
21
28
35
O bser vation
42
49
56
63
Individual V alue
25
U C L=23.34
20
15
_
X=12.45
10
5
LC L=1.56
1
14
21
28
35
O bser vation
42
16
M oving Range
16-82
2 X
49
56
63
U C L=13.38
12
8
__
M R=4.09
4
0
LC L=0
1
14
21
28
35
O bser vation
16-58
42
49
56
63
b) I-MR chart for crude oil percentage of retail price. It appears there have been a couple of shifts
in the process around observations 16 and 56.
I-MR Chart of crude
1
Individual V alue
55
1
1
1
1
50
U C L=50.45
45
_
X=44.63
40
1
35
1 1
14
21
LC L=38.82
1 1
28
35
O bser vation
42
49
56
63
M oving Range
U C L=7.139
6
4
__
M R=2.185
2
0
LC L=0
1
14
21
28
35
O bser vation
42
49
56
63
c)
I-MR Chart of refine_v
1
Individual V alue
60
1
U C L=41.85
40
_
X=25.41
20
LC L=8.97
1
14
24
21
28
35
O bser vation
42
49
56
63
1
1
M oving Range
U C L=20.20
18
12
__
M R=6.18
6
0
LC L=0
1
14
21
28
35
O bser vation
16-59
42
49
56
63
Individual V alue
40
U C L=37.02
30
_
X=19.56
20
10
LC L=2.09
0
1
14
21
28
35
O bser vation
42
49
56
63
24
M oving Range
U C L=21.46
18
12
__
M R=6.57
6
0
LC L=0
1
16-83
14
21
28
35
O bser vation
42
49
56
63
Let p denote the probability that a point plots outside of the control limits when the mean has shifted from 0 to
= 0 + 1.5. Then:
= P 0 3
< X < 0 + 3
| = 0 + 1.5
n
n
1.5
X 1.5
= P
3<
<
+ 3
/ n / n
/ n
1.5
X 1.5
= P
2 <
<
+ 2 when n = 4
/ n / n
/ n
= P(5 < Z < 1) = 0.1587 0 = 0.1587
Therefore, the probability the shift is undetected for three consecutive samples is (1 - p)3 = 0.15873 = 0.004.
16-84
LCL = 0 k / n
CL = 0
UCL = 0 + k / n
a)
16-60
| = 0 +
= 1 P 0 k
< X < 0 + k
n
n
1 = 1 P k
<
<k
/ n / n
/ n
= 1 P (k n < Z < k n )
= 1 [ (k n ) ( k n ]
where (Z) is the standard normal cumulative distribution function.
16-85
LCL = 0 k / n
CL = 0
UCL = 0 + k / n
a) ARL = 1/p where p is the probability a point plots outside of the control limits. Then,
X 0
1 p = P(LCL < X < UCL | 0 ) = P k <
< k | 0
/ n
= P( k < Z < k ) = (k ) (k ) = 2 (k ) 1
where (Z) is the standard normal cumulative distribution function. Therefore, p = 2 2(k) and ARL =
1/[22(k)]. The mean time until a false alarm is 1/p hours.
b) ARL = 1/p where
< X <k
1 p = P LCL < X < UCL | 1 = 0 + = P k
/ n
/ n
= P k n < Z < k n
= (k n ) ( k n )
and
p = 1 ( k n ) + ( k n )
c) ARL = 1/p where 1 - p = P(-3<Z<3)= 0.9973. Thus, ARL = 1/0.0027 = 370.4.
If k = 2, 1 - p = P(-2<Z<2) = 0.9545 and ARL = 1/p = 22.0.
The 2-sigma limits result in a false alarm for every 22 points on the average. This is a high number of false
alarms for the routine use of a control chart.
d) From part (b), ARL = 1/p, where 1 - p = P( 3 5 < Z < 3 5 ) = 0.7764 . ARL = 4.47 assuming 3sigma control limits.
16-86
16-61
p (1 p )
pc
P pc
n
| p = pc
<
pc (1 pc )
pc (1 pc )
n
n
k p (1 p )
k p (1 p )
p pc
p pc
<Z <
pc (1 pc )
pc (1 pc )
pc (1 pc )
pc (1 pc )
n
n
p k p (1 p ) pc
n
= P
<
pc (1 pc )
= P
p+k
Use the fact that if pc > p then the probability is approximately equal to the probability that Z is less than the
upper limit.
P Z <
(
1
)
k
p
p
p pc
pc (1 pc )
pc (1 pc )
k p (1 p )
p pc
=
pc (1 pc )
pc (1 pc )
n
16-87
Solving for n, n =
k 2 p(1 p)
( p pc )2
pk
p(1 p )
n
The LCL is
pk
16-88
p (1 p )
k 2 (1 p )
= 0 or n =
n
p
The P(LCL < P < UCL| p = 0.08) is desired. Now, using the normal approximation:
LCL = p 3
p (1 p )
0.05(0.95)
= 0.05 3
= 0.015 0
n
100
UCL = p + 3
p (1 p )
0.05(0.95)
= 0.05 + 3
= 0.115
n
100
P 0.08
= P
<
0.08(1 0.08)
100
0.115 0.08
= P( Z < 1.29) = 0.90
0.08(1 0.08)
100
16-62
Therefore, the probability of detecting shift on the first sample following the shift is 1 0.90 = 0.10.
The probability of detecting a shift by at least the third sample following the shift can be determined from the
geometric distribution to be 0.10 + 0.90(0.10) + 0.902(0.10) = 0.27
16-89
The process should be centered at the middle of the specifications; that is, at 100.
For an x chart:
X 105 85 105
X 105 115 105
P ( X < 85) + P ( X > 115) = P
<
>
+ P
5
5
5
5
Sample Count
16-90
3.0SL=54.70
45
NP=40.00
35
-3.0SL=25.30
25
0
10
Sample Number
c) Note that
np 3 np (1 p ) < nP < np + 3 np (1 p )
p3
p (1 p )
p (1 p )
<P< p+3
n
n
16-63
20
Therefore, the np control chart always provides results equivalent to the p chart.
16-91
a)
The center line C
The UCL=16.49
The LCL=0
=8
C Chart of Number of defects
18
UCL=16.49
16
Sample Count
14
12
10
_
C=8
8
6
4
2
0
LCL=0
2
10
12
Sample
14
16
18
20
b)Yes.
16-92
3 <
P p
p (1 p )
p (1 p )
< 3 or p 3
< Pi < p + 3
n
n
p (1 p )
n
a point is in control on this chart if and only if the point is in control on the original p chart.
For unequal sample sizes, the p control chart can be used with the value of n equal to the size of each sample.
That is,
Zi =
P p
p (1 p )
ni
3
2
1
zi
16-93
X=0.000
0
-1
-2
-3
-3.0SL=-3.000
-3
0
Observation Number
16-64
10
16-65
CHAPTER 16
Note to the Instructor: Many of the control charts and control chart summaries were created using
Statgraphics. Minitab can be used to provide similar results.
Section 16-5
16-1
a)
x=
7805
= 223
35
chart
r =
1200
= 34.286
35
chart
= x = 223
=
16-2
a)
x=
r
34.286
=
= 14.74
d2
2.326
362.75
= 14.510
25
chart
r=
8.60
3.64
= 0.344 s =
= 0.1456
25
25
x
UCL = CL + A2 r = 14.510 + 0.577(0.344) = 14.708
CL = 14.510
LCL = CL A2 r = 14.510 0.577(0.344) = 14.312
R
chart
UCL = D4 r = 2.115(0.344) = 0.728
CL = 0.344
LCL = D3 r = 0(0.344) = 0
16-1
b) c4 = 0.94
chart
UCL = CL + 3
s
c4 n
= 14.510 + 3
0.1456
0.94 25
= 14.603
CL = 14.510
LCL = CL 3
s
c4 n
= 14.510 3
0.1456
0.94 25
= 14.417
chart
UCL = s + 3
s
0.1456
2
1 c 42 = 0.1456 + 3
1 0.94 = 0.3041
c4
0
.
94
CL = 0.1456
s
0.1456
2
1 c 42 = 0.1456 3
1 0.94 = 0.0129 0
c4
0.94
4460
271.6
a) x =
= 223
s=
= 13.58
20
20
LCL = s 3
16-3
chart
UCL = CL + 3
13.58
= 223 + 3
= 245.11
n
0.9213 4
13.58
= 200.89
= 223 3
n
0.9213 20
c4
CL = 223
LCL = CL 3
c4
chart
UCL = s + 3
s
13.58
1 c 42 = 13.58 + 3
0.1512 = 30.77
c4
0.9213
CL = 13.58
LCL = s 3
s
13.58
1 c 42 = 13.58 3
0.1512 = 3.61 0
c4
0.9213
= x = 223
s
13.58
=
= 14.74
c 4 0.9213
16-2
16-4
a)
x = 20.0
r
= 1 .4
d2
x
chart
UCL = CL + A2 r = 20.0 + 0.483(3.5476) = 21.71
CL = 20.0
LCL = CL A2 r = 20.0 0.483(3.5476) = 18.29
R
chart
UCL = D4 r = 2.004(3.53476) = 7.11
CL = 3.5476
LCL = D3 r = 0(3.5476) = 0
x
chart
where s / c 4 = 1.5
UCL = CL + 3
s
c4 n
CL = 20.0
LCL = CL 3
s
c4 n
UCL = CL + 3
s
1 c 42 = 1.427 + 3(1.5) 0.0946 = 2.811
c4
CL = 1.427
LCL = CL 3
16-5
s
1 c 42 = 1.427 3(1.5) 0.0946 = 0.043
c4
a)
X-bar and Range - Initial Study
Problem 16-1
X-bar
----UCL: +
3.0 sigma = 37.5789
Centerline
= 34.32
LCL: 3.0 sigma = 31.0611
| Range
| ----| UCL: +
3.0 sigma = 11.9461
| Centerline
= 5.65
| LCL: 3.0 sigma = 0
|
out of limits = 1
| out of limits = 0
-------------------------------------------------------------------------------Chart: Both
Normalize: No
20 subgroups, size 5
0 subgroups excluded
Estimated
process mean = 34.32
process sigma = 2.42906
mean Range
= 5.65
16-3
Problem 16-5
39
37.5789
37
X -bar
35
34.32
33
31.0611
31
0
12
16
20
12
10
8
R ange
6
4
2
0
11.9461
5.65
0
0
8
12
subgroup
16
20
b)
Charting xbar
| Range
| ----| UCL: +
3.0 sigma = 12.1297
| Centerline
= 5.73684
| LCL: 3.0 sigma = 0
|
out of limits = 0
| out of limits = 0
-------------------------------------------------------------------------------20 subgroups, size 5
1 subgroups excluded
Estimated
process mean = 34.0947
process sigma = 2.4664
mean Range
= 5.73684
X-bar
----UCL: +
3.0 sigma = 37.4038
Centerline
= 34.0947
LCL: 3.0 sigma = 30.7857
Problem 16-5
X-bar
40
38
36
34
32
30
37.4038
34.0947
30.7857
0
R ange
12
16
20
15
12
9
6
3
0
12.1297
5.73684
0
0
8
12
subgroup
16-4
16
20
16-6
a)
X-bar and Range - Initial Study
Charting Problem 16-8
X-bar
| Range
----| ----UCL: +
3.0 sigma = 7.44253
| UCL: +
3.0 sigma = 2.9153
Centerline
= 6.2836
| Centerline
= 1.1328
LCL: 3.0 sigma = 5.12467
| LCL: 3.0 sigma = 0
|
out of limits = 0
| out of limits = 0
Estimated
process mean = 6.2836
process sigma = 0.669108
mean Range
= 1.1328
Problem 16-6
X-bar
7.5
7.1
6.7
6.3
5.9
5.5
5.1
7.44253
6.2836
5.12467
0
10
15
20
25
3
2.5
2
Range
1.5
1
0.5
0
2.9153
1.1328
0
0
10
15
subgroup
20
25
There are no points beyond the control limits. The process appears to be in control.
b) No points fell beyond the control limits. The limits do not need to be revised.
16-7 a)
5.792
2.25
0
Test Results: X-bar One point more than 3.00 sigmas from center line.
Test Failed at points: 4 6 7 10 12 15 16 20
Test Results for R Chart: One point more than 3.00 sigmas from center line.
16-5
Sample Mean
20
UCL=17.40
Mean=15.09
15
1
LCL=12.79
10
1
Subgroup
10
20
Sample Range
UCL=5.792
5
4
3
R=2.25
2
1
0
LCL=0
.
b) Removed points 4, 6, 7, 10, 12, 15, 16, 19, and 20 and revised the control limits. The control limits are not
as wide after being revised: X-bar UCL=17.96, CL=15.78, LCL=13.62 and R UCL = 5.453, R-bar = 2.118,
LCL=0.
18
UCL=17.95
17
16
Mean=15.78
15
14
LCL=13.62
13
Subgroup
Sample Range
10
UCL=5.453
5
4
3
R=2.118
2
1
LCL=0
c)
3.051
1.188
0
Test Results: X-bar One point more than 3.00 sigmas from center line.
Test Failed at points: 4 6 7 10 12 15 16 20
Test Results for S Chart:One point more than 3.00 sigmas from center line.
Test Failed at points: 19
16-6
Sample Mean
20
UCL=17.42
Mean=15.09
15
1
LCL=12.77
10
1
Subgroup
10
20
Sample StDev
1
UCL=3.051
3
2
S=1.188
1
0
LCL=0
Removed points 4, 6, 7, 10, 12, 15, 16, 19, and 20 and revised the control limits. The control limits
are not as wide after being revised: X-bar UCL = 17.95, CL =1 5.78, LCL=13.62 and S UCL =
2.848, S-bar=1.109, LCL=0.
18
UCL=17.95
17
16
Mean=15.78
15
14
LCL=13.62
13
Subgroup
Sample StDev
10
UCL=2.848
2
1
S=1.109
LCL=0
16-7
16-8
a) The average range is used to estimate the standard deviation. Sample 2, 9, and 17 are out-of-control
Xbar-R Chart of x1, ..., x5
Sample M ean
4.98
U C L=4.97331
4.97
_
_
X=4.96065
4.96
4.95
4.94
LC L=4.94799
1
Sample Range
0.060
10
Sample
12
14
16
18
20
U C L=0.04641
0.045
0.030
_
R=0.02195
0.015
0.000
LC L=0
2
10
Sample
12
14
16
18
20
b)
Xbar-R Chart of x1_1, ..., x5_1
U C L=4.97284
Sample M ean
4.970
4.965
_
_
X=4.96124
4.960
4.955
4.950
LC L=4.94963
2
10
12
14
16
Sample
U C L=0.04254
Sample Range
0.04
0.03
_
R=0.02012
0.02
0.01
0.00
LC L=0
2
10
Sample
16-8
12
14
16
c) For X -S chart
Xbar-S Chart of x1, ..., x5
4.98
Sample M ean
U C L=4.97333
4.97
_
_
X=4.96065
4.96
4.95
4.94
LC L=4.94797
1
10
Sample
12
14
16
18
20
Sample StDev
0.020
U C L=0.01855
0.015
_
S =0.00888
0.010
0.005
0.000
LC L=0
2
10
Sample
12
14
16
18
20
4.970
4.965
_
_
X=4.96124
4.960
4.955
4.950
LC L=4.94942
2
10
12
14
16
Sample
Sample StDev
0.020
U C L=0.01730
0.015
0.010
_
S =0.00828
0.005
0.000
LC L=0
2
10
12
14
16
Sample
16-9 a)
The control limits for the following chart were obtained from an estimate of obtained from the
pooled standard deviation from each subgroup of size 5.
16-9
0.0640
Sample M ean
0.0635
UC L=0.063505
0.0630
_
_
X=0.062938
0.0625
LC L=0.062370
1
0.0620
10
12
14
Sample
16
18
20
22
24
Sample Range
0.0024
UC L=0.002080
0.0018
0.0012
_
R=0.000984
0.0006
0.0000
LC L=0
2
10
12
14
Sample
16
18
20
22
24
b) The test failed at points 1, 6, 14, 15, 17, 21 and 22. After eliminating these points, the test failed
at points 4, 6, 9 and 15 (new sample numbers based on remaining samples). After eliminating
these points, the test failed at point 4. After eliminating this point, the points are in control:
Xbar-R Chart of C6
0.0634
Sample Mean
UCL=0.0633116
0.0632
_
_
X=0.0629877
0.0630
0.0628
LCL=0.0626638
0.0626
1
7
Sample
10
11
12
13
Sample Range
0.0012
UCL=0.001187
0.0009
_
R=0.000562
0.0006
0.0003
0.0000
LCL=0
1
7
Sample
16-10
10
11
12
13
Xbar-S Chart of C6
1
0.0640
Sample Mean
0.0635
UCL=0.063461
0.0630
_
_
X=0.062938
0.0625
LCL=0.062414
1
0.0620
11
13
Sample
15
17
19
21
23
25
0.00100
Sample StDev
UCL=0.000766
0.00075
0.00050
_
S=0.000367
0.00025
0.00000
LCL=0
1
11
13
Sample
15
17
19
21
23
25
Removal of these points left points 4, 6, 9 and 15 out of control (newly labeled).
Removal of these points left point 4 out of control. After the removal of this point, the
remaining points were in control.
Xbar-S Chart of C6
0.0634
Sample Mean
UCL=0.0633197
0.0632
_
_
X=0.0629877
0.0630
0.0628
LCL=0.0626556
0.0626
1
7
Sample
10
11
12
13
UCL=0.0004860
Sample StDev
0.00048
0.00036
_
S=0.0002326
0.00024
0.00012
LCL=0
0.00000
1
7
Sample
10
11
12
13
16-10
It violates the rule of two out of three consecutive points plot beyond a 2-sigma limit.
16-11
The sample standard deviation is 2.956. It is larger than the estimate obtained from the revised
chart, because the mean of the process has been shifted in the original data.
16-12
The sample standard deviation is 0.0127. It is larger than the estimate 0.010301 obtained from the revised
and R chart, because the mean of the process has been shifted in the original data.
Section 16-6
16-11
and R
Xbar-S Chart of C6
1
0.0640
Sample Mean
0.0635
UCL=0.063461
0.0630
_
_
X=0.062938
0.0625
LCL=0.062414
1
0.0620
11
13
Sample
15
17
19
21
23
25
0.00100
Sample StDev
UCL=0.000766
0.00075
0.00050
_
S=0.000367
0.00025
0.00000
LCL=0
1
11
13
Sample
15
17
19
21
23
25
Removal of these points left points 4, 6, 9 and 15 out of control (newly labeled).
Removal of these points left point 4 out of control. After the removal of this point, the
remaining points were in control.
Xbar-S Chart of C6
0.0634
Sample Mean
UCL=0.0633197
0.0632
_
_
X=0.0629877
0.0630
0.0628
LCL=0.0626556
0.0626
1
7
Sample
10
11
12
13
UCL=0.0004860
Sample StDev
0.00048
0.00036
_
S=0.0002326
0.00024
0.00012
LCL=0
0.00000
1
7
Sample
10
11
12
13
16-10
It violates the rule of two out of three consecutive points plot beyond a 2-sigma limit.
16-11
The sample standard deviation is 2.956. It is larger than the estimate obtained from the revised
chart, because the mean of the process has been shifted in the original data.
16-12
The sample standard deviation is 0.0127. It is larger than the estimate 0.010301 obtained from the revised
and R chart, because the mean of the process has been shifted in the original data.
Section 16-6
16-11
and R
16-13
a)
Individuals and MR(2) - Initial Study
-----------------------------------------------------------------------------Charting Problem 16-13
Ind.x
----UCL: +
3.0 sigma = 60.8887
Centerline
= 53.05
LCL: 3.0 sigma = 45.2113
| MR(2)
| ----| UCL: +
3.0 sigma = 9.63382
| Centerline
= 2.94737
| LCL: 3.0 sigma = 0
|
out of limits = 0
| out of limits = 0
-----------------------------------------------------------------------------Chart: Both
Normalize: No
20 subgroups, size 1
0 subgroups excluded
Estimated
process mean = 53.05
process sigma = 2.61292
mean MR
= 2.94737
There are no points beyond the control limits. The process appears to be in control.
Individual Value
60
55
Mean=53.05
50
LCL=45.21
45
Subgroup
10
20
Moving Range
10
UCL=9.630
5
R=2.947
0
LCL=0
= x = 53.05
16-14
mr 2.94737
=
= 2.613
d2
1.128
a) The process appears to be in statistical control. There are no points beyond the control limits.
16-12
Individual Value
Subgroup
UCL=19.15
Mean=15.99
LCL=12.83
0
10
20
Moving Range
30
UCL=3.887
3
2
1
R=1.190
LCL=0
= x = 15.99
mr 1.190
=
= 1.05
d 2 1.128
16-15
a)
Ind.x and MR(2) - Initial Study
-----------------------------------------------------------------------------Charting diameter
Ind.x
----UCL: +
3.0 sigma = 10.5358
Centerline
= 10.0272
LCL: 3.0 sigma = 9.51856
| MR(2)
| ----| UCL: +
3.0 sigma = 0.625123
| Centerline
= 0.19125
| LCL: 3.0 sigma = 0
|
out of limits = 0
| out of limits = 0
-----------------------------------------------------------------------------Chart: Both
Normalize: No
25 subgroups, size 1
0 subgroups excluded
Estimated
process mean = 10.0272
process sigma = 0.169548
mean MR(2)
= 0.19125
16-13
Individual Value
UCL=10.54
10.0
Mean=10.03
9.5
LCL=9.519
Moving Range
Subgroup
10
15
20
25
0.7
0.6
0.5
0.4
0.3
0.2
0.1
0.0
UCL=0.6249
R=0.1912
LCL=0
There are no points beyond the control limits. The process appears to be in control.
b) Estimated process mean and standard deviation
= x = 10.0272
16-16
mr 0.19125
=
= 0.16955
d2
1.128
a)
Ind.x and MR(2) - Initial Study
-------------------------------------------------------------------------------Charting Problem 16-16
Ind.x
----UCL: +
3.0 sigma = 552.112
Centerline
= 500.6
LCL: 3.0 sigma = 449.088
| MR(2)
| ----| UCL: +
3.0 sigma = 63.308
| Centerline
= 19.3684
| LCL: 3.0 sigma = 0
|
out of limits = 0
| out of limits = 0
-------------------------------------------------------------------------------Chart: Both
Normalize: No
20 subgroups, size 1
Estimated
process mean = 500.6
process sigma = 17.1706
mean MR(2)
= 19.3684
0 subgroups excluded
16-14
Individual Value
UCL=552.1
500
Mean=500.6
450
LCL=449.1
Moving Range
Subgroup
10
20
70
60
50
40
30
20
10
0
UCL=63.28
R=19.37
LCL=0
= x = 500.6
mr 19.3684
=
= 17.17
d2
1.128
16-17
a)
I-MR Chart of SIZE
Individual V alue
140
U C L=130.50
120
_
X=100.78
100
80
LC L=71.06
2
10
12
14
O bser vation
16
18
20
22
24
40
M oving Range
U C L=36.51
30
20
__
M R=11.18
10
0
LC L=0
2
10
12
14
O bser vation
16-15
16
18
20
22
24
Individual V alue
U C L=127.08
120
_
X=99.48
100
80
LC L=71.88
2
10
12
14
O bser vation
16
18
20
22
24
U C L=33.91
M oving Range
30
20
__
M R=10.38
10
LC L=0
2
10
12
14
O bser vation
16
18
20
22
24
16-18. a)
PCR =
USL LSL 40
=
= 1.1111
6
36
2 (20 / 6) = 0.0858%
b)
PCR =
USL LSL 40
=
= 1.1111
6
36
Since the mean of the process is not centered at the nominal dimension,
14 26
PCRk = min , = 0.7778
18 18
The small PCRK indicates that the process is likely to produce units outside the specification limits.
The fraction defective is
16-19
a) PCR
USL LSL 12
=
= 1.5
6
6
so
= 1.3333
b) (20+32)/2=26 When the process is centered at the nominal dimension, the fraction defective is minimized
for any .
16-16
Individual V alue
U C L=127.08
120
_
X=99.48
100
80
LC L=71.88
2
10
12
14
O bser vation
16
18
20
22
24
U C L=33.91
M oving Range
30
20
__
M R=10.38
10
LC L=0
2
10
12
14
O bser vation
16
18
20
22
24
16-18. a)
PCR =
USL LSL 40
=
= 1.1111
6
36
2 (20 / 6) = 0.0858%
b)
PCR =
USL LSL 40
=
= 1.1111
6
36
Since the mean of the process is not centered at the nominal dimension,
14 26
PCRk = min , = 0.7778
18 18
The small PCRK indicates that the process is likely to produce units outside the specification limits.
The fraction defective is
16-19
a) PCR
USL LSL 12
=
= 1.5
6
6
so
= 1.3333
b) (20+32)/2=26 When the process is centered at the nominal dimension, the fraction defective is minimized
for any .
16-16
16-20
PCR =
USL LSL 20
=
= 1.6667
6
12
PCR =
USL LSL 20
=
= 1.6667
6
12
Since the mean of the process is not centered at the nominal dimension,
13 7
PCRk = min , = 1.1667
6 6
The fraction defective is
a) If the process uses 66.7% of the specification band, then 6 = 0.667(USL-LSL). Because the process is
centered
4.5 4.5
,
= 1 .5
= min
3
3
Because PCR and PCRk exceed unity, the natural tolerance limits are inside the specification limits and few
defective units should be produced.
b) Assuming a normal distribution with 6 = 0.667(USL LSL) and a centered process, then
3 = 0.667(USL ). Consequently, USL = 4.5 and LSL = 4.5
4.5
= 11 = 0
a) If the process uses 85% of the spec band then 6 = 0.85(USL LSL) and
PCR =
USL LSL
1
=
= 1.18
0.85(USL LSL ) 0.85
Therefore,
3.53 3.53
PCRk = min
,
= 1.18
3
3
16-17
Because PCR and PCRk exceed unity, the natural tolerance limits are inside the specification limits and few
defective units should be produced.
b) Assuming a normal distribution with 6 = 0.85(USL LSL) and a centered process, then
3 = 0.85(USL ). Consequently, USL = 3.5 and LSL = 3.5
3.5
= 1 0.999767 = 0.000233
34.286
= 14.74
2.326
LSL
180 223
= 0.00175
USL
260 223
14.74
PCR =
USL x x LSL
,
PCRK = min
3
3
260 223 223 180
= min
,
3(14.74) 3(14.74)
= min[0.837,0.972]
= 0.837
r
0.344
=
= 0148
.
a) Assume a normal distribution with = 14.510 and =
d 2 2.326
LSL
14.00 14.51
0.148
= 0.00028
USL
15.00 14.51
0.148
16-18
b)
PCR =
USL x x LSL
,
PCRK = min
3
3
15.00 14.51 14.51 14.00
,
= min
3(0.148)
3(0.148)
= min[1.104,1.15]
= 1.104
Because PCR and PCRk exceed unity, the natural tolerance limits are inside the specification limits and few
defective units should be produced.
Because PCRK PCR the process appears to be centered.
16-25
s
13.58
=
= 14.74
c4 0.9213
LSL
170 223
= 0.00016
USL
270 223
14.75
Probability of producing a part outside the specification limits is 0.00016 + 0.00074 = 0.0009
PCR =
PCR K
3(14.75) 3(14.75)
= min[1.06,1.19]
= 1.06
Because PCR and PCRk exceed unity, the natural tolerance limits are inside the specification limits and few
defective units should be produced. The estimated proportion nonconforming is given by P(X < LSL) + P(X
> USL) = 0.00016 + 0.00074 = 0.0009
16-19
16-26
PCR =
USL LSL 25 15
=
= 1.19
6( )
6(1.4)
PCRK
USL LSL
,
= min
3
3
25 20 20 15
,
= min
3(1.4) 3(1.4)
= min[1.19,1.19]
= 1.19
16-27
a)
r
5.737
=
= 2.446 = 0.0002466
d 2 2.326
USL LSL 0.5045 0.5025
=
PCR =
= 1.35
6( )
6(0.0002466)
PCRK
USL x x LSL
,
= min
3
3
0.5045 0.5034 0.5034 0.5025
,
= min
3(0.0002466)
3(0.0002466)
= min[1.489,1.217]
= 1.217
Because PCR and PCRk exceed unity, the natural tolerance limits are inside the specification limits and few
defective units should be produced.
Because PCRK PCR the process is slightly off center.
b) Assume a normal distribution with = 0.5034 and = 0.0002466
LSL
USL
= 11 = 0
Therefore, the proportion nonconforming is given by
P(X<LSL) + P(X>USL) = 0.00013 + 0 = 0.00013
16-20
16-28
PCR =
USL LSL
75
=
= 0.50
6( )
6(0.669)
1.1328
= 0.669
1.693
USL x x LSL
PCRK = min
,
3
3
7 6.284 6.284 5
,
= min
3(0.669) 3(0.669)
= min[0.357,0.640]
= 0.357
16-29
r
2.25
=
= 1.329
d 2 1.693
x = 15.09
P( X > 15) + P ( X < 5)
15 15.09
5 15.09
= P Z >
+ P Z <
1.693
1.693
15 5
= 0.985
6(1.693)
Because the estimated PCR is less than unity, the process capability is not good.
16-30
PCR =
USL x x LSL
PCRK = min
,
3
3
525 500.6 500.6 475
= min
,
3(17.17)
3(17.17)
= min[0.474,0.50]
= 0.474
Because the process capability ratios are less than unity, the process capability appears to be poor.
16-21
Section 16-8
16-31
a) This process is out of control statistically
P Chart of C1
1
0.14
0.12
0.10
Proportion
UCL=0.0835
0.08
_
P=0.0645
0.06
LCL=0.0455
0.04
1
1
0.02
1
1
10
12
Sample
14
16
18
20
b)
P Chart of C1
1
0.14
UCL=0.1382
0.12
Proportion
0.10
0.08
_
P=0.0645
0.06
0.04
0.02
0.00
LCL=0
2
10
12
Sample
14
16
18
20
The process is still out of control, but not as many points fall outside of the control limits. The
control limits are wider for smaller values of n.
Test Failed at points: 5 and 7.
16-22
0.12
Proportion
0.10
0.08
_
P=0.0561
0.06
0.04
0.02
0.00
LCL=0
2
10
Sample
12
14
16
18
c) The larger sample size leads to smaller standard deviation of proportions and thus narrower
control limits.
16-32
a)
P C h a rt f o r x
Proportion
0 .2
U C L = 0 .1 8 9 1
0 .1
P = 0 .0 9 9 3 3
L C L = 0 .0 0 9 6 0 1
0 .0
0
10
20
S a m p le N u m b e r
16-23
30
sigma
= 0.0159937
Problem 16-28
0.4
0.3
0.2
0.198553
0.150571
0.10259
0.1
0
0
12
16
20
24
subgroup
The samples with out-of-control points are 1, 2, 3, 6, 7, 8, 11, 12, 13, 15, 17, and 20. The control limits need
to be revised.
P Chart - Revised Limits
Charting Problem 16-33
P Chart
----UCL: +
3.0 sigma = 0.206184
Centerline
= 0.157333
LCL: 3.0 sigma = 0.108482
out of limits = 0
Estimated
mean P = 0.157333
sigma
= 0.0162837
Problem 16-28
Revised Limits
0.4
0.3
0.206184
0.2
0.157333
0.108482
0.1
0
0
12
16
20
24
subgroup
There are no further points out of control for the revised limits.
The process does not appear to be in control.
0.02
Sample Count
16-34
1
UCL=0.01670
0.01
U=0.008143
LCL=0
0.00
0
10
15
20
25
16-24
30
35
16-35
a)
Sample Count
3.0SL=3.811
3
2
U=1.942
1
-3.0SL=0.07217
0
0
10
15
20
25
Sample Number
Samples 5 and 24 are points beyond the control limits. The limits need to be revised.
b)
U Chart for defects_
4
UCL=3.463
Sample Count
2
U=1.709
1
LCL=0
0
10
20
Sample Number
The control limits are calculated without the out-of-control points. There are no further points out of control
for the revised limits.
16-36
a)
16-25
40
Sample Count Per Unit
35
1
1
UCL=32.77
30
25
_
U=19.51
20
15
10
5
LCL=6.26
10
20
30
40
50
60
Sample
70
80
90
100
a) (112-100)/3=4
b)
88 96 X 112 96
P (88 < X < 112) = P
<
<
= P ( 2 < Z < 4 )
X
4
4
a)
+3
= UCL
100 + 3
= 106
2
3
= (106 100) = 4
b)
x =
4
= 2,
2
= 96
94 96 X 106 96
= P( 1 < Z < 5)
P(94 < X < 106) = P
<
<
2
2
ARL =
1
1
=
= 6.301
p 0.1587
16-26
40
Sample Count Per Unit
35
1
1
UCL=32.77
30
25
_
U=19.51
20
15
10
5
LCL=6.26
10
20
30
40
50
60
Sample
70
80
90
100
a) (112-100)/3=4
b)
88 96 X 112 96
P (88 < X < 112) = P
<
<
= P ( 2 < Z < 4 )
X
4
4
a)
+3
= UCL
100 + 3
= 106
2
3
= (106 100) = 4
b)
x =
4
= 2,
2
= 96
94 96 X 106 96
= P( 1 < Z < 5)
P(94 < X < 106) = P
<
<
2
2
ARL =
1
1
=
= 6.301
p 0.1587
16-26
16-39
a) x = 74.01
= 74.01
x = 0.0045
= P
<
<
x
0.0045
0.0045
16-40
b)
ARL =
a)
1
1
=
= 4 .6
p 0.2177
R 0.344
=
= 0.148
d 2 2.326
x =
0.148
= 0.066 ,
5
= 14.6
0
.
066
0
.
066
x
16-41
b)
ARL =
a)
1
1
=
= 19.8
p 0.0505
R
= 14.74
d2
x =
14.74
5
= 6.592 , = 210
16-42
b)
ARL =
a)
1
1
=
= 6 .6
p 0.1515
R
= 1.4 x
d2
1.4
6
= 0.5715 , = 17
18.29 17 X 21.71 17
ARL =
1
1
=
= 1.012
p 0.9881
16-27
16-43
2.4664
= 1.103 , = 36
n
5
30.78 36 X 37.404 36
a) x =
16-44
b)
ARL =
a)
1
1
=
= 9.8
p 0.102
R
2.25
=
= 1.329
d 2 1.693
x =
1.329
3
= 0.767 , = 13
12.79 13 X 17.4 13
16-45
b)
ARL =
a)
1
1
=
= 2.54
p 0.3936
R 0.000924
=
= 0.000397
d2
2.326
x =
0.000397
5
= 0.000178 , = 0.0625
1
1
=
= 3.48
p 0.2877
R
0.669
= 0.669 x =
=
= 0.386 , = 5.5
a) =
d2
n
3
b)
16-46
ARL =
16-28
b)
ARL =
1
1
=
= 6.02
p 0.16603
Section 16-10
16-47
a) Yes, this process is in-control. CUSUM chart with h = 4 and k = 0.5 is shown.
CUSUM Chart of viscosity
4
UCL=3.875
Cumulative Sum
2
1
0
-1
-2
-3
LCL=-3.875
-4
1
8
9
Sample
10
11
12
13
14
15
b) Yes, this process has shifted out-of-control. For the CUSUM estimated from all the data
observation 20 exceeds the upper limit.
16-29
b)
ARL =
1
1
=
= 6.02
p 0.16603
Section 16-10
16-47
a) Yes, this process is in-control. CUSUM chart with h = 4 and k = 0.5 is shown.
CUSUM Chart of viscosity
4
UCL=3.875
Cumulative Sum
2
1
0
-1
-2
-3
LCL=-3.875
-4
1
8
9
Sample
10
11
12
13
14
15
b) Yes, this process has shifted out-of-control. For the CUSUM estimated from all the data
observation 20 exceeds the upper limit.
16-29
Cumulative Sum
5.0
UCL=3.88
2.5
0.0
-2.5
LCL=-3.88
-5.0
2
10
12
Sample
14
16
18
20
For a CUSUM with standard deviation estimated from the moving range of the first 15 observation, the
moving range is 1.026 and the standard deviation estimate is 0.9096. If this standard deviation is used with a
target of 14.1, the following CUSUM is obtained.
CUSUM Chart of C1
7.5
Cumulative Sum
5.0
UCL=3.64
2.5
0.0
-2.5
LCL=-3.64
-5.0
1
16-48
11
Sample
16-30
13
15
17
19
3.2
Cumulative Sum
3
2
1
0
-1
-2
-3
-3.2
Lower CUSUM
10
20
The CUSUM control chart for purity does not indicate an out-of-control situation. The SH values do not plot
beyond the values of H and H.
b) CUSUM Control chart with k = 0.5 and h = 4
CUSUM Chart for New Purity Data
Upper CUSUM
3.2
Cumulative Sum
3
2
1
0
-1
-2
-3
-3.2
Lower CUSUM
10
15
Subgroup Number
= 0.1736
b) CUSUM Control chart with k = 0.5 and h = 4
16-31
20
25
Cumulative Sum
0.678191
0.5
0.0
-0.5
-6.8E-01
Lower CUSUM
10
15
20
25
Subgroup Number
32
Cumulative Sum
30
20
10
0
-10
-20
-30
-32
Lower CUSUM
10
20
Subgroup Number
0 51 50
= 0.5 standard deviations. From Table 16-9, ARL = 38.0
=
2
0 51 50
= 1 standard deviation. From Table 16-9, ARL =
b) If n = 4, the shift to 51 is a shift of
=
/ n
2/ 4
10.4
a) A shift to 51 is a shift of
16-32
EWMA Chart of C1
91.0
UCL=90.800
EWMA
90.5
_
_
X=90
90.0
89.5
LCL=89.200
89.0
2
10
12
Sample
14
16
18
20
EWMA Chart of C1
91.5
UCL=91.386
91.0
90.5
EWMA
16-52
_
_
X=90
90.0
89.5
89.0
LCL=88.614
88.5
2
10
12
Sample
14
16
18
20
c) For part (a), there is no evidence that the process has shifted out of control.
16-33
EWMA Chart of C1
91.0
UCL=90.800
EWMA
90.5
_
_
X=90
90.0
89.5
LCL=89.200
89.0
2
10
12 14
Sample
16
18
20
22
24
For part b), there is no evidence that the process has shifted out of control.
EWMA Chart of C1
91.5
UCL=91.386
91.0
EWMA
90.5
_
_
X=90
90.0
89.5
89.0
LCL=88.614
88.5
2
16-53
10
12 14
Sample
16
16-34
18
20
22
24
EWMA Chart of C1
10.2
UCL=10.1695
EWMA
10.1
_
_
X=10
10.0
9.9
LCL=9.8305
9.8
2
10
12 14
Sample
16
18
20
22
24
UCL=10.2937
10.2
EWMA
10.1
_
_
X=10
10.0
9.9
9.8
LCL=9.7063
9.7
2
16-54
10
12 14
Sample
16-35
16
18
20
22
24
EWMA Chart of C1
110
UCL=108.00
EWMA
105
_
_
X=100
100
95
LCL=92.00
90
2
10
12
Sample
14
16
18
20
UCL=113.86
110
EWMA
105
_
_
X=100
100
95
90
LCL=86.14
85
2
10
12
Sample
14
16
18
20
0.5 , smaller is preferred to detect the shift fast. so the chart in part (a)
= 0.5
2
2/
= 3 for n gives us the required sample size
n
of 9.
16-36
16-56
102 100
= 0.5 standard deviations.
4
From Table 16-9, ARL = 38. The hours of production are 2(38) = 76.
b) The ARL = 38. However, the time to obtain 38 samples is now 0.5(38) = 19.
c) From Table 16-9, the ARL when there is no shift is 465. Consequently, the time between false alarms is
0.5(465) = 232.5 hours. Under the old interval, false alarms occurred every 930 hours.
d) If the process shifts to 102, the shift is 0 = 102 100 = 1 standard deviation. From Table 16-9, the
/ n
4/ 4
ARL for this shift is 10.4. Therefore, the time to detect the shift is 2(10.4) = 20.8 hours. Although this time
is slightly longer than the result in part (b), the time between false alarms is 2(465) = 930 hours, which is
better than the result in part (c).
Supplementary Exercises
a)
X-bar and Range - Initial Study
-------------------------------------------------------------------------------X-bar
| Range
---| ----UCL: +
3.0 sigma = 64.0181
| UCL: +
3.0 sigma = 0.0453972
Centerline
= 64
| Centerline
= 0.01764
LCL: 3.0 sigma = 63.982
| LCL: 3.0 sigma = 0
|
out of limits = 0
| out of limits = 0
-------------------------------------------------------------------------------Chart: Both
Normalize: No
Estimated
process mean = 64
process sigma = 0.0104194
mean Range
= 0.01764
Sample Mean
64.02
UCL=64.02
64.01
Mean=64.00
64.00
63.99
LCL=63.98
63.98
Subgroup
10
0.05
Sample Range
16-57
25
UCL=0.04541
0.03
0.02
R=0.01764
0.01
LCL=0
c)
20
0.04
0.00
R 0.01764
=
= 0.0104
d2
1.693
USL LSL 64.02 63.98
PCR =
=
= 0.641
6
6(0.0104)
b) = x = 64
15
16-37
16-56
102 100
= 0.5 standard deviations.
4
From Table 16-9, ARL = 38. The hours of production are 2(38) = 76.
b) The ARL = 38. However, the time to obtain 38 samples is now 0.5(38) = 19.
c) From Table 16-9, the ARL when there is no shift is 465. Consequently, the time between false alarms is
0.5(465) = 232.5 hours. Under the old interval, false alarms occurred every 930 hours.
d) If the process shifts to 102, the shift is 0 = 102 100 = 1 standard deviation. From Table 16-9, the
/ n
4/ 4
ARL for this shift is 10.4. Therefore, the time to detect the shift is 2(10.4) = 20.8 hours. Although this time
is slightly longer than the result in part (b), the time between false alarms is 2(465) = 930 hours, which is
better than the result in part (c).
Supplementary Exercises
a)
X-bar and Range - Initial Study
-------------------------------------------------------------------------------X-bar
| Range
---| ----UCL: +
3.0 sigma = 64.0181
| UCL: +
3.0 sigma = 0.0453972
Centerline
= 64
| Centerline
= 0.01764
LCL: 3.0 sigma = 63.982
| LCL: 3.0 sigma = 0
|
out of limits = 0
| out of limits = 0
-------------------------------------------------------------------------------Chart: Both
Normalize: No
Estimated
process mean = 64
process sigma = 0.0104194
mean Range
= 0.01764
Sample Mean
64.02
UCL=64.02
64.01
Mean=64.00
64.00
63.99
LCL=63.98
63.98
Subgroup
10
0.05
Sample Range
16-57
25
UCL=0.04541
0.03
0.02
R=0.01764
0.01
LCL=0
c)
20
0.04
0.00
R 0.01764
=
= 0.0104
d2
1.693
USL LSL 64.02 63.98
PCR =
=
= 0.641
6
6(0.0104)
b) = x = 64
15
16-37
The process does not meet the minimum capability level of PCR 1.33.
d)
USL x x LSL
PCRk = min
,
3
3
64.02 64 64 63.98
= min
,
3(0.0104) 3(0.0104)
= min[0.641,0.641]
= 0.641
e) In order to make this process a six-sigma process, the variance 2 would have to be decreased such that
x LSL
= 2.0 for :
PCRk = 2.0. The value of the variance is found by solving PCRk =
3
64 63.98
= 2.0
3
6 = 64.0 63.98
64.0 63.98
=
6
= 0.0033
Therefore, the process variance would have to be decreased to 2 = (0.0033)2 = 0.000011.
f) x = 0.0104
ARL =
1
1
=
= 5.87
p 0.1705
16-38
a)
Sample Mean
64.02
UCL=64.02
64.01
Mean=64.00
64.00
63.99
LCL=63.98
63.98
Subgroup
10
15
20
25
UCL=0.02382
Sample StDev
16-58
0.02
0.01
S=0.009274
0.00
LCL=0
b) = x = 64
c) Same as 16-57
s
0.009274
=
= 0.0104
c4
0.8862
USL LSL 64.02 63.98
PCR =
=
= 0.641
6
6(0.0104)
The process does not meet the minimum capability level of PCR 1.33.
d) Same as 16-57
USL x x LSL
PCRk = min
,
3
3
64.02 64 64 63.98
= min
,
3(0.0104) 3(0.0104)
= min[0.641,0.641]
= 0.641
e) Same as 16-57 e). In order to make this process a six-sigma process, the variance 2 would have to be
x LSL
decreased such that PCRk = 2.0. The value of the variance is found by solving PCRk =
= 2.0 for :
3
64 63.98
= 2.0
3
6 = 64.0 63.98
64.0 63.98
=
6
= 0.0033
16-39
ARL =
a)
Proportion
0.2
P=0.11
0.1
LCL=0.01613
0.0
0
10
20
Sample Number
There are no points beyond the control limits. The process is in control.
b)
1
UCL=0.1764
Proportion
16-59
1
1
=
= 5.87
p 0.1705
0.14
P=0.11
0.09
LCL=0.04363
0.04
0
10
20
Sample Number
There is one point beyond the upper control limit. The process is out of control. The revised limits are:
16-40
Proportion
0.14
P=0.1063
0.09
LCL=0.04093
0.04
0
10
20
Sample Number
1
1
UCL=1.503
Sample Count
16-60
U=0.528
LCL=0
0
10
15
20
25
Sample Number
Points14 and 23 are beyond the control limits. The process is out of control.
16-41
Sample Count
UCL=1.302
1.0
0.5
U=0.4261
LCL=0
0.0
0
10
20
Sample Number
1
1
Sample Count
U C L= 0.7514
0.5
U = 0.264
LC L= 0
0.0
0
10
15
20
25
S am ple N um ber
Sample Count
0.6
0.5
0.4
0.3
U=0.2130
0.2
0.1
LCL=0
0.0
0
10
20
Sample Number
16-42
I-MR Chart of C1
Individual V alue
60.3275
U C L=60.327362
60.3270
_
X=60.32641
60.3265
60.3260
60.3255
LC L=60.325458
2
10
12
O bser vation
14
16
18
20
U C L=0.001169
M oving Range
0.00100
0.00075
0.00050
__
M R=0.000358
0.00025
LC L=0
0.00000
2
10
12
O bser vation
14
16
18
20
b) The chart is identical to the chart in part (a) except for the scale of the individuals chart.
I-MR Chart of C1
Individual V alue
0.0015
U C L=0.001362
0.0010
_
X=0.00041
0.0005
0.0000
-0.0005
LC L=-0.000542
2
10
12
O bser vation
14
16
18
20
U C L=0.001169
0.00100
M oving Range
16-61
0.00075
0.00050
__
M R=0.000358
0.00025
LC L=0
0.00000
2
10
12
O bser vation
14
16
18
USL LSL
0.002
=
= 1.0505
6
6(0.0003173)
0.0009 0.0011
,
PCRk = min
= 0.9455
3
3
PCR =
16-43
20
16-62
a)
Individual V alue
7000
U C L=6669
_
X=5832
6000
LC L=4995
5000
1
4000
3
12
15
18
O bser vation
21
24
27
30
U C L=1028
M oving Range
1000
750
500
__
M R=315
250
0
LC L=0
3
12
15
18
O bser vation
21
24
27
30
b) The data does not appear to be generated from an in-control process. The average tends to drift
to larger values and then drop back off over the last 5 values.
a)
Trial control limits :
S chart
UCL= 170.2482
CL = 86.4208
LCL = 2.59342
600
400
xbar
800
X bar chart
UCL= 670.0045
CL = 558.766
LCL = 447.5275
10
15
20
25
15
20
25
50
150
Index
16-63
10
Index
16-44
b) An estimate of is given by
Based on the capability ratios above (both <1), the process is operating off-center and will result
in a large number of non-conforming units.
c) To determine the new variance, solve
558.77 330
in this exercise, we find =38.128 or 2=1453.77.
3
670
.
0045
600
447
.
5275
600
=
P (447.5275 < X < 670.0045) = P
<Z<
90.8259
90.8259
6
6
b) =84.78839
USL x x LSL
,
PCRK = min
3
3
830 551.95 551.95 330
,
= min
3(84.79)
3(84.79)
= 0.8725
c)
USL x x LSL
PCRK = min
,
3
3
551.95 330
=
3( )
=2
So = 36.9917
16-45
d) In-control distribution
~ N(551.9477, 34.6 )
2
P[448.1
=0.9463
Out-of-control ARL=
16-64
1
= 18.6 19.
1 0.9463
CL =
UCL = + 2
, LCL = 2
n
n
X
P X > + 2
> 2 = P( Z > 2) = 1 P( Z < 2) = 1 0.97725 = 0.02275
= P
n
/ n
and
X
P X < 2
< 2 = P( Z < 2) = 1 P( Z < 2) = 1 0.97725 = 0.02275
= P
n
/ n
The answer is 0.02275 + 0.02275 = 0.0455. The answer for 3-sigma control limits is 0.0027. The 3-sigma
control limits result in many fewer false alarms.
16-65
a) The following control chart use the average range from 25 subgroups of size 3 to estimate the process
standard deviation. Minitab uses a pooled estimate of variance as the default method for an EWMA control
chart so that the range method was selected from the options. Points are clearly out of control.
EWMA Chart of C3, ..., C5
64.7
64.6
64.5
EWMA
64.4
64.3
UCL=64.2759
64.2
_
_
X=64.1334
64.1
64.0
LCL=63.9910
63.9
2
10
12 14
Sample
16-46
16
18
20
22
24
b) The following control chart use the average range from 25 subgroups of size 3 to estimate the process
standard deviation. There is a big shift in the mean at sample 10 and the process is out of control at
this point.
EWMA Chart of C3, ..., C5
65.75
65.50
65.25
EWMA
65.00
64.75
64.50
UCL=64.380
_
_
X=64.133
64.25
64.00
LCL=63.887
2
10
12 14
Sample
16
18
20
22
24
16-66
a) The data appears to be generated from an out-of-control process.
EWMA Chart of C1
6600
6400
6200
UCL=6111
EWMA
6000
_
_
X=5832
5800
5600
LCL=5553
5400
5200
5000
3
12
15
18
Sample
21
16-47
24
27
30
EWMA Chart of C1
7000
6500
UCL=6315
EWMA
6000
_
_
X=5832
5500
LCL=5349
5000
4500
3
12
15
18
Sample
21
24
27
30
60.3267
EWMA
60.3266
60.3265
_
_
X=60.32641
60.3264
60.3263
60.3262
60.3261
LCL=60.3260927
2
10
12
Sample
14
16
18
20
UCL=60.326960
60.32675
EWMA
16-67
60.32650
_
_
X=60.32641
60.32625
60.32600
LCL=60.325860
2
10
12
Sample
16-48
14
16
18
20
16-68
CUSUM Chart of C1
UCL=5.27
5.0
Cumulative Sum
2.5
0.0
-2.5
-5.0
LCL=-5.27
3
12
15
18
Sample
21
24
27
30
Process standard deviation is estimated using the average moving range of size 2 with MR/d2, where d2 =
1.128. The estimate is 1.05. Recommendation for k and h are 0.5 and 4 or 5, respectively for n =1. For this
chart h = 5 was used.
The process is not in control.
K = k = 1, so that k = 0.5
H = h = 10, so that h = 5
Upper CUSUM
40
Cumulative Sum
16-69
30
20
10
10
0
-10
-10
Lower CUSUM
10
Subgroup Number
16-49
20
16-70
85.8529
Cumulative Sum
80
-80
-85.8529
Lower CUSUM
10
20
Subgroup Number
Process standard deviation is estimated using the average moving range of size 2 with MR/d2, where d2 =
1.128 for a moving range of 2. The estimate is 17.17. Recommendation for k and h are 0.5 and 4 or 5,
respectively, for n = 1.
a)
CUSUM Chart of x
0.020
0.015
Cumulative Sum
16-71
UCL=0.01152
0.010
0.005
0.000
-0.005
-0.010
LCL=-0.01152
2
10
12 14
Sample
16
18
20
22
24
is estimated using the moving range: 0.0026/1.128=0.0023. H and K were computed using
k=0.5 and h=5. The process is not in control.
b) EWMA gives similar results.
16-50
EWMA Chart of x
0.404
UCL=0.403489
0.403
EWMA
0.402
_
_
X=0.401184
0.401
0.400
0.399
LCL=0.398879
2
16-72
10
12 14
Sample
16
18
20
22
24
a) Let p denote the probability that a point plots outside of the control limits when the mean has shifted from
0 to = 0 + 1.5. Then,
3
3
n
n
1.5
X
1.5
= P
3<
<
+ 3
/ n / n
/ n
Therefore, the probability the shift is undetected for three consecutive samples is (1 p)3 =
(0.5)3 = 0.125.
b) If 2-sigma control limits were used, then
2
2
1.5
1.5
X
= P
2<
<
+ 2
/ n / n
/ n
= P (5 < Z < 1) = P ( Z < 1) P ( Z < 5)
= 1 0.84134 [0] = 0.15866
Therefore, the probability the shift is undetected for three consecutive samples is (1 p)3 = (0.15866)3 =
0.004.
c) The 2-sigma limits are narrower than the 3-sigma limits. Because the 2-sigma limits have a smaller
probability of a shift being undetected, the 2-sigma limits would be better than the 3-sigma limits for a mean
shift of 1.5. However, the 2-sigma limits would result in more signals when the process has not shifted
(false alarms).
16-51
16-73.
ARL = 1/p where p is the probability a point falls outside the control limits.
a) = 0 + and n = 1
0 +
0
=P Z>
/ n
0
0
n
+ P Z <
/ n
= P( Z > 3 n ) + P( Z < 3 n )
= P( Z > 2) + P( Z < 4)
= 1 0.97725 + [0]
= 0.02275
when n = 1
0 +
0 2
=P Z>
/ n
0
0 2
n
+ P Z <
/ n
= P ( Z > 3 2 n ) + P ( Z < 3 2 n )
= P( Z > 1) + P ( Z < 5)
when n = 1
= 1 0.84134 + [0]
= 0.15866
Therefore, ARL = 1/p = 1/0.15866 = 6.30.
c) = 0 + 3
0 +
0 3
=P Z>
/ n
0
0 3
n
+ P Z <
/ n
= P( Z > 3 3 n ) + P( Z < 3 3 n )
= P( Z > 0) + P( Z < 6)
when n = 1
= 1 0.50 + [0]
= 0.50
Therefore, ARL = 1/p = 1/0.50 = 2.00.
d) The ARL is decreasing as the magnitude of the shift increases from to 2 to 3. The ARL decrease as
the magnitude of the shift increases since a larger shift is more likely to be detected earlier than a smaller
shift.
16-52
16-74 a) Because ARL = 370, on the average we expect there to be one false alarm every 370 hours. Each 30-day
month contains 30 24 = 720 hours of operation. Consequently, we expect 720/370 = 1.9 false alarms each month
= P( Z > 1) + P( Z < 3)
= 1 P( Z < 1) + [1 P( Z < 3)]
= 1 0.84134 + 1 0.99865
= 0.160
Therefore, ARL=1/p = 1/0.160 = 6.25. The 2-sigma limits reduce the ARL for detecting a shift in the
mean of magnitude . However, the next part of this solution shows that the number of false alarms increases
with 2-sigma limits.
c) 2 limits
a)
X-bar and Range - Initial Study
Charting xbar
X-bar
| Range
----| ----UCL: +
3.0 sigma = 140.168
| UCL: +
3.0 sigma = 2.48437
Centerline
= 139.49
| Centerline
= 1.175
LCL: 3.0 sigma = 138.812
| LCL: 3.0 sigma = 0
out of limits = 9
| out of limits = 0
Estimated
process mean = 139.49
process sigma =
0.505159
mean Range
=
1.175
Pro b lem 16-51
141
140.168
140
X-b ar
139.49
139
138.812
138
137
0
12
16
20
2.5
2.48437
2
R an g e
1.5
1.175
1
0.5
0
0
0
8
12
su b g rou p
16
20
There are points beyond the control limits. The process is out of control. The points are 8, 10, 11,
12, 13, 14, 15, 16, and 19.
16-53
Xb ar C h art
140.5
U C L= 140.4
140.0
xbar
M ean= 1 39.7
139.5
139.0
LC L= 139.0
0
10
S u b g ro u p N u m b e r
R C h a rt
3 .0
U C L = 2 .5 9 6
1 .5
M e a n= 1 .2 2 7
LC L=0
0
0
10
S u b g ro u p N u m b e r
d2
2.326
3
3
= min[1.45,1.08] = 1.08
Because the process capability ratios are less than unity, the process capability appears to be poor. PCR is
slightly larger than PCRk indicating that the process is somewhat off center.
16-54
d) In order to make this process a six-sigma process, the variance 2 would have to be decreased such that
PCRk = 2.0. The value of the variance is found by solving PCRk =
139.709 138
= 2.0
3
6 = 139.709 138
139.709 138
=
6
= 0.2848
x LSL
= 2.0 for :
3
= P
<
<
x
0.528
0.528
The probability that this shift will be detected on the next sample is 1 p = 10.8197 = 0.1803.
ARL =
1
1
=
= 5.55
1 p 0.1803
16-76
a) The probability of having no signal is
16-78
LCL = p 3
p (1 p )
0.05(1 0.05)
= 0.05 3
= 0.015 0
n
100
UCL = p + 3
p (1 p )
0.05(1 0.05)
= 0.05 + 3
= 0.115
n
100
16-55
0.115 0.08
P 0.08
P (0 P 0.115) = P ( P 0.115) = P
0.08(0.92)
0.08(0.92)
100
100
= P ( Z 1.29) = 0.90
Using the normal approximation to the distribution of P . Therefore, the probability of detecting the shift
on the first sample following the shift is 1 0.90 = 0.10.
b) The probability that the control chart detects a shift to 0.08 on the second sample, but not the first, is
(0.90)0.10= 0.09. This uses the fact that the samples are independent to multiply the probabilities.
c) p = 0.10
0.115 0.10
P 0.10
P (0 P 0.115) = P ( P 0.115) = P
0.10(0.90)
0.10(0.90)
100
100
= P ( Z 0.5) = 0.69146
from the normal approximation to the distribution of P . Therefore, the probability of detecting the shift on
the first sample following the shift is 1 0.69146 = 0.30854.
The probability that the control chart detects a shift to 0.10 on the second sample after the shift, but not the
first, is 0.69146(0.30854) = 0.2133.
d) A larger shift is generally easier to detect. Therefore, we should expect a shift to 0.10 to be detected
quicker than a shift to 0.08.
16-79
u=8
a) n = 4
UCL = u + 3
u
8
=8+3
= 12.24
n
4
u
8
=83
= 3.76
n
4
12.24 16
P( U > 12.24 when = 16) = P Z >
16
LCL = u 3
3.76 16
< 3.78) = P Z <
16
16-56
UCL = u + 3
8
u
= 8+3
= 10.68
10
n
LCL = u 3
8
u
= 83
= 5.32
10
n
10.68 16
= P(Z > 4.22) = 1
P(U > 10.68 when = 16) = P Z >
16
10
So the probability is 1.
16-80
u = 10
a) n = 1
UCL = u + 3
10
u
= 10 + 3
= 19.49
1
n
LCL = u 3
10
u
= 10 3
= 0.51
1
n
19.94 14
P( U > 19. 94 when = 14) = P Z >
14
= P(Z > 1.47)= 1 P(Z < 1.47) = 1 0.9292 = 0.0708
and
P( U < 0.51) =
0.51 14
P Z <
=0
14
b) n = 4
UCL = u + 3
u
10
= 10 + 3
= 14.74
n
4
u
10
= 10 3
= 5.26
n
4
14
.
74
14
P( U > 14.74 when = 14) =
= P(Z >0.40) = 1 0.6554 = 0.3446
P Z>
14
5
.
26
14
= P(Z < 4.67) = 0
P( U < 5.26 when = 14) = P Z <
14
LCL = u 3
16-57