Final Exam (listed) for 2008: December 2 Due Day: December 9 (9:00AM)
-
Upload
mckenzie-harmon -
Category
Documents
-
view
29 -
download
2
description
Transcript of Final Exam (listed) for 2008: December 2 Due Day: December 9 (9:00AM)
1 1 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Final Exam (listed) for 2008: Final Exam (listed) for 2008: December 2 December 2
Due Day: December 9 Due Day: December 9 (9:00AM)(9:00AM)
Exam Materials: All the Exam Materials: All the Topics After Mid Term ExamTopics After Mid Term Exam
2 2 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
EMGT 501
HW Solutions
Chapter 15 - SELF TEST 9
Chapter 15 - SELF TEST 35
3 3 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
15-9a.a.
Method Forecast MSE
3-Quarter 80.73 2.53
4-Quarter 80.55 2.81
The 3-quarter moving average forecast is The 3-quarter moving average forecast is betterbetter
because it has the smallest MSE. because it has the smallest MSE.
4 4 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
b.b. Method Forecast MSE
= .4 80.40 2.40
= .5 80.57 2.01
The a = .5 smoothing constant is better The a = .5 smoothing constant is better becausebecause
it has the smallest MSE. it has the smallest MSE.
c.c. The exponential smoothing is better The exponential smoothing is better because it has the smallest MSE. because it has the smallest MSE.
5 5 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
15-35a.a.
xi yi xiyi xi2
20 21 420 400
20 19 380 400
40 15 600 1600
30 16 480 900
60 14 840 3600
40 17 680 1600
210 102 3400 8500
1 0ˆ35 17 0.1478 22.713 22.713 0.1478 x y b b y x
6 6 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
= 22.173 - 0.1478(50) = 14.783 or = 22.173 - 0.1478(50) = 14.783 or approximately 15 approximately 15 defective parts defective parts
y
b.b.
7 7 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
)4()bxa(y
)3()bxa(y
have we(2), to(1) From
)2()bxa(y)bxa(y2
1
)1()bxa(y)bxa(y2
1
)deviations negative and positive
ing(represent variablesnew introduce We
)bxa(yMin
iiii
iiii
iiiii
iiiii
n
1iii
8 8 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
0, URS,:,
)~1(..
)(Min
or
)~1()(..
)(Min
have we(4), and (3) From
n
1i
n
1i
ii
iiii
ii
iiii
ii
ba
niybxats
nibxayts
9 9 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
17b40b40aa
14b60b60aa
16b30b30aa
15b40b40aa
19b20b20aa
21b20b20aa s.t.
)( Min
66
55
44
33
22
11
6
1iii
10 10 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
14.783) :(LS 15.25
0.125(50)-21.5y
)b(
0.1478x)-22.713y :(LS
x125.0500.21
bx-ay
)a(
11 11 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Chapter 16Chapter 16Markov ProcessesMarkov Processes
Transition ProbabilitiesTransition Probabilities Steady-State ProbabilitiesSteady-State Probabilities Absorbing StatesAbsorbing States Transition Matrix with Transition Matrix with
SubmatricesSubmatrices Fundamental MatrixFundamental Matrix
12 12 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Markov ProcessesMarkov Processes
Markov process modelsMarkov process models are useful in are useful in studying the evolution of systems over studying the evolution of systems over repeated trials or sequential time periods repeated trials or sequential time periods or stages.or stages.
They have been used to describe the They have been used to describe the probability that:probability that:
•a machine that is functioning in one a machine that is functioning in one period will continue to function or break period will continue to function or break down in the next period.down in the next period.
•A consumer purchasing brand A in one A consumer purchasing brand A in one period will purchase brand B in the next period will purchase brand B in the next period.period.
13 13 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Transition ProbabilitiesTransition Probabilities
Transition probabilitiesTransition probabilities govern the govern the manner in which the state of the manner in which the state of the system changes from one stage to system changes from one stage to the next. These are often the next. These are often represented in a represented in a transition matrixtransition matrix. .
14 14 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Transition ProbabilitiesTransition Probabilities
A system has a A system has a finite Markov chainfinite Markov chain with with stationary transition probabilitiesstationary transition probabilities if: if:
•there are a finite number of states,there are a finite number of states,
•the transition probabilities remain the transition probabilities remain constant from stage to stage, andconstant from stage to stage, and
•the probability of the process being in the probability of the process being in a particular state at stage a particular state at stage n+n+1 is 1 is completely determined by the state completely determined by the state of the process at stage of the process at stage nn (and not the (and not the state at stage state at stage n-n-1). This is referred to 1). This is referred to as the as the memory-less propertymemory-less property..
15 15 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Steady-State ProbabilitiesSteady-State Probabilities
The The state probabilitiesstate probabilities at any stage at any stage of the process can be recursively of the process can be recursively calculated by multiplying the initial calculated by multiplying the initial state probabilities by the state of state probabilities by the state of the process at stage the process at stage nn..
The probability of the system being The probability of the system being in a particular state after a large in a particular state after a large number of stages is called a number of stages is called a steady-state probabilitysteady-state probability. .
16 16 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Steady-State ProbabilitiesSteady-State Probabilities
Steady state probabilitiesSteady state probabilities can be can be found by solving the system of found by solving the system of equations equations PP = = together with together with the condition for probabilities that the condition for probabilities that ii = 1. = 1.
•Matrix Matrix PP is the transition is the transition probability matrixprobability matrix
•Vector Vector is the vector of steady is the vector of steady state probabilities.state probabilities.
17 17 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Absorbing StatesAbsorbing States
An An absorbing stateabsorbing state is one in which is one in which the probability that the process the probability that the process remains in that state once it enters remains in that state once it enters the state is 1.the state is 1.
If there is more than one absorbing If there is more than one absorbing state, then a steady-state state, then a steady-state condition independent of initial condition independent of initial state conditions does not exist.state conditions does not exist.
18 18 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Transition Matrix with SubmatricesTransition Matrix with Submatrices
If a Markov chain has both absorbing and If a Markov chain has both absorbing and nonabsorbing states, the states may be nonabsorbing states, the states may be rearranged so that the transition matrix can be rearranged so that the transition matrix can be written as the following composition of four written as the following composition of four submatrices: submatrices: II,, 0 0, , RR, and , and QQ: :
II 00
RR QQ
19 19 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Transition Matrix with SubmatricesTransition Matrix with Submatrices
II = an identity matrix indicating one always = an identity matrix indicating one always remains in an absorbing state once it is remains in an absorbing state once it is
reachedreached
00 = a zero matrix representing 0 probability of = a zero matrix representing 0 probability of
transitioning from the absorbing states to transitioning from the absorbing states to the the
nonabsorbing statesnonabsorbing states
RR = the transition probabilities from the = the transition probabilities from the nonabsorbing states to the nonabsorbing states to the
absorbing statesabsorbing states
QQ = the transition probabilities between the = the transition probabilities between the nonabsorbing states nonabsorbing states
20 20 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Fundamental MatrixFundamental Matrix
The The fundamental matrixfundamental matrix, , NN, is the inverse of , is the inverse of the difference between the identity matrix the difference between the identity matrix and the and the QQ matrix. matrix.
NN = ( = (II - - Q Q ))-1-1
21 21 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
NRNR Matrix Matrix
The The NRNR matrix matrix is the product of the is the product of the fundamental (fundamental (NN) matrix and the ) matrix and the R R matrix. matrix.
It gives the probabilities of eventually moving It gives the probabilities of eventually moving from each nonabsorbing state to each from each nonabsorbing state to each absorbing state. absorbing state.
Multiplying any vector of initial nonabsorbing Multiplying any vector of initial nonabsorbing state probabilities by state probabilities by NRNR gives the vector of gives the vector of probabilities for the process eventually probabilities for the process eventually reaching each of the absorbing states. Such reaching each of the absorbing states. Such computations enable economic analyses of computations enable economic analyses of systems and policies.systems and policies.
22 22 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: North’s HardwareExample: North’s Hardware
Henry, a persistent salesman, calls Henry, a persistent salesman, calls North'sNorth's
Hardware Store once a week hoping to Hardware Store once a week hoping to speakspeak
with the store's buying agent, Shirley.with the store's buying agent, Shirley.
If Shirley does not accept Henry'sIf Shirley does not accept Henry's
call this week, the probability shecall this week, the probability she
will do the same next week is .35.will do the same next week is .35.
On the other hand, if she acceptsOn the other hand, if she accepts
Henry's call this week, the probabilityHenry's call this week, the probability
she will not do so next week is .20. she will not do so next week is .20.
23 23 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: North’s HardwareExample: North’s Hardware
Transition MatrixTransition Matrix
Next Week's Next Week's CallCall
Refuses AcceptsRefuses Accepts
ThisThis Refuses .35 Refuses .35 .65 .65
Week'sWeek's CallCall Accepts .20 Accepts .20
.80 .80
24 24 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Steady-State ProbabilitiesSteady-State Probabilities
QuestionQuestion
How many times per year can Henry How many times per year can Henry expect to talk to Shirley?expect to talk to Shirley?
AnswerAnswer
To find the expected number of accepted To find the expected number of accepted calls per year, find the long-run proportion calls per year, find the long-run proportion (probability) of a call being accepted and (probability) of a call being accepted and multiply it by 52 weeks.multiply it by 52 weeks.
continued . . .continued . . .
Example: North’s HardwareExample: North’s Hardware
25 25 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Steady-State ProbabilitiesSteady-State Probabilities
AnswerAnswer (continued)(continued)
Let Let 11 = long run proportion of refused = long run proportion of refused callscalls
22 = long run proportion of accepted = long run proportion of accepted callscalls
Then, Then,
.35 .65 .35 .65
[[ ] = [] = [ ]]
.20 .80 .20 .80
continued . . .continued . . .
Example: North’s HardwareExample: North’s Hardware
26 26 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Steady-State ProbabilitiesSteady-State Probabilities
Answer (continued)Answer (continued)
+ + = = (1) (1)
+ + = = (2) (2)
+ + = 1 (3) = 1 (3)
Solve for Solve for and and
continued . . .continued . . .
Example: North’s HardwareExample: North’s Hardware
27 27 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: North’s HardwareExample: North’s Hardware
Steady-State ProbabilitiesSteady-State Probabilities
Answer (continued)Answer (continued)
Solving using equations (2) and (3). Solving using equations (2) and (3). (Equation 1 is redundant.) Substitute (Equation 1 is redundant.) Substitute = 1 - = 1 - into (2) to give:into (2) to give:
.65(1 - .65(1 - 22) + ) + = = 22
This gives This gives = .76471. Substituting back = .76471. Substituting back into equation (3) gives into equation (3) gives = .23529. = .23529.
Thus the expected number of accepted Thus the expected number of accepted calls per year is:calls per year is:
(.76471)(52) = 39.76 or about 40(.76471)(52) = 39.76 or about 40
28 28 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
State ProbabilityState Probability
QuestionQuestion
What is the probability Shirley will What is the probability Shirley will accept Henry's next two calls if she does not accept Henry's next two calls if she does not accept his call this week?accept his call this week?
Example: North’s HardwareExample: North’s Hardware
29 29 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: North’s HardwareExample: North’s Hardware
State ProbabilityState Probability
AnswerAnswer
P = .35(.35) = .1225P = .35(.35) = .1225
P = .35(.65) = .2275P = .35(.65) = .2275
P = .65(.20) = .1300P = .65(.20) = .1300RefusesRefuses
RefusesRefuses
RefusesRefuses
RefusesRefuses
AcceptsAccepts
AcceptsAccepts
AcceptsAccepts
.35.35
.35.35
.65.65
.20.20
.80.80
.65.65P = .65(.80) = .5200P = .65(.80) = .5200P = .65(.80) = .5200P = .65(.80) = .5200
30 30 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
State ProbabilityState Probability
QuestionQuestion
What is the probability of Shirley What is the probability of Shirley accepting exactly one of Henry's next two calls accepting exactly one of Henry's next two calls if she accepts his call if she accepts his call this week?this week?
Example: North’s HardwareExample: North’s Hardware
31 31 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: North’s HardwareExample: North’s Hardware
State ProbabilityState Probability
AnswerAnswer
The probability of exactly one of the next The probability of exactly one of the next two calls being accepted if this week's call is two calls being accepted if this week's call is accepted can be found by adding the accepted can be found by adding the probabilities of (accept next week and refuse probabilities of (accept next week and refuse the following week) and (refuse next week and the following week) and (refuse next week and accept the following week) = accept the following week) =
.12 + .16 = .28.12 + .16 = .28
32 32 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
The vice president of personnel at JetairThe vice president of personnel at Jetair
Aerospace has noticed that yearly shifts in personnelAerospace has noticed that yearly shifts in personnel
can be modeled by a Markov process. The transitioncan be modeled by a Markov process. The transition
matrix is:matrix is:
Next YearNext Year Same Pos. Promotion Retire Quit FiredSame Pos. Promotion Retire Quit Fired
Current YearCurrent Year Same Same
Position .55 .10 .05 .20 .10Position .55 .10 .05 .20 .10 Promotion .70 .20 0 .10 0Promotion .70 .20 0 .10 0 Retire Retire 0 0 1 0 0 0 0 1 0 0 Quit Quit 0 0 0 1 0 0 0 0 1 0 Fired Fired 0 0 0 0 0 0 0 0
1 1
Example: Jetair AerospaceExample: Jetair Aerospace
33 33 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: Jetair AerospaceExample: Jetair Aerospace
Transition MatrixTransition Matrix
Next YearNext Year Retire Quit Fired Same Retire Quit Fired Same
PromotionPromotion Current YearCurrent Year Retire Retire 1 0 0 0 01 0 0 0 0 Quit Quit 0 1 0 0 0 1 0 0
0 0 Fired Fired 0 0 1 0 0 0 1 0
0 0
Same Same .05 .20 .10 .55 .05 .20 .10 .55 .10 .10
Promotion 0 .10 0 .70 .20Promotion 0 .10 0 .70 .20
34 34 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Fundamental MatrixFundamental Matrix
-1 -1 -1 -1
1 0 .55 .10 1 0 .55 .10 .45 -.10 .45 -.10
N N = ( = (II - - Q Q ) ) -1 -1 = = = =
0 1 .70 .20 0 1 .70 .20 -.70 .80 -.70 .80
Example: Jetair AerospaceExample: Jetair Aerospace
35 35 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: Jetair AerospaceExample: Jetair Aerospace
Fundamental MatrixFundamental Matrix
The determinant, The determinant, dd = = aaaa - - aaaa
= (.45)(.80) - (-.70)(-.10) = = (.45)(.80) - (-.70)(-.10) = .29 .29
Thus, Thus,
.80/.29 .10/.29 2.76 .34 .80/.29 .10/.29 2.76 .34
NN = = = =
.70/.29 .45/.29 2.41 .70/.29 .45/.29 2.41 1.551.55
36 36 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: Jetair AerospaceExample: Jetair Aerospace
NR NR MatrixMatrix
The probabilities of eventually moving to The probabilities of eventually moving to the absorbing states from the nonabsorbing the absorbing states from the nonabsorbing states are given by:states are given by:
2.76 .34 2.76 .34 .05 .20 .10 .05 .20 .10
NRNR = = x x
2.41 1.55 0 .10 2.41 1.55 0 .10 0 0
37 37 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: Jetair AerospaceExample: Jetair Aerospace
NR NR Matrix (continued)Matrix (continued)
Retire Quit FiredRetire Quit Fired
Same .14 .59 .28 Same .14 .59 .28
NRNR = =
Promotion .12 .64 .24Promotion .12 .64 .24
38 38 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: Jetair AerospaceExample: Jetair Aerospace
Absorbing StatesAbsorbing States
QuestionQuestion
What is the probability of someone who What is the probability of someone who was just was just promoted eventually retiring? . . . promoted eventually retiring? . . . quitting? . . . quitting? . . . being fired? being fired?
39 39 Slide
Slide
© 2005 Thomson/South-Western© 2005 Thomson/South-Western
Example: Jetair AerospaceExample: Jetair Aerospace
Absorbing States (continued)Absorbing States (continued)
AnswerAnswer
The answers are given by the bottom The answers are given by the bottom row of the row of the NRNR matrix. The answers are matrix. The answers are therefore:therefore:
Eventually Retiring = .12Eventually Retiring = .12
Eventually Quitting = .64Eventually Quitting = .64
Eventually Being Fired = .24Eventually Being Fired = .24