1 chapter 17 markov chains. 2 description sometimes we are interested in how a random variable...

50
1 Chapter 17 Markov Chains

Upload: phillip-bryant

Post on 18-Dec-2015

219 views

Category:

Documents


3 download

TRANSCRIPT

Page 1: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

1

Chapter 17Markov Chains

Page 2: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

2

Description

Sometimes we are interested in how a random variable changes over time.

The study of how a random variable evolves over time includes stochastic processes.

An explanation of stochastic processes – in particular, a type of stochastic process known as a Markov chain is included.

We begin by defining the concept of a stochastic process.

Page 3: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

3

5.1 What is a Stochastic Process?• Suppose we observe some characteristic of a system at

discrete points in time.

• Let Xt be the value of the system characteristic at time t. In most situations, Xt is not known with certainty before time t and may be viewed as a random variable.

• A discrete-time stochastic process is simply a description of the relation between the random variables X0, X1, X2 …..

• Example: Observing the price of a share of Intel at the beginning of each day

• Application areas: education, marketing, health services, finance, accounting, and production

Page 4: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

4

• A continuous –time stochastic process is simply the stochastic process in which the state of the system can be viewed at any time, not just at discrete instants in time.

• For example, the number of people in a supermarket t minutes after the store opens for business may be viewed as a continuous-time stochastic process.

Page 5: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

5

5.2 What is a Markov Chain?

• One special type of discrete-time is called a Markov Chain.

• Definition: A discrete-time stochastic process is a Markov chain if, for t = 0,1,2… and all statesP(Xt+1 = it+1|Xt = it, Xt-1=it-1,…,X1=i1, X0=i0)

=P(Xt+1=it+1|Xt = it)

• Essentially this says that the probability distribution of the state at time t+1 depends on the state at time t(it) and does not depend on the states the chain passed through on the way to it at time t.

Page 6: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

6

• In our study of Markov chains, we make further assumption that for all states i and j and all t, P(Xt+1 = j|Xt = i) is independent of t.

• This assumption allows us to write P(Xt+1 = j|Xt = i) = pij where pij is the probability that given the system is in state i at time t, it will be in a state j at time t+1.

• If the system moves from state i during one period to state j during the next period, we call that a transition from i to j has occurred.

Page 7: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

7

• The pij’s are often referred to as the transition

probabilities for the Markov chain.• This equation implies that the probability law relating

the next period’s state to the current state does not change over time.

• It is often called the Stationary Assumption and any Markov chain that satisfies it is called a stationary Markov chain.

• We also must define qi to be the probability that the chain is in state i at the time 0; in other words, P(X0=i) = qi.

Page 8: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

8

• We call the vector q= [q1, q2,…qs] the initial probability distribution for the Markov chain.

• In most applications, the transition probabilities are displayed as an s x s transition probability matrix P. The transition probability matrix P may be written as

ssss

s

s

ppp

ppp

ppp

P

21

22221

11211

Page 9: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

9

• For each I

• We also know that each entry in the P matrix must be nonnegative.

• Hence, all entries in the transition probability matrix are nonnegative, and the entries in each row must sum to 1.

sj

jijp

1

1

Page 10: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

10

The Gambler’s Ruin Problem

• At time 0, I have $2. At times 1, 2, …, I play a game in which I bet $1, with probabilities p, I win the game, and with probability 1 – p, I lose the game. My goal is to increase my capital to $4, and as soon as I do, the game is over. The game is also over if my capital is reduced to 0.– Let Xt represent my capital position after the time t game

(if any) is played– X0, X1, X2, …. May be viewed as a discrete-time stochastic

process

Page 11: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

11

The Gambler’s Ruin Problem

• $0 $1 $2 $3 $4

P =

10000

0100

0010

0001

00001

pp

pp

pp

Page 12: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

12

5.3 n-Step Transition Probabilities

• A question of interest when studying a Markov chain is: If a Markov chain is in a state i at time m, what is the probability that n periods later the Markov chain will be in state j?

• This probability will be independent of m, so we may writeP(Xm+n =j|Xm = i) = P(Xn =j|X0 = i) = Pij(n)

where Pij(n) is called the n-step probability of a transition from state i to state j.

• For n > 1, Pij(n) = ijth element of Pn

• Pij(2) is the (i, j)th element of matrix P2 = P1 P1

• Pij(n) is the (i, j)th element of matrix Pn = P1 Pn-1

Page 13: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

13

The Cola Example

• Suppose the entire cola industry produces only two colas.

• Given that a person last purchased cola 1, there is a 90% chance that their next purchase will be cola 1.

• Given that a person last purchased cola 2, there is an 80% chance that their next purchase will be cola 2.

1. If a person is currently a cola 2 purchaser, what is the probability that they will purchase cola 1 two purchases from now?

2. If a person is currently a cola 1 a purchaser, what is the probability that they will purchase cola 1 three purchases from now?

Page 14: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

14

The Cola Example• We view each person’s purchases as a Markov chain

with the state at any given time being the type of cola the person last purchased.

• Hence, each person’s cola purchases may be represented by a two-state Markov chain, where– State 1 = person has last purchased cola 1

– State 2 = person has last purchased cola 2

• If we define Xn to be the type of cola purchased by a person on her nth future cola purchase, then X0, X1, … may be described as the Markov chain with the following transition matrix:

Page 15: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

15

The Cola Example

We can now answer questions 1 and 2.

1. We seek P(X2 = 1|X0 = 2) = P21(2) = element 21 of P2:

21

80.20.

10.90.

2

1ColaCola

Cola

ColaP

66.34.

17.83.

80.20.

10.90.

80.20.

10.90.2P

Page 16: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

16

The Cola Example

– Hence, P21(2) =.34. This means that the probability is .34 that two purchases in the future a cola 2 drinker will purchase cola 1.

2. We seek P11(3) = element 11 of P3:

Therefore, P11(3) = .781

562.438.

219.781.

66.34.

17.83.

80.20.

10.90.)( 23 PPP

Page 17: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

17

• Many times we do not know the state of the Markov chain at time 0. Then we can determine the probability that the system is in state i at time n by using the reasoning.

• Probability of being in state j at time n

• where q=[q1, q2, … q3].

• Hence, qn = qopn = qn-1p

• Example, q0 = (.4,.6)– q1= (.4,.6)

q1 = (.48,.52)

)(1

nPq ij

si

ii

80.20.

10.90.

Page 18: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

18

• To illustrate the behavior of the n-step transition probabilities for large values of n, we have computed several of the n-step transition probabilities for the Cola example.

• This means that for large n, no matter what the initial state, there is a .67 chance that a person will be a cola 1 purchaser.

Page 19: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

19

5.4 Classification of States in a Markov Chain

• To understand the n-step transition in more detail, we need to study how mathematicians classify the states of a Markov chain.

• The following transition matrix illustrates most of the following definitions. A graphical representation is shown in the book (State-Transition diagram)

2.8.000

1.4.5.00

07.3.00

0005.5.

0006.4.

P

Page 20: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

20

• Definition: Given two states of i and j, a path from i to j

is a sequence of transitions that begins in i and ends in j, such that each transition in the sequence has a positive probability of occurring.

• Definition: A state j is reachable from state i if there is a path leading from i to j.

• Definition: Two states i and j are said to communicate if j is reachable from i, and i is reachable from j.

• Definition: A set of states S in a Markov chain is a closed set if no state outside of S is reachable from any state in S.

Page 21: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

21

• Definition: A state i is an absorbing state if pij=1.• Definition: A state i is a transient state if there exists a state j that is

reachable from i, but the state i is not reachable from state j.• Definition:If a state is not transient, it is called a recurrent state.• Definition:A state i is periodic with period k > 1 if k is the smallest

number such that all paths leading from state i back to state i have a length that is a multiple of k. If a recurrent state is not periodic, it is referred to as aperiodic.

• If all states in a chain are recurrent, aperiodic, and communicate with each other, the chain is said to be ergodic.

• The importance of these concepts will become clear after the next two sections.

Page 22: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

22

5.5 Steady-State Probabilities and Mean First Passage Times

• Steady-state probabilities are used to describe the long-run behavior of a Markov chain.

• Theorem 1: Let P be the transition matrix for an s-state ergodic chain. Then there exists a vector π = [π1 π2 … πs] such that

s

s

s

n

nP

21

21

21

lim

Page 23: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

23

• Theorem 1 tells us that for any initial state i,

• The vector π = [π1 π2 … πs] is often called the steady-state distribution, or equilibrium distribution, for the Markov chain. Hence, they are independent of the initial probability distribution defined over the states

jijn

nP

)(lim

Page 24: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

24

Transient Analysis & Intuitive Interpretation

• The behavior of a Markov chain before the steady state is reached is often call transient (or short-run) behavior.

• An interpretation can be given to the steady-state probability equations.

• This equation may be viewed as saying that in the steady-state, the “flow” of probability into each state must equal the flow of probability out of each state.

jk

kjkjjj pp )1(

Page 25: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

25

Steady-State Probabilities

• The vector = [1, 2, …. , s ] is often known as the steady-state distribution for the Markov chain

• For large n and all i,

• Pij(n+1) Pij(n) j

• In matrix form = P

• For any n and any i,

• Pi1(n) + Pi2(n) + … + Pis(n) = 1

• As n, we have 1 + 2 + …. + s = 1

s

kkjkj

kjsk

kikij

p

PnPnP

1

1)()1(

Page 26: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

26

An Intuitive Interpretation of Steady-State Probabilities

• Consider

• Subtracting jpjj from both sides of the above equation, we have

• Probability that a particular transition enters state j = probability that a particular transition leaves state j

s

kkjkj p

1

jk

kjkjjj pp )1(

Page 27: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

27

Use of Steady-State Probabilities in Decision Making

• In the Cola Example, suppose that each customer makes one purchase of cola during any week.

• Suppose there are 100 million cola customers. • One selling unit of cola costs the company $1 to

produce and is sold for $2.• For $500 million/year, an advertising firm

guarantees to decrease from 10% to 5% the fraction of cola 1 customers who switch after a purchase.

• Should the company that makes cola 1 hire the firm?

Page 28: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

28

• At present, a fraction π1 = ⅔ of all purchases are cola 1 purchases, since:

π1 = .90π1+.20π2

π2 = .10π1+.80π2

and using the following equation by π1 + π2 = 1

• Each purchase of cola 1 earns the company a $1 profit. We can calculate the annual profit as $3,466,666,667 [2/3(100 million)(52 weeks)$1].

• The advertising firm is offering to change the P matrix to

80.20.

05.95.1P

Page 29: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

29

• For P1, the steady-state equations becomeπ1 = .95π1+.20π2

π2 = .05π1+.80π2

• Replacing the second equation by π1 + π2 = 1 and solving, we obtain π1=.8 and π2 = .2.

• Now the cola 1 company’s annual profit will be $3,660,000,000 [.8(100 million)(52 weeks)$1-($500 million)].

• Hence, the cola 1 company should hire the ad agency.

Page 30: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

30

Inventory Example• A camera store stocks a particular model camera that can be ordered

weekly. Let D1, D2, … represent the demand for this camera (the number of units that would be sold if the inventory is not depleted) during the first week, second week, …, respectively. It is assumed that the Di’s are independent and identically distributed random variables having a Poisson distribution with a mean of 1. Let X0 represent the number of cameras on hand at the outset, X1 the number of cameras on hand at the end of week 1, X2 the number of cameras on hand at the end of week 2, and so on. – Assume that X0 = 3. – On Saturday night the store places an order that is delivered in time for the

next opening of the store on Monday. – The store using the following order policy: If there are no cameras in stock,

3 cameras are ordered. Otherwise, no order is placed. – Sales are lost when demand exceeds the inventory on hand

Page 31: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

31

Inventory Example• Xt is the number of Cameras in stock at the end of week t (as defined

earlier), where Xt represents the state of the system at time t• Given that Xt = i, Xt+1 depends only on Dt+1 and Xt (Markovian property)• Dt has a Poisson distribution with mean equal to one. This means that

P(Dt+1 = n) = e-11n/n! for n = 0, 1, …• P(Dt = 0 ) = e-1 = 0.368• P(Dt = 1 ) = e-1 = 0.368• P(Dt = 2 ) = (1/2)e-1 = 0.184• P(Dt 3 ) = 1 – P(Dt 2) = 1 – (.368 + .368 + .184) = 0.08• Xt+1 = max(3-Dt+1, 0) if Xt = 0 and Xt+1 = max(Xt – Dt+1, 0) if Xt 1, for t = 0, 1, 2, ….

Page 32: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

32

Inventory Example: (One-Step) Transition Matrix• P03 = P(Dt+1 = 0) = 0.368

• P02 = P(Dt+1 = 1) = 0.368

• P01 = P(Dt+1 = 2) = 0.184

• P00 = P(Dt+1 3) = 0.080

44434241

34333231

23221211

04030201

3

2

1

0

3210

pppp

pppp

pppp

pppp

Page 33: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

33

Inventory Example: Transition Diagram

0 1

2 3

Page 34: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

34

Inventory Example: (One-Step) Transition Matrix

368.368.184.080.3

0368.368.264.2

00368.632.1

368.368.184.080.0

3210

Page 35: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

35

Transition Matrix: Two-Step

• P(2) = PP

165.300.286.249.3

097.233.319.351.2

233.233.252.283.1

165.300.286.249.0

3210

)2( P

Page 36: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

36

Transition Matrix: Four-Step

• P(4) = P(2)P(2)

164.261.286.289.3

171.263.283.284.2

166.268.285.282.1

164.261.286.289.0

3210

)4( P

Page 37: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

37

Transition Matrix: Eight-Step

• P(8) = P(4)P(4)

166.264.285.286.3

166.264.285.286.2

166.264.285.286.1

166.264.285.286.0

3210

)8( P

Page 38: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

38

Steady-State Probabilities

• The steady-state probabilities uniquely satisfy the following steady-state equations

0 = 0p00 + 1p10 + 2p20 + 3p30

1 = 0p01 + 1p11 + 2p21 + 3p31

2 = 0p02 + 1p12 + 2p22 + 3p32

3 = 0p03 + 1p13 + 2p23 + 3p33

• 1 = 0 + 1 + 2 + 3

1

......s 2, 1, 0, jfor

0

0

s

jj

s

iijij p

Page 39: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

39

Steady-State Probabilities: Inventory Example

0 = .0800 + .6321 + .2642+ .0803

1 = .1840 + .3681 + .3682 + .1843

2 = .3680 + .3682 + .3683

3 = .3680 + .3683

• 1 = 0 + 1 + 2 + 3

0 = .286, 1 = .285, 2 = .263, 3 = .166• The numbers in each row of matrix P(8) match the

corresponding steady-state probability

Page 40: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

40

Mean First Passage Times

• For an ergodic chain, let mij = expected number of transitions before we first reach state j, given that we are currently in state i; mij is called the mean first passage time from state i to state j.

• In the example, we assume we are currently in state i. Then with probability pij, it will take one transition to go from state i to state j. For k ≠ j, we next go with probability pik to state k. In this case, it will take an average of 1 + mkj transitions to go from i and j.

Page 41: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

41

• This reasoning implies

• By solving the linear equations of the equation above, we find all the mean first passage times. It can be shown that

jk

kjikij mpm 1

iiim

1

Page 42: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

42

• For the cola example, π1=2/3 and π2 = 1/3

• Hence, m11 = 1.5 and m22 = 3

• m12 = 1 + p11m12 = 1 + .9m12

m21 = 1 + p22m21 = 1 + .8m21

Solving these two equations yields,

• m12 = 10 and m21 = 5

Page 43: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

43

Solving for Steady-State Probabilities and Mean First Passage Times on the Computer

• Since we solve steady-state probabilities and mean first passage times by solving a system of linear equations, we may use LINDO to determine them.

• Simply type in an objective function of 0, and type the equations you need to solve as your constraints.

Page 44: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

44

5.6 Absorbing Chains

• Many interesting applications of Markov chains involve chains in which some of the states are absorbing and the rest are transient states.

• This type of chain is called an absorbing chain.• To see why we are interested in absorbing chains

we consider the following accounts receivable example.

Page 45: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

45

Accounts Receivable Example• The accounts receivable situation of a firm is often

modeled as an absorbing Markov chain.• Suppose a firm assumes that an account is uncollected

if the account is more than three months overdue.• Then at the beginning of each month, each account may

be classified into one of the following states:• State 1 New account

• State 2 Payment on account is one month overdue

• State 3 Payment on account is two months overdue

• State 4 Payment on account is three months overdue

• State 5 Account has been paid

• State 6 Account is written off as bad debt

Page 46: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

46

• Suppose that past data indicate that the following Markov chain describes how the status of an account changes from one month to the next month:

100000

010000

3.7.0000

06.4.000

05.05.00

04.006.0New

1 month

2 months

3 months

Paid

Bad Debt

New 1 month 2 months 3 months Paid Bad Debt

Page 47: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

47

• To simplify our example, we assume that after three months, a debt is either collected or written off as a bad debt.

• Once a debt is paid up or written off as a bad debt, the account if closed, and no further transitions occur.

• Hence, Paid or Bad Debt are absorbing states. Since every account will eventually be paid or written off as a bad debt, New, 1 month, 2 months, and 3 months are transient states.

Page 48: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

48

• A typical new account will be absorbed as either a collected debt or a bad debt.

• What is the probability that a new account will eventually be collected?

• To answer this questions we must write a transition matrix. We assume s – m transient states and m absorbing states. The transition matrix is written in the form of m

columns

10

RQs-m rows

m rows

s-mcolumns

P =

Page 49: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

49

• The transition matrix for this example is

• Then s =6, m =2, and Q and R are as shown.

100000

010000

3.7.0000

06.4.000

05.05.00

04.006.0New

1 month

2 months

3 months

Paid

Bad Debt

New 1 month 2 months 3 months Paid Bad Debt

Q R

Page 50: 1 Chapter 17 Markov Chains. 2 Description Sometimes we are interested in how a random variable changes over time. The study of how a random variable evolves

50

1. What is the probability that a new account will eventually be collected? (.964)

2. What is the probability that a one-month overdue account will eventually become a bad debt? (.06)

3. If the firm’s sales average $100,000 per month, how much money per year will go uncollected?

From answer 1, only 3.6% of all debts are uncollected. Since yearly accounts payable are $1,200,000 on the average, (0.036)(1,200,000) = $43,200 per year will be uncollected.