<< Chapter < Page | Chapter >> Page > |
Remarks
The transition diagram and the transition matrix
The previous examples suggest that a Markov chain is a dynamic system, evolving in time. On the other hand, the stochastic behavior of a homogeneous chain is determined completelyby the probability distribution for the initial state and the one-step transition probabilities as presented in the transition matrix P . The time-invariant transition matrix may convey a static impression of the system.However, a simple geometric representation, known as the transition diagram , makes it possible to link the unchanging structure, represented by thetransition matrix, with the dynamics of the evolving system behavior.
Definition . A transition diagram for a homogeneous Markov chain is a linear graph with one node for each state and one directededge for each possible one-step transition between states (nodes).
We ignore, as essentially impossible, any transition which has zero transition probability. Thus, the edges on the diagram correspond topositive one-step transition probabilities between the nodes connected. Since for some pair of states, we may have but , we may have a connecting edge between two nodes in one direction, but none in the other. The system can be viewed as an object jumping from state to state (node tonode) at the successive transition times . As we follow the trajectory of this object, we achieve a sense of the evolution of the system.
Consider, again, the transition matrix P for the inventory problem (rounded to three decimals).
Figure 1 shows the transition diagram for this system. At each node corresponding to one of the possible states, the state value is shown. In this example, the state value is one less than the state number. For convenience, we refer to the node for state , which has state value k , as node k . If the state value is zero, there are four possibilities: remain in that condition with probability 0.080; move to node 1 with probability 0.184; move to node 2with probability 0.368; or move to node 3 with probability 0.368. These are represented by the “self loop” and a directed edge to each of the nodes representingstates. Each of these directed edges is marked with the (conditional) transition probability. On the other hand, probabilities of reaching state value 0 from each of the others isrepresented by directed edges into the node for state value 0. A similar situation holds for each other node. Note that the probabilities on edges leaving a node (including a selfloop) must total to one, since these correspond to the transition probability distribution from that node. There is no directed edge from the node 2 to node 3, since the probabilityof a transition from value 2 to value 3 is zero. Similary, there is no directed edge from node 1 to either node 2 or node 3.
Notification Switch
Would you like to follow the 'Applied probability' conversation and receive update notifications?