<< Chapter < Page Chapter >> Page >
Unifilar source for Example 4
Unifilar source for [link] .

An irreducible component with period at least 2 is called periodic. In contrast, a component with period 1 is called aperiodic or ergodic, because the probabilities in being ineach of the states eventually converge to the statistical averages. Note that a periodic component cannot yield a stationary distribution.

An irreducible component E contains asymptotic state probabilities given by the solution of the equations,

s ' E q ( s ' ) q ( s | s ' ) = q ( s ) , s E .

Additionally,

lim t Pr { S t = s | S 1 = s ' } = q ( s ) ,

for aperiodic E . The convergence of Pr { S t = s | S 1 = s ' } to q ( s ) is exponential in t , and depends on the eigen-values of the transition probability matrix. In periodic components we donot have stationarity, but the following weaker property still holds,

lim t 1 t i = 1 t Pr { S i = s } = q ( s ) .

The entropy rate of the output of a unifilar source at state s is given by,

H ( X | S = s ) = - a A p ( a | s ) · log p ( a | s ) .

To see this, we begin with a lemma.

Lemma 1 For a unifilar source,

H ( X n | X 1 n - 1 , S 1 = s ' ) = s S Pr { S n = s | S 1 = s ' } H ( X | s ) .

Proof Pr { X n | X 1 n - 1 , S 1 } = Pr { X n | S n , X 1 n - 1 , S 1 } , because S n is deterministic in S n - 1 , X n - 1 . Therefore, owing to the Markov property,

Pr ( X n | X 1 n - 1 , S 1 ) = Pr ( X n | S n ) .

We can now compute the entropy,

H ( X n | X 1 n - 1 , S 1 = s 1 ) = - ( X 1 n , S n ) Pr ( X 1 n , S n | S 1 ) log ( Pr ( X n | X 1 n - 1 , S 1 ) ) = - ( X 1 n , S n ) Pr ( X 1 n - 1 , S n | S 1 ) X n Pr ( X n | S n ) log ( Pr ( X n | S n ) ) = S Pr ( S n = s | S 1 ) · H ( X | S n = s ) .

Having proved the lemma, we can move forward with the computation of entropy rate. Note that

1 n H ( X 1 , . . . , X n | S 1 ) = 1 n l = 1 n H ( X l | X 1 l - 1 , S 1 ) = 1 n l = 1 n s S Pr ( S l = s ) · H ( X | s ) = s S 1 n l = 1 n Pr ( S l = s ) H ( X | S = s ) ,

and

1 n l = 1 n Pr ( S l = s ) q ( s ) .

Therefore,

lim n 1 n H ( X 1 n | S 1 ) = s S q ( s ) H ( X | s ) .

It is not complicated to show that this is the entropy rate of the unifilar source X , provided that it is irreducible and aperiodic.

Tree Sources : Tree sources are very useful in modeling and processing text and data files in general.Consequently, tree sources have been studied in great detail. We will introduce these source models here and later in [link] discuss algorithms for universal lossless compression of these sources.

In a tree source, states are arranged as contexts , where a context consists of the last several symbols generatedby the source. The unique property of a context tree source is that states are arranged as leaves of a tree, and the state can be identified by following labels of the branches of the tree based onprevious symbols, until we reach a leaf.

[link] shows a simple tree source. For this source, a typical input might be

x = 011001010 .

This input sequence corresponds to the following sequence of states,

s = s 0 s 01 s 11 s 0 s 0 s 01 s 0 s 01 s 0 .
Tree source for Example 5
Tree source for  [link] .

As before, for every state (leaf) there exist conditional probabilities for generating the next symbol. That said, and maybe surprisingly, a tree source is not necessarily a unifilar source. On the one hand, states can be observed from the source symbols. On the other hand, when transitioning from a “shallow" leaf to a deep leaf in the tree, the symbol currently being generated does not contain enough information to define the new state. To address this problem, extra states can be added to the tree. On the other hand, adding extra states to a tree source might make the context tree less efficient,because it increases the number of states; see for example the detailed discussion of expansion of tree machines by Martín et al.  [link] .

Hidden Markov models : Consider the following state transition probabilities,

Pr ( X t = a , S t = s | S t - 1 = s ' , S t - 2 , X t - 1 , . . . ) = Pr ( X t = a , S t = s | S t - 1 = s ' ) = Pr ( a , s | s ' ) .

The parameter space can be defined,

θ = { Pr ( a , s | s ' ) } ,

there are M ( M · r - 1 ) parameters in total.

Note that

Pr ( X 1 n , S 1 n | S 0 ) = t = 1 n Pr ( X t , S t | S t - 1 )

and

Pr ( X 1 n | S 0 ) = s 1 n Pr ( X 1 n , S 1 n | S 0 ) = s 1 n t = 1 n Pr ( X t , S t | S t - 1 ) .

It is important to emphasize that the next state is not deterministic. This type of behavior is called a hidden Markov model (HMM).

What are the typical sets for HMM sources? Recall that for an i.i.d. source we saw

Pr ( X 1 n ) = a A Pr ( a ) n x ( a ) ,

this is the main concept underlying the method of types. For a unifilar source,

Pr ( X 1 n ) = t = 1 n Pr ( X t | S t ) = A × S Pr ( a | s ) n X ( a | s ) .

For an HMM, there is no such partition into typical sets.

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Universal algorithms in signal processing and communications. OpenStax CNX. May 16, 2013 Download for free at http://cnx.org/content/col11524/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Universal algorithms in signal processing and communications' conversation and receive update notifications?

Ask