<< Chapter < Page Chapter >> Page >

We have discussed several parametric sources, and will now start developing mathematical tools in order to investigate properties of universal codes that offer universal compression w.r.t.a class of parametric sources.

Preliminaries

Consider a class Λ of parametric models, where the parameter set θ characterizes the distribution for a specific source within this class, { p θ ( · ) , θ Λ } .

Consider the class of memoryless sources over an alphabet α = { 1 , 2 , . . . , r } . Here we have

θ = { p ( 1 ) , p ( 2 ) , . . . , p ( r - 1 ) } .

The goal is to find a fixed to variable length lossless code that is independent of θ , which is unknown, yet achieves

E θ l ( X 1 n ) n n H θ ¯ ( X ) ,

where expectation is taken w.r.t. the distribution implied by θ . We have seen for

p ( x ) = 1 2 p 1 ( x ) + 1 2 p 2 ( x )

that a code that is good for two sources (distributions) p 1 and p 2 exists, modulo the one bit loss [link] . As an expansion beyond this idea, consider

p ( x ) = Λ d w ( θ ) p θ ( X ) ,

where w ( θ ) is a prior.

Let us revisit the memoryless source, choose r = 2 , and define the scalar parameter

θ = Pr ( X i = 1 ) = 1 - Pr ( X i = 0 ) .

Then

p θ ( x ) = θ n X ( 1 ) · ( 1 - θ ) n X ( 0 )

and

p ( x ) = 0 1 d θ · θ n X ( 1 ) · ( 1 - θ ) n X ( 0 ) .

Moreover, it can be shown that

p ( x ) = n X ( 0 ) ! n X ( 1 ) ! ( n + 1 ) ! ,

this result appears in Krichevsky and Trofimov  [link] .

Is the source X implied by the distribution p ( x ) an ergodic source? Consider the event lim n 1 n i = 1 n X i 1 2 . Owing to symmetry, in the limit of large n the probability of this event under p ( x ) must be 1 2 ,

Pr lim n 1 n i = 1 n X i 1 2 = 1 2 .

On the other hand, recall that an ergodic source must allocate probability 0 or 1 to this flavor of event. Therefore, the source implied by p ( x ) is not ergodic.

Recall the definitions of p θ ( x ) and p ( x ) in [link] and [link] , respectively. Based on these definitions, consider the following,

H θ ( X 1 n ) = - X 1 n A n p θ ( X 1 n ) log p θ ( X 1 n ) = H ( X 1 n | Θ = θ ) , H ( X 1 n ) = - X 1 n p ( X 1 n ) log p ( X 1 n ) , H ( X 1 n | Θ ) = Λ d w ( θ ) · H ( X 1 n | Θ = θ ) .

We get the following quantity for mutual information between the random variable Θ and random sequence X 1 N ,

I ( Θ ; X 1 n ) = H ( X 1 n ) - H ( X 1 n | Θ ) .

Note that this quantity represents the gain in bits that the parameter θ creates; more about this quantity will be mentioned later.

Redundancy

We now define the conditional redundancy ,

r n ( l , θ ) = 1 n [ E θ ( l ( X 1 n ) ) - H θ ( X 1 n ) ] ,

this quantifies how far a coding length function l is from the entropy where the parameter θ is known. Note that

l ( X 1 n ) = Λ d w ( θ ) E θ ( l ( X 1 n ) ) H ( X 1 n | θ ) .

Denote by c n the collection of lossless codes for length- n inputs, and define the expected redundancy of a code l C n by

R n - ( w , l ) = Λ d w ( θ ) r n ( l , θ ) , R n - ( w ) = inf l C n R n - ( w , l ) .

The asymptotic expected redundancy follows,

R - ( w ) = lim n R n - ( w ) ,

assuming that the limit exists.

We can also define the minimum redundancy that incorporates the worst prior for parameter,

R n - = sup w W R n - ( w ) ,

while keeping the best code. Similarly,

R - = lim n R n - .

Let us derive R n - ,

R n - = sup w inf l Λ d w ( θ ) 1 n E θ ( l ( X 1 n ) ) - H ( X 1 n | Θ = θ ) = sup w inf l 1 n E p l ( X 1 n ) - H ( X 1 n | Θ ) = sup w 1 n H ( X 1 n ) - H ( X 1 n | Θ ) = sup w 1 n I ( Θ ; X 1 n ) = C n n ,

where C n is the capacity of a channel from the sequence x to the parameter  [link] . That is, we try to estimate the parameter from the noisy channel.

In an analogous manner, we define

R n + = inf l sup θ Λ r n ( l , θ ) = inf l sup θ 1 n E θ log p θ ( x n ) 2 - l ( x n ) = inf Q sup θ 1 n D ( P θ | | Q ) ,

Questions & Answers

material that allows electric current to pass through
Deng Reply
material which don't allow electric current is called
Deng
insulators
Covenant
how to study physic and understand
Ewa Reply
what is conservative force with examples
Moses
what is work
Fredrick Reply
the transfer of energy by a force that causes an object to be displaced; the product of the component of the force in the direction of the displacement and the magnitude of the displacement
AI-Robot
why is it from light to gravity
Esther Reply
difference between model and theory
Esther
Is the ship moving at a constant velocity?
Kamogelo Reply
The full note of modern physics
aluet Reply
introduction to applications of nuclear physics
aluet Reply
the explanation is not in full details
Moses Reply
I need more explanation or all about kinematics
Moses
yes
zephaniah
I need more explanation or all about nuclear physics
aluet
Show that the equal masses particles emarge from collision at right angle by making explicit used of fact that momentum is a vector quantity
Muhammad Reply
yh
Isaac
A wave is described by the function D(x,t)=(1.6cm) sin[(1.2cm^-1(x+6.8cm/st] what are:a.Amplitude b. wavelength c. wave number d. frequency e. period f. velocity of speed.
Majok Reply
what is frontier of physics
Somto Reply
A body is projected upward at an angle 45° 18minutes with the horizontal with an initial speed of 40km per second. In hoe many seconds will the body reach the ground then how far from the point of projection will it strike. At what angle will the horizontal will strike
Gufraan Reply
Suppose hydrogen and oxygen are diffusing through air. A small amount of each is released simultaneously. How much time passes before the hydrogen is 1.00 s ahead of the oxygen? Such differences in arrival times are used as an analytical tool in gas chromatography.
Ezekiel Reply
please explain
Samuel
what's the definition of physics
Mobolaji Reply
what is physics
Nangun Reply
the science concerned with describing the interactions of energy, matter, space, and time; it is especially interested in what fundamental mechanisms underlie every phenomenon
AI-Robot
Got questions? Join the online conversation and get instant answers!
Jobilize.com Reply

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Universal algorithms in signal processing and communications. OpenStax CNX. May 16, 2013 Download for free at http://cnx.org/content/col11524/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Universal algorithms in signal processing and communications' conversation and receive update notifications?

Ask