<< Chapter < Page Chapter >> Page >

Here's another everyday example. Someone living in Ithaca (New York) would be completely unsurprisedthat the weather forecast called for rain, and such a prediction would convey little real information since it rains frequently.On the other hand, to someone living in Reno (Nevada), a forecast of rain would be very surprising, and would convey that very unusualmeteorological events were at hand. In short, it would convey considerable information. Again, the amount of informationconveyed is inversely proportional to the probabilities of the events.

To transform this informal argument into a mathematical statement, consider a set of N possible events x i , for i = 1 , 2 , ... , N . Each event represents one possible outcome of an experiment, such as the flipping of a coin or the transmissionof a symbol across a communication channel. Let p ( x i ) be the probability that the i th event occurs, and suppose that some event must occur. When flipping the coin, it cannot roll into the corner and stand on its edge; each flip results in either H or T . This means that i = 1 N p ( x i ) = 1 . The goal is to find a function I ( x i ) that represents the amount of information conveyed by each outcome.

The following three qualitative conditions all relate the probabilities of events with the amount of information they convey:

( i ) p ( x i ) = p ( x j ) I ( x i ) = I ( x j ) ; ( i i ) p ( x i ) < p ( x j ) I ( x i ) > I ( x j ) ; ( i i i ) p ( x i ) = 1 I ( x i ) = 0 .

Thus, receipt of the symbol x i should

  1. give the same information as receipt of x j if they are equally likely,
  2. give more information if x i is less likely than x j , and
  3. convey no information if it is known a priori that x i is the only alternative.

What kinds of functions I ( x i ) fulfill these requirements? There are many. For instance, I ( x i ) = 1 p ( x i ) - 1 and I ( x i ) = 1 - p 2 ( x i ) p ( x i ) both fulfill (i)–(iii).

To narrow down the possibilities, consider what happens when a series of experiments are conducted, or equivalently, when a series ofsymbols are transmitted. Intuitively, it seems reasonable that if x i occurs at one trial and x j occurs at the next, then the total information in the twotrials should be the sum of the information conveyed by receipt of x i and the information conveyed by receipt of x j ; that is, I ( x i ) + I ( x j ) . This assumes that the two trials are independent of each other,that the second trial is not influenced by the outcome of first (and vice versa).

Formally, two events are defined to be independent if the probability that both occur is equal to the product of the individual probabilities—that is, if

p ( x i and x j ) = p ( x i ) p ( x j ) ,

where p ( x i and x j ) means that x i occurred in the first trial and x j occurred in the second. This additivity requirement for the amount ofinformation conveyed by the occurrence of independent eventsis formally stated in terms of the information function as

( i v ) I ( x i and x j ) = I ( x i ) + I ( x j )

when the events x i and x j are independent.

Combining the additivity in ( i v ) with the three conditions ( i ) - - ( i i i ) , there is one (and only one) possibility for I ( x i ) :

I ( x i ) = log 1 p ( x i ) = - log ( p ( x i ) ) .

It is easy to see that ( i ) - - ( i i i ) are fulfilled, and ( i v ) follows from the properties of the log (recall that log ( a b ) = log ( a ) + log ( b ) ). Therefore,

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Software receiver design. OpenStax CNX. Aug 13, 2013 Download for free at http://cnx.org/content/col11510/1.3
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Software receiver design' conversation and receive update notifications?

Ask