<< Chapter < Page | Chapter >> Page > |
In Example [link] , messages of the same length from the first source give lessinformation than those from the second source. Hence, sources with the same number of symbols but different probabilitiescan have different entropies. The key is to design a system to maximize entropy since this will have the largest throughput, or largestaverage flow of information. But how can this be achieved?
First, consider the simple case in which there are two symbols in the alphabet, with probability , and with probability . (Think of a coin that is weighted so as to give heads with higher probability than tails.)Applying the definition [link] shows that the entropy is
This is plotted as a function of in [link] . For all allowable values of , is positive. As approaches either zero or one, approaches zero, which represent the symmetric cases in which either occurs all the time or occurs all the time, and no information is conveyed. reaches its maximum in the middle, at . For this example, entropy is maximized whenboth symbols are equally likely.
Show that is maximized at by taking the derivative and setting it equal to zero.
The next result shows that an -symbol source cannot have entropy larger than , and that this bound is achieved when all the symbols are equally likely.Mathematically, , which is demonstrated by showing that . First,
since . Gathering terms, this can be rewritten
and changing the base of the logarithm (using , where ), gives
If all symbols are equally likely, , then and . Hence . On the other hand, if the symbols are not equally likely, then the inequality (which holds for ) implies that
Rearranging [link] gives the desired bound on the entropy, that . This says that, all else being equal, it is preferableto choose a code in which each symbol occurs with the same probability.Indeed, Example [link] provides a concrete source for which the equal probability case has higherentropy than the unequal case.
"Redundancy" showed how letters in the text of natural languages do not occur with equal probability. Thus,naively using the letters will not lead to an efficient transmission. Rather, the letters must be carefully translatedinto equally probable symbols in order to increase the entropy. A method for accomplishing this translation is given in "Source Coding" , but "Channel Capacity" examines the limits of attainable performance when transmitting symbols across a noisy (but otherwise perfect) channel.
Notification Switch
Would you like to follow the 'Software receiver design' conversation and receive update notifications?