<< Chapter < Page | Chapter >> Page > |
The base of the logarithm can be any (positive) number. The most common choice is base 2, in which case the measurementof information is called bits . Unless otherwise stated explicitly, all logs in this chapter are assumed to be base 2.
Suppose there are symbols in the alphabet, which are transmitted with probabilities , , and . Then the information conveyed by receiving is 1 bit, since
Similarly, the information conveyed by receiving either or is bits.
Suppose that a length binary sequence is transmitted, with all symbols equally probable. Thus , is the binary representation of the th symbol for , and . The information contained in the receipt of any given symbol is
Consider a standard six-sided die. Identify , , and . How many bits of information are conveyed if a 3 is rolled. Now roll two dice, and suppose thetotal is 12. How many bits of information does this represent?
Consider transmitting a signal with values chosen from the six-level alphabet .
The 8-bit binary ASCII representation of any letter (or any character
of the keyboard) can be found using the M
atlab command
dec2bin(text)
where
text
is any string.
Using ASCII, how much information is contained in the letter “a,”assuming that all the letters are equally probable?
Consider a decimal representation of Calculate the information (number of bits) required to transmit successive digits of , assuming that the digits are independent.Identify , , and . How much information is contained in the first million digits of ?
There is an alternative definition of information (in common usage in the mathematical logic and computerscience communities) which defines information in terms of the complexity of representation,rather than in terms of the reduction in uncertainty. Informally speaking, this alternative defines thecomplexity (or information content) of a message by the length of the shortest computer program thatcan replicate the message. For many kinds of data, such as a sequence of random numbers, the two measuresagree because the shortest program that can represent the sequence is just a listing of thesequence. But in other cases, they can differ dramatically. Consider transmitting the first million digits of thenumber . Shannon's definition gives a large information content (as in [link] ), while the complete sequence can, in principle, be transmittedwith a very short computer program.
All the examples in the previous section presume that there is no relationshipbetween successive symbols. (This was the independence assumption in [link] .) This section shows by example that real messages often have significantcorrelation between symbols, which is a kind of redundancy. Consider the following sentence from Shannon's paper A Mathematical Theory of Communication :
Notification Switch
Would you like to follow the 'Software receiver design' conversation and receive update notifications?