This module presents a quantification of information by the use of entropy. Entropy, or average self-information, measures the uncertainty of a source and hence provides a measure of the information it could reveal.
Information sources take very different forms. Since the
information is not known to the destination, it is then bestmodeled as a random process, discrete-time or continuous time.
Here are a few examples:
Digital data source (
e.g. , a text)
can be modeled as a discrete-time and discrete valued randomprocess
,
,…,
where
with a particular
,
,…,
and a specific
,
,…,
and
,
,…,
etc.
Video signals can be modeled as a continuous time random
process. The power spectral density is bandlimited toaround 5 MHz (the value depends on the standards used to
raster the frames of image).
Audio signals can be modeled as a continuous-time random
process. It has been demonstrated that the power spectraldensity of speech signals is bandlimited between 300 Hz and
3400 Hz. For example, the speech signal can be modeled as aGaussian process with the
shown power spectral density over a small observation period.
These analog information signals are bandlimited. Therefore, if
sampled faster than the Nyquist rate, they can be reconstructedfrom their sample values.
A speech signal with bandwidth of 3100 Hz can be sampled at
the rate of 6.2 kHz. If the samples are quantized with a 8level quantizer then the speech signal can be represented with
a binary sequence with the rate of
The sampled real values can be quantized to create a discrete-time
discrete-valued random process. Since any bandlimited analoginformation signal can be converted to a sequence of discrete
random variables, we will continue the discussion only for discreterandom variables.
The random variable
takes the value of 0 with probability 0.9 and the value of 1 with
probability 0.1. The statement that
carries more information than the statement that
.
The reason is that
is expected to be 0, therefore, knowing that
is more surprising news!! An intuitive definition of
information measure should be larger when the probability issmall.
The information content in the statement about the temperature
and pollution level on July 15th in Chicago should be the sumof the information that July 15th in Chicago was hot and
highly polluted since pollution and temperature could beindependent.
An analog source is modeled as a continuous-time random
process with power spectral density bandlimited to the bandbetween 0 and 4000 Hz. The signal is sampled at the Nyquist
rate. The sequence of random variables, as a result ofsampling, are assumed to be independent. The samples are
quantized to 5 levels
.
The probability of the samples taking the quantized values are
,
respectively. The entropy of the random variables are
There are 8000 samples per second. Therefore, the source
produces
of information.
The joint entropy of two discrete random variables
(
,
) is defined by
The joint entropy for a random vector
is defined as
Conditional Entropy
The conditional entropy of the random variable
given the random variable
is defined by
It is easy to show that
and
If
,
,…,
are mutually independent it is easy to show that
Entropy Rate
The entropy rate of a stationary discrete-time random process
is defined by
The limit exists and is equal to
The entropy rate is a measure of the uncertainty of
information content per output symbol of the source.
Entropy is closely tied to
source coding . The extent to which a
source can be compressed is related to its entropy. In 1948,Claude E. Shannon introduced a theorem which related the
entropy to the number of bits per second required to representa source without much loss.