Introduces the Karhuenen-Loeve transform, with applications.
Define the random vector
with mean zero and covariance matrix
; this matrix is symmetric and positive semidefinite.
Lemma 1 Every eigenvalue of
is real and non-negative.
Let e be an eigenvector of
with eigenvalue
.
The last statement falls out by the definite of positive semi-definite. We have
. Since
, it follows that
, i.e. all the eigenvalues are non-negative.
The eigenvectors of the matrix
provide an orthonormal basis
, which can be collected into an orthonormal basis matrix
. Then let
. We have:
Let us look at the adjoint of
:
If we take the adjoint again, we get
Going back to our derivation of
:
The matrix
is known as the KLT matrix defined by
. The transformation given by the KLT matrix provides a set of random variables
that are uncorrelated.
Example 1 (Whitening Filter) For a random vector
,
has positive eigenvalues. Let us write
and
where
. We have
The matrix
is known as a "whitening filter", as it maps an arbitrary random vector
to a “white Gaussian noise” vector
.
Example 2 (Transform Coding) Let
is a unitary operator. Assume we have a signal
that we want to send it through a channel by only sending
numbers or "items", where
; in words, we wish to compress the signal
. The block diagram for the compression/transmission system is given in
[link] .
We want to minimize
given
by choosing the optimal transformation
. We know
which implies
since U is unitary. Therefore,
This means that we can minimize
in place of
. For simplicity, we choose a basic means of compression that preserves only the first
entries of
:
We then have
. Therefore,
It turns out that the choice of transform basis
that minimizes this amount is provided by the eigendecomposition of
, as specified by the following theorem.
Theorem 1 Let X be a length-n random vector with covariance matrix
that has eigenvalues
and matching eigenvectors
. Let
be the orthogonal projection of x onto a subspace
of dimension
. Then
we have that
. Plugging this into
[link] , we have
Now, denote
, and see that
as all
are unit-norm. Now, we have that
Since the
are monotonically decreasing, we have that
If we set
, (i.e.,
) then it is easy to check that
proving the theorem.
Example 3 (Transform Coding) Transform coding is a common scheme for data compression that leverages the Karhuenen-Loève transform. Examples include JPEG and MP3. In particular, JPEG can be broadly described as follows:
Take the image
and create tiles of size
. We assume that the tiles are draws from a random variable
, i.e., the tiles
with
Compute the KLT of the tile random variable
from
by obtaining its eigendecomposition
.
Compute KLT coefficients for each block as
.
Pick as many coefficients of
as allowed by communications or storage constraints; save them as the compressed image.
Load saved coefficients and append zeros to build coefficient vector
.
Run inverse KLT to obtain the decompressed tiles
.
Reassemble the image from the decompressed tiles.
In practice, it is not desirable to recompute the KLT for each individual image. Thus, the JPEG algorithm employs the discrete cosine transform (DCT). It turns out that the DCT is a good approximation of the KLT for tiles of natural images. Additionally, instead of selecting a subset of the coefficients, they are quantized to varying quality/error according to their index and the total amount of bits available.