<< Chapter < Page Chapter >> Page >
The module will explain Autocorrelation and its function and properties. Also, examples will be provided to help you step through some of the more complicated statistical analysis.

Before diving into a more complex statistical analysis of random signals and processes , let us quickly review the idea of correlation . Recall that the correlation of two signals or variables is the expectedvalue of the product of those two variables. Since our focus will be to discover more about a random process, a collection ofrandom signals, then imagine us dealing with two samples of a random process, where each sample is taken at a different pointin time. Also recall that the key property of these random processes is that they are now functions of time; imagine themas a collection of signals. The expected value of the product of these two variables (or samples) will now depend on how quickly theychange in regards to time . For example, if the two variables are taken from almost the same time period,then we should expect them to have a high correlation. We will now look at a correlation function that relates a pair of randomvariables from the same process to the time separations between them, where the argument to this correlation function will bethe time difference. For the correlation of signals from two different random process, look at the crosscorrelation function .

Autocorrelation function

The first of these correlation functions we will discuss is the autocorrelation , where each of the random variables we will deal with come from the same random process.

Autocorrelation
the expected value of the product of a random variable or signal realization with a time-shifted version of itself
With a simple calculation and analysis of the autocorrelation function, we can discover a few important characteristicsabout our random process. These include:
  • How quickly our random signal or processes changes with respect to the time function
  • Whether our process has a periodic component and what the expected frequency might be
As was mentioned above, the autocorrelation function is simply the expected value of a product. Assume we have a pair ofrandom variables from the same process, X 1 X t 1 and X 2 X t 2 , then the autocorrelation is often written as
R xx t 1 t 2 X 1 X 2 x 1 x 2 x 1 x 2 f x 1 x 2
The above equation is valid for stationary and nonstationary random processes. For stationary processes , we can generalize this expression a little further. Given a wide-sensestationary processes, it can be proven that the expected values from our random process will be independent of theorigin of our time function. Therefore, we can say that our autocorrelation function will depend on the time differenceand not some absolute time. For this discussion, we will let τ t 2 t 1 , and thus we generalize our autocorrelation expression as
R xx t t τ R xx τ X t X t τ
for the continuous-time case. In most DSP course we will be more interested in dealing with real signal sequences, and thuswe will want to look at the discrete-time case of the autocorrelation function. The formula below will prove to bemore common and useful than :
R xx n n m n x n x n m
And again we can generalize the notation for ourautocorrelation function as
R xx n n m R xx m X n X n m

Properties of autocorrelation

Below we will look at several properties of the autocorrelation function that hold for stationary random processes.

  • Autocorrelation is an even function for τ R xx τ R xx τ
  • The mean-square value can be found by evaluating the autocorrelation where τ 0 , which gives us R xx 0 X 2
  • The autocorrelation function will have its largest value when τ 0 . This value can appear again, for example in a periodic function at the values of the equivalentperiodic points, but will never be exceeded. R xx 0 R xx τ
  • If we take the autocorrelation of a period function, then R xx τ will also be periodic with the same frequency.

Estimating the autocorrleation with time-averaging

Sometimes the whole random process is not available to us. In these cases, we would still like to be able to find outsome of the characteristics of the stationary random process, even if we just have part of one sample function.In order to do this we can estimate the autocorrelation from a given interval, 0 to T seconds, of the sample function.

Ř xx τ 1 T τ t T τ 0 x t x t τ
However, a lot of times we will not have sufficientinformation to build a complete continuous-time function of one of our random signals for the above analysis. If thisis the case, we can treat the information we do know about the function as a discrete signal and use the discrete-timeformula for estimating the autocorrelation.
Ř xx m 1 N m n N m 1 0 x n x n m

Examples

Below we will look at a variety of examples that use the autocorrelation function. We will begin with a simple exampledealing with Gaussian White Noise (GWN) and a few basic statistical properties that will prove very useful in theseand future calculations.

We will let x n represent our GWN. For this problem, it is important to remember the following fact about the mean of aGWN function: x n 0

Gaussian density function. By examination, can easily see that the above statement is true - the mean equals zero.

Along with being zero-mean , recall that GWN is always independent . With these two facts, we are now ready to do the short calculationsrequired to find the autocorrelation. R xx n n m x n x n m Since the function, x n , is independent, then we can take the product of the individual expected values of both functions. R xx n n m x n x n m Now, looking at the above equation we see that we can break it up further into two conditions: one when m and n are equal and one when they are not equal. When they are equal we can combine theexpected values. We are left with the following piecewise function to solve: R xx n n m x n x n m m 0 x n 2 m 0 We can now solve the two parts of the above equation. Thefirst equation is easy to solve as we have already stated that the expected value of x n will be zero. For the second part, you should recall from statistics that the expected value of the squareof a function is equal to the variance. Thus we get the following results for the autocorrelation: R xx n n m 0 m 0 σ 2 m 0 Or in a more concise way, we can represent the results as R xx n n m σ 2 δ m

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Ece 454 and ece 554 supplemental reading. OpenStax CNX. Apr 02, 2012 Download for free at http://cnx.org/content/col11416/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Ece 454 and ece 554 supplemental reading' conversation and receive update notifications?

Ask