<< Chapter < Page | Chapter >> Page > |
We derived the minimum mean-squared error estimator in the previous section with no constraint on the form of the estimator. Depending on the problem, thecomputations could be a linear function of the observations (which is always the case in Gaussian problems) ornonlinear. Deriving this estimator is often difficult, which limits its application. We consider here a variation of MMSEestimation by constraining the estimator to be linear while minimizing the mean-squared estimation error. Such linear estimators may not be optimum; the conditional expected value may be nonlinear and it always has the smallest mean-squared error. Despite this occasionalperformance deficit, linear estimators have well-understood properties, they interact will with other signal processingalgorithms because of linearity, and they can always be derived, no matter what the problem.
Let the parameter estimate be expressed as where is a linear operator: where , are scalars. Although all estimators of this form are obviously linear, the term linear estimator denotes that member of this family that minimizes the mean-squarederror.
Because of the transformation's linearity, the theory of linear vector spaces can be fruitfully used to derive the estimator andto specify its properties. One result of that theoretical framework is the well-known Orthogonality Principle (Papoulis, pp. 407-414) The linear estimator is that particular linear transformation that yieldsan estimation error orthogonal to all linear transformations of the data. The orthogonality of the error to all linear transformations is termed the universality constraint. This principle provides us not only with a formal definition of the linear estimator butalso with the mechanism to derive it. To demonstrate this intriguing result, let denote the absract inner product between two vectors and the associated norm.
The estimation error for the minimum mean-squared linear estimator can be calculated to some degree without knowledge ofthe form of the estimator. The mean-squared estimation error is given by
Note that the definition of the minimum mean-squared error
linear estimator makes no explicit
assumptions about the parameter estimation problem beingsolved. This property makes this kind of estimator attractive in
many applications where neither the
Express the first example in vector notation so that the observation vector is written as where the matrix has the form . The expected value of the parameter is zero. The linear estimator has the form , where is a matrix. The orthogonality Principle states that the linear estimator satisfies To use the Orthogonality Principle to derive an equation implicitly specifying the linear estimator, the "for alllinear transformations" phrase must be interpreted. Usually the quantity specifying the linear transformation must beremoved from the constraining inner product by imposing a very stringent but equivalent condition. In this example, thisphrase becomes one about matrices. The elements of the matrix can be such that each element of the observation vector multiplies each elementof the estimation error. Thus, in this problem the Othogonality Principle means that the expected value of thematrix consisting of all pairwise priducts of these elements must be zero. Thus, two terms must equal each other: . The second term equals as the additive noise and the parameter are assumed to be statistically independent quantities. The quantity in the first term is the correlation matrix of the observations, which is given by . Here, is the noise covariance matrix, and is the parameter's variance. The quantity is a matrix with each element equaling 1. The noise vector has independent components; the covariance matrix thusequals . The equation that must satisfy is therefore given by The components of are equal and are given by . Thus, the minimum mean-squared error linear estimator has the form
Note that this result equals the minimum mean-squared error estimate derived earlier under the condition that . Mean-squared error, linear estimators, and Gaussian problems are intimately related to each other. The linearminimum mean-squared error solution to a problem is optimal if the underlying distributions are Gaussian.
Notification Switch
Would you like to follow the 'Estimation theory' conversation and receive update notifications?