<< Chapter < Page | Chapter >> Page > |
In this section, we consider only real-valued wavelet functions that form an orthogonal basis, hence and . We saw in Orthogonal Bases from Multiresolution analysis and wavelets how a given function belonging to could be represented as a wavelet series. Here, we explain how to use a wavelet basis to construct a nonparametric estimator for the regression function in the model
where are equispaced design points and the errors are i.i.d. Gaussian, .
A wavelet estimator can be linear or nonlinear . The linear wavelet estimator proceeds by projecting the data onto a coarse level space. This estimator is of a kernel-type, see "Linear smoothing with wavelets" . Another possibility for estimating is to detect which detail coefficients convey the important information about the function and to put equal to zero all the other coefficients. This yields a nonlinear wavelet estimator as described in "Nonlinear smoothing with wavelets" .
Suppose we are given data coming from the model [link] and an orthogonal wavelet basis generated by . The linear wavelet estimator proceeds by choosing a cutting level and represents an estimation of the projection of onto the space :
with the coarsest level in the decomposition, and where the so-called empirical coefficients are computed as
The cutting level plays the role of a smoothing parameter: a small value of means that many detail coefficients are left out, and this may lead to oversmoothing. On the other hand, if is too large, too many coefficients will be kept, and some artificial bumps will probably remain in the estimation of .
To see that the estimator [link] is of a kernel-type, consider first the projection of onto :
where the (convolution) kernel is given by
Härdle et al. [link] studied the approximation properties of this projection operator. In order to estimate [link] , Antoniadis et al. [link] proposed to take:
Approximating the last integral by , we find back the estimator in [link] .
By orthogonality of the wavelet transform and Parseval's equality, the risk (or integrated mean square error IMSE) of a linear wavelet estimator is equal to the risk of its wavelet coefficients:
where
are called `theoretical' coefficients in the regression context. The term in [link] constitutes the stochastic bias whereas is the deterministic bias. The optimal cutting level is such that these two bias are of the same order. If is Hölder continuous, it is easy to see that the optimal cutting level is . The resulting optimal IMSE is of order . In practice, cross-validation methods are often used to determine the optimal level [link] , [link] .
Notification Switch
Would you like to follow the 'An introduction to wavelet analysis' conversation and receive update notifications?