<< Chapter < Page | Chapter >> Page > |
where the infimum is taken over all measurable estimators of Similarly, we define the linear minimax risk as
where the infimum is now taken over all linear estimators Obviously, We first state some definitions.
In order to fix the idea, we consider only the risk in the remaining part of this section, thus .
In [link] , [link] , the authors found that the optimal rate of convergence attainable by an estimator when the underlying function belongs to the Sobolev class is , hence . We saw in "Linear smoothing with wavelets" that linear wavelet estimators attain the optimal rate for Hölder function in case of the risk (also called `IMSE'). For a Sobolev class , the same result holds provided that . More precisely, we have the two following situations.
The second result is due to the spatial variability of functions in Sobolev spaces with small index . Linear estimators are based on the idea of spatial homogeneity of the function and hence do perform poorly in the presence of non-homogeneous functions. In contrast, even if , the SureShrink estimator attains the minimax rate [link] . The same type of results holds for more general Besov spaces, see for example [link] , Chapter 10.
We just saw that a nonlinear wavelet estimator is able to estimate in an optimal way functions ofinhomogeneous regularity. However, it may not be sufficient to know that for belonging to a given space, the estimator performs well. Indeed, in general we do not know which space the function belongs to. Hence it is ofgreat interest to consider a scale of function classes and to look for an estimator that attains simultaneously the best rates of convergence across the whole scale. For example, the Sobolev scale is a set of Sobolev function classes indexed by the parameters and , see [link] for the definition of a Sobolev class. We now formalize the notion of an adaptive estimator.
Let be a given set and let be the scale of functional classes indexed by Denote the minimax risk over for the loss:
The estimator is called adaptive up to a logarithmic factor for the loss and the scale of classes if for any there exist and such that
Thus, adaptive estimators have an optimal rate of convergence and behave as if they know in advance in which class the function to be estimated lies.
The VisuShrink procedure is adaptive up to a logarithmic factor for the loss over every Besov, Hölder and Sobolev class that is contained in , see Theorem 1.2 in [link] . The SureShrink estimator does better: it is adaptive for the loss, for a large scale of Besov, Hölder and Sobolev classes, see Theorem 1 in [link] .
In this chapter, we saw the basic properties of standard wavelet theory and explained how these are related to the construction of wavelet regression estimators.
Notification Switch
Would you like to follow the 'An introduction to wavelet analysis' conversation and receive update notifications?