This Demonstration considers one of the simplest nonparametric-regression problems: let
be a smooth real-valued function over the interval
; recover (or estimate)
when one only knows
that satisfy the model
are independent, standard normal random variables. Sometimes one assumes that the noise level
is known. Such a curve estimation problem is also called a signal denoising problem.
Perhaps the simplest "solution" to this problem is by the classical and widely used kernel-smoothing method, which is a particular case of the Loess method (locally weighted scatterplot smoothing); see the Details below.
In the kernel-smoothing method (like in any similar nonparametric method, e.g. the smoothing spline method) one or several smoothing parameters have to be appropriately chosen to obtain a "good" estimate of the true curve
. For the kernel method, it is known that choosing a good value for the famous bandwidth parameter (see Details) is crucial, much more than the choice of the class of the kernels, which is fixed here.
The curve estimate corresponding to a given value
for the bandwidth parameter is then denoted
. Notice that
, and the
's only through the data
Three very popular methods are available for choosing
: cross-validation (also called the "leave-one-out" principle; see PRESS statistic
), generalized cross validation (GCV), and Mallows’
(also sometimes denoted by
or UBR for unbiased risk estimate). In fact, cross-validation and GCV coincide in our context, where periodic end conditions are assumed. See  for a review; and see  for the definition and an analysis of
. Notice that GCV (in contrast to the
method) does not require that
This Demonstration provides interactive assessments of the statistical efficiency of GCV and
smoothing-parameter selectors; and this can be done for rather large
could be taken as large as
with reasonably fast interactivity on a current personal computer).
Here six examples for
(the "true curve") can be tried:
is plotted (in blue) in the third of the three possible views using the tabs:
1. The data and the curve-estimate
, where you choose
by the trial-bandwidth slider, show that the choice of the bandwidth is crucial.
2. The two-curve estimates given by GCV and
are very often so similar that they can not be distinguished.
3. The third tab displays quantities related to the "truth": the ASE-optimal choice yields the green curve, this is the
-value that minimizes a global discrepancy between
, defined by
where ASE stands for average of the squared errors. (Note that this ASE-optimal choice is a target that seems to be unattainable in practice, since we do not know the true curve
In the third view, the curve-estimate associated with the automatic
choice is again plotted (purple curve): we can then assess the typical efficiency of
(or the quite close GCV choice) by comparing this curve estimate with the targeted ASE-optimal curve.
This third view also displays the two associated ASE values, in addition to the one associated with the bandwidth chosen by-eye. This shows that it is difficult to choose by eye a trial-bandwidth in the first view (thus without the help of the true curve plot) that will turn to be as least as good as the GCV or
choice, when "better" still means a lower ASE-value.
Staying in the third view and only increasing the size of the data set, one typically observes that the
curve estimate and the ASE-optimal curve generally become very close, and the two associated ASE distances often become relatively similar. This agrees with the known asymptotic theory, see  and the references therein.
The rate at which the relative difference between the
(or GCV) bandwidth and the optimal bandwidth converges to zero can be assessed in the additional panel, top right, in the third view, which shows the two kernels (precisely only the right-hand part of each of the two kernels, since they are even functions) associated with the two bandwidths: the differences between the two bandwidths can be scrutinized with a much better precision than by looking only at the two associated curves.
By varying the seed that generates the data, you can also observe that, in certain cases, there sometimes remains non-negligible differences between the
(or GCV) choice and the optimal bandwidth.