The theoretical foundations for the thin-plate smoothing spline are described in Duchon (1976, 1977) and Meinguet (1979). Further results and applications are given in: Wahba and Wendelberger (1980); Hutchinson and Bischof (1983); Seaman and Hutchinson (1985).
Suppose that is a space of functions whose partial derivatives of total order m are in , where is the domain of .
Now, consider the data model
where .
Using the notation from the section Penalized Least Squares Estimation, for a fixed , estimate f by minimizing the penalized least squares function
is the penalty term to enforce smoothness on f. There are several ways to define . For the thin-plate smoothing spline, with of dimension d, define as
where . Under this definition, gives zero penalty to some functions. The space that is spanned by the set of polynomials that contribute zero penalty is called the polynomial space. The dimension of the polynomial space M is a function of dimension d and order m of the smoothing penalty, .
Given the condition that , the function that minimizes the penalized least squares criterion has the form
where and are vectors of coefficients to be estimated. The M functions are linearly independent polynomials that span the space of functions for which is zero. The basis functions are defined as
When d = 2 and m = 2, then , , , and . is as follows:
For the sake of simplicity, the formulas and equations that follow assume m = 2. See Wahba (1990) and Bates et al. (1987) for more details.
Duchon (1976) showed that can be represented as
where for d = 2. For derivations of for other values of d, see Villalobos and Wahba (1987).
If you define with elements and with elements , the goal is to find vectors of coefficients and that minimize
A unique solution is guaranteed if the matrix is of full rank and .
If and , the expression for becomes
The coefficients and can be obtained by solving
To compute and , let the QR decomposition of be
where is an orthogonal matrix and is an upper triangular, with (Dongarra et al. 1979).
Since , must be in the column space of . Therefore, can be expressed as for a vector . Substituting into the preceding equation and multiplying through by gives
or
The coefficient can be obtained by solving
The influence matrix is defined as
and has the form
Similar to the regression case, if you consider the trace of as the degrees of freedom for the model and the trace of as the degrees of freedom for the error, the estimate can be represented as
where is the residual sum of squares. Theoretical properties of these estimates have not yet been published. However, good numerical results in simulation studies have been described by several authors. For more information, see O’Sullivan and Wong (1987); Nychka (1986a, 1986b, 1988); Hall and Titterington (1987).
Viewing the spline model as a Bayesian model, Wahba (1983) proposed Bayesian confidence intervals for smoothing spline estimates as
where is the ith diagonal element of the matrix and is the quantile of the standard normal distribution. The confidence intervals are interpreted as intervals "across the function" as opposed to pointwise intervals.
For SCORE data sets, the hat matrix is not available. To compute the Bayesian confidence interval for a new point , let
and let be an vector with ith entry
When d = 2 and m = 2, is computed with
is a vector of evaluations of by the polynomials that span the functional space where is zero. The details for , , and are discussed in the previous section. Wahba (1983) showed that the Bayesian posterior variance of satisfies
where
Suppose that you fit a spline estimate that consists of a true function f and a random error term to experimental data. In repeated experiments, it is likely that about of the confidence intervals cover the corresponding true values, although some values are covered every time and other values are not covered by the confidence intervals most of the time. This effect is more pronounced when the true surface or surface has small regions of particularly rapid change.
The quantity is called the smoothing parameter, which controls the balance between the goodness of fit and the smoothness of the final estimate.
A large heavily penalizes the mth derivative of the function, thus forcing close to 0. A small places less of a penalty on rapid change in , resulting in an estimate that tends to interpolate the data points.
The smoothing parameter greatly affects the analysis, and it should be selected with care. One method is to perform several analyses with different values for and compare the resulting final estimates.
A more objective way to select the smoothing parameter is to use the "leave-out-one" cross validation function, which is an approximation of the predicted mean squares error. A generalized version of the leave-out-one cross validation function is proposed by Wahba (1990) and is easy to calculate. This generalized cross validation (GCV) function is defined as
The justification for using the GCV function to select relies on asymptotic theory. Thus, you cannot expect good results for very small sample sizes or when there is not enough information in the data to separate the model from the error component. Simulation studies suggest that for independent and identically distributed Gaussian noise, you can obtain reliable estimates of for n greater than 25 or 30. Note that, even for large values of n (say, ), in extreme Monte Carlo simulations there might be a small percentage of unwarranted extreme estimates in which or (Wahba 1983). Generally, if is known to within an order of magnitude, the occasional extreme case can be readily identified. As n gets larger, the effect becomes weaker.
The GCV function is fairly robust against nonhomogeneity of variances and non-Gaussian errors (Villalobos and Wahba 1987). Andrews (1988) has provided favorable theoretical results when variances are unequal. However, this selection method is likely to give unsatisfactory results when the errors are highly correlated.
The GCV value might be suspect when is extremely small because computed values might become indistinguishable from zero. In practice, calculations with or near 0 can cause numerical instabilities that result in an unsatisfactory solution. Simulation studies have shown that a with is small enough that the final estimate based on this almost interpolates the data points. A GCV value based on a might not be accurate.