Ex. 5.7
Ex. 5.7
Derivation of smoothing splines (Nonparametric Regression and Generalized Linear Models). Suppose that \(N\ge 2\), and that \(g\) is the natural cubic spline interpolant to the pairs \({x_i, z_i}_1^N\), with \(a<x_1<\dots<x_N<b\). This is a natural spline with a knot at every \(x_i\); being an \(N\)-dimensional space of functions, we can determine the coefficients such that it interpolates the sequence \(z_i\) exactly. Let \(\tilde g\) be any other differentiable function on \([a,b]\) that interpolates the \(N\) pairs.
(a)
Let \(h(x) = \tilde g(x) - g(x)\). Use integration by parts and the fact that \(g\) is a natural cubic spline to show that
(b)
Hence show that
and the equality can only hold if \(h\) is identically zero in \([a,b]\).
(c)
Consider the penalized least squares problem
Use (b) to argue that the minimizer must be a cubic spline with knots at each of the \(x_i\).
Soln. 5.7
(a)
Using integration by parts, we obtain
where the second equation and the last equation follow from the fact that \(g\) is a natural cubic spline thus linear on \([a, x_1]\) and \([x_N, b]\). Again, since \(g\) is a natural cubic spline, so that \(g'''\) is a constant on each interval \((x_j, x_{j+1})\) for \(j=1,...,N-1\). By definition of integration we obtain the item above is equal to
Since both \(\tilde g\) and \(g\) interpolate the \(N\) pairs \(\{x_i, z_i\}_1^N\), we know \(h(x_j)=0\) for \(j=1,...,N\). Thus the proof is complete.
(b)
From (a), we have
so that
Then, we start with \(\int_a^b h''(t)^2dt\ge 0\).
Thus we have
and the equality can only hole if \(h''\) is zero in \([a,b]\). Thus \(h\) is linear in \([a,b]\). However by definition \(h(x_j)=0\) for all \(j=1, ..., N\), we conclude that \(h\) must be identically zero in \([a,b]\).
(c)
Consider a minimizer \(f_0\). We can always construct a natural cubic spline \(f\) such that \(f_0\) and \(f\) have the same values at each of \(\{x_i, i=1,...,N\}\), therefor we have \(\sum_{i=1}^N(y_i-f_0(x_i))^2 = \sum_{i=1}^N(y_i-f(x_i))^2\). From (b) we know that
Since \(f_0\) is a minimize, the above inequality is actually a equality, therefore by (b) again, \(f_0=f\) in \([a, b]\). The proof is complete.