Ex. 3.15
Ex. 3.15
Verify expression (3.64), and hence show that the partial least squares directions are a compromise between the ordinary regression coefficient and the principal component directions.
Soln. 3.15
Note that
We are essentially solving
where \(\bb{S}=\bX^T\bX\) is the sample covariance matrix of the \(\bx_j\).
We start with the case \(m=1\), which immediately gives what we call the first canonical covariance variable (see Ex. 3.20) with
Note that \(\hat\alpha_1 \propto \hat\varphi_1\) in Algorithm 3.3 in the text.
The second canonical covariance variable, namely \(\hat\alpha_2\), has to maximize the same objective with additional constraint \(\hat\alpha_2^T\bb{S}\hat\alpha_1=0\). It turns out
To see that, we first verify that
Second, for \(\alpha_2\) satisfying \(\alpha_2^T \bb{S}\hat\alpha_1=0\), that is, \(\alpha_2^T \bb{S}\bX^T\by=0\), we have the objective to maximize
Therefore we see \(\eqref{eq:3-15a}\) holds. Note that \(\hat\alpha_2\propto \hat\varphi_2\) in the Algorithm 3.3 in the text. Continuing this, we are able to derive \(\hat\varphi_m\) for \(m\ge 1\).
Now we are ready to show that partial least squares (PLS) directions are a compromise between the ordinary regression coefficient (OLS) and the principal component directions (PCR). The regressors for OLS, PCR and PLS may be referred to as canonical correlation, canonical variance and canonical covariance variables respectively. A generalized criterion that encompasses all three methods is
where \(r\in [0, 1).\) When \(r =0\) we recover OLS, and when \(r \ra 1\) we get PCR. The case when \(r=1/2\) gives PLS. Note that this generalized regression is referred to as continuum regression. See paper Continuum Regression and Ridge Regression for more details.