Ex. 13.6
Ex. 13.6
Here we consider the problem of shape averaging. In particular, \(\bb{L}_i, i=1,...,M\) are each \(N\times 2\) matrices of points in \(\mathbb{R}^2\), each sampled from corresponding positions of handwritten (cursive) letters. We seek an affine invariant average \(\bb{V}\), also \(N\times 2\), \(\bb{V}^T\bb{V}=I\), of the \(M\) letters \(\bb{L}_i\) with the following property: \(\bb{V}\) minimizes
Characterize the solution.
This solution can suffer if some of the letters are big and dominate the average. An alternative approach is to minimize instead:
Derive the solution to this problem. How do the criteria differ? Use the SVD of the \(\bb{L}_j\) to simplify the comparison of the two approaches.
Soln. 13.6
This exercise is similar to Ex. 14.10. Note the norm \(\|\cdot\|\) here is understood as Frobenius norm \(\|M\|^2 = \text{trace}[M^TM]\) for matrix \(M.\)
First consider
with \(\bb{V}\) fixed. By definition, we have
By properties of trace operator (\cite{matbook}), we have
so that we know the optimal \(\hat{\bb{A}}_j = \bb{V}^T\bb{L}_j\). Plug it into the original problem, we then need to minimize
subject to \(\bb{V}^T\bb{V}=\bb{I}\).
Note that
the problem reduces to
Consider the Lagrangian function
We know that
where
are both symmetric.
Therefore we know that the optimal \(\hat{\bb{V}}\) satisfies
which is an invariant subspace equation. The equation allow \(\hat{\bb{V}}\) to be an arbitrary orthogonal basis for the rank-2 subspace. Therefore, \(\hat{\bb{V}}\) is the \(N\times 2\) matrix formed from the 2 largest eigenvectors of \(\bb{L}=\frac{1}{M}\sum_{j=1}^M\bb{L}_j\bb{L}_j^T\).
Now we turn to the second part of the exercise where the objective becomes
Following the same logic above, we are able to show that the optimal \(\hat{\bb{V}}\) is the \(N\times 2\) matrix formed from the 2 largest eigenvectors of \(\bar{\bb{L}} = \frac{1}{M}\sum_{j=1}^M\bb{L}_j(\bb{L}_j^T\bb{L}_j)^{-1}\bb{L}_j^T\). For a detailed proof, see Ex. 14.10.
The second criteria is a normalized version of the first one.