Orthogonal Projections Onto Higher-Dimensional Sub-Spaces
Orthogonal Projections Onto Higher-Dimensional Sub-Spaces
The
projection πU (x) of x ∈ R3 onto U can be expressed as a linear combination of
b1 , b2 and the displacement vector x − πU (x) is orthogonal to both b1 and b2 .
x
x − πU (x)
U
b2
πU (x)
0 b1
hb1 , x − pi = b>
1 (x − p) = 0 (3)
1
..
. (4)
hbm , x − pi = b>
m (x − p) = 0 (5)
b>
1 (x − Bλ) = 0 (6)
..
. (7)
b>
m (x − Bλ) = 0 (8)
The last expression is called normal equation. Since b1 , . . . , bm are a ba- normal equa-
sis of U and, therefore, linearly independent, B > B ∈ Rm×m is regular tion
and can be inverted. This allows us to solve for the optimal coefficients/
coordinates
The matrix (B > B)−1 B > is also called the pseudo-inverse of B, which can pseudo-
be computed for non-square matrices B. It only requires that B > B is inverse
positive definite, which is the case if B is full rank.1
2. Find the projection πU (x) = p ∈ U . We already established that p = Bλ.
Therefore, with (11)
3. Find the projection matrix P π . From (12) we can immediately see that
the projection matrix that solves P π x = p must be
to guarantee increase numerical stability and positive definiteness. This “ridge” can be rigorously
derived using Bayesian inference.
2
Remark. The projections πU (x) are still vectors in Rn although they lie in an
m-dimensional subspace U ⊆ Rn . However, to represent a projected vector
we only need the m coordinates λ1 , . . . , λm with respect to the basis vectors
b1 , . . . , bm of U .
Remark. In vector spaces with general inner products, we have to pay attention
when computing angles and distances, which are defined by means of the inner
product.
We can find
Projections allow us to look at situations where we have a linear system approximate
Ax = b without a solution. Recall that this means that b does not lie in the solutions to
span of A, i.e., the vector b does not lie in the subspace spanned by the columns unsolvable lin-
of A. Given that the linear equation cannot be solved exactly, we can find an ear equation
approximate solution. The idea is to find the vector in the subspace spanned by systems using
the columns of A that is closest to b, i.e., we compute the orthogonal projection projections.
of b onto the subspace spanned by the columns of A. This problem arises often
in practice, and the solution is called the least squares solution (assuming the least squares
dot product as the inner product) of an overdetermined system. solution