Math 202A: Review Lecture 5

A key concept is that there is a third fundamentally important subspace associated to every morphism A \in \mathrm{Hom}(V,W). Unlike \mathrm{Ker}(A) \leq V and \mathrm{Im}(A) \leq W, this subspace is geometric rather than algebraic in nature: it is defined as the set

\mathrm{Opt}(A) = \{v \in V \colon \|Av\|=\|A\|\|v\|\}

of vectors in V which saturate the operator norm inequality for A. It is not obvious that this actually is a vector space, and the following result is both crucial and nontrivial.

Theorem 5.1. \mathrm{Opt}(A) is a nonzero subspace of V.

The fact that \mathrm{Opt}(A) contains a unit vector follows from the extreme value theorem. We found a way to use the parallelogram law to show that \mathrm{Opt}(A) is a subspace of V. The essence of the singular value decomposition is understanding the relationship between the three fundamental subspaces

\mathrm{Ker}(A),\ \mathrm{Opt}(A),\ \mathrm{Im}(A),

a geometric endeavor which goes beyond the algebraic relationship between kernel and image given by the rank-nullity theorem. In order to succeed in this question, one needs the following fact.

Orthogonality Lemma. Two nonzero vectors v_1,v_2 \in V are orthogonal if and only if \|v_1\| \leq \|v_1+\lambda v_2\| for all \lambda \in \mathbb{C}.

Proof: Geometrically, the Orthogonality Lemma says that v_1,v_2 are orthogonal if and only if the point on the affine line v_1 + \mathbb{C}v_2 closes to 0_V is v_1. In fact, this characterization remains correct even in the infnite-dimensional setting.

Suppose first that v_1 and v_2 are orthogonal. Then, for any scalar \lambda we have

\|v_1+\lambda v_2\|^2 = \|v_1\|^2 + |\lambda|^2\|v_2\|^2 \geq \|v_1\|^2.

Conversely, suppose that the nonnegative function

f(\lambda) =\|v_1+\lambda v_2\|^2

is minimized by taking \lambda =0, the minimum being f(0)=\|v_1\|^2. Suppose \alpha = \langle v_1,v_2 \rangle \neq 0, and look at

\lambda_0 := -\frac{\alpha}{\|v_2\|^2}.

We then compute

f(\lambda_0) = \|v_1\|^2 -\frac{|\alpha|^2}{\|v_2\|^2} < \|v_1\|^2,

a contraction. \square

Armed with the Orthogonality Lemma, we turn to an examination of the geometric relationships between the three fundamental subspaces of a linear transformation. The first such relationship is the nice fact that maximally stretched vectors are orthogonal to squashed ones.

Theorem 5.2. Assuming A \neq 0_{\mathrm{Hom}(V,W)}, we have

\mathrm{Opt}(A) \leq \mathrm{Ker}(A)^\perp.

Proof: Since \mathrm{Opt}(A) is a nonzero subspace of V, we can choose a nonzero vector v_1 \in \mathrm{Opt}(A). If the kernel of A is trivial, the statement we want to prove is obvious. If not, we can choose a nonzero vector v_2 \in \mathrm{Ker}(A). Now use the Orthogonality Lemma on the affine line v_1+\mathbb{C}v_2. \square

Now let us see how Theorem 5.2 refines what the rank-nullity theorem tells us. We have the orthogonal decompositions

V= \mathrm{Ker}(A)^\perp \oplus \mathrm{Ker}(A) \quad\text{ and }\quad W = \mathrm{Im}(A) \oplus \mathrm{Im}(A)^\perp.

Rank-nullity says that the restriction of A to a morphism A_1 \in \mathrm{Hom}(\mathrm{Ker}(A)^\perp, \mathrm{Im}(A) is an isomorphism. Theorem 5.2 tells us that sitting inside \mathrm{Ker}(A)^\perp we have the space \mathrm{Opt}(A) of maximally stretched vectors. The best case scenario is that \mathrm{Opt}(A) exhausts \mathrm{Ker}(A)^\perp. Then, we learn that the isomorphism A_1 \in \mathrm{Hom}(\mathrm{Ker}(A)^\perp,\mathrm{Im}(A)) is almost an isometric isomorphism: it is \sigma_1 times the isometric isomorphism U_1 \in \mathrm{Hom}(\mathrm{Ker}(A)^\perp,\mathrm{Im}(A)) given by U_1=\frac{1}{\sigma_1}U_1. If we are not so lucky, then we have

\mathrm{Ker}(A)^\perp = \mathrm{Opt}(A) \oplus V_2,

where V_2, the orthogonal complement of \mathrm{Opt}(A) inside \mathrm{Ker}(A)^\perp, is a nontrivial subspace of \mathrm{Ker}(A)^\perp. Now we have to try again: this is the start of an iterative process which terminates after finitely many steps, the output being the Singular Value Decomposition, which is adequately treated in the main notes. If you understand this process, you understand everything about linear algebra that you need to know for Math 202A.

Leave a Reply