In this lecture we will discuss the adjoint and transpose. You probably know all about these matrix operations, and may feel that you are too enlightened to participate in this discussion. That feeling is called “ignorance”. Before enlightenment: take adjoints, transpose matrices. After enlightenment: take adjoints, transpose matrices.
In the beginning (of Math 202A), there was quantization,
sends a finite set
to the free Hilbert space
containing
as an orthonormal basis. It also sends a set function
to the linear transformation
whose action on this basis is
is not the only functor from sets to spaces. As the Good Book says, the adjoint
was with quantization, and the adjoint was quantization. Indeed, matches
on sets,
but acts differently on functions: instead of pushing a point of forward onto a point of
via
pulls a point of
back onto its preimage under
Thus the adjoint of
is a contravariant functor,
It is worth noting that while is a linear transformation which contains the same information as the original set function
, presented in the same way,
has no direct counterpart in
In the classical computational category, you can identify a function
with the ordered list of its fibers, but the fibers of
are not points of its domain, they are subobjects of it. In the quantum computational category you can consider
which is literally a point of the domain of
It is precisely this fact which allowed us to establish the Singular Value Decomposition for quantized set functions with no additional technology.
One more remark about the relationship between and
. A function
can be viewed as a bipartite graph. The transformation
encodes the edges of the bipartite graph
, but not quite in the usual way graph theorists do. A graph theorist would encode adjacency in
as an endomorphism of
namely as the linear operator whose matrix in the ordered orthonormal basis
is
There is a very concise way to capture the relationship between the pushforward and the pullback
using scalar products.
Problem 17.1. Let and
; write
and
Show that
Now we look at Problem 16.1 and ask ourselves the same question we asked for the Singular Value Decomposition: can we untether this from combinatorics? If so, what does the untethered relation mean geometrically? More precisely, given any two finite-dimensional Hilbert spaces and any linear transformation
we seek a linear transformation
which satisfies
Assuming such an exists, we want a geometric interpretation of it.
Problem 17.2. Show that if exists, it is unique.
The proof of the existence of is very straightforward. Let
and
be any two orthonormal bases; these are guaranteed to exist by Gram-Schmidt and we have
and
. However, it is not necessarily true that we can select
such that our pre-selected
maps every point of
to a point of
; indeed, a necessary condition for this to be possible is that
be an isometry. Nevertheless, we can simply define
by declaring its matrix elements to be what we want them to be. More precisely, set
Then, for any latex and
we have
and this coincides with
thanks to our definition of . This is completely true, and defines the adjoint
of a linear transformation
in a manner which is logically unassailable and yet totally unsatisfying for all but the most soulless of algebraists.
A natural geometric approach to reversing and arbitrary linear transformation can be developed using the Singular Value Decomposition, which gives us orthogonal decompositions
positive numbers
and isometric isomorphisms
such that
By abuse of notation, we write
where is the nullity of
and
is the corank of
so that
is the zero transformation. Now let us define a corresponding transformation
by
which is shorthand for declaring that the restriction of to
is
for
and the restriction of
to
is the zero transformation into
Problem 17.3. Prove that Deduce
and
One advantage to the geometric interpretation of as “reversing” the singular value decomposition of
is that it immediately leads you to wonder about “inverting” the singular value decomposition of
More precisely, the the linear transformation in
defined by
is known as the Moore-Penrose pseudoinverse of Indeed,
is defined by inverting the invertible part of
in the following sense.
Problem 17.4. Show that the restriction of to
is an isomorphism onto
with inverse
. Prove that
holds in
and that
holds in
Now let us discuss the transpose of No doubt you are familiar with the transpose of a matrix, so I will make the following remark: the transpose of the matrix of
is the matrix of
with respect to orthonormal bases adapted to the singular value decomposition, i.e. with respect to orthonormal bases of
and
consisting of left and right singular vectors, respectively. This is a singular consequence of the fact that the singular values of
are real. For general orthonormal bases of
and
the corresponding matrices of
and
are related by conjugate transpose, not transpose.
The intrinsic meaning of the matrix transpose is not to be found in the adjoint operation, but in a different operation coming from Riesz duality in Recall that for every finite dimensional Hilbert space
we have an antilinear isomorphism with
, the space of linear functionals on
, given by the Riesz mapping
n particular, this allows us to promote from a vector space to a Hilbert space by defining the scalar product
This scalar product is on is precisely what we used to prove that the Frobenius scalar product on
is basis-independent via an argument inductive in the dimension of
If
is an orthonormal basis, we get a corresponding orthonormal basis
consisting of the linear functionals
We now declare the transpose of to be the morphism
defined by
Problem 17.5. Verify that really is a linear transformation
Taking orthonormal bases
and
show that the matrix of
relative to the dual bases
and
is the transpose of the matrix of
relative to
and
oh man. the definition of the Moore-Penrose pseudoinverse was satisfying. Each time I encountered it, it was just $A^+ = (A^T A)^{-1} A^T$, whatever that means! It makes much more sense as the inverse of the invertible part as per the SVD. Thanks!
I’m happy this was helpful. Now get out there and pseudo-invert some matrices.