Throughout this lecture, is a finite-dimensional complex vector space equipped with a scalar product
, which is by convention a linear function of its second argument. The scalar product on
determines various symmetry classes in
the algebra of linear operators
and the study of these symmetry classes is one of the main components of Math 202A. In particular, recall that an operator
is said to be selfajdoint (or Hermitian) if
Hermitian operators are the subject of the Spectral Theorem, which may be formulated as follows.
Theorem 1: If is selfadjoint, then there exists an orthonormal basis
of
such that
, where
are real numbers.
Theorem 1 is the eigenvalue/eigenvector form of the Spectral Theorem. We will give a proof of this result which iteratively constructs the eigenvalues/eigenvectors of as the maxima/maximizers of the (necessarily real-valued) quadratic form
on a nested sequence of compact sets in
. Instead of invoking the Fundamental Theorem of Algebra to claim the existence of a root of the characteristic polynomial, this argument calls the Extreme Value Theorem as a subroutine.
Lemma 1: If is a selfadjoint operator such that the affiliated quadratic form
is nonnegative, then any zero of
is in the kernel of
.
Proof: In fact, the argument does not use the finite-dimensionality of . Let
be such that
Consider a real perturbation of this zero in an arbitrary direction, i.e. consider
, where
is arbitrary and
is a fixed unit vector. We then have
The first term is zero by hypothesis. The middle terms are the real number times
Thus by hypothesis the function
is a nonnegative function of the real variable . This function has the asymptotics
as , so it must be the case that
, otherwise
would change sign at
, which it does not, being a nonnegative function. Repeating the above with a pure imaginary perturbation
, we get that
as well. Thus
is orthogonal to every unit vector
, whence it must be the zero vector,
— Q.E.D.
Proof of Theorem 1: Since is finite-dimensional, its unit sphere
is compact, and the function is continuous. Thus, by the Extreme Value Theorem, there exists a unit vector
at which
attains its maximum value on
, i.e. we have
for all unit vectors . But this is equivalent to
for all unit vectors , and in this form it is clear that the inequality actually holds for arbitrary vectors
, i.e. we have
for all where
Note that is selfadjoint (indeed, selfadjoint operators in
form a real vector space), and the above says that the associated quadratic form
is nonnegative. Since
by construction, we may invoke Lemma 1 to conclude that
is in the kernel of
, i.e.
is an eigenvector of
with eigenvalue
.
Consider now the orthogonal complement
i.e. the -dimensional hyperplane in
perpendicular to the line through
. Let
be the unit sphere in
, let
be the maximum of the quadratic form
on
, and let
be a point at which the maximum is achieved. By exactly the same argument as above, we get that
is an eigenvector of
with eigenvalue
. Indeed, we can iterate the argument
times to obtain an orthonormal set
of eigenvectors with eigenvalues
.
— Q.E.D.
2. The subspace lattice
Let denote the set of all subspaces of
, including the trivial subspace
and
itself. This set carries a natural partial order: given
we declare
if and only if
is a subspace of
. In fact,
is a graded poset in which the rank of a subspace
is its dimension
. The graded poset
is moreover a lattice: the largest subspace contained in each of two given subspaces
and
is their intersection, and the smallest subspaces containing both of them is their span. Finally, the lattice
is self-dual: the map
sending a subspace to its orthogonal complement is an order-reversing involution.
In formulating the above, we have just encountered our first example of a functor: we can view as a covariant functor from the category of vector spaces to the category of posets. Indeed, if
are vector spaces and
is a linear transformation, we get an order-preserving function
defined by
In this course we will be considering quite a few more functors, most of which will be endofunctors from the category of vector spaces to itself. The most basic of these are tensor powers, symmetric powers, and exterior powers, which are the basic operations of multilinear algebra and are closely related to the physical concept of Fock space, which is intended to capture the notion of symmetry classes of elementary particles in quantum mechanics — bosons and fermions and such.
3. Flags and Grassmannians
Very loosely speaking, vector spaces are continuous versions of sets. A finite set might represent the possible states of some physical system, while the free vector space
over this set allows for superpositions of states, i.e. linear combinations of them. Associated to
is its lattice of subsets
, where meet and join are intersection and union. This is called a Boolean lattice, and it is graded by cardinality rather than dimension. Sometimes
is referred to as a hypercube, because its elements can be thought of as length
bitstrings, and these may in turn be thought of as the vertices of a polytope in
which is a cube when
.
In a general poset, a chain is a totally ordered subset, while an antichain is a subset which does not contain any comparable pairs of elements. A chain or antichain is said to be saturated if adding any new element will cause it to no longer be a chain or antichain. It is easy to spot saturated chains in the Boolean lattice — they are just sequences
of subsets of
such that
, and there are factorially many of these. It is equally easy to spot
saturated antichains, namely the collections
consisting of all subsets of
of cardinality
. The cardinality of
itself is then
, and it is a famous result in extremal combinatorics that the size of the largest antichain in the Boolean lattice
is the largest number in the
th row of Pascal’s triangle.
The lattice of subspaces of a finite-dimensional vector space is in many ways analogous to the lattice of subsets of a finite set. Indeed, if one could define a field with one element in a satisfactory way, then the two notions would coincide for vector spaces over . This is problematic, but one can analyze the subspace lattice of a vector space defined over
without difficulty, and count chains, antichains, etc. For example, the antichain consisting of
-dimensional subspaces is the Gaussian binomial coefficient
, which degenerates to the ordinary binomial coefficient in the
limit, almost as if
really did exist…
Anyway, we are working over , so everything is infinite. Chains in
are typically referred to as flags, though this term is more often than not reserved for chains that contain the trivial subspace and the
itself. Saturated chains are called complete flags: they are sequences
such that The analogue of the antichain
in the Boolean lattice consisting of subsets of cardinality
is the collection of
-dimensional subspaces of
; these antichains are called Grassmannians and denoted
. We are going to discuss Grassmannians of finite-dimensional complex vector spaces in some detail, and I hope that we will have time to discuss a certain special infinite-dimensional Grassmannian known as the Sato Grassmannian. For now, we concentrate on the following fundamental relationship between eigenvalues of selfadjoint operators and Grassmannians.
3. Eigenvalues as extremizers
Let be a selfadjoint operator, and define a function
by
Thus, assigns to each subspace
the maximum of the quadratic form associated to
on the unit sphere in
. In particular, we know from our proof of the Spectral Theorem above that the largest eigenvalue of
is given by
The following result, known as the Courant-Fisher minimax theorem, generalizes this by giving a variational characterization of each individual eigenvalue of
.
Theorem 2: If is a selfadjoint operator with eigenvalues
then for each
we have
That is, the th largest eigenvalue of
is the minimum value of
over the Grassmannian of
-dimensional subspaces of
.
Proof: Let be an orthonormal basis of
such that
,
For each
, consider the space
spanned by the eigenvectors corresponding to the largest eigenvalues of
and also set
. In the proof of Theorem 1, we showed that
so that indeed the th largest eigenvalue
is given by evaluating the function
at a point of the Grassmannian
, namely
It remains to show that the evaluation of
on every other subspace of dimension
is at least
.
To achieve this end, we establish the following squeeze inequality for the values of the quadratic form on the subspace spanned by the first
eigenvectors: for all
we have
Indeed, let be any vector in
. Using orthonormality of the eigenvectors we calculate
and the lower bound is established in exactly the same way.
Now let be arbitrary. Since
,
the intersection of the subspaces and
is at least one-dimensional (see Problem set 1). Let
be a unit vector in
. Since
, we have
by the above inequality. But also
, which shows that
, as required.
— Q.E.D.
4. Adjoint orbits
Any operator which preserves the scalar product,
is called a unitary operator. It is clear that the set of unitary operators on
is a group, indeed a subgroup of the general linear group
. Note that the former depends on the scalar product but the latter does not; one may equivalently say that invertible operators map bases to bases, while unitary operators map orthonormal bases to orthonormal bases.
Let be given real numbers, and let
be the set of selfadjoint operators on
with this given spectrum.
Theorem 3: Choose any orthonormal basis , and let
be the selfadjoint operator defined by
Then the isospectral set is given by
Proof: First we check that consists entirely of selfadjoint operators having spectrum
. The selfadjointness of
follows immediately from the selfadjointness of
together with the fact that
. To see that
has spectrum
, let
be the orthonormal basis of
defined by
Thent is an eigenvector of
with eigenvalue
.
Now let be a given selfadjoint operator with eigenvalues
. By the Spectral Theorem, there exists an orthonormal basis
such that
If
is the unitary operator sending
to
, then
which is equivalent to .
— Q.E.D.
In due time, we shall see that both Grassmannians and adjoint orbits
are complex projective varieties.
The numbering of the subtitles goes 2. 3. 3. 4. Could you format the links to open in a new tab?