This week we are looking at a landmark paper by Grigori Olshanski and Anatoly Vershik, “Ergodic Unitarily Invariant Measures on the Space of Infinite Hermitian Matrices.” This paper is among the first to consider characteristic functions in the context of random matrix theory. Namely, it considers the problem of approximating the function
defined by
where the integration is over the unitary group of rank against Haar measure. We have referred to this function as the “quantum Fourier kernel,” but as we have stressed it is also the classical characteristic function of the diagonal
of
of a uniformly random Hermitian matrix with deterministic eigenvalues
The paper of Olshanski-Vershik is about approximating
as
which probabilistically means that we are trying to understand how the joint distribution of the diagonal matrix elements of a very large uniformly random Hermitian matrix with prescribed eigenvalues
is determined by the vector
This is an example of a high-dimensional problem: we are trying to approximate a function: our goal is to approximate a function
of real variables as the number of variables goes to infinity. Typically such problems are much harder than traditional problems of asymptotic analysis, which consider the approximation of some parameter-dependent family of functions defined on a fixed domain as the parameter approaches a given value.
One way to simplify the question is to consider the restriction of to the subspace
with
fixed. That is, we consider the characteristic function
of only the first diagonal matrix elements of the random Hermitian matrix
Recall that we have previously shown that all diagonal matrix elements of
are identically distributed, but not necessarily independent. The paper of Olshanski and Vershik shows that if the eigenvalues
, scaled by a factor of
converge to a point configuration on
as
then the random variables
converge to independent random variables whose characteristic function can be calculated explicitly in terms of this point configuration. We will try to work through this result, which I believe to be the first high-dimensional limit theorem in RMT obtained via characteristic functions.
The Olshanski-Vershik approach of computing the asymptotics of the characteristic function
is based on series expansion of this matrix integral. I am a big proponent of this approach, but it does require some quite extensive background from representation theory and algebraic combinatorics which is not traditionally found in the realm of probability. We will try not to get too bogged down in the algebra, and in order to do this it will be necessary to accept some results from representation theory as black boxes.
First, let us analytically continue the characteristic function to a function
by declaring
where is a complex number and
are complex vectors. Thus, by compactness of
, we have that
is an entire function of
complex variables
and the Maclaurin series of this function, which is a multivariate Taylor series centered at the origin, converges uniformly absolutely on compact subsets of
We recover our original characteristic function
by setting
and restricting
to
.
The analytic function has various symmetries that allow us to present its Maclaurin series in a more advantageous form. More precisely,
is invariant under independent permutations of
and
, and also stable under swapping these two sets of variables. This means that the Macluarin series of
can be presented in the form
where is a homogeneous degree
symmetric polynomial in
and also a homogeneous degree
symmetric polynomial in
. Therefore, by Newton’s theorem on symmetric polynomials, we can write this partial derivative as
where is the set of integer partitions of the degree
which we identify with the set of Young diagrams with
cells, and
and
are the Newton power-sum symmetric polynomials corresponding to
For example, if
and
then
Our objective then is to calculate the numerical coefficients of the polynomial
which we will refer to as its Newton coefficients.
Problem 9.1. Show that and use this to show that
What is good about the Newton polynomials is that they connect us to point configurations, which is what we want. More precisely, if is a real vector, then we can view it not as a point in
-dimensional space but as a configuration of particles
on the real line. The empirical distribution
of
is the discrete measure on
which places unit mass at each particle, and the degree
moment of this measure is
This means that writing the Maclaurin expansion of $K_N(z,A,B)$ in terms of Newton symmetric polynomials is exactly what we want to do, because it is expressing this function in terms of the empirical distributions of its arguments , and the empirical distribution of a point configuration is a natural way to describe the “shape” of the configuration in the infinite particle limit where
Unfortunately, there is no obvious way to calculate the Newton coefficients What we can do is compute the Maclaurin series of
by repeatedly differentiating under the integral sign with respect to
. Equivalently, we just expand the exponential function of
in the inetegrand and then integrate term-by-term, which gives
This gives us the formula
which at least reduces our initial problem of computing an exponential integral over to computing a polynomial integral over
but still there is no obvious path forward.
At this point we should ask ourselves what integrals over we actually can compute. To make the discussion as simple as possible, take
so
is just the unit circle in
One thing we can confidently state is the orthogonality relation
Representation theory gives us analogous formulas for all namely a system of multivariate polynomials which are orthogonal polynomials for integration with respect to Haar measure on
Let
be the set of all Young diagrams with at most
rows.
Defintion 9.1. For each the corresponding Schur symmetric polynomial is defined by
where are the row lengths of
Out of the many equivalent definitions of Schur polynomials, we are taking this one because it is a fairly explicit and elementary formula. Indeed, the determinant in the denominator is the Vandermonde determinant, which has the property that it divides any other antisymmetric polynomial in to yield a symmetric polynomial.
Problem 9.2. Let be the single-cell Young diagram. Compute the Schur polynomial
directly from Definition 9.1. Do the same for a few more small Young diagrams.
The disadvantage of Definition 9.1 is that it makes the orthogonality relations for Schur polynomials, which are really consequences of the fact that these polynomials are characters of irreducible representations of into a theorem whose proof would take us to far afield. So we will have to treat this property as a black box. Given a matrix
let us write
for the evaluation
of the Schur polynomial
on the eigenvalues of
Theorem 9.1 (Schur orthogonality). For any two Young diagrams and any two matrices
we have
and
where is the number of semi standard Young tableaux of shape
with entries from
We also need a remarkable “multinomial formula” due to Frobenius. Let be the set of Young diagrams with exactly
cells and at most
rows.
Theorem 9.2 (Frobenius formula). We have
where is the number of standard Young tableaux of shape
The reason the tableaux counts and
are denoted this way is that the former is the dimension of an irreducible representation of the symmetric group
and the latter is the dimension of an irreducible representation of the unitary group
At this point we have enough material to explicitly calculate a series expansion of
. This expansion is the starting point of Olshanski and Vershik’s asymptotic analysis, and is stated as Theorem 5.1 in their paper.
Theorem 9.3 (Schur expansion of ). We have
Proof: From the Theorem 9.2, we have
and applying the first integration formula in Theorem 9.1 this becomes
– QED
Problem 9.3. Compute the Schur expansion of up to
This is straightforward but will give you a feel for the structure of the formula.
Let us include the analogue of Theorem 9.3 for the quantum Bessel kernel
where and the integration is against Haar measure on the product group
Here
are real vectors viewed as diagonal matrices in
Since
is invariant under signed permutations of the coordinates of
and
, we can assume without loss in generality that these are non increasing lists of nonnegative numbers. The quantum Bessel kernel is the characteristic function of a uniformly random
rectangular matrix with singular values
Once again we can analytically continue this to
where and
are viewed as complex diagonal
matrices in the integral. The following expansion of
is not in the Olshanski-Vershik paper, but it can be obtained in essentially the same way; a more general result is proved here.
Theorem 9.4 (Schur expansion of ). We have
We close this lecture with a pointer to interesting work in the engineering literature which considers exactly these integrals in an applied context.