Abstract
The free multiplicative Brownian motion \(b_{t}\) is the large-N limit of the Brownian motion on \(\mathsf {GL}(N;\mathbb {C}),\) in the sense of \(*\)-distributions. The natural candidate for the large-N limit of the empirical distribution of eigenvalues is thus the Brown measure of \(b_{t}\). In previous work, the second and third authors showed that this Brown measure is supported in the closure of a region \(\Sigma _{t}\) that appeared in the work of Biane. In the present paper, we compute the Brown measure completely. It has a continuous density \(W_{t}\) on \(\overline{\Sigma }_{t},\) which is strictly positive and real analytic on \(\Sigma _{t}\). This density has a simple form in polar coordinates:
where \(w_{t}\) is an analytic function determined by the geometry of the region \(\Sigma _{t}\). We show also that the spectral measure of free unitary Brownian motion \(u_{t}\) is a “shadow” of the Brown measure of \(b_{t}\), precisely mirroring the relationship between the circular and semicircular laws. We develop several new methods, based on stochastic differential equations and PDE, to prove these results.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
1.1 The Brown measure
Let \((\mathcal {A},\tau )\) be a tracial von Neumann algebra: a von Neumann algebra \(\mathcal {A}\) together with a faithful, normal, tracial state \(\tau :\mathcal {A}\rightarrow \mathbb {C}.\) Such algebras frequently arise from large-N limits of random matrix models, with \(\tau \) playing the role of the normalized trace for matrices. For an element a of \(\mathcal {A}\), the notion of the empirical eigenvalue distribution of a matrix is then played by the Brown measure of a, defined as follows [5]. We let
which is defined as a finite real number for almost every \(\lambda .\) (The quantity \(s_{a}(\lambda )\) is twice the logarithm of the Fuglede–Kadison determinant [9, 10] of \(a-\lambda .\)) Then \(s_{a}\) is a subharmonic function and the Brown measure \(\mu _{a}\) of a is defined in terms of the distributional Laplacian of \(s_{a}\):
The definition of the Brown measure is the operator-algebra counterpart to Girko’s method [11] of computing the empirical eigenvalue distribution of a random matrix.
By regularizing the right-hand side of (1.1), one can construct the Brown measure \(\mu _{a}\) as a weak limit,
where \(\Delta _{\lambda }\) is the Laplacian with respect to \(\lambda \) and \(d\lambda \) is the Lebesgue measure on the plane (see [23, Section 11.5] and [18, Eq. (2.11)]). Here, the positive parameter \(\varepsilon \) regularizes the logarithm, so that \(\tau [\log [(a-\lambda )^{*}(a-\lambda )+\varepsilon )]\) is a smooth function of \(\lambda \in \mathbb {C}\).
In general, \(\mu _{a}\) is a probability measure supported on the spectrum of a. If a happens to be a normal operator, \(\mu _{a}\) coincides with the law (or distribution) of a. That is to say, if a is normal, \(\mu _{a} (E)=\tau (P_{a}(E)),\) where \(P_{a}\) is the projection-valued measure associated to a by the spectral theorem.
1.2 The free unitary and multiplicative Brownian motions
Let \(\sigma _{t}\) be a free semicircular Brownian motion (e.g., [4, Section 1.1]) and let \(c_{t}\) be a free circular Brownian motion, which may be constructed as \(c_{t}=(x_{t}+iy_{t})/\sqrt{2},\) where \(x_{t}\) and \(y_{t}\) are freely independent semicircular Brownian motions. These are the large-N limits, in the sense of \(*\)-distribution, of Brownian motions in the space of Hermitian \(N\times N\) matrices and in the space of all \(N\times N\) matrices, respectively. We then introduce the free unitary Brownian motion \(u_{t}\) and the free multiplicative Brownian motion \(b_{t}\) given by the free stochastic differential equations
both starting at 1. These processes are the large-N limits of Brownian motions in the unitary group and in the general linear group, respectively. (For the free unitary Brownian motion, this limiting result is due to Biane, while for the free multiplicative Brownian motion it was conjectured by Biane [3] and proved by the third author [20]).
Biane also computed the law \(\nu _{t}\) of \(u_{t}.\) We now record this result, since it relates directly to the results of the present paper (Sect. 2.2). Let \(f_{t}\) denote the holomorphic function on \(\mathbb {C}{\setminus }\{1\}\) defined by
Then \(f_{t}\) has a holomorphic inverse \(\chi _{t}\) in the open unit disk, and \(\chi _{t}\) extends continuously to the closed unit disk. Biane showed that
where \(\psi _{u_{t}}(z)=\tau [(1-zu_{t})^{-1}]-1\) is the (recentered) moment-generating function of \(u_{t}\). From this (and other SDE computations) he determined the following result.
Theorem 1.1
(Biane [2, 3]). The spectral measure \(\nu _{t}\) of the free unitary Brownian motion \(u_{t}\) is supported in the arc
for \(t<4\), and is fully supported on the circle for \(t\ge 4\). The measure \(\nu _{t}\) has a continuous density \(\kappa _{t}\), which is real analytic on the interior of its support arc, given by
See, for example, p. 275 in [3]. In the present paper, we compute the Brown measure \(\mu _{b_{t}}\) of the free multiplicative Brownian motion and show a direct relationship between \(\mu _{b_{t}}\) and the law \(\nu _{t}\) of the free unitary Brownian motion \(u_{t}\) (Sect. 2.2).
1.3 The Brown measure of \(b_{t}\)
The main result of this paper is a formula for the Brown measure \(\mu _{b_{t}}\) of the free multiplicative Brownian motion \(b_{t}.\)
A previous result [18] of the second and third authors showed that the support of \(\mu _{b_{t}}\) is contained in closure of a certain region \(\Sigma _{t}\) introduced by Biane in [3]; see Fig. 1 and Definition 2.1. (We reprove that result in the present paper by a different method; see Theorem 6.2 in Sect. 6.2). Nonrigorous results on the support of the Brown measure were also obtained in the physics literature by Gudowska-Nowak et al. [13] and then by Lohmayer et al. [22]. None of the results mentioned in this paragraph say anything about the actual Brown measure itself—only about its support.
We conjecture that the Brown measure of \(b_{t}\) coincides with the limiting eigenvalue distribution of the corresponding Brownian motion \(B_{t}^{N}\) in the general linear group (see Fig. 2). Proving such results is, however, well known to be a difficult problem, which we do not address here.
Since the first version of this paper appeared on the arXiv, four subsequent works have appeared that use the techniques developed here to analyze Brown measures of other operators. First, work of Ho and Zhong [19] has extended the results of the present paper to the case of a free multiplicative Brownian motion with an arbitrary unitary initial condition. This means that they compute the Brown measure of \(ub_{t},\) where u is a unitary element freely independent of \(b_{t}.\) Ho and Zhong also compute the Brown measure of \(x_{0}+c_{t},\) where \(c_{t}\) is a free circular Brownian motion and \(x_{0}\) is a self-adjoint element freely independent of \(c_{t}.\) Second, Demni and Hamdi [7] have computed the support of the Brown measure of \(u_{t}P,\) where \(u_{t}\) is the free unitary Brownian motion and P is a projection freely independent of \(u_{t}\). Third, Hall and Ho [16] have computed the Brown measure of \(x_{0}+i\sigma _{t},\) where \(\sigma _{t}\) is the free semicircular Brownian motion and \(x_{0}\) is a self-adjoint element freely independent of \(x_{t}\). Last, Hall and Ho [17] have computed the Brown of \(ub_{s,\tau }\), where \(b_{s,\tau }\) is a family of multiplicative Brownian motions allowing different diffusion rates in the Hermitian and skew-Hermitian directions and where u is freely independent of \(b_{s,\tau }.\)
The reader may also consult the expository article [15] by the second author, which provides a nontechnical introduction to the techniques used in the present paper. See also the paper [12] of Grela, Nowak, and Tarnowski that explains the PDE method from a physical perspective.
2 Statement of main results
2.1 A formula for the Brown measure
To state our main result, we need to briefly describe the regions \(\Sigma _{t}.\) For each \(t>0,\) consider the holomorphic function \(f_{t}\) on \(\mathbb {C}{\setminus }\{1\}\) defined by (1.3). It is easily verified that if \(\left| \lambda \right| =1\) then \(\left| f_{t} (\lambda )\right| =1.\) There are, however, other points where \(\left| f_{t}(\lambda )\right| =1.\) We then define
and
Definition 2.1
For each \(t>0,\) we define \(\Sigma _{t}\) to be the connected component of the complement of \(E_{t}\) containing 1.
We will show (Theorem 3.1) that \(\Sigma _{t}\) may also be characterized as
where \(T(\lambda )=\left| \lambda -1\right| ^{2}\log (\left| \lambda \right| ^{2})/(\left| \lambda \right| ^{2}-1).\) Each region \(\Sigma _{t}\) is invariant under the maps \(\lambda \mapsto 1/\lambda \) and \(\lambda \mapsto \bar{\lambda }.\) If we consider a ray from the origin with angle \(\theta ,\) if this ray intersects \(\Sigma _{t}\) at all, it does so in an interval of the form \(1/r_{t}(\theta )<r<r_{t}(\theta )\) for some \(r_{t} (\theta )>1\) (see Figs. 3 and 4). See Sect. 3 for more information.
We are now ready to state our main result.
Theorem 2.2
For all \(t>0,\) the Brown measure \(\mu _{b_{t}}\) of \(b_{t}\) is absolutely continuous with respect to the Lebesgue measure on the plane and \(\mu _{b_{t}}(\Sigma _{t})=1.\) In \(\Sigma _{t},\) the density \(W_{t}\) of \(\mu _{b_{t}}\) with respect to the Lebesgue measure is strictly positive and real analytic, with the following form in polar coordinates:
for a certain even function \(w_{t}.\) This function may be computed as
where \(r_{t}(\theta )\) is the larger of the two radii where the ray with angle \(\theta \) intersects the boundary of \(\Sigma _{t}.\)
Since \(\Sigma _{t}\) is invariant under \(\lambda \mapsto \bar{\lambda },\) the function \(r_{t}(\theta )\) is an even function of \(\theta \), from which it is easy to check that the second term on the right-hand side of of (2.5) is also an even function of \(\theta .\) Although we will customarily let \(r_{t}(\theta )\) denote the larger of the the two radii where the ray with angle \(\theta \) intersects the boundary of \(\Sigma _{t}\), we note that
is invariant under \(r\mapsto 1/r.\) Thus, the value of \(w_{t}\) does not actually depend on which radius is used. It is noteworthy that the one nonexplicit part of the formula for \(w_{t},\) namely the second term on the right-hand side of (2.5), is computable entirely in terms of the geometry of the region \(\Sigma _{t}.\) According to Proposition 7.5, \(w_{t}\) can also be computed as a logarithmic derivative along the boundary of \(\Sigma _{t}\) of the function \(f_{t}\) in (1.3).
It follows from (2.3) that the function T equals t on the boundary of \(\Sigma _{t}.\) It is then possible to use implicit differentiation in the equation \(T(\lambda )=t\) to compute \(dr_{t}(\theta )/d\theta \) as a function of \(r_{t}(\theta )\) and \(\theta .\) We may then use this computation to rewrite (2.5) in a form that no longer involves a derivative with respect to \(\theta ,\) as follows.
Proposition 2.3
The function \(w_{t}\) in Theorem 2.2 may also be computed in the form
Here
where
Thus, to compute \(w_{t}(\theta ),\) we evaluate \(\omega /(2\pi t)\) on the boundary of \(\Sigma _{t}\) and then parametrize the boundary by the angle \(\theta \); see Fig. 5. Using Proposition 2.3, we can derive small- and large-t asymptotics of \(w_{t}(\theta )\) as follows:
See Sect. 7 for details.
The following simple consequences of Theorem 2.2 help to explain the significance of the factor of \(1/r^{2}\) in the formula (2.4) for \(W_{t}\).
Corollary 2.4
The Brown measure \(\mu _{b_{t}}\) of \(b_{t}\) has the following properties.
-
(1)
\(\mu _{b_{t}}\) is invariant under the maps \(\lambda \mapsto 1/\lambda \) and \(\lambda \mapsto \bar{\lambda }.\)
-
(2)
Let \(\Xi _{t}\) denote the image of \(\Sigma _{t}{\setminus }(-\infty ,0)\) under the complex logarithm map, using the standard branch cut along the negative real axis. We write points \(z\in \Xi _{t}\) as \((\rho ,\theta ).\) Then for points in \(\Xi _{t},\) the pushforward of \(\mu _{b_{t} }\) by the logarithm map has density \(\omega _{t}(\rho ,\theta )\) given by
$$\begin{aligned} \omega _{t}(\rho ,\theta )=w_{t}(\theta ), \end{aligned}$$independent of \(\rho .\)
Proof
As we have stated above, the region \(\Sigma _{t}\) is invariant under the maps \(\lambda \mapsto 1/\lambda \) and \(\lambda \mapsto \bar{\lambda }.\) The invariance of \(\mu _{b_{t}}\) under \(\lambda \mapsto \bar{\lambda }\) follows from the fact that \(w_{t}\) is even. Now, we may compute \(\mu _{b_{t}}\) in polar coordinates as
where \(\log r\) and \(\theta \) are the real and imaginary parts of the complex logarithm of \(\lambda =re^{i\theta },\) as claimed in Point 2. The invariance of the measure under \(\lambda \mapsto 1/\lambda ,\) that is, under \((r,\theta )\mapsto (1/r,-\theta )\) is now evident. \(\square \)
Plots of \(w_{t}(\theta )\) are shown in Fig. 6. Note that for \(t<4,\) not all angles \(\theta \) actually occur in the domain \(\Sigma _{t}\). Thus, for \(t<4,\) the function \(w_{t}(\theta )\) is only defined for \(\theta \) in a certain interval \((-\theta _{\max }(t),\theta _{\max }(t))\)—where, as shown in Sect. 3, \(\theta _{\max }(t)=\cos ^{-1}(1-t/2).\) Plots of \(W_{t}\) for \(t=1\) and \(t=4\) are then shown in Fig. 7. Actually, when \(t=1,\) the function \(w_{t}\) is almost constant. Thus, the variation in \(W_{t}\) in the top part of Fig. 7 comes almost entirely from the variation in the factor of \(1/r^{2}\) in (2.4).
We also observe that, by Point 1 of Corollary 2.4, half the mass of \(\mu _{b_{t}}\) is contained in the unit disk and half in the complement of the unit disk. Thus, although the density \(W_{t}\) becomes large near the origin when \(t=4,\) it is not correct to say that most of the mass of \(\mu _{b_{t}}\) is near the origin.
2.2 A connection to free unitary Brownian motion
It follows easily from Theorem 2.2 that the distribution of the argument of \(\lambda \) with respect to \(\mu _{b_{t}}\) has a density given by
where, as in Theorem 2.2, we take \(r_{t}(\theta )\) to be the outer radius of the domain (with \(r_{t}(\theta )>1\)). After all, the Brown measure in the domain is computed in polar coordinates as \((1/r^{2})w_{t}(\theta )r~dr~d\theta \). Integrating with respect to r from \(1/r_{t}(\theta )\) to \(r_{t}(\theta )\) then gives the claimed density for \(\theta \).
Recall from Theorem 1.1 that the limiting eigenvalue distribution \(\nu _{t}\) for Brownian motion in the unitary group was determined by Biane. We now claim that the distribution in (2.8) is related to Biane’s measure \(\nu _{t}\) by a natural change of variables. To each angle \(\theta \) arising in the region \(\Sigma _{t},\) we associate another angle \(\phi \) by the formula
where \(f_{t}\) is as in (1.3). (Recall that, by Definition 2.1, the boundary of \(\Sigma _{t}\) maps into the unit circle under \(f_{t}.\)) We then have the following remarkable direct connection between the Brown measure of \(b_{t}\) and Biane’s measure \(\nu _{t}.\)
Proposition 2.5
If \(\theta \) is distributed according to the density in (2.8) and \(\phi \) is defined by (2.9), then \(\phi \) is distributed as Biane’s measure \(\nu _{t}.\)
We may think of this result in a more geometric way, as follows. Define a map
by requiring (a) that \(\Phi _{t}\) should agree with \(f_{t}\) on the boundary of \(\Sigma _{t},\) and (b) that \(\Phi _{t}\) should be constant along each radial segment inside \(\overline{\Sigma }_{t},\) as in Fig. 8. (This specification makes sense because \(f_{t}\) has the same value at the two boundary points on each radial segment). Explicitly, \(\Phi _{t}\) may computed as
Then Proposition 2.5 gives the following result, which may be summarized by saying that the distribution \(\nu _{t}\) of free unitary Brownian motion is a “shadow” of the Brown measure of \(b_{t}\).
Proposition 2.6
The push-forward of the Brown measure of \(b_{t}\) under the map \(\Phi _{t}\) is Biane’s measure \(\nu _{t}\) on \(S^{1}.\) Indeed, the Brown measure of \(b_{t}\) is the unique measure \(\mu \) on \(\overline{\Sigma }_{t}\) with the following two properties: (1) the push-forward of \(\mu \) by \(\Phi _{t}\) is \(\nu _{t}\) and (2) \(\mu \) is absolutely continuous with respect to Lebesgue measure with a density W having the form
in polar coordinates, for some continuous function g.
Now, the results of [3, 18] already suggest a relationship between the free unitary Brownian motion \(u_{t}\) (whose law is \(\nu _{t}\)) and the free multiplicative Brownian motion \(b_{t}\) (whose Brown measure we are studying in this paper). It is nevertheless striking to see such a direct relationship between \(\mu _{b_{t}}\) and \(\nu _{t}\). Indeed, Proposition 2.6 precisely mirrors the relationship between the semicircle law and the circular law. If \(c_{t}\) is a circular random variable of variance t, and \(x_{t}\) is semicircular of variance t, then the distribution of \(x_{t}\) (the semicircle law on the interval \([-2\sqrt{t},2\sqrt{t}]\)) is the push-forward of the Brown measure of \(c_{t}\) (the uniform probability measure on the disk \(\overline{\mathbb {D}}(\sqrt{t})\) of radius \(\sqrt{t}\)) under a similar “shadow map”: first project the disk onto its upper boundary circle via \((x,y)\mapsto (x,\sqrt{t-x^{2}})\), and then use the conformal map \(z\mapsto z+\frac{t}{z}\) from \(\mathbb {C}{\setminus }\overline{\mathbb {D}}(\sqrt{t})\) onto \(\mathbb {C}{\setminus }[-2\sqrt{t},2\sqrt{t}]\). The net result of these two operations is the map \((x,y)\mapsto 2x,\) and the push-forward of the uniform measure on \(\overline{\mathbb {D}}(\sqrt{t})\) under this map is the semicircular measure on \([-2\sqrt{t},2\sqrt{t}].\)
2.3 Deriving the formula
We now briefly indicate the method we will use to compute the Brown measure \(\mu _{b_{t}}\). Following the general construction of the Brown measure in (1.2), we consider the function S defined by
for \(\lambda \in \mathbb {C}\) and \(\varepsilon >0,\) where \(b_{t}\) is the free multiplicative Brownian motion and \(\tau \) is the trace in the von Neumann algebra in which \(b_{t}\) lives. It is easily verified that as \(\varepsilon \) decreases with t and \(\lambda \) fixed, \(S(t,\lambda ,\varepsilon )\) also decreases. Hence, the limit
exists, possibly with the value \(-\infty .\)
The general theory developed by Brown [5] shows that \(s_{t}(\lambda )\) is a subharmonic function of \(\lambda \) for each fixed t, so that the Laplacian (in the distribution sense) of \(s_{t}(\lambda )\) with respect to \(\lambda \) is a positive measure. If this measure happens to be absolutely continuous with respect to the Lebesgue measure, then the density \(W(t,\lambda )\) of the Brown measure is computed in terms of the value of \(s_{t}(\lambda ),\) as follows:
See also Chapter 11 in [23] and Section 2.3 in [18] for general information on Brown measures.
The first major step toward proving Theorem 2.2 is the following result.
Theorem 2.7
The function S in (2.10) satisfies the following PDE:
with the initial condition
We emphasize that \(S(t,\lambda ,\varepsilon )\) is only defined for \(\varepsilon >0.\) Although, as we will see, \(\lim _{\varepsilon \rightarrow 0^{+} }S(t,\lambda ,\varepsilon )\) is finite, \(\partial S/\partial \varepsilon \) develops singularities in this limit. Thus, it is not correct to formally set \(\varepsilon =0\) in (2.12) to obtain \(\partial s_{t}/\partial t=0.\) (Actually, it will turn out that \(s_{t}(\lambda )\) is independent of t for as long as \(\lambda \) remains outside \(\overline{\Sigma }_{t},\) but not after this time; see Sect. 6.2).
After verifying this equation (Sect. 4), we will use the Hamilton–Jacobi formalism to analyze the solution (Sect. 5). In the remaining sections, we will then analyze the limit of the solution as \(\varepsilon \) tends to zero and compute the Laplacian in (2.11). The expository article [15] of the second author provides an introduction to the methods used in the present paper.
By way of comparison, we mention that a similar PDE appeared in Biane’s paper [1]. There he studies the spectral measure \(\mu _{t}\) of \(x_{0}+x_{t}\), the free additive Brownian motion with a nonconstant initial distribution \(x_{0}\) freely independent from \(x_{t}\). Biane studies the Cauchy transform G of \(\mu _{t}\):
and shows that G satisfies the complex inviscid Burger’s equation
The measure \(\mu _{t}\) may then be recovered, up to a constant, as \(\lim _{\varepsilon \rightarrow 0^{+}}{\text {Im}}G(t,x+i\varepsilon )\).
In our paper, we similarly use a first-order, nonlinear PDE whose solution in a certain limit gives the desired measure. We note, however, that the PDE (2.14) is not actually the main source of information about \(\mu _{t}\) in [1]. By contrast, our analysis of the Brown measure of the free multiplicative Brownian motion \(b_{t}\) is based entirely on the PDE in Theorem 2.7.
Finally, we mention that, for the case of the circular Brownian motion \(c_{t},\) a PDE similar to the one in Theorem 2.12 appeared in work of Burda et al. [6, Equation (9)].
3 Properties of \(\Sigma _{t}\)
We now verify some important properties of the regions \(\Sigma _{t}\) in Definition 2.1. Define
Since the function
has a removable singularity at \(x=1\) with a limiting value of 1, we interpret \(T(\lambda )\) as equaling \(\left| \lambda -1\right| ^{2}\) when \(\left| \lambda \right| ^{2}=1.\) Then \(T(\lambda )\) is a real analytic function on all of \(\mathbb {C}{\setminus }\{0\}.\) Since, also,
we see that \(T(\lambda )\rightarrow +\infty \) as \(\lambda \rightarrow 0.\) By checking the three cases \(\left| \lambda \right| >1,\) \(\left| \lambda \right| =1,\) and \(\left| \lambda \right| <1,\) we may verify that \(T(\lambda )\ge 0\) for all \(\lambda ,\) with equality only if \(\lambda =1.\)
Theorem 3.1
For all \(t>0,\) the region \(\Sigma _{t}\) may be expressed as
and the boundary of \(\Sigma _{t}\) may be expressed as
Thus, each fixed \(\lambda \in \mathbb {C}\) will be outside \(\overline{\Sigma } _{t}\) until \(t=T(\lambda )\) and will be inside \(\Sigma _{t}\) for all \(t>T(\lambda ).\) We may therefore say that \(T(\lambda )\) is the time that the domain \(\Sigma _{t}\) gobbles up \(\lambda .\) See Figs. 9 and 10.
Theorem 3.2
For each \(t>0,\) the region \(\Sigma _{t}\) has the following properties.
-
(1)
For \(t\le 4,\) we have \(\left| \arg \lambda \right| \le \cos ^{-1}(1-t/2)\) for all \(\lambda \in \overline{\Sigma }_{t},\) with equality precisely for the points on the unit circle with \(\cos \theta =1-t/2.\)
-
(2)
Consider the ray from the origin with angle \(\theta \); if \(t\le 4,\) assume \(\left| \theta \right| <\cos ^{-1} (1-t/2)\). Then this ray intersects \(\Sigma _{t}\) precisely in an open interval of the form \(1/r_{t}(\theta )<r<r_{t}(\theta )\) for some \(r_{t}(\theta )>1.\)
-
(3)
The boundary of \(\Sigma _{t}\) is smooth for all \(t>0\) with \(t\ne 4.\) When \(t=4,\) the boundary of \(\Sigma _{t}\) is smooth except at \(\lambda =-1,\) near which it looks like the transverse intersection of two smooth curves.
-
(4)
The region \(\Sigma _{t}\) is invariant under \(\lambda \mapsto 1/\lambda \) and under \(\lambda \mapsto \bar{\lambda }.\)
-
(5)
The region \(\Sigma _{t}\) coincides with the one defined by Biane in [3].
We now begin working toward the proofs of Theorems 3.1 and 3.2.
Lemma 3.3
For \(\lambda \in \mathbb {C}\) with \(\left| \lambda \right| \ne 1,\) we have \(\left| f_{t}(\lambda )\right| =1\) if and only if \(T(\lambda )=t.\)
Proof
Since \(f_{t}(0)=0,\) we must have \(\lambda \ne 0\) if \(\left| f_{t} (\lambda )\right| \) is going to equal 1. For nonzero \(\lambda ,\) we compute that
Thus, for nonzero \(\lambda ,\) the condition \(\left| f_{t}(\lambda )\right| =1\) is equivalent to
When \(\left| \lambda \right| \ne 1,\) this condition simplifies to
as claimed. \(\square \)
We now state some important properties of the function \(r_{t}\) occurring in the statement of Theorem 2.2; the proof is given on p. 15.
Proposition 3.4
Consider a real number \(t>0\) and an angle \(\theta \in (-\pi ,\pi ],\) where if \(t\le 4,\) we require \(\left| \theta \right| <\cos ^{-1}(1-t/2).\) Then there exist exactly two radii \(r\ne 1\) for which \(\left| f_{t}(re^{i\theta })\right| =1,\) and these radii have the form \(r=r_{t}(\theta )\) and \(r=1/r_{t}(\theta )\) with \(r_{t}(\theta )>1.\) Furthermore, \(r_{t}(\theta )\) depends analytically on \(\theta \) and if \(t\le 4,\) then \(r_{t}(\theta )\rightarrow 1\) as \(\theta \rightarrow \pm \cos ^{-1}(1-t/2).\)
If \(t\le 4\) and \(\theta \in (-\pi ,\pi ]\) satisfies \(\left| \theta \right| \ge \cos ^{-1}(1-t/2),\) then there are no radii \(r\ne 1\) with \(\left| f_{t}(r)\right| =1.\)
Using the proposition, we can now compute the sets \(F_{t}\) and \(E_{t} =\overline{F}_{t}\) that enter into the definition of \(\Sigma _{t}\). (Recall (2.1) and (2.2)).
Corollary 3.5
For \(t\le 4,\) the set \(F_{t}\) consists of points of the form \(r_{t}(\theta )e^{i\theta }\) and \((1/r_{t}(\theta ))e^{i\theta }\) for \(-\cos ^{-1}(1-t/2)<\theta <\cos ^{-1}(1-t/2).\) In this case, the closure of \(F_{t}\) consists of \(F_{t}\) together with the points \(e^{i\theta }\) on the unit circle with \(\cos \theta =1-t/2.\) There are two such points when \(t<4\) and one such point when \(t=4,\) namely \(-1.\)
For \(t>4,\) the set \(F_{t}\) consists of points of the form \(r_{t} (\theta )e^{i\theta }\) and \((1/r_{t}(\theta ))e^{i\theta },\) where \(\theta \) ranges over all possible angles, and this set is closed.
We now set out to prove Proposition 3.4. In the proof, we will always rewrite the equation \(\left| f_{t}(\lambda )\right| =1\), for \(\left| \lambda \right| \ne 1\), as \(T(\lambda )=t\) (Lemma 3.3).
Lemma 3.6
Let us write the function T in (3.1) in polar coordinates. Then for each \(\theta ,\) the function \(r\mapsto T(r,\theta )\) is strictly decreasing for \(0<r<1\) and strictly increasing for \(r>1.\) For each \(\theta ,\) the minimum value of \(T(r,\theta )\), achieved at \(r=1\), is \(2(1-\cos \theta ),\) and we have
Proof
We will show in Proposition 5.13 that the function \(T(\lambda )\) is the limit of another function \(t_{*}(\lambda _{0},\varepsilon _{0})\) as \(\varepsilon _{0}\) goes to zero. Explicitly, this amounts to saying that \(T(r,\theta )=g_{\theta }(\delta ),\) where g is defined in (5.58) and \(\delta =r+1/r.\) Now, \(\delta \) is decreasing for \(0<r<1\) and increasing for \(r>1.\) Thus, the claimed monotonicity of T follows if \(g_{\,\theta }(\delta )\) an increasing function \(\delta \) for each \(\theta ,\) which we will show in the proof of Proposition 5.16.
For the convenience of the reader, we briefly outline how the argument goes in the context of the function \(T(r,\theta ).\) We note that
where if we assign \(\log (r^{2})/(r^{2}-1)\) the value 1 at \(r=1,\) then T is analytic except at \(r=0.\) We then compute that, after simplification,
We then claim that for all \(\theta ,\) we have \(\partial T/\partial r>0\) for \(r>1\) and \(\partial T/\partial r<0\) for \(r<1.\) Note that for each fixed r, the right-hand side of (3.2) depends linearly on \(\cos \theta .\) Thus, if, for a fixed r, if \(\partial T/\partial r\) is positive when \(\cos \theta =1\) and when \(\cos \theta =-1,\) it will be positive for all \(\theta \). Specifically, we may say that
It is now an elementary (if slightly messy) computation to check that the right-hand side of (3.3) is strictly positive for all \(r>1.\) A similar argument then shows that \(\partial T/\partial r\) is negative for all \(\theta \) and all \(0<r<1.\)
We conclude that for each \(\theta ,\) the function \(r\mapsto T(r,\theta )\) is decreasing for \(0<r<1\) and increasing for \(r>1.\) The minimum value therefore occurs at \(r=1,\) and this value is the value of \(r^{2}+1-2\cos \theta \) at \(r=1,\) namely \(2(1-\cos \theta ).\) Finally, we can easily see that for r approaching zero, we have
and for r approaching infinity, we have
as claimed. \(\square \)
Proof of Proposition 3.4
The minimum value of \(T(r,\theta ),\) achieved at \(r=1,\) is \(2-2\cos \theta .\) This value is always less than t, as can be verified separately in the cases \(t>4\) (all \(\theta \)) and \(t\le 4\) (\(\left| \theta \right| <\cos ^{-1}(1-t/2)\)). Thus, Lemma 3.6 tells us that the equation \(T(r,\theta )=t\) has exactly one solution for r with \(0<r<1\) and exactly one solution for \(r>1.\) Since, as is easily verified, \(T(1/r,\theta )=T(r,\theta ),\) the two solutions are reciprocals of each other, and we let \(r_{t}(\theta )\) denote the solution with \(r>1.\) Since \(\partial T/\partial r\) is nonzero for all \(r\ne 1,\) the implicit function theorem tells us that \(r_{t}(\theta )\) depend analytically on \(\theta .\)
Now, if \(t\le 4\) and \(\theta \) approaches \(\pm \cos ^{-1}(1-t/2),\) the minimum value of \(2-2\cos \theta \)—achieved at \(r=1\)—approaches \(2-2(1-t/2)=t.\) It should then be plausible that \(r_{t}(\theta )\) will approach \(r=1.\) To make this claim rigorous, we need to show that \(T(r,\theta )\) increases rapidly enough as r increases from 1 that the \(T(r,\theta )=t\) is achieved close to \(r=1.\) To that end, let g(r) denote the function on the right-hand side of (3.3), which is continuous everywhere and strictly positive for \(r>1.\) Then for \(r>1,\) we have
Now, G(r) is continuous and strictly increasing for \(r\ge 1,\) with \(G(1)=0.\) Thus, G it has a continuous inverse function satisfying \(G^{-1}(0)=1.\)
For \(\varepsilon >0,\) choose \(\delta >0\) so that \(G^{-1}(R)<1+\varepsilon \) when \(0<R<\delta .\) Then take \(\theta \) sufficiently close to \(\pm \cos ^{-1}(1-t/2)\) that \(2-2\cos \theta \) is within \(\delta \) of t. Then
which is to say that there is an R with \(1<R<1+\varepsilon \) such that
From (3.4) we can then see that \(T(R,\theta )>t.\) Thus, \(r_{t} (\theta )\) will satisfy
We have therefore shown that \(r_{t}(\theta )\rightarrow 1\) as \(\theta \rightarrow \pm \cos ^{-1}(1-t/2).\)
Finally, if \(t\le 4\) and \(\theta \in (-\pi ,\pi ]\) satisfies \(\left| \theta \right| \ge \cos ^{-1}(1-t/2),\) the minimum value of \(T(r,\theta ),\) achieved at \(r=1,\) is \(2-2\cos \theta \ge t.\) Thus, there are no values of \(r\ne 1\) where \(T(r,\theta )=t.\) \(\square \)
We are now ready for the proofs of our main results about \(\Sigma _{t}.\)
Proof of Theorem 3.1
We first claim that the set \(E_{t}=\overline{F_{t}}\) is precisely the set where \(T(\lambda )=t.\) To see this, first note that \(F_{t}\) is, by Lemma 3.3, the set of \(\lambda \) with \(\left| \lambda \right| \ne 1\) where \(T(\lambda )=t.\) Then by Corollary 3.5, the closure of \(F_{t}\) is obtained by adding in the points on the unit circle (zero, one, or two such points, depending on t) where \(\cos \theta =1-t/2.\) But these points are easily seen to be the points on the unit circle where \(T(\lambda )=t.\)
Using Corollary 3.5, we see that the complement of the set \(E_{t}=\left\{ \lambda |T(\lambda )=t\right\} \) has two connected components when \(t<4\) and three connected components when \(t\ge 4.\) Since \(T(1)=0<t,\) we have \(T(\lambda )<t\) on the entire connected component of \(E_{t}^{c}\) containing 1, which is, by definition, the region \(\Sigma _{t}.\) The remaining components of \(E_{t}^{c}\) are the unbounded component and (for \(t\ge 4\)) the component containing 0. Since \(T(\lambda )\) tends to \(+\infty \) at zero and at infinity, we see that \(T(\lambda )>t\) on these regions, so that \(T(\lambda )<t\) precisely on \(\Sigma _{t}.\)
It is also clear from Corollary 3.5 that the boundary of the region \(\Sigma _{t}\) (i.e., the connected component of \(E_{t}^{c}\) containing 1) contains the entire set \(E_{t}=\) \(\left\{ \lambda |T(\lambda )=t\right\} .\) \(\square \)
Proof of Theorem 3.2
Point 1 follows easily from Corollary 3.5. For Point 2, we note that by Proposition 3.4, we have \(T(r,\theta )<t\) for \(1/r_{t}(\theta )<r<r_{t}(\theta ),\) and \(T(r,\theta )\ge t\) for \(0<r\le 1/r_{t}(\theta )\) and for \(r\ge r_{t}(\theta ).\) Thus, by Theorem 3.1, the ray with angle \(\theta \) intersects \(\Sigma _{t}\) precisely in the claimed interval.
For Point 3, we have already shown that \(\partial T/\partial r\) is nonzero except when \(r=1.\) When \(r=1,\) we know from (3.1) that
Thus, when \(r=1,\) we have \(\partial T/\partial \theta =2\sin \theta ,\) which is nonzero except when \(\theta =0\) or \(\theta =\pi .\) Thus, the gradient of \(T(\lambda )\) is nonzero except when \(\lambda =0\) (where \(T(\lambda )\) is undefined), when \(\lambda =1\), and when \(\lambda =-1.\) Since 0 is never in \(\Sigma _{t}\) and 1 is always in \(\Sigma _{t},\) the only possible singular point in the boundary of \(\Sigma _{t}\) is at \(\lambda =-1.\) Since \(T(r,\theta )=2-2\cos \pi =4\) when \(r=1\) and \(\theta =\pi ,\) the point \(\lambda =-1\) belongs to the boundary of \(\Sigma _{4}.\)
Meanwhile, the Taylor expansion of T to second order at \(\lambda =-1\) is easily found to be \(T(\lambda )\approx 4+({\text {Re}}\lambda +1)^{2}/3-({\text {Im}}\lambda )^{2}.\) By the Morse lemma, we can then make a smooth change of variables so that in the new coordinate system,
Thus, near \(\lambda =-1,\) the set \(T(\lambda )=4\) is the union of the curves \(u+v=0\) and \(u-v=0.\)
The invariance of \(\Sigma _{t}\) under \(\lambda \mapsto 1/\lambda \) and under \(\lambda \mapsto \bar{\lambda }\) follows from the easily verified invariance of \(T(\lambda )\) under these transformations. Finally, we verify that the domain \(\Sigma _{t},\) as we have defined it, coincides with the one originally introduced by Biane [3]. Let us start with the case \(t<4.\) According to the discussion at the bottom of p. 273 in [3], the boundary of Biane’s domain \(\Sigma _{t}\) consists in this case of two analytic arcs. The interior of one arc lies in the open unit disk and the interior of the other arc lies in the complement of the closed unit disk, while the endpoints of both arcs lie on the unit circle. The first arc is then computed by applying a certain holomorphic function \(\chi (t,\cdot )\) to the support of Biane’s measure \(\nu _{t}\) in the unit circle. Now, \(\chi (t,\cdot )\) satisfies \(f_{t}(\chi (t,z))=z\) on the closed unit disk. (Combine the identity involving \(\kappa \) on p. 266 of [3] with the definition of \(\chi \) on p. 273). We see that the interior of the first arc consists of points with \(\left| \lambda \right| \ne 1\) but \(\left| f_{t}(\lambda )\right| =1.\) This arc must, therefore, coincide with the arc of points with radius \(1/r_{t}(\theta ).\) The second arc is obtained from the first by the map \(\lambda \mapsto 1/\lambda \) and therefore coincides with the points of radius \(r_{t}(\theta ).\) We can now see that the boundary of Biane’s domain coincides with the boundary of the domain we have defined. A similar analysis applies to the cases \(t>4\) and \(t=4,\) using the description of the boundary of \(\Sigma _{t}\) in those cases at the top of p. 274 in [3]. \(\square \)
4 The PDE for S
In this section, we will verify the PDE for S in Theorem 2.7. The claimed initial condition (2.13) holds because \(b_{0}=1.\) We now proceed to verify the Eq. (2.12) itself. We let \(b_{t}\) be the free multiplicative Brownian motion, which satisfies the free stochastic differential equation
Throughout the rest of this section, we use the notation
Lemma 4.1
The function S in (2.10) satisfies
Proof
The basic tools for computing with SDEs involving the free circular Brownian motion \(c_{t}\) are the free Itô formulas, which may be stated informally as
for a continuous adapted process \(g_{t}.\) Free stochastic calculus was developed by Biane and Speicher [4] and extended by Kümmerer and Speicher [21]. We will specifically use the free stochastic product rule and free functional Itô formula developed by Nikitopoulos [24]. (An earlier version of this paper, available on the arXiv, used a power series argument in place of Nikitopoulos’s result).
For each \(\lambda \in \mathbb {C},\) define a self-adjoint element \(m_{t}\) by
Then by the free stochastic product rule [24, Thm. 3.2.5] and the free SDE for \(b_{t},\) we have
Then by the free functional Itô formula [24, Thm. 3.5.3], we have
where
Noting that \(S=\tau [\log (m_{t}+\varepsilon )],\) and using (4.2) and (4.3), (4.4) becomes
Equation (4.5) is actually just of Eq. (33) in Example 3.5.5 of [24] with \(n=1\),
To compute further, we note that
Multiplying by \((b_{t,\lambda }^{*}b_{t,\lambda }+\varepsilon )^{-1}\) on the right and \((b_{t,\lambda }b_{t,\lambda }^{*}+\varepsilon )^{-1}\) on the left gives a useful identity:
Replacing \(b_{t,\lambda }\) by its adjoint gives another version of the identity:
We also claim that
This result can be verified for large \(\varepsilon \) by expanding both sides in powers of \(1/\varepsilon \) and checking the identity term by term. The result for general \(\varepsilon \) then follows by analyticity of both sides in \(\varepsilon .\)
We now use (4.7) to show that
Thus, (4.5) becomes the claimed formula (4.1) for \(\partial S/\partial t.\) \(\square \)
Lemma 4.2
We have the following formulas for the derivatives of S with respect to \(\varepsilon \) and \(\lambda \):
Proof
The lemma follows easily from the formula for the derivative of the trace of a logarithm (Lemma 1.1 in [5]):
(We emphasize that there is no such simple formula for the derivative of \(\log (f(u))\) without the trace, unless df/du commutes with f(u)). \(\square \)
We are now ready for the proof of our main result.
Proof of Theorem 2.7
We start from the formula for \(\partial S/\partial t\) in Lemma 4.1. Noting that
we expand the second factor on the right-hand side of (4.1) as
We then simplify the first term by writing \(b_{t,\lambda }b_{t,\lambda }^{*}=b_{t,\lambda }b_{t,\lambda }^{*}+\varepsilon -\varepsilon .\) In the middle two terms, we use (4.6), (4.7), and cyclic invariance of the trace. Using also (4.8), we get
Thus,
All terms on the right-hand side of (4.10) are expressible using Lemma 4.2 in terms of derivatives of S, and the claimed differential equation follows. \(\square \)
5 The Hamilton–Jacobi method
5.1 Setting up the method
The Eq. (2.12) is a first-order, nonlinear PDE of Hamilton–Jacobi type. (The reader may consult, for example, Section 3.3 in the book of Evans [8], but we will give a brief self-contained account of the theory in the proof of Proposition 5.3). We consider a Hamiltonian function obtained from the right-hand side of (2.12) by replacing each partial derivative with momentum variable, with an overall minus sign. Thus, we define
We then consider Hamilton’s equations for this Hamiltonian. That is to say, we consider this system of six coupled ODEs:
As convenient, we will let
The initial conditions for a, b, and \(\varepsilon \) are arbitrary:
while those for \(p_{a},\) \(p_{b},\) and \(p_{\varepsilon }\) are determined by those for a, b, and \(\varepsilon \) as follows:
where
The motivation for (5.4) is that the momentum variables \(p_{a},\) \(p_{b},\) and \(p_{\varepsilon }\) will correspond to the derivatives of S along the curves \((a(t),b(t),\varepsilon (t))\); see (5.8). Thus, the initial momenta are simply the derivatives of the initial value (2.13) of S, evaluated at \((a_{0},b_{0},\varepsilon _{0}).\)
For future reference, we record the value \(H_{0}\) of the Hamiltonian at time \(t=0\).
Lemma 5.1
The value of the Hamiltonian at \(t=0\) is
Proof
Plugging \(t=0\) into (5.1) and using (5.4 ) gives
which simplifies to
But using the formula (5.5) for \(p_{0},\) we see that \(a_{0} ^{2}-2a_{0}+b_{0}^{2}+\varepsilon _{0}\) equals \(1/p_{0}-1,\) from which (5.6) follows. \(\square \)
The main result of this section is the following; the proof is given on p. 22.
Theorem 5.2
Assume \(\lambda _{0}\ne 0\) and \(\varepsilon _{0}>0.\) Suppose a solution to the system (5.2) with initial conditions (5.3) and (5.4) exists with \(\varepsilon (t)>0\) for \(0\le t<T.\) Then we have
for all \(t\in [0,T).\) Furthermore, the derivatives of S with respect to a, b, and \(\varepsilon \) satisfy
Note that \(S(t,\lambda ,\varepsilon )\) is only defined for \(\varepsilon >0.\) Thus, (5.7) and (5.8) only make sense as long as the solution to (5.2) exists with \(\varepsilon (t)>0.\)
Since our objective is to compute \(\Delta s_{t}(\lambda )=\partial ^{2} s_{t}/\partial a^{2}+\partial ^{2}s_{t}/\partial ^{2}b^{2},\) the formula (5.8) for the derivatives of S will ultimately be of as great importance as the formula (5.7) for S itself. We emphasize that we are not using the Hamilton–Jacobi method to construct a solution to (2.12); the function \(S(t,\lambda ,\varepsilon )\) is already defined in (2.10) in terms of free probability and is known (Theorem 2.7) to satisfy (2.12). Rather, we are using the Hamilton–Jacobi method to analyze a solution that is already known to exist.
We begin by briefly recapping the general form of the Hamilton–Jacobi method.
Proposition 5.3
Fix an open set \(U\subset \mathbb {R}^{n},\) a time-interval [0, T], and a function \(H(\mathbf {x},\mathbf {p}).\) Consider a function \(S(t,\mathbf {x})\) satisfying
Consider a pair \((\mathbf {x}(t),\mathbf {p}(t))\) with \(\mathbf {x}(t)\in U,\) \(\mathbf {p}(t)\in \mathbb {R}^{n},\) and \(t\in [0,T_{1}]\) with \(T_{1}\le T.\) Assume this pair satisfies Hamilton’s equations:
with initial conditions
Then we have
and
Again, we are not trying to construct solutions to (5.9), but rather to analyze a solution that is already assumed to exist.
Proof
Take an arbitrary (for the moment) smooth curve \(\mathbf {x}(t)\) and note that
where we use the Einstein summation convention. Let us use the notation
that is \(p_{j}(t)=\partial S/\partial x_{j}(t,\mathbf {x}(t)).\) Then (5.13) may be rewritten as
If we can choose \(\mathbf {x}(t)\) so that \(\mathbf {p}(t)\) is somehow computable, then the right-hand side of (5.14) would be known and we could integrate to get \(S(t,\mathbf {x}(t)).\)
To see how we might be able to compute \(\mathbf {p}(t),\) we try differentiating:
Now, from (5.9), we have
Thus, (5.15) becomes (suppressing the dependence on the path)
If we now take \(\mathbf {x}(t)\) to satisfy
the second term on the right-hand side of (5.16) vanishes, and we find that \(\mathbf {p}(t)\) satisfies
With this choice of \(\mathbf {x}(t),\) (5.14) becomes
because H is constant along the solutions to Hamilton’s equations.
Note that not all solutions \((\mathbf {x}(t),\mathbf {p}(t))\) to Hamilton’s Eqs. (5.17) and (5.18) will arise by the above method. After all, we are assuming that \(\mathbf {p}(t)=(\nabla _{\mathbf {x}}S)(t,\mathbf {x} (t)),\) from which it follows that the initial conditions \((\mathbf {x} _{0},\mathbf {p}_{0})\) will be of the form in (5.10).
On the other hand, suppose we take a pair \((\mathbf {x}_{0},\mathbf {p}_{0})\) as in (5.10). Let us then take \(\mathbf {x}(t)\) to be the solution to
where since S is a fixed, “known” function, this ODE for \(\mathbf {x}(t)\) will have unique solutions for as long as they exist. If we set \(\mathbf {p}(t)=(\nabla _{\mathbf {x} }S)(t,\mathbf {x}(t)),\) then \(\mathbf {p}(0)=\mathbf {p}_{0}\) as in (5.10) and (5.19) says that the pair \((\mathbf {x} (t),\mathbf {p}(t))\) satisfies the first of Hamilton’s equations. Applying (5.16) with this choice of \(\mathbf {x}(t)\) shows that the pair \((\mathbf {x}(t),\mathbf {p}(t))\) also satisfies the second of Hamilton’s equations. Thus, \((\mathbf {x}(t),\mathbf {p}(t))\) must be the unique solution to Hamilton’s equations with the given initial condition \((\mathbf {x}_{0},\mathbf {p}_{0}).\)
We conclude that for any solution to Hamilton’s equations with initial conditions of the form (5.10), the formula (5.14) holds. Since, also, H is constant along solutions to Hamilton’s equations, we may replace \(H(\mathbf {x}(t),\mathbf {p}(t))\) by \(H(\mathbf {x} _{0},\mathbf {p}_{0})\) in (5.14), at which point, integration with respect to t gives (5.11). Finally, (5.12) holds by the definition of \(\mathbf {p}(t).\) \(\square \)
We are now ready for the proof of Theorem 5.2.
Proof of Theorem 5.2
We apply Proposition 5.3 with \(n=3\) and the open set U consisting of triples \((a,b,\varepsilon )\) with \(\varepsilon >0\). The PDE (2.12) is of the type in (5.9), with H given by (5.1). The initial conditions (5.4) are obtained by differentiating the initial condition \(S(0,\lambda ,\varepsilon )=\log (\left| \lambda -1\right| ^{2}+\varepsilon ).\)
We let \(\mathbf {x}(t)=(a(t),b(t),\varepsilon (t))\) and \(\mathbf {p} (t)=(p_{a}(t),p_{b}(t),p_{\varepsilon }(t)).\) For the case of the Hamiltonian (5.1), a simple computation shows that
Thus, the general formula (5.11) becomes, in this case,
But we also may compute that
Thus,
If we now plug in the value of \(S(0,\mathbf {x}_{0})=S(0,\lambda _{0} ,\varepsilon _{0})\) and use Lemma 5.1 along with the definition (5.5) of \(p_{0},\) we obtain (5.7). Finally, (5.8) is just the general formula (5.12), applied to the case at hand. \(\square \)
5.2 Constants of motion
We now identify several constants of motion for the system (5.2), from which various useful formulas can be derived. Throughout the section, we assume we have a solution to (5.2) with the initial conditions (5.3) and (5.4), defined on a time-interval of the form \(0\le t<T.\) We continue the notation \(\lambda (t)=a(t)+ib(t).\)
Proposition 5.4
Along any solution of (5.2), following quantities remain constant:
-
(1)
The Hamiltonian H,
-
(2)
The “angular momentum” in the (a, b) variables, namely \(ap_{b}-bp_{a},\) and
-
(3)
The argument of \(\lambda ,\) assuming \(\lambda _{0}\ne 0.\)
Proof
For any system of the form (5.2), the Hamiltonian H itself is a constant of motion, as may be verified easily from the equations. The conservation of the angular momentum is a consequence of the invariance of H under simultaneous rotations of (a, b) and \((p_{a},p_{b})\); see Proposition 2.30 and Conclusion 2.31 in [14]. This result can also be verified by direct computation from (5.2).
Finally, note from (5.21) that if \(\lambda _{0}\ne 0,\) then \(\log \left| \lambda (t)\right| \) remains finite as long as the solution to (5.2) exists, so that \(\lambda (t)\) cannot pass through the origin. We then compute that
(If \(a=0,\) we instead compute the time-derivative of \(\cot (\arg \lambda ),\) which also equals zero). \(\square \)
Proposition 5.5
The Hamiltonian H in (5.1) in invariant under the one-parameter group of symplectic linear transformations given by
with \(\sigma \) varying over \(\mathbb {R}.\) Thus, the generator of this family of transformations, namely,
is a constant of motion for the system (5.2). The constant \(\Psi \) may be computed in terms of \(\varepsilon _{0}\) and \(\lambda _{0}\) as
where \(p_{0}\) is as in (5.5).
Proof
The claimed invariance of H is easily checked from the formula (5.1). One can easily check that \(\Psi \) is the generator of this family. That is to say, if we replace H by \(\Psi \) in (5.2), the solution is given by the map in (5.22). Thus, by a simple general result, \(\Psi \) will be a constant of motion; see Conclusion 2.31 in [14]. Of course, one can also check by direct computation that the function in (5.23) is constant along solutions to (5.2). The expression (5.24) then follows easily from the initial conditions in (5.4). \(\square \)
Proposition 5.6
For all t, we have
where \(C=2\Psi -1\) and \(\Psi \) is as in (5.23). The constant C in (5.25) may be computed in terms of \(\varepsilon _{0}\) and \(\lambda _{0}\) as
Proof
We compute that
and then that
This result simplifies to
The unique solution to this equation is (5.25). The expression (5.26) is obtained by evaluating \(\Psi \) at \(t=0,\) using the initial conditions (5.4), and simplifying. \(\square \)
We now make an important application of preceding results.
Theorem 5.7
Suppose a solution to (5.2) exists with \(\varepsilon (t)>0\) for \(0\le t<t_{*},\) but that \(\lim _{t\rightarrow t^{*}}\varepsilon (t)=0.\) Then
where \(C=2\Psi -1\) is as in Proposition 5.6. Furthermore, we have
Equation (5.29) is a key step in the derivation of our main result; see Sect. 6.1. We will write (5.28) in a more explicit way in Proposition 5.12, after the time \(t_{*}\) has been determined. We note also from Proposition 5.6 that since \(\varepsilon (t)\) approaches zero as t approaches \(t_{*},\) then \(p_{\varepsilon }(t)\) must be blowing up, so that \(\varepsilon (t)p_{\varepsilon }(t)^{2}\) can remain positive in this limit.
Proof
Using the constant of motion \(\Psi \) in (5.23), we can rewrite the Hamiltonian H as
Now, by assumption, the variable \(\varepsilon \) approaches zero as t approaches \(t_{*}.\) Furthermore, by Proposition 5.6, \(\varepsilon p_{\varepsilon }^{2}\) remains finite in this limit, so that \(\varepsilon p_{\varepsilon }=\sqrt{\varepsilon }\sqrt{\varepsilon p_{\varepsilon }^{2}}\) tends to zero. Thus, in the \(t\rightarrow t_{*}\) limit, the \(\varepsilon p_{\varepsilon }\) terms in (5.30) vanish while \(\varepsilon p_{\varepsilon }^{2}\) remains finite, leaving us with
Since H is a constant of motion, we may write this result as
where we have used Lemma 5.1 in the second equality and Proposition 5.6 in the third. The formula (5.28) follows.
Meanwhile, as t approaches \(t_{*},\) the \(\varepsilon p_{\varepsilon }\) term in the formula (5.23) for \(\Psi \) vanishes and we find, using (5.28), that
as claimed in (5.29), where we have used (5.28) in the last equality. \(\square \)
5.3 Solving the equations
We now solve the system (5.2) subject to the initial conditions (5.3) and (5.4). The formula in Proposition 5.6 for \(\varepsilon (t)p_{\varepsilon }(t)^{2}\) will be a key tool. Although we are mainly interested in the case \(\varepsilon _{0}>0,\) we will need in Sect. 6.2 to allow \(\varepsilon _{0}\) to be slightly negative.
We begin by with the following elementary lemma.
Lemma 5.8
Consider a number \(a^{2}\in \mathbb {R}\) and let a be either of the two square roots of \(a^{2}.\) Then the solution to the equation
subject to the initial condition \(y(0)=y_{0}>0\) is
If \(a^{2}\ge y_{0}^{2},\) the solution exists for all \(t>0.\) If \(a^{2} <y_{0}^{2}\), then y(t) is a strictly increasing function of t until the first positive time \(t_{*}\) at which the solution blows up. This time is given by
Here, we use the principal branch of the inverse hyperbolic tangent, with branch cuts \((-\infty ,-1]\) and \([1,\infty )\) on the real axes, which corresponds to using the principal branch of the logarithm. When \(a=0,\) we interpret the right-hand side of (5.33) or (5.34) as having its limiting value as a approaches zero, namely \(1/y_{0}\).
In passing from (5.33) to (5.34), we have used the standard formula for the inverse hyperbolic tangent,
In (5.32), we interpret \(\sinh (at)/a\) as having the value t when \(a=0.\) If \(a^{2}<0,\) so that a is pure imaginary, one can rewrite the solution in terms of ordinary trigonometric functions, using the identities \(\cosh (i\alpha )=\cos \alpha \) and \(\sinh (i\alpha )=i\sin \alpha .\) For each fixed t, the solution is an even analytic function of a and therefore an analytic function of \(a^{2}.\)
Proof
If a is nonzero and real, we may integrate (5.31) to obtain
It is then straightforward to solve for y(t) and simplify to obtain (5.32). Similar computations give the result when a is zero (recalling that we interpret \(\sinh (at)/a\) as equaling t when \(a=0\)) and when a is nonzero and pure imaginary. Alternatively, one may check by direct computation that the function on the right-hand side of (5.32) satisfies the Eq. (5.31) for all \(a\in \mathbb {C}.\)
Now, if \(a^{2}\ge y_{0}^{2}>0,\) the denominator in (5.32) is easily seen to be nonzero for all t and there is no singularity. If \(a^{2}\) is positive but less than \(y_{0}^{2},\) the denominator remains positive until it becomes zero when \(\tanh (at)=a/y_{0}.\) If \(a^{2}\) is negative, so that \(a=i\alpha \) for some nonzero \(\alpha \in \mathbb {R}\), we write the solution using ordinary trigonometric functions as
The denominator in (5.36) becomes zero at \(\alpha t=\tan ^{-1} (\alpha /y_{0})<\pi /2.\) Finally, if \(a^{2}=0,\) the solution is \(y(t)=y_{0} /(1-y_{0}t),\) which blows up at \(t=1/y_{0}.\)
It is then not hard to check that for all cases with \(a^{2}<y_{0}^{2},\) the blow-up time can be computed as \(t_{*}=\frac{1}{a}\tanh ^{-1}(a/y_{0}),\) where we use the principal branch of the inverse hyperbolic tangent, with branch cuts \((-\infty ,-1]\) and \([1,\infty )\) on the real axis. (At \(a=0\) we have a removable singularity with a value of \(1/y_{0}.\)) This recipe corresponds to using the principal branch of the logarithm in the last expression in (5.34). \(\square \)
We now apply Lemma 5.8 to compute the \(p_{\varepsilon } \)-component of the solution to (5.2). We use the following notations, some of which have been introduced previously:
We now make the following standing assumptions:
We note that under these assumptions, \(y_{0}\) is positive. Furthermore, we may compute that
from which we obtain
so that \(a^{2}<y_{0}^{2}.\) Now, the assumptions \(p_{0}>0\) and \(\delta >0\) can be written as \(\varepsilon _{0}>-\left| \lambda _{0}-1\right| ^{2}\) and \(\varepsilon _{0}>-(1+\left| \lambda _{0}\right| ^{2}).\) Thus, for \(\lambda _{0}\ne 0,\) the assumptions (5.42) are always satisfied if \(\varepsilon _{0}>0.\) Furthermore, except when \(\lambda _{0}=1,\) some negative values of \(\varepsilon _{0}\) are allowed.
Proposition 5.9
Under the assumptions (5.42), the \(p_{\varepsilon }\)-component of the solution to (5.2) subject to the initial conditions (5.3) and (5.4) is given by
for as long as the solution to the system (5.2) exists. Here we write a as in (5.43) and we use the same choice of \(\sqrt{\delta ^{2}-4}\) in the computation of a as in the two times \(\sqrt{\delta ^{2}-4}\) appears explicitly in (5.45). If \(\delta =2,\) we interpret \(\sinh (at)/\sqrt{\delta ^{2}-4}\) as equaling \(\frac{1}{2} p_{0}\left| \lambda _{0}\right| t.\)
If \(\varepsilon _{0}\ge 0,\) the numerator in the fraction on the right-hand side of (5.45) is positive for all t. Hence when \(\varepsilon _{0}\ge 0,\) we see that \(p_{\varepsilon }(t)\) is positive for as long as the solution exists and \(1/p_{\varepsilon }(t)\) extends to a real-analytic function of t defined for all \(t\in \mathbb {R}.\)
The first time \(t_{*}(\lambda _{0},\varepsilon _{0})\) at which the expression on the right-hand side of (5.45) blows up is
where \(\theta _{0}=\arg \lambda _{0}\) and \(\sqrt{\delta ^{2}-4}\) is either of the two square roots of \(\delta ^{2}-4.\) The principal branch of the inverse hyperbolic tangent should be used in (5.46), with branch cuts \((-\infty ,-1]\) and \([1,\infty )\) on the real axis, which corresponds to using the principal branch of the logarithm in (5.47). When \(\delta =2,\) we interpret \(t_{*}(\lambda _{0},\varepsilon _{0})\) as having its limiting value as \(\delta \) approaches 2, namely \(\delta -2\cos \theta _{0}.\)
Note that the expression
is an even function of a with b fixed, with a removable singularity at \(a=0.\) This expression is therefore an analytic function of \(a^{2}\) near the origin. In particular, the value of \(t_{*}(\lambda _{0},\varepsilon _{0})\) does not depend on the choice of square root of \(\delta ^{2}-4.\)
Proof of Proposition 5.9
We assume at first that \(\varepsilon _{0}\ne 0.\) We recall from Proposition 5.6 that \(\varepsilon (t)p_{\varepsilon }(t)^{2}\) is equal to \(\varepsilon _{0}p_{0} ^{2}e^{-Ct},\) which is never zero, since we assume \(\varepsilon _{0}\) is nonzero and \(p_{0}\) is positive. Thus, as long as the solution to the system (5.2) exists, both \(\varepsilon (t)\) and \(p_{\varepsilon }(t)\) must be nonzero—and must have the same signs they had at \(t=0.\) Using (5.27) and the fact that H is a constant of motion, we obtain
But \(\varepsilon _{0}p_{0}^{2}/\varepsilon (t)=p_{\varepsilon }(t)^{2}e^{Ct}\) and we obtain
Then if \(y(t)=e^{Ct}p_{\varepsilon }(t)+C/2,\) we find that y satisfies (5.31). Thus, we obtain \(p_{\varepsilon } (t)=(y(t)-C/2)e^{-Ct},\) where y(t) is as in (5.32), which simplifies to the claimed formula for \(p_{\varepsilon }.\) The same formula holds for \(\varepsilon _{0}=0\), by the continuous dependence of the solutions on initial conditions. (It is also possible to solve the system (5.2 ) with \(\varepsilon _{0}=0\) by postulating that \(\varepsilon (t)\) is identically zero and working out the equations for the other variables).
In this paragraph only, we assume \(\varepsilon _{0}\ge 0.\) Then \(a^{2}\ge 0,\) with \(a=0\) occurring only if \(\varepsilon _{0}=0\) and \(\left| \lambda _{0}\right| =1,\) so that \(\delta =2.\) In that case, the numerator on the right-hand side of (5.45) is identically equal to 1. If \(a^{2}>0,\) then the numerator will always be positive provided that
which is equivalent to
But a computation shows that
and we are assuming \(\varepsilon _{0}\ge 0.\) Now, since the numerator in (5.45) is always positive, we conclude that \(p_{\varepsilon }\) remains positive until it blows up.
For any value of \(\varepsilon _{0},\) the blow-up time for the function on the right-hand side of (5.45) is computed by plugging the expression (5.44) for \(a/y_{0}\) into the formula (5.34), giving
After computing that
we obtain the claimed formula (5.46) for \(t_{*}(\lambda _{0},\varepsilon _{0}).\) \(\square \)
Remark 5.10
If \(\varepsilon _{0}<0,\) then numerator on the right-hand side of (5.45) can become zero. The time \(\sigma \) at which this happens is computed using (5.43) and (5.48) as
By considering separately the cases \(\left| \lambda _{0}\right| \ne 1\) and \(\left| \lambda _{0}\right| =1,\) we can verify that \(\sigma \) tends to infinity, locally uniformly in \(\lambda _{0},\) as \(\varepsilon _{0}\) tends to zero from below. Thus, for small negative values of \(\varepsilon _{0},\) the function on the right-hand side of (5.45) will remain positive until the time \(t_{*}(\lambda _{0},\varepsilon _{0})\) at which it blows up.
We now show that the whole system (5.2) has a solution up to the time at which the function on the right-hand side of (5.45) blows up.
Proposition 5.11
Assume that \(\varepsilon _{0}\) and \(\lambda _{0}\) satisfy the assumptions (5.42). Assume further that if \(\varepsilon _{0}<0,\) then \(\left| \varepsilon _{0}\right| \) is sufficiently small that \(p_{\varepsilon }\) remains positive until it blows up, as in Remark 5.10. Then the solution to the system (5.2) exists up to the time \(t_{*}(\lambda _{0},\varepsilon _{0})\) in Proposition 5.9.
For any \(\varepsilon _{0},\) we have
If \(\varepsilon _{0}=0,\) the solution has \(\varepsilon (t)\equiv 0\) and \(\lambda (t)\equiv \lambda _{0}.\)
Proof
Let T be the maximum time such that the solution to (5.2) exists on [0, T). We now compute formulas for the solution on this interval. Recall from Proposition 5.9 that if \(\varepsilon _{0}\ge 0,\) then \(p_{\varepsilon }(t)\) remains positive for as long as the solution exists; by Remark 5.10, the same assertion holds if \(\varepsilon _{0}\) is small and negative.
Now, since \(\varepsilon p_{\varepsilon }^{2}=\varepsilon _{0}p_{0}^{2}e^{-Ct}\), we see that
Since \(p_{\varepsilon }(t)\) remains positive until it blows up, \(\varepsilon (t)\) remains bounded until time \(t_{*}(\lambda _{0},\varepsilon _{0}),\) at which time \(\varepsilon (t)\) approaches zero, as claimed in (5.49 ). We recall from Proposition 5.4 that the argument of \(\lambda (t)\) remains constant. Then as in shown in (5.21), we have
Finally,
which is a first-order, linear equation for \(p_{a},\) which can be solved using an integrating factor. A similar calculation applies to \(p_{b}.\)
Suppose now that the existence time T of the whole system were smaller than the time \(t_{*}(\lambda _{0},\varepsilon _{0})\) at which the right-hand side of (5.45) blows up. Then from the formulas (5.50), (5.51), and (5.52), we see that all functions involved would remain bounded up to time T. But then by a standard result, T could not actually be the maximal time. The solution to the system (5.2) must therefore exist all the way up to time \(t_{*}(\lambda _{0} ,\varepsilon _{0}).\)
Finally, we note that when \(\varepsilon _{0}=0,\) (5.50) gives \(\varepsilon (t)\equiv 0\) and (5.51) gives \(\left| \lambda (t)\right| \equiv \left| \lambda _{0}\right| .\) Since also the argument of \(\lambda (t)\) is constant, we see that \(\lambda (t)\equiv \lambda _{0}.\) \(\square \)
5.4 More about the lifetime of the solution
In light of Propositions 5.9 and 5.11, the lifetime of the solution to the system (5.2) is \(t_{*} (\lambda _{0},\varepsilon _{0}),\) as computed in (5.46) or (5.47). In this subsection, we (1) analyze the behavior of \(\log \left| \lambda (t)\right| \) as t approaches \(t_{*} (\lambda _{0},\varepsilon _{0}),\) (2) analyze the behavior of \(t_{*} (\lambda _{0},\varepsilon _{0})\) as \(\varepsilon _{0}\) approaches zero, and (3) show that \(t_{*}(\lambda _{0},\varepsilon _{0})\) is an increasing function of \(\varepsilon _{0}\) with \(\lambda _{0}\) fixed.
Proposition 5.12
Assume that \(\varepsilon _{0}\) and \(\lambda _{0}\) satisfy the assumptions (5.42). Then
where \(\delta \) is as in (5.38).
Notice that there is a strong similarity between the formula (5.47) for \(t_{*}(\lambda _{0},\varepsilon _{0})\) and the expression on the right-hand side of (5.53).
Proof
By (5.28) in Theorem 5.7, we have \(\lim _{t\rightarrow t_{*}(\lambda _{0},\varepsilon _{0})}\log \left| \lambda (t)\right| =Ct_{*}(\lambda _{0},\varepsilon _{0})/2,\) where by (5.26),
From this result and the second expression (5.47) for \(t_{*}(\lambda _{0},\varepsilon _{0})\), (5.53) follows easily. \(\square \)
Proposition 5.13
If \(t_{*}(\lambda _{0},\varepsilon _{0})\) is defined by (5.47), then for all nonzero \(\lambda _{0}\) we have
where the function T is defined in (3.1). Furthermore, when \(\varepsilon _{0}=0,\) we have
Recall that the formula for \(t_{*}(\lambda _{0},\varepsilon _{0})\) is defined under the standing assumptions in (5.42). Note that for all \(\lambda _{0}\ne 0,\) the value \(\varepsilon _{0}=0\) satisfies these assumptions.
Since \(\log (x)/(x-1)\rightarrow 1\) as \(x\rightarrow 1,\) we see that \(t_{*}(\lambda _{0},0)\) is a continuous function of \(\lambda _{0}\in \mathbb {C}^{*}.\) Comparing the formula for \(t_{*}(\lambda _{0},0)\) to Theorem 3.1, we have the following consequence.
Corollary 5.14
For \(\lambda _{0}\in \Sigma _{t},\) we have \(t_{*} (\lambda _{0},0)<t,\) while for \(\lambda _{0}\in \partial \Sigma _{t},\) we have \(t_{*}(\lambda _{0},0)=t\), and for \(\lambda _{0}\notin \overline{\Sigma }_{t},\) we have \(t_{*}(\lambda _{0},0)>t.\)
Proof of Proposition 5.13
In the limit as \(\varepsilon _{0}\rightarrow 0,\) we have
and
so that
In the case \(\left| \lambda _{0}\right| =1,\) the limiting value of \(\delta \) is 2. We then make use of the elementary limit
Thus, using (5.47), we obtain in this case,
which agrees with the value of \(T(\lambda _{0})\) when \(\left| \lambda _{0}\right| =1.\)
In the case \(\left| \lambda _{0}\right| \ne 1,\) we note that the quantity \((1/b)\log ((a+b)/(a-b))\) is an even function of b with a fixed. We may therefore choose the plus sign on the right-hand side of (5.55), regardless of the sign of \(\left| \lambda _{0}\right| ^{2}-1.\) We then obtain, using (5.47),
A similar calculation, beginning from (5.53), establishes (5.54). \(\square \)
Remark 5.15
If we began with (5.46) instead of (5.47), we would obtain by similar reasoning
Using (5.35), this expression is easily seen to agree with \(T(\lambda _{0})\) but is more transparent in its behavior at \(\left| \lambda _{0}\right| =1.\)
Proposition 5.16
For each \(\lambda _{0},\) the function \(t_{*}(\lambda _{0},\varepsilon _{0})\) is a strictly increasing function of \(\varepsilon _{0}\) for \(\varepsilon _{0}\ge 0,\) and
Proof
We note that the quantity \(\delta \) in (5.38) is an increasing function of \(\varepsilon _{0}\) with \(\lambda _{0}\) fixed, with \(\delta \) tending to infinity as \(\varepsilon _{0}\) tends to infinity. We note also that if \(\varepsilon _{0}\ge 0,\) then
It therefore suffices to show that for each angle \(\theta _{0},\) the function
is strictly increasing, non-negative, continuous function of \(\delta \) for \(\delta \ge 2\) that tends to \(+\infty \) as \(\delta \) tends to infinity. Here when \(\delta =2,\) we interpret \(g_{\theta _{0}}(\delta )\) as having the value \(2-2\cos \theta _{0},\) in accordance with the limit (5.56).
Throughout the proof, we use the notation
We note that
Meanwhile, for large \(\delta ,\) we have
whereas
Thus, \(g_{\theta _{0}}(\delta )\) grows like \(\log (\delta ^{2})\) as \(\delta \rightarrow \infty \).
Our definition of \(g_{\theta _{0}}(\delta )\) for \(\delta =2,\) together with (5.56), shows that \(g_{\theta _{0}}\) is non-negative and continuous there. To show that \(g_{\theta _{0}}\) is an increasing function of \(\delta ,\) we show that \(\partial g_{\theta _{0}}/\partial \delta \) is positive for \(\delta >2.\) The derivative is computed, after simplification, as
Since this expression depends linearly on \(\cos \theta _{0}\) with \(\delta \) fixed, if it is positive when \(\cos \theta _{0}=1\) and also when \(\cos \theta _{0}=-1,\) it will be positive always. Thus, it suffices to verify the positivity of the functions
and
Now, (5.59) is clearly positive for all \(\delta >2.\) Meanwhile, a computation shows that
and
from which we conclude that (5.60) is also positive for all \(\delta >2.\) \(\square \)
5.5 Surjectivity
In Sect. 6.3, we will compute \(s_{t}(\lambda ):=\lim _{\varepsilon \rightarrow 0^{+}}S(t,\lambda ,\varepsilon )\) for \(\lambda \) in \(\Sigma _{t}.\) We will do so by evaluating S (and its derivatives) along curves of the form \((t,\lambda (t),\varepsilon (t))\) and then the taking the limit as we approach the time \(t_{*}\) when \(\varepsilon (t)\) becomes zero. For this method to be successful, we need the following result, whose proof appears on p. 34.
Theorem 5.17
Fix \(t>0.\) Then for all \(\lambda \in \Sigma _{t},\) there exists a unique \(\lambda _{0}\in \mathbb {C}\) and \(\varepsilon _{0}>0\) such that the solution to (5.2) with these initial conditions exists on [0, t) with \(\lim _{u\rightarrow t^{-}}\varepsilon (u)=0\) and \(\lim _{u\rightarrow t^{-}}\lambda (u)=\lambda .\) For all \(\lambda \in \Sigma _{t},\) the corresponding \(\lambda _{0}\) also belongs to \(\Sigma _{t}.\)
Define functions \(\Lambda _{0}^{t}:\Sigma _{t}\rightarrow \Sigma _{t}\) and \(E_{0}^{t}:\Sigma _{t}\rightarrow (0,\infty )\) by letting \(\Lambda _{0} ^{t}(\lambda )\) and \(E_{0}^{t}(\lambda )\) be the corresponding values of \(\lambda _{0}\) and \(\varepsilon _{0},\) respectively. Then \(\Lambda _{0}^{t}\) and \(E_{0}^{t}\) extend to continuous maps of \(\overline{\Sigma }_{t}\) into \(\overline{\Sigma }_{t}\) and \([0,\infty )\), respectively, with the continuous extensions satisfying \(\Lambda _{t}(\lambda )=\lambda \) and \(E_{0}^{t} (\lambda )=0\) for \(\lambda \in \partial \Sigma _{t}.\)
We first recall that we have shown (Proposition 5.16) that the lifetime of the path to be a strictly increasing function of \(\varepsilon _{0}\ge 0\) with \(\lambda _{0}\) fixed. If \(\lambda _{0}\) is outside \(\Sigma _{t},\) then by Theorem 3.1 and Proposition 5.13, the lifetime is at least t, even at \(\varepsilon _{0}=0.\) (That is to say, \(T(\lambda _{0})=t_{*}(\lambda _{0},0)\ge t\) for \(\lambda _{0}\) outside \(\Sigma _{t}.\)) Thus, for \(\lambda _{0}\) outside \(\Sigma _{t}\), the lifetime cannot equal t for \(\varepsilon _{0}>0.\) On the other hand, if \(\lambda _{0}\in \Sigma _{t},\) then \(t_{*}(\lambda _{0},0)<t\) and Proposition 5.16 tells us that there is a unique \(\varepsilon _{0}>0\) with \(t_{*}(\lambda _{0},\varepsilon _{0})=t.\)
Lemma 5.18
Fix \(t>0.\) Define maps
as follows. For \(\lambda _{0}\in \Sigma _{t},\) we let \(\varepsilon _{0} ^{t}(\lambda _{0})\) denote the unique positive value of \(\varepsilon _{0}\) for which \(t_{*}(\lambda _{0},\varepsilon _{0})=t.\) Then we set
where \(\lambda (\cdot )\) is computed with initial conditions \(\lambda (0)=\lambda _{0}\) and \(\varepsilon (0)=\varepsilon _{0}^{t}(\lambda _{0}).\) Then both \(\varepsilon _{0}^{t}\) and \(\lambda _{t}\) extend continuously from \(\Sigma _{t}\) to \(\overline{\Sigma }_{t},\) with the extended maps satisfying \(\varepsilon _{0}^{t}(\lambda _{0})=0\) and \(\lambda _{t}(\lambda _{0})=\lambda _{0}\) for \(\lambda _{0}\in \partial \Sigma _{t}.\) The extended map \(\lambda _{t}\) is a homeomorphism of \(\overline{\Sigma }_{t}\) to itself.
We note that the desired function \(\Lambda _{0}^{t}\) in Theorem 5.17 is the inverse function to \(\lambda _{t}\) and that \(E_{0}^{t}(\lambda )=\varepsilon _{0}^{t}(\lambda _{t}^{-1}(\lambda )).\)
Recall from Proposition 5.4 that the argument of \(\lambda (t)\) is constant. By the formula (5.28) in Theorem 5.7 together with the expression (5.26) for the constant C, we can write
where we have used that \(t_{*}(\lambda _{0},\varepsilon _{0}^{t}(\lambda _{0}))=t.\) As noted in the proof of Proposition 5.12, this formula can also be written as
where \(\delta =(\left| \lambda _{0}\right| ^{2}+1+\varepsilon _{0} ^{t}(\lambda _{0}))/\left| \lambda _{0}\right| .\)
Proof
We start by trying to compute the function \(\varepsilon _{0}^{t},\) which we will do by finding the correct value of \(\delta \) and then solving for \(\varepsilon _{0}^{t}.\) Recall that the lifetime \(t_{*}(\lambda _{0},\varepsilon _{0})\) is computed as \(g_{\theta _{0}}(\delta ),\) where \(\delta \) is as in (5.38) and \(g_{\theta _{0}}\) is as in (5.58). As we have computed in (5.57), we have
Assume, then, that the ray with angle \(\theta _{0}\) intersects \(\Sigma _{t}\) and let \(r_{t}(\theta _{0})\) be the outer (for definiteness) radius at which this ray intersects the boundary of \(\Sigma _{t}.\) Then Theorem 3.1 tells us that \(T(r_{t}(\theta _{0})e^{i\theta _{0}})=t,\) and we conclude that
Consider, then, some \(\lambda _{0}\in \Sigma _{t}\) with \(\arg (\lambda _{0} )=\theta _{0}.\) By the formula (5.47), to find \(\varepsilon _{0}\) with \(t_{*}(\lambda _{0},\varepsilon _{0})=t,\) we first find \(\delta \) so that \(g_{\theta _{0}}(\delta )=t.\) (Note that the value of \(\delta \) depends only on the argument of \(\lambda _{0}.\)) We then adjust \(\varepsilon _{0}\) so that \((\left| \lambda _{0}\right| ^{2}+\varepsilon _{0}+1)/\left| \lambda _{0}\right| =\delta .\) Since the correct value of \(\delta \) is given in (5.63), this means that we should choose \(\varepsilon _{0}\) so that
We can solve this relation for \(\varepsilon _{0}\) to obtain
Now, we have shown that \(r_{t}(\theta )\) is continuous for the full range of angles \(\theta \) occurring in \(\overline{\Sigma }_{t}.\) Since 0 is not in \(\overline{\Sigma }_{t}\), we can then see that the formula (5.64) is well defined and continuous on all of \(\overline{\Sigma }_{t}.\) For \(\lambda _{0}\in \partial \Sigma _{t},\) we have that \(\left| \lambda _{0}\right| \) equals \(r_{t}(\arg \lambda _{0})\) or \(1/r_{t}(\arg \lambda _{0}),\) so that \(\varepsilon _{0}^{t}(\lambda _{0})\) equals zero.
Now, the point 0 is always outside \(\overline{\Sigma }_{t},\) while the point 1 is always in \(\Sigma _{t}\) and therefore not on the boundary of \(\Sigma _{t}.\) Thus, since \(\varepsilon _{0}^{t}\) is continuous on \(\overline{\Sigma }_{t}\) and zero precisely on the boundary, we see from (5.61) that \(\lambda _{t}\) is continuous on \(\overline{\Sigma }_{t}.\) Furthermore, on \(\partial \Sigma _{t},\) we compute \(\lambda _{t}(\lambda _{0})\) by putting \(\varepsilon _{0}^{t}(\lambda _{0})=0\) in (5.61). Suppose now that \(\lambda _{0}\) is in \(\partial \Sigma _{t}.\) Then \(\varepsilon _{0} ^{t}(\lambda _{0})=0\) and, by Theorem 3.1, the function \(T(\lambda _{0})\) in (3.1) has the value t, so that
Thus, from (5.61), we see that \(\lambda _{t}(\lambda _{0})=\lambda _{0}.\)
Consider an angle \(\theta _{0}\) for which the ray \(\mathrm {Ray}(\theta _{0})\) with angle \(\theta _{0}\) intersects \(\overline{\Sigma }_{t}\) and let \(\delta \) be chosen so that \(g_{\theta _{0}}(\delta )=t,\) noting again that the value of \(\delta \) depends only on \(\theta _{0}=\arg \lambda _{0}.\) We now observe from (5.62) that \(\left| \lambda _{t}(\lambda _{0})\right| \) is a strictly increasing function of \(\left| \lambda _{0}\right| \) with \(\delta \) fixed. Thus, \(\lambda _{t}\) is a strictly increasing function of the interval \(\mathrm {Ray}(\theta _{0})\cap \overline{\Sigma }_{t}\) into \(\mathrm {Ray}(\theta _{0})\) that fixes the endpoints. Thus, actually, \(\lambda _{t}\) maps this interval bijectively into itself. Since this holds for all \(\theta _{0},\) we conclude that \(\lambda _{t}\) maps \(\overline{\Sigma }_{t}\) bijectively into itself. The continuity of the inverse then holds because \(\lambda _{t}\) is continuous and \(\overline{\Sigma }_{t}\) is compact. \(\square \)
Proof of Theorem 5.17
We have noted before the statement of Lemma 5.18 that if the desired pair \((\lambda _{0},\varepsilon _{0})\) exists, \(\lambda _{0}\) must be in \(\Sigma _{t}.\) The lemma then tells us that a unique pair \((\lambda _{0},\varepsilon _{0})\) exists with \(\lambda _{0}\in \Sigma _{t}.\) We compute \(\Lambda _{0}^{t}(\lambda )\) as \(\lambda _{t}^{-1}(\lambda )\) and \(E_{0} ^{t}(\lambda )\) as \(\varepsilon _{0}^{t}(\lambda _{t}^{-1}(\lambda )),\) both of which extend continuously to \(\overline{\Sigma }_{t}\). For \(\lambda \in \partial \Sigma _{t},\) we have \(\lambda _{t}^{-1}(\lambda )=\lambda \) and \(\varepsilon _{0}^{t}(\lambda _{t}^{-1}(\lambda ))=\varepsilon _{0}^{t} (\lambda )=0.\) \(\square \)
6 Letting \(\varepsilon \) tend to zero
6.1 Outline
Our goal is to compute the Laplacian with respect to \(\lambda \) of the function \(s_{t}(\lambda ):=\lim _{\varepsilon \rightarrow 0^{+}}S(t,\lambda ,\varepsilon ),\) using the Hamilton–Jacobi method of Theorem 5.2. We want the curve \(\varepsilon (\cdot )\) occurring in (5.7) and (5.8) to approach zero at time t; a simple way we might try to accomplish this is to let the initial condition \(\varepsilon _{0}\) approach zero. Suppose, then, that \(\varepsilon _{0}\) is very small. Using various formulas from Sect. 5.3, we then find that for as long as the solution to the system (5.2) exists, the whole curve \(\varepsilon (\cdot )\) will be small and the whole curve \(\lambda (\cdot )\) will be approximately constant. Thus, by taking \(\varepsilon _{0}\approx 0\) and \(\lambda _{0}\approx \lambda ,\) we obtain a curve with \(\varepsilon (t)\approx 0\) and \(\lambda (t)\approx \lambda .\) We may then hope to compute \(s_{t}(\lambda )\) by letting \(\lambda _{0}\) and \(\lambda (t)\) approach \(\lambda \) and \(\varepsilon _{0}\) approach zero in the Hamilton–Jacobi formula (5.7), with the result that
It is essential to note, however, that this approach is only valid if the solution to system (5.2) exists up to time t. Corollary 5.14 tells us that for \(\varepsilon _{0}\approx 0,\) the solution will exist beyond time t provided \(\lambda \) is outside \(\overline{\Sigma }_{t}.\) Thus, we expect that for \(\lambda \) outside \(\overline{\Sigma }_{t},\) the function \(s_{t}\) will be given by (6.1) and therefore that \(\Delta s_{t}\) will be zero. (The function \(\log (\left| \lambda -1\right| ^{2})\) is harmonic except at the point \(\lambda =1,\) which is always inside \(\Sigma _{t}.\))
To analyze \(s_{t}(\lambda )\) for \(\lambda \) inside \(\Sigma _{t},\) we first make use of the surjectivity result in Theorem 5.17. The theorem says that for each \(t>0\) and \(\lambda \in \Sigma _{t},\) there exist \(\varepsilon _{0}>0\) and \(\lambda _{0}\in \Sigma _{t}\) such that \(\varepsilon (u)\) approaches 0 and \(\lambda (u)\) approaches \(\lambda \) as u approaches t. We then use the formula (5.29) in Theorem 5.7. In light of the second Hamilton–Jacobi formula (5.8), we can write (5.29) as
Once we have established enough regularity in the function \(S(t,\lambda ,\varepsilon )\) near \(\varepsilon =0,\) we will be able to identify the left-hand side of (6.2) with the corresponding derivative of \(s_{t}\), giving the following explicit formula for one of the derivatives of \(s_{t}\):
We now compute in logarithmic polar coordinates, with \(\rho =\log \left| \lambda \right| \) and \(\theta =\arg \lambda .\) We may recognize the left-hand side of (6.3) as the derivative of \(s_{t}\) with respect to \(\rho ,\) giving
for points inside \(\Sigma _{t}.\) Remarkably, \(\partial s_{t}/\partial \rho \) is independent of \(\theta \)! Thus,
meaning that \(\partial s_{t}/\partial \theta \) is independent of \(\rho .\)
Now, we will show in Sect. 6.4 that the first derivatives of \(s_{t}\) have the same value as we approach a point \(\lambda \in \partial \Sigma _{t}\) from the inside as when we approach \(\lambda \) from the outside. We can therefore give a complete description of the function \(\partial s_{t}/\partial \theta \) on \(\Sigma _{t}\) as follows. It is the unique function on \(\Sigma _{t}\) that is independent of \(\rho \) (or, equivalently, independent of \(r=\left| \lambda \right| \)) and whose boundary values agree
Since the points on the outer boundary of \(\Sigma _{t}\) have the polar form \((r_{t}(\theta ),\theta ),\) we conclude that
From this result, the expression (6.4), and the formula for the Laplacian in logarithmic polar coordinates, we obtain
for points inside \(\Sigma _{t},\) accounting for the formula in Theorem 2.2.
We now briefly discuss what is needed to make the preceding arguments rigorous. If \(\lambda \) is outside \(\overline{\Sigma }_{t}\) and \(\varepsilon \) is small and positive, we need to know that we can find a \(\lambda _{0}\) close to \(\lambda \) and a small, positive \(\varepsilon _{0}\) such that with these initial conditions, \(\varepsilon (t)=\varepsilon \) and \(\lambda (t)=\lambda .\) To show this, we apply the inverse function theorem to the map \(U_{t}(\lambda _{0},\varepsilon _{0}):=(\lambda (t),\varepsilon (t))\) in a neighborhood of the point \((\lambda _{0},\varepsilon _{0})=(\lambda ,0).\)
For \(\lambda \) inside \(\Sigma _{t},\) we need to know first that \(S(t,\lambda ,\varepsilon )\) is continuous—in all three variables—up to \(\varepsilon =0.\) After all, \(s_{t}(\lambda )\) is defined letting \(\varepsilon \) tend to zero in the expression \(S(t,\lambda ,\varepsilon ),\) with t and \(\lambda \) fixed. But the Hamilton–Jacobi formula (5.7) gives a formula for \(S(u,\lambda (u),\varepsilon (u))\), in which the first two variables in S are not remaining constant. Furthermore, we want to apply also the Hamilton–Jacobi formula (5.8) for the derivatives of S, which means we need also continuity of the derivatives of S with respect to \(\lambda \) up to \(\varepsilon =0.\) Using another inverse function theorem argument, we will show that after making the change of variable \(z=\sqrt{\varepsilon },\) the function S will extend smoothly up to \(\varepsilon =z=0,\) from which the needed regularity will follow.
We use the following notation throughout the section.
Notation 6.1
We will let
denote the \(\varepsilon \)-component of the solution to (5.2) with \(\lambda (0)=\lambda _{0}\) and \(\varepsilon (0)=\varepsilon _{0}\) (and with initial values of the momenta given by (5.4)), and similarly for the other components of the solution.
6.2 Outside \(\overline{\Sigma }_{t}\)
The goal of this subsection is to prove the following result.
Theorem 6.2
Fix a pair \((t,\lambda )\) with \(\lambda \) outside \(\overline{\Sigma }_{t}.\) Then
Thus,
whenever \(\lambda \) is outside \(\overline{\Sigma }_{t}.\)
As we have discussed in Sect. 6.1, the idea is that for \(\lambda \) outside \(\overline{\Sigma }_{t}\) and \(\varepsilon \) small and positive, we should try to find a \(\lambda _{0}\) close to \(\lambda \) and a small, positive \(\varepsilon _{0}\) such that \(\varepsilon (u)\) and \(\lambda (u)\) will approach 0 and \(\lambda ,\) respectively, as u approaches t. To that end, we define, for each \(t>0,\) a map \(U_{t}\) from an open subset of \(\mathbb {R}\times \mathbb {C}\) into \(\mathbb {R}\times \mathbb {C}\) by
We wish to evaluate the derivative of this map at the point \((\lambda _{0},\varepsilon _{0})=(\lambda ,0).\) For this idea to make sense, \(\lambda (t;\lambda _{0},\varepsilon _{0})\) and \(\varepsilon (t;\lambda _{0},\varepsilon _{0})\) must be defined in a neighborhood of \((\lambda ,0)\); it is for this reason that we have allowed \(\varepsilon _{0}\) to be negative in Sect. 5.3.
The domain of \(U_{t}\) consists of pairs \((\lambda _{0},\varepsilon _{0})\) such that (1) the assumptions (5.42) are satisfied; (2) the function \(p_{\varepsilon }(\cdot )\) remains positive until it blows up, as in Remark 5.10; and (3) we have \(t_{*}(\lambda _{0},\varepsilon _{0})>t.\) We note that these conditions allow \(\varepsilon _{0}\) to be slightly negative and that all the results of Sect. 5.3 hold under these conditions. We also note that by Proposition 5.11, if \(\varepsilon _{0}=0\), then \(\varepsilon (t)\equiv 0\) and \(\lambda (t)\equiv \lambda _{0}\); thus,
We now fix a pair \((t,\lambda )\) with \(\lambda \) outside of \(\overline{\Sigma }_{t}\) (so that \(\lambda \ne 1\)). By Corollary 5.14, we then have \(t_{*}(\lambda ,0)>t.\)
Lemma 6.3
The Jacobian of \(U_{t}\) at \((\lambda ,0)\) has the form
with \(\partial \varepsilon /\partial \varepsilon _{0}(t;\lambda ,0)>0.\) In particular, the inverse function theorem applies at \((\lambda ,0).\)
Proof
The claimed form of the second column of \(U_{t}^{\prime }(\lambda ,0)\) follows immediately from (6.7). We then compute from (5.50) that
which is positive. \(\square \)
Proof of Theorem 6.2
We note that the inverse of the matrix in (6.8) will have a positive entry in the bottom right corner, meaning that \(U_{t}^{-1}\) has the property that \(\partial \varepsilon _{0} /\partial \varepsilon >0.\) It follows that the \(\varepsilon _{0}\)-component of \(U_{t}^{-1}(\lambda ,\varepsilon )\) will be positive for \(\varepsilon \) small and positive. In that case, the solution to the system (5.2) will have \(\varepsilon (u)>0\) up to the blow-up time. The blow-up time, in turn, exceeds t for all points in the domain of \(U_{t}.\)
We may, therefore, apply the Hamilton–Jacobi formula (5.7), which we write as follows. We let \(\mathrm {HJ}\) denote the right-hand side of the Hamilton–Jacobi formula (5.7):
and we then have
If \(\varepsilon \) is small and positive, we therefore obtain
where we note that by definition \(\lambda (t;U_{t}^{-1}(\lambda ,\varepsilon ))=\lambda .\)
Now, in the limit \(\varepsilon \rightarrow 0^{+}\) with \(\lambda \) fixed, the inverse function theorem tells us that \(U_{t}^{-1}(\varepsilon ,\lambda )\rightarrow (0,\lambda ).\) Thus, the limit (6.6) may be computed by putting \(\lambda (t;\lambda _{0},\varepsilon _{0})=\lambda \) in (6.11) and letting \(\varepsilon _{0}\) tend to zero and \(\lambda _{0}\) tend to \(\lambda .\) This process gives (6.6).
Finally, when \(\lambda _{0}=0,\) we can use continuous dependence of the solutions on the initial conditions. The formula for \(p_{\varepsilon }(t)\) in Proposition 5.9 has a limit as \(\left| \lambda _{0}\right| \) tends to zero, so that \(\delta \) tends to \(+\infty .\) From (5.44), we find that \(a^{2}=y_{0}^{2},\) so that from (5.32), \(y(t)\equiv y_{0}.\) We then obtain
which remains nonsingular for all t. We can then continue to use the formula (5.50) for \(\varepsilon (t).\) Furthermore, by exponentiating (5.51) and letting \(\left| \lambda _{0}\right| \) tend to zero, we find that \(\lambda (t)\equiv 0.\) We then continue to use the remaining formulas in the proof of Proposition 5.11 and find that the solution to the system exists for all time.
When \(\lambda _{0}=0,\) we apply the Hamilton–Jacobi formula in the form (5.20), which is to say that we replace the last two terms in (5.7) by \(\int _{0}^{t}\varepsilon (s)p_{\varepsilon }(s)~ds.\) We then compute as in (6.9) that the derivative of \(\varepsilon (t;0,\varepsilon _{0})\) with respect to \(\varepsilon _{0}\) is positive at \(\varepsilon _{0}=0.\) Thus, by the inverse function theorem, for small positive \(\varepsilon ,\) we can find a small positive \(\varepsilon _{0}\) that gives \(\varepsilon (t;0,\varepsilon _{0})=\varepsilon .\) We then apply (5.20 ) with \(\lambda _{0}=0\) and \(\lambda (t)=0,\) and let \(\varepsilon \) tend to zero, which means that \(\varepsilon _{0}\) also tends to zero. As \(\varepsilon _{0}\) tends to zero, the function
tends to zero uniformly and we obtain (6.6). \(\square \)
6.3 Inside \(\Sigma _{t}\)
In this subsection, we establish the needed regularity of \(S(t,\lambda ,\varepsilon )\) as \(\varepsilon \) tends to zero, for \(\lambda \) in \(\Sigma _{t}.\) This result, whose proof is on p. 41, together with Theorem 5.7, will allow us to understand the structure of \(s_{t}\) and its derivatives on \(\Sigma _{t}.\)
Theorem 6.4
Define
Fix a pair \((\sigma ,\mu )\) with \(\mu \) in \(\Sigma _{\sigma }.\) Then \(\tilde{S}(t,\lambda ,z),\) initially defined for \(z>0,\) extends to a real-analytic function in a neighborhood of \((\sigma ,\mu ,0)\) inside \(\mathbb {R} \times \mathbb {C}\times \mathbb {R}.\)
We emphasize that the analytically extended \(\tilde{S}\) does not satisfy the identity \(\tilde{S}(t,\lambda ,z)=S(t,\lambda ,z^{2}).\) Indeed, since \(\sqrt{\varepsilon (t)}p_{\varepsilon }(t)\) is always bounded away from zero (Proposition 5.6), the second Hamilton–Jacobi formula (5.8) tells us that \(\partial \tilde{S}/\partial z(t,\lambda ,z)=2\sqrt{\varepsilon }\partial S/\partial \varepsilon (t,\lambda ,z^{2})\) has a nonzero limit as z tends to zero, ruling out a smooth extension that is even in z.
Corollary 6.5
Fix a pair \((\sigma ,\mu )\) with \(\mu \) in \(\Sigma _{\sigma }.\) Then the functions
all have extensions that are continuous in all three variables to the set of \((t,\lambda ,\varepsilon )\) with \(\lambda \in \Sigma _{t}\) and \(\varepsilon \ge 0.\) Furthermore, for each \(t>0,\) the function \(s_{t}\) is infinitely differentiable on \(\Sigma _{t},\) and its derivatives with respect to a and b agree with the \(\varepsilon \rightarrow 0^{+}\) limit of \(\partial S/\partial a\) and \(\partial S/\partial b.\) If we let \(t_{*}\) be short for \(t_{*} (\lambda _{0},\varepsilon _{0}),\) then for all \(\lambda _{0}\) and \(\varepsilon _{0}>0,\) we have
and
Proof
We note that the four functions in (6.12) may be computed as
respectively, and that \(S(t,\lambda ,0)=\tilde{S}(t,\lambda ,0).\) The first claim then follows from Theorem 6.4. Now that the continuity of S and its derivatives has been established, we may let t approach \(t_{*}(\lambda _{0},\varepsilon _{0})\) in the Hamilton–Jacobi formulas (5.7) and (5.8) to obtain the second claim. \(\square \)
Corollary 6.6
Let us write \(\lambda \in \Sigma _{t}\) in logarithmic polar coordinates, with \(\rho =\log \left| \lambda \right| \) and \(\theta =\arg \lambda .\) Then for each pair \((t,\lambda )\) with \(\lambda \in \Sigma _{t},\) we have
Furthermore, \(\partial s_{t}/\partial \theta \) is independent of \(\rho \); that is,
for some smooth function \(m_{t}.\) Thus,
for some smooth function \(m_{t}\), and
In Sect. 6.4, we will obtain a formula for the function \(m_{t}(\theta )\) appearing in Corollary 6.6.
Proof
The derivative \(\partial /\partial \rho \) may be computed in ordinary polar coordinates as \(r\partial /\partial r\) or in rectangular coordinates as \(a\partial /\partial a+b\partial /\partial b.\) It then follows from the Hamilton–Jacobi formula (5.8) that
Now, for each pair \((t,\lambda )\) with \(\lambda \in \Sigma _{t},\) Theorem 5.17 tells us that we can find \((\lambda _{0},\varepsilon _{0})\) so that
In light of (6.13), the formula (6.14) then follows from the formula (5.29) in Theorem 5.7.
Now, \(\partial s_{t}/\partial \rho \) is manifestly independent of \(\theta .\) Since, by Corollary 6.5, \(s_{t}\) is an analytic, hence \(C^{2},\) function on \(\Sigma _{t},\) we conclude that
showing that \(\partial s_{t}/\partial \theta \) is independent of \(\rho .\) The formula (6.15) then follows by differentiating (6.14) with respect to \(\rho .\) Finally, if we use the standard formula for the Laplacian in polar coordinates,
we obtain (6.16) from (6.15). \(\square \)
We now begin preparations for the proof of Theorem 6.4. Recall from Proposition 5.6 that \(\varepsilon (t)p_{\varepsilon }(t)^{2}=\varepsilon _{0}p_{0}^{2}e^{-Ct},\) where \(C=2\Psi -1\) is a constant computed from \(\varepsilon _{0}\) and \(\lambda _{0}\) as in (5.26). Recall also from Proposition 5.9 that for \(\varepsilon _{0}>0,\) the function \(1/p_{\varepsilon }(t)\) extends to real analytic function of t defined for all \(t\in \mathbb {R}.\) We then define, for \(\varepsilon _{0}>0,\)
for all \(t\in \mathbb {R}.\) For \(t<t_{*}(\lambda _{0},\varepsilon _{0}),\) the function \(z(t;\lambda _{0},\varepsilon _{0})\) is positive and satisfies
while for \(t=t_{*}(\lambda _{0},\varepsilon _{0}),\) we have \(z(t;\lambda _{0},\varepsilon _{0})=0\) and for \(t>t_{*}(\lambda _{0},\varepsilon _{0}),\) the function \(z(t;\lambda _{0},\varepsilon _{0})\) is negative.
Furthermore, using (5.51) and Point 3 of Proposition 5.4, we see that
where by Proposition 5.6, we have
Since \(1/p_{\varepsilon }(s)\) extends to an analytic function of \(s\in \mathbb {R}\), we see that \(\lambda (t)\) extends to an analytic function of \(t\in \mathbb {R}.\) We may therefore define a map
for all \(t\in \mathbb {R},\) \(\lambda _{0}\in \mathbb {C},\) and \(\varepsilon _{0}>0.\)
Proposition 6.7
Suppose \((t,\lambda _{0},\varepsilon _{0})\) has the property that \(\lambda _{0}\ne 0\) and \(t_{*}(\lambda _{0},\varepsilon _{0})=t,\) so that \(z(t;\lambda _{0},\varepsilon _{0})=0.\) Then the Jacobian matrix of V at \((t,\lambda _{0},\varepsilon _{0})\) is invertible.
Proof
We make some convenient changes of variables. First, we replace \((t,\lambda _{0},\varepsilon _{0})\) with \((t,\lambda _{0},\delta ),\) where \(\delta \) is as in (5.38). This change has a smooth inverse, since we can recover \(\varepsilon _{0}\) from \(\delta \) as
Then we write \(\lambda _{0}\) in terms of its polar coordinates, \((r_{0} ,\theta _{0}).\) Finally, we write \(\lambda (t;\lambda _{0},\varepsilon _{0})\) in logarithmic polar coordinates,
where by Point 3 of Proposition 5.4, \(\theta (t;\lambda _{0},\varepsilon _{0} )=\theta _{0}.\)
Thus, to prove the proposition, it suffices to verify that the Jacobian matrix of the map
is invertible. We observe that this Jacobian has the form
where
Now, by Proposition 5.9, the lifetime is independent of \(r_{0}\) with \(\delta \) and \(\theta _{0}\) fixed. Thus, if we start at a point with \(t_{*}(\lambda _{0},\varepsilon _{0})=t\) and vary \(r_{0},\) the lifetime will remain equal to t and \(z(t;\varepsilon _{0},\lambda _{0})\) will remain equal to 0. Thus, at the point in question, \(\partial z/\partial r_{0}=0.\) Meanwhile, Proposition 5.12 gives a formula for the value of \(\rho (t;\lambda _{0},\varepsilon _{0})\) at \(t=t_{*}(\lambda _{0} ,\varepsilon _{0}),\) from which we can easily see that \(\partial \rho /\partial r_{0}>0.\) It therefore remains only to verify that \(\partial z/\partial \delta \) is nonzero.
Now, \(z(t_{*}(\lambda _{0},\varepsilon _{0});\lambda _{0},\varepsilon _{0})=0.\) If we differentiate this relation with respect to \(\varepsilon _{0}\) with \(\lambda _{0}\) fixed, we find that
The derivative \(\partial t_{*}/\partial \varepsilon _{0}\) may be computed as
where \(g_{\theta _{0}}\) is as in (5.58). But the proof of Proposition 5.16 shows that \(\partial g_{\theta }/\partial \delta >0\) for all \(\delta >2,\) while from the formula (5.38) for \(\delta ,\) we see that \(\partial \delta /\partial \varepsilon _{0}>0.\) (Note also that \(\delta >2\) whenever \(\varepsilon _{0}>0.\)) Thus, \(\partial t_{*}(\lambda _{0},\varepsilon _{0})/\partial \varepsilon _{0}>0.\)
Meanwhile, from (6.17) and (5.45), we have
If we differentiate with respect to t and evaluate at the time \(t_{*}\) when the last factor is zero, the product rule gives
which is negative because \(\delta /\sqrt{\delta ^{2}-4}>1\) and the denominator is positive (Proposition 5.9). Thus, from (6.18), we conclude that \(\partial z/\partial \varepsilon _{0}>0.\) \(\square \)
We are now ready for the proof of the main result of this section.
Proof of Theorem 6.4
By (6.11), we have
where \(\mathrm {HJ}\) is as in (6.10), whenever \(t_{*}(\lambda _{0},\varepsilon _{0})>\sigma .\) Fix a point \((\sigma ,\mu )\) with \(\mu \in \Sigma _{\sigma }.\) Then by Theorem 5.17, we can find a pair \((\lambda _{0},\varepsilon _{0})\) with \(t_{*}(\lambda _{0},\varepsilon _{0} )=t\)—so that \(z(t;\lambda _{0},\varepsilon _{0})=0\)—and \(\lambda (t;\lambda _{0},\varepsilon _{0})=\lambda .\) We now construct a local inverse \(V^{-1}\) to V around the point \(V(t,\lambda _{0},\varepsilon _{0} )=(t,\lambda ,0).\)
For any triple \((t,\lambda ,z)\) in the domain of \(V^{-1},\) we write \(V^{-1}(t,\,\lambda ,z)\) as \((t,\lambda _{0},\varepsilon _{0}).\) We note that if \(z>0\) then \(t_{*}(\lambda _{0},\varepsilon _{0})\) must be greater than t, because if we had \(t_{*}(\lambda _{0},\varepsilon _{0})\le t,\) then \(z(t;\lambda _{0},\varepsilon _{0})=z\) would be zero or negative. Thus, we may apply (6.19) at \((t,\lambda _{0},\varepsilon _{0})=V^{-1} (t,\lambda ,z)\) to obtain
whenever \((t,\lambda ,z)\) is in the domain of \(V^{-1}\) and \(z>0.\)
Recall now that \(\lambda (t;\lambda _{0},\varepsilon _{0})\) extends to an analytic function of \(t\in \mathbb {R}.\) Thus, the function \(\mathrm {HJ}\) in (6.10) extends to a smooth function of \(t\in \mathbb {R},\) \(\lambda _{0}\in \mathbb {C}{\setminus }\{0\},\) and \(\varepsilon _{0}>0,\) defined even if \(t_{*}(\lambda _{0},\varepsilon _{0})<t.\) Therefore, the right-hand side of (6.20) provides the claimed smooth extension of \(\tilde{S}.\) \(\square \)
6.4 Near the boundary of \(\Sigma _{t}\)
We start by considering what is happening right on the boundary of \(\Sigma _{t}.\)
Remark 6.8
Neither the method of Sect. 6.2 nor the method of Sect. 6.3 allows us to compute the value of \(s_{t}(\lambda )\) for \(\lambda \) in the boundary of \(\Sigma _{t}.\) Although we expect that this value will be \(\log (\left| \lambda -1\right| ^{2}),\) the question is irrelevant to the computation of the Brown measure. After all, we are supposed to consider \(\Delta s_{t}\) computed in the distribution sense, that is, the distribution whose value on a test function \(\psi \) is
The value of (6.21) is unaffected by the value of \(s_{t}(\lambda )\) for \(\lambda \) in \(\partial \Sigma _{t},\) which is a set of measure zero in \(\mathbb {C}.\)
It is nevertheless essential to understand the behavior of \(s_{t}(\lambda )\) as \(\lambda \) approaches the boundary of \(\Sigma _{t}.\)
Definition 6.9
We say that a function \(f:\mathbb {C}\rightarrow \mathbb {R}\) is analytic up to the boundary from inside \(\Sigma _{t}\) if the following conditions hold. First, f is real analytic on \(\Sigma _{t}\). Second, for each \(\lambda \in \partial \Sigma _{t},\) we can find an open set U containing \(\lambda \) and a real analytic function g on U such that g agrees with f on \(U\cap \Sigma _{t}.\) We may similarly define what it means for f to be analytic up to the boundary from outside \(\Sigma _{t}.\)
Proposition 6.10
For each \(t>0,\) the function \(s_{t}\) is analytic up to the boundary from inside \(\Sigma _{t}\) and analytic up to the boundary from outside \(\Sigma _{t}.\)
Note that the proposition is not claiming that \(s_{t}\) is an analytic function on all of \(\mathbb {C}.\) Indeed, our main results tell us that \(\frac{1}{4\pi }\Delta s_{t}(\lambda )\) is identically zero for \(\lambda \) outside \(\Sigma _{t}\) but approaches a typically nonzero value as \(\lambda \) approaches a boundary point from the inside. As we approach from the inside a boundary point with polar coordinates \((r,\theta ),\) the limiting value of \(\frac{1}{4\pi }\Delta s_{t}(\lambda )\) is \(w_{t}(\theta )/r^{2}.\) This quantity certainly cannot always be zero, or the Brown measure of \(b_{t}\) would be identically zero. Actually, we will see in Sect. 7.1 that \(w_{t}(\theta )\) is strictly positive except when \(t=4\) and \(\theta =\,\pi .\)
Proof
We have shown that \(s_{t}(\lambda )=\log (\left| \lambda -1\right| ^{2})\) for \(\lambda \) in \((\overline{\Sigma }_{t})^{c}.\) Since \(1\in \Sigma _{t},\) we see that \(s_{t}\) is analytic from the outside of \(\Sigma _{t}.\)
To address the analyticity from the inside, first note that by applying (6.20) with \(z=0,\) we have
where \(\mathrm {HJ}\) is as in (6.10). But if \(\varepsilon _{0} ^{t}:\Sigma _{t}\rightarrow \mathbb {R}\) and \(\lambda _{t}:\Sigma _{t} \rightarrow \mathbb {C}\) are as in Lemma 5.18, then we can see that
and we conclude that
We now claim that the function \(\varepsilon _{0}^{t}(\lambda _{0}),\) initially defined for \(\lambda _{0}\in \overline{\Sigma }_{t},\) extends to an analytic function in a neighborhood of \(\overline{\Sigma }_{t}.\) For \(t\ge 4,\) we can simply use the formula (5.64) for all nonzero \(\lambda _{0}.\) For \(t<4,\) however, the formula (5.64) becomes undefined in a neighborhood of a point where \(\partial \Sigma _{t}\) intersects the unit circle.
Nevertheless, we can make a general argument as follows. To compute \(\varepsilon _{0}^{t}(\lambda _{0}),\) we solve the equation \(t_{*} (\lambda _{0},\varepsilon _{0})=t\) for \(\varepsilon _{0}\) as a function of \(\lambda _{0}.\) To do this, we first solve the equation \(g_{\theta _{0}} (\delta )=t\) for \(\delta _{\theta _{0},t}\) and then solve for \(\varepsilon _{0}\) in terms of \(\delta \) as \(\varepsilon _{0}=\left| \lambda _{0}\right| \delta -\left| \lambda _{0}\right| ^{2}-1.\) Now, we know from the proof of Proposition 5.16 that \(g_{\theta _{0}}(\delta )=t\) has a solution when \(\left| \theta _{0}\right| \le \theta _{\max }(t)=\cos ^{-1}(1-t/2),\) with the solution being \(\delta =2\) when \(\theta _{0}=\pm \theta _{\max }(t).\) We can also verify that \(\partial g_{\theta _{0}} /\partial \delta >0\) for all \(\delta \ge 2.\) This was verified for \(\delta >2\) in the proof of Proposition 5.16. To see that the result holds even when \(\delta =2,\) it suffices to verify that the expressions in (5.59) and (5.60) have positive limits as \(\delta \rightarrow 2^{+}\). We omit this verification and simply note that the limits have the values 1 and 1/3, respectively. It then follows from the implicit function theorem that (1) the solution \(\delta _{\theta _{0},t}\) continues to exist (with \(\delta <2\)) for \(\left| \theta _{0}\right| \) slightly larger than \(\theta _{\max }(t),\) and (2) the solution \(\delta _{\theta _{0},t}\) depends analytically on \(\theta _{0}.\) Then, the expression
makes sense and is analytic for all nonzero \(\lambda _{0}\) with \(\left| \arg \lambda _{0}\right| <\theta _{\max }(t)+\alpha _{t},\) for some positive quantity \(\alpha _{t}.\) We note that in this expression, \(\varepsilon _{0} ^{t}(\lambda _{0})\) can be negative—for example if \(\left| \lambda _{0}\right| =1\) and \(\arg \lambda _{0}>\theta _{\max }(t).\)
We now consider the function \(\lambda _{t},\) defined as
and we recall that \(\lambda _{t}(\lambda _{0})=\lambda _{0}\) for \(\lambda _{0} \in \partial \Sigma _{t}.\) Although \(\lambda _{t}\) was initially defined for \(\lambda _{0}\) in \(\overline{\Sigma }_{t},\) it has an analytic extension to a neighborhood of \(\overline{\Sigma }_{t}\), namely the set of \(\lambda _{0}\) in the domain of the extended function \(\varepsilon _{0}^{t}\) for which the pair \((\lambda _{0},\varepsilon _{0})\) satisfy the assumptions in (5.42). We now claim that the derivative of \(\lambda _{t}(\lambda _{0})\) is invertible at each point in its domain. We use polar coordinates in both domain and range. Since \(\arg (\lambda _{t}(\lambda _{0}))=\arg \lambda _{0},\) the derivative will have the form
and it therefore suffices to check that \(\partial \left| \lambda _{t}\right| /\partial r\) is nonzero. To see this, we use the formula (5.62), where \(\delta =\delta _{\theta _{0},t}\) as in the previous paragraph. Since \(\delta \) is independent of \(\left| \lambda _{0}\right| \) with t and \(\arg \lambda _{0}\) fixed, we can easily verify from (5.62) that \(\partial \left| \lambda _{t}\right| /\partial r>0.\)
Now, we have already established that \(s_{t}\) is analytic in the interior of \(\Sigma _{t}.\) Consider, then, a point \(\lambda \) in \(\partial \Sigma _{t},\) so that \(\lambda _{t}(\lambda )=\lambda .\) Since \(\lambda _{t}^{\prime }(\lambda )\) is invertible, it has a analytic local inverse \(\lambda _{t}^{-1}\) defined near \(\lambda .\) Then the formula (6.22) gives an analytic extension of \(s_{t}\) to a neighborhood of \(\lambda .\) \(\square \)
Proposition 6.11
Fix a point \(\mu \) on the boundary of \(\Sigma _{t}.\) Then the functions
all approach the same value when \(\lambda \) approaches \(\mu \) from inside \(\Sigma _{t}\) as when \(\lambda \) approaches \(\mu \) from outside \(\overline{\Sigma }_{t}.\)
Proof
We begin by considering \(s_{t}\) itself. The limit as \(\lambda \) approaches \(\mu \) from the inside may be computed by using (6.22). By Lemma 5.18, as \(\lambda \) approaches \(\mu \) from the inside, \(\lambda _{t}^{-1}(\lambda )\) approaches \(\lambda _{t}^{-1}(\mu )=\mu ,\) and \(\varepsilon _{0}^{t}(\lambda _{t}^{-1}(\lambda ))\) approaches 0. Thus, the limiting value of \(s_{t}\) from the inside is
where \(\mathrm {HJ}\) is given by (6.10) and were we have used that \(\lambda (t;\mu ,0)=\mu .\) (See the last part of Proposition 5.11). Since \(s_{t}(\lambda )=\log (\left| \lambda -1\right| ^{2})\) outside \(\overline{\Sigma }_{t},\) the limit of \(s_{t}\) from the outside agrees with the limit from the inside.
Next we consider the derivatives, which we compute in logarithmic polar coordinates \(\rho =\log \left| \lambda \right| \) and \(\theta =\arg \lambda \). By (6.14), we have
for \(\lambda \in \Sigma _{t}.\) Letting \(\lambda \) approach \(\mu \) from the inside gives the value \(\log (\left| \mu \right| ^{2})/t+1.\) Since \(\mu \) is on the boundary of \(\Sigma _{t},\) Theorem 3.1 says that \(T(\mu )=t,\) so that
Taking the corresponding derivative of the “outside” function \(\log (\left| \lambda -1\right| ^{2})\) and letting \(\lambda \) tend to \(\mu \) from the outside gives the same result.
Finally, we recall from Proposition 5.4 that \(ap_{b}-bp_{a}\) is a constant of motion. Thus, by the second Hamilton–Jacobi formula (5.8) and the initial conditions (5.4), we have
If we choose \(\varepsilon _{0}\) and \(\lambda _{0}\) so that \(t_{*}(\lambda _{0},\varepsilon _{0})=t\) we can use the regularity result in Corollary 6.5 to let u tend to t. This gives
where now \(\lambda _{0}=\lambda _{t}^{-1}(\lambda )\) and \(\varepsilon _{0}=\varepsilon _{0}^{t}(\lambda _{t}^{-1}(\lambda )).\) As \(\lambda \) approaches \(\mu \), Theorem 5.17 says that the value of \(\lambda _{0}\) approaches \(\mu \) and \(\varepsilon _{0}\) approaches 0, so we get
Taking the corresponding derivative of \(\log (\left| \lambda -1\right| ^{2})\) and letting \(\lambda \) tend to \(\mu \) from the outside gives the same result. \(\square \)
6.5 Proof of the main result
In this subsection, we prove our first main result, Theorem 2.2. Proposition 2.3 will then be proved in Sect. 7.1, while Propositions 2.5 and 2.6 will be proved in Sect. 7.2.
Proposition 6.12
For each fixed t, the restriction to \(\Sigma _{t}\) of the function
is the unique function that on \(\Sigma _{t}\) that (1) extends continuously to the boundary, (2) agrees with the \(\theta \)-derivative of \(\log (\left| \lambda -1\right| ^{2})\) on the boundary, and (3) is independent of \(r=\left| \lambda \right| .\) Thus, the function \(m_{t}\) in Corollary 6.6 is given by
where \(r_{t}(\theta )\) is the outer radius of the domain \(\Sigma _{t}\) (Fig. 3).
Proof
We have already established in Corollary 6.6 that \(\partial s_{t}/\partial \theta \) is independent of \(\rho \) (or equivalently, of r) in \(\Sigma _{t}.\) Then Propositions 6.10 and 6.11 tell us that \(\partial s_{t}/\partial \theta \) is continuous up to the boundary and agrees there with the angular derivative of \(\log (\left| \lambda -1\right| ^{2}).\) Thus, to compute \(\partial s_{t}/\partial \theta \) at a point in \(\Sigma _{t},\) we travel along a radial segment (in either direction) until we hit the boundary at radius \(r_{t}(\theta )\) or \(1/r_{t}(\theta )\). We then evaluate the angular derivative of \(\log (\left| \lambda -1\right| ^{2}),\) as in (6.5), giving the claimed expression for \(\partial s_{t}/\partial \theta =m_{t} (\theta ).\) \(\square \)
Proposition 6.13
For each \(t>0,\) the distributional Laplacian of \(s_{t}(\lambda )\) with respect to \(\lambda \) may be computed as follows. Take the pointwise Laplacian of \(s_{t}\) outside \(\overline{\Sigma }_{t}\) (giving zero), take the pointwise Laplacian of \(s_{t}\) inside \(\Sigma _{t}\) (giving the expression (6.16) in Corollary 6.6) and ignore the boundary of \(\Sigma _{t}.\)
Proof
Since, by Proposition 6.10, \(s_{t}\) is analytic up to the boundary of \(\Sigma _{t}\) from the inside, Green’s second identity says that
for any test function \(\psi ,\) where in the last integral, the limiting value of \(\nabla s_{t}\) from the inside should be used. This identity holds because the boundary of \(\Sigma _{t}\) is smooth for \(t\ne 4\) and piecewise smooth when \(t=4\) (Point 3 of Theorem 3.2). We also have similar formula for the integral over the complement of \(\overline{\Sigma }_{t},\) provided that \(\psi \) is compactly supported, but with the direction of the unit normal reversed. Proposition 6.11 then tells us that the boundary terms in the two integrals cancel, giving
where the integral over \((\overline{\Sigma }_{t})^{c}\) is actually zero, since \(\Delta s_{t}(\lambda )=0\) there. The formula (6.23) says that the distributional Laplacian of \(s_{t}\) may be computed by taking the ordinary, pointwise Laplacian in \(\overline{\Sigma }_{t}\) and in \(\Sigma _{t}\) and ignoring the boundary of \(\Sigma _{t}.\) \(\square \)
We now have all the ingredients for a proof of Theorem 2.2.
Proof of Theorem 2.2
Proposition 6.13 tells us that we can compute the distributional Laplacian of \(s_{t}\) separately inside \(\Sigma _{t}\) and outside \(\overline{\Sigma }_{t},\) ignoring the boundary. Theorem 6.2 tells us that the Laplacian outside \(\overline{\Sigma }_{t}\) is zero. Corollary 6.6 gives us the form of \(\Delta s_{t}\) inside \(\Sigma _{t},\) while Proposition 2.6 identifies the function \(m_{t}\) appearing in Corollary 6.6. The claimed formula for the Brown measure therefore holds. \(\square \)
7 Further properties of the Brown measure
7.1 The formula for \(\omega \)
In this subsection, we derive the formula for \(w_{t}\) given in Proposition 2.3 in terms of the density \(\omega .\) Throughout, we will write the function T in (3.1) in polar coordinates as
We start with a simple rewriting of the expression for \(w_{t}\) in Theorem 2.2.
Lemma 7.1
The density \(w_{t}(\theta )\) in Theorem 2.2 may also be written as
where
Proof
We start by noting that the point with polar coordinates \((r_{t} (\theta ),\theta )\) is on the boundary of \(\Sigma _{t}.\) Thus, by Theorem 3.1, we have \(T(r_{t}(\theta ),\theta )=t,\) from which we obtain
Thus, we may write
from which the claimed formula follows easily from the expression in Theorem 2.2. \(\square \)
We now formulate the main result of this subsection, whose proof is on p. 47.
Theorem 7.2
Consider the function \(\omega (r,\theta )\) defined in (2.7). Although the right-hand side of (2.7) is indeterminate at \(r=1,\) the function \(\omega \) has a smooth extension to all \(r>0\) and all \(\theta .\) The function \(w_{t}(\theta )\) in Theorem 2.2 can then be expressed as
The function \(\omega \) has the following properties.
-
(1)
We have \(\omega (1/r,\theta )=\omega (r,\theta )\) for all \(r>0\) and all \(\theta .\)
-
(2)
When \(r=1,\) we have
$$\begin{aligned} \omega (1,\theta )=3\frac{1+\cos \theta }{2+\cos \theta }. \end{aligned}$$In particular, \(\omega (1,0)=2\) and \(\omega (1,\pi )=0.\)
-
(3)
The density \(\omega (r,\theta )\) is strictly positive except when \(r=1\) and \(\theta =\pm \pi .\) Furthermore, \(\omega (r,\theta )\le 2\) with equality precisely when \(r=1\) and \(\theta =0.\)
-
(4)
We have
$$\begin{aligned} \lim _{r\rightarrow 0}\omega (r,\theta )=1, \end{aligned}$$where the limit is uniform in \(\theta .\)
We now derive consequences for \(w_{t}.\) For \(t\le 4,\) the density \(w_{t}(\theta )\) is only defined for \(-\theta _{\max }(t)<\theta <\theta _{\max }(t),\) where \(\theta _{\max }(t)=\cos ^{-1}(1-t/2),\) while for \(t>4,\) the density \(w_{t}(\theta )\) is defined for all \(\theta .\) (Recall Theorem 3.2).
Corollary 7.3
(Positivity). If \(t>4,\) then \(w_{t}(\theta )\) is strictly positive for all \(\theta .\) If \(t<4,\) then \(w_{t}(\theta )\) is strictly positive for \(-\theta _{\max }(t)<\theta <\theta _{\max }(t)\) and the limit as \(\theta \) approaches \(\pm \theta _{\max }(t)\) of \(w_{t}(\theta )\) is strictly positive. Finally, if \(t=4,\) then \(w_{t}(\theta )\) is strictly positive for \(-\pi<\theta <\pi ,\) but \(\lim _{\theta \rightarrow \pm \pi }w_{t}(\theta )=0.\)
Proof
The only time \(\omega (r,\theta )\) equals zero is when \(r=1\) and \(\theta =\pm \pi .\) When \(t>4,\) the function \(r_{t}(\theta )\) is continuous and and greater than 1 for all \(\theta ,\) so that \(w_{t}(\theta )\) is strictly positive in this case. When \(t\le 4,\) we know from Proposition 3.4 that \(r_{t}(\theta )\) is greater than 1 for \(\left| \theta \right| <\theta _{\max }(t)\) and approaches 1 when \(\theta \) approaches \(\pm \theta _{\max }(t).\) Thus, \(w_{t}(\theta )=\omega (r_{t}(\theta ),\theta )\) is strictly positive for \(\left| \theta \right| <\theta _{\max }(t).\) When \(t<4,\) we have \(\theta _{\max }(t)=\cos ^{-1}(1-t/2)<\pi \) and the limiting value of \(w_{t}(\theta )\)—namely \(\omega (1,\theta _{\max })\)—will be positive. Finally, when \(t=4,\) we have \(\theta _{\max }(t)=\pi \) and the limiting value of \(w_{t}(\theta )\) is \(\omega (1,\pi )=0.\) \(\square \)
Corollary 7.4
(Asymptotics). The density \(w_{t}(\theta )\) has the property that
for small t. More precisely, for all sufficiently small t and all \(\theta \in (-\theta _{\max }(t),\theta _{\max }(t)),\) the quantity \(\pi tw_{t}(\theta )\) is close to 1. Furthermore,
for large t. More precisely, for all sufficiently large t and all \(\theta ,\) the quantity \(2\pi tw_{t}(\theta )\) is close to 1.
The small- and large-t behavior of the region \(\Sigma _{t}\) can also be determined using the behavior of the function \(T(\lambda )\) near \(\lambda =1\) (small t) and near \(\lambda =0\) (large t), together with the invariance of the region under \(\lambda \mapsto 1/\lambda .\) For small t, the region resembles a disk of radius \(\sqrt{t}\) around 1, while for large t, the region resembles an annulus with inner radius \(e^{-t/2}\) and outer radius \(e^{t/2}.\) In particular, the expected behavior of the Brown measure for small t can be observed: it resembles the uniform probability measure on a disk of radius \(\sqrt{t}\) centered at 1.
Proof
When t is small, the entire boundary of \(\Sigma _{t}\) will be close to \(\lambda =1,\) since this is the only point where \(T(\lambda )=0.\) Furthermore, when t is small, \(\theta _{\max }(t)=\cos ^{-1}(1-t/2)\) is close to zero. When t is small, therefore, the quantity
will be close to \(\omega (1,0)/2=1\) for all \(\theta \in (-\theta _{\max } (t),\theta _{\max }(t)),\) by Point 2 of Theorem 7.2.
When t is large (in particular, greater than 4), the inner boundary of the domain will be close to \(\lambda =0,\) since this is the only point in the unit disk where \(T(\lambda )\) is large. Thus, for large t, the inner radius \(1/r_{t}(\theta )\) of the domain will be uniformly small, and therefore
will be uniformly close to 1, by Point 4 of Theorem 7.2. \(\square \)
Proof of Theorem 7.2
We note that the function T in (7.1) can be written as
so that
Applying implicit differentiation to the identity \(T(r_{t}(\theta ),\theta )=t\) then gives
By the chain rule and (7.2), \(\frac{d}{d\theta }[h(r_{t} (\theta ))\sin \theta ]=q(r_{t}(\theta ),\theta ),\) where
After computing that
it is a straightforward but tedious exercise to simplify (7.3) and obtain the claimed formula (2.7).
Since \(h(1/r)=h(r),\) we may readily verify Point (1); both numerator and denominator in the fraction on the right-hand side of (2.7) change by a factor of \(1/r^{2}\) when r is replaced by 1/r.
To understand the behavior of \(\omega \) at \(r=1,\) we need to understand the function h better. We may easily calculate that h has a removable singularity at \(r=1\) with \(h(1)=1\), \(h^{\prime }(1)=0,\) and \(h^{\prime \prime }(1)=-1/3.\) We also claim that h satisfies \(0<h(r)\le 1,\) with \(h(r)=1\) only at \(r=1\). To verify the claim, we first compute that \(\lim _{r\rightarrow 0}h(r)=0\) and that
Using the Taylor expansion of logarithm, we may then compute that
for \(0<r<1.\) Thus, h(r) increases from 0 to 1 on [0, 1].
We now write h in the form
for some analytic function c(r), with \(c(1)=1/6.\) The minus sign in (7.4) is convenient because h has a strict global maximum at 1, which means c(r) is strictly positive everywhere.
Now, since \(h(1)=1,\) the fraction on the right-hand side of (2.7) is of 0/0 form when \(r=1.\) To rectify this situation, we observe that \(\alpha \) and \(\beta \) may be written as
Thus, we can take a factor of \((r-1)^{2}\) out of numerator and denominator to obtain
where \(\tilde{\alpha }(r)=1+2rc(r)\) and \(\tilde{\beta }(r)=1-(r^{2}+1)c(r).\) This expression is no longer of 0/0 form at \(r=1.\) Indeed, since \(h(1)=1\) and \(c(1)=1/6,\) we may easily verify the claimed formula for \(\omega (1,\theta )\) in Point 2 of the theorem. We will shortly verify that the denominator in the fraction on the right-hand side of (7.5) is positive for all \(r>0\) and all \(\theta ,\) from which the claimed smooth extension of \(\omega \) follows.
To verify the claimed positivity of \(\omega ,\) we first observe that \(\tilde{\beta }(r)z+\tilde{\alpha }(r)\) is positive when \(z=1\) (with a value of \(2-(r-1)^{2}c(r)=1+h(r)\)) and also positive when \(z=-1\) (with a value of \((r+1)^{2}c(r)\)), and hence positive for all \(-1\le z\le 1.\) Thus, the denominator in the fraction on the right-hand side of (7.5) is never zero. We then compute that
for all r and \(\theta .\) Thus, \((\tilde{\alpha }(r)z+\tilde{\beta } (r))/(\tilde{\beta }(r)z+\tilde{\alpha }(r))\) increases from \(-1\) to 1 as z increases from \(-1\) to 1. Since h(r) is positive, we conclude that
for all r and \(\theta ,\) with equality when \(\cos \theta =-1\) in the first case and when \(\cos \theta =1\) in the second case. Since \(h(r)\le 1\) with equality only at \(r=1,\) we see that \(\omega (r,\theta )\) is positive except when \(r=1\) and \(\cos \theta =-1.\) Similarly, \(\omega (r,\theta )\le 2\) with equality only if \(r=1\) and \(\cos \theta =1.\)
Finally, from the definition (7.4) and the fact that \(\lim _{r\rightarrow 0}h(r)=0,\) we find that \(\lim _{r\rightarrow 0}c(r)=1.\) Thus, as \(r\rightarrow 0,\) we have \(\tilde{\alpha }(r)\rightarrow 1\) and \(\tilde{\beta }(r)\rightarrow 0.\) In this limit, the fraction on the right-hand side of (7.5) converges uniformly to \(\cos \theta ,\) while h(r) tends to zero, giving Point 4. \(\square \)
7.2 The connection to free unitary Brownian motion
Recall from Theorem 1.1 that the spectral measure \(\nu _{t}\) of the free unitary Brownian motion \(u_{t}\) was computed by Biane. In this subsection, we prove Proposition 2.5, which connects the Brown measure of \(b_{t}\) to Biane’s measure \(\nu _{t}\). The support of \(\nu _{t}\) is a proper subset of the unit circle for \(t<4\) and the entire unit circle for \(t\ge 4.\) For \(t<4,\) the support of \(\nu _{t}\) consists of points with angles \(\phi \) satisfying \(\left| \phi \right| \le \phi _{\max }(t),\) where
Recall the definition in (1.3) of the function \(f_{t}\). Then \(f_{t}\) maps the boundary of \(\Sigma _{t}\) into the unit circle. (This is true by the definition (2.1) for points in \(\partial \Sigma _{t}\) outside the unit circle and follows by continuity for points in \(\partial \Sigma _{t}\) in the unit circle). Indeed, let the outer boundary of \(\Sigma _{t},\) denoted \(\partial \Sigma _{t}^{\mathrm {out}},\) be the portion of \(\partial \Sigma _{t}\) outside the open unit disk. Then \(f_{t}\) is a homeomorphism of \(\partial \Sigma _{t}^{\mathrm {out}}\) to the support of \(\nu _{t}\):
In particular, for \(t<4,\) let us define
so that the two points in \(\partial \Sigma _{t}\cap S^{1}\) have angles \(\pm \theta _{\max }(t)\) (Theorem 3.2). Then
as may be verified by direct computation from the definition of \(f_{t}.\) (Use the formula (7.10) below with \(r=1\) and \(\cos \theta =1-t/2.\))
We now describe the map (7.6) more concretely. We denote by \(\lambda _{t}(\theta )\) the point at angle \(\theta \) in \(\partial \Sigma _{t}^{\mathrm {out}}\):
where for \(t<4,\) we require \(\left| \theta \right| \le \theta _{\max }(t).\) Then the map in (7.6) can be thought of as a map of \(\theta \) to \(\phi \) determined by the relation
We now observe a close relationship between the density \(w_{t}(\theta )\) in Theorem 2.2 and the map in (7.7).
Proposition 7.5
Let \(\phi \) and \(\theta \) be related as in (7.7), where if \(t<4,\) we require \(\left| \phi \right| \le \phi _{\max }(t)\) and \(\left| \theta \right| \le \theta _{\max }(t).\) Then the density \(w_{t}\) in Theorem 2.2 may be computed as
We may also write this formula as a logarithmic derivative of \(f_{t}\) along the outer boundary of \(\Sigma _{t}\):
Proof
We compute that
Thus, using the definition (1.3) of \(f_{t},\) we find that
Evaluating this expression at the point \(\lambda _{t}(\theta )\) gives
(Strictly speaking, \(\phi \) and \(\theta \) are only defined “mod \(2\pi ,\)” but for any local continuous version of \(\theta ,\) the last expression in (7.11) gives a local continuous version of \(\phi .\)) Thus,
and the formula (7.8) follows easily by recalling the definition (2.5) of \(w_{t}.\) The expression (7.9) is then obtained by noting that \(\phi =\frac{1}{i}\log f_{t}(\lambda _{t}(\theta )).\) \(\square \)
Proposition 7.6
Biane’s measure \(\nu _{t}\) may be computed as
or as
Here, as usual, \(r_{t}(\theta )\) is the outer radius of the domain \(\Sigma _{t}\) and \(\theta \) is viewed as a function of \(\phi \) by inverting the relationship (7.7). When \(t<4,\) the formula should be used only for \(\left| \phi \right| \le \phi _{\max }(t).\)
Proof
We make use of the expression for \(\nu _{t}\) in Theorem 1.1. If \(\phi \) is in the interior of the support of \(\nu _{t},\) then \(\chi _{t} (e^{i\phi })\) is in the open unit disk, so that the density of \(\nu _{t}\) is nonzero at this point. Now, since \(\chi _{t}\) is an inverse function to \(f_{t}\) we see that \(\chi _{t}(e^{i\phi })\) is (for \(\phi \) in the interior of the support of \(\nu _{t}\)) the unique point \(\lambda \) with \(\left| \lambda \right| <1\) for which \(f_{t}(\lambda )=e^{i\phi }.\) Thus,
which reduces to (7.12). Finally, since \(T(\lambda )=t\) on \(\partial \Sigma _{t}\) (Theorem 3.1), we have
which allows us to obtain (7.13) from (7.12). \(\square \)
We are now ready for the proof of Proposition 2.5.
Proof of Proposition 2.5
The distribution of \(\arg \lambda \) with respect to the Brown measure of \(b_{t}\) is given in (2.8) as \(2\log (r_{t}(\theta ))w_{t}(\theta )~d\theta ,\) which we write using Proposition 7.5 and Proposition 7.6 as
as claimed. \(\square \)
Proof of Proposition 2.6
The value \(\Phi _{t}(\lambda )\) is computed by first taking the argument of \(\lambda \) to obtain \(\theta \) and then applying the map in (7.7) to obtain \(\phi .\) Thus, the first result is just a restatement of Proposition 2.5. For the uniqueness claim, suppose a measure \(\mu \) on \(\Sigma _{t}\) has the form
Then the distribution of the argument \(\theta \) of \(\lambda \) will be, by integrating out the radial variable, \(2\log (r_{t}(\theta ))g(\theta )~d\theta .\) The distribution of \(\phi \) will then be
The only way this can reduce to Biane’s measure as computed in (7.13) is if g coincides with \(w_{t}.\) \(\square \)
References
Biane, P.: On the free convolution with a semi-circular distribution. Indiana Univ. Math. J. 46, 705–718 (1997)
Biane, P.: Free Brownian motion, free stochastic calculus and random matrices. In: Free Probability Theory (Waterloo, ON, 1995). Fields Institute Communications 12, pp. 1–19. American Mathematical Society, Providence, RI (1997)
Biane, P.: Segal–Bargmann transform, functional calculus on matrix spaces and the theory of semi-circular and circular systems. J. Funct. Anal. 144, 232–286 (1997)
Biane, P., Speicher, R.: Stochastic calculus with respect to free Brownian motion and analysis on Wigner space. Probab. Theory Relat. Fields 112, 373–409 (1998)
Brown, L.G.: Lidskiĭ’s theorem in the type II case. In: Geometric Methods in Operator Algebras (Kyoto. 1983), Pitman Research Notes in Mathematics Series, vol. 123, pp. 1–35. Longman Scientific and Technical, Harlow (1986)
Burda, Z., Grela, J., Nowak, M.A., Tarnowski, W., Warchoł, P.: Dysonian dynamics of the Ginibre ensemble. Phys. Rev. Lett. 113, 104102 (2014)
Demni, N., Hamdi, T.: Support of the Brown measure of the product of a free unitary Brownian motion by a free self-adjoint projection. J. Funct. Anal. 282, 109362 (2022)
Evans, L.C.: Partial differential equations. In: Graduate Studies in Mathematics, vol. 19, 2nd edn. American Mathematical Society, Providence, RI (2010)
Fuglede, B., Kadison, R.V.: On determinants and a property of the trace in finite factors. Proc. Natl. Acad. Sci. USA 37, 425–431 (1951)
Fuglede, B., Kadison, R.V.: Determinant theory in finite factors. Ann. Math. 55, 520–530 (1952)
Girko, V.L.: The circular law. Teor. Veroyatnost. i Primenen. 29, 669–679 (1984). (Russian)
Grela, J., Nowak, M.A., Tarnowski, W.: Eikonal formulation of large dynamical random matrix models. Phys. Rev E 104, 054111 (2021)
Gudowska-Nowak, E., Janik, R.A., Jurkiewicz, J., Nowak, M.A.: Infinite products of large random matrices and matrix-valued diffusion. Nucl. Phys. B 670, 479–507 (2003)
Hall, B.C.: Quantum theory for mathematicians. In: Graduate Texts in Mathematics, 267. Springer, New York (2013)
Hall, B.C.: PDE methods in random matrix theory. In: Harmonic Analysis and Applications, pp. 77–124. Springer (2021)
Hall, B.C., Ho, C.-W.: The Brown measure of the sum of a self-adjoint element and an imaginary multiple of a semicircular element. Lett. Math. Phys. 112, 19 (2022)
Hall, B.C., Ho, C.-W.: The Brown measure of a family of free multiplicative Brownian motions. arXiv:2104.07859 [math.PR]
Hall, B.C., Kemp, T.: Brown measure support and the free multiplicative Brownian motion. Adv. Math. 355, 106771 (2019)
Ho, C.-W., Zhong, P.: Brown measures of free circular and multiplicative Brownian motions with self-adjoint and unitary initial conditions. J. Eur. Math. Soc., to appear
Kemp, T.: The large-\(N\) limits of Brownian motions on \(\sf GL_{N}\). Int. Math. Res. Not. 2016, 4012–4057 (2016)
Kümmerer, B., Speicher, R.: Stochastic integration on the Cuntz algebra \(O_{\infty }\). J. Funct. Anal. 103, 372–408 (1992)
Lohmayer, R., Neuberger, H., Wettig, T.: Possible large-N transitions for complex Wilson loop matrices. J. High Energy Phys. 2008(11), 053 (2008)
Mingo, J.A., Speicher, R.: Free probability and random matrices. In: Fields Institute Monographs, vol. 35. Springer, New York; Fields Institute for Research in Mathematical Sciences, Toronto, ON (2017)
Nikitopoulos, E.A.: Itô’s formula for noncommutative \(C^{2}\) functions of free Itô processes with respect to circular Brownian motion. arXiv:2011.08493 [math.OA]
Acknowledgements
The authors thank Ching-Wei Ho and Maciej Nowak for useful discussions. We also thank the referee for helping us to streamline the presentation. The figures were generated in Mathematica and the Mathematica notebook file is available online at https://www.notebookarchive.org/driver-hall-kemp-plots-and-simulations--2019-05-bjnnq2h/.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Hall’s research supported in part by a grant from the Simons Foundation.
Kemp’s research supported in part by NSF CAREER Award DMS-1254807, NSF Award DMS-1800733, and NSF Award DMS-2055340.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Driver, B.K., Hall, B. & Kemp, T. The Brown measure of the free multiplicative Brownian motion. Probab. Theory Relat. Fields 184, 209–273 (2022). https://doi.org/10.1007/s00440-022-01142-z
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00440-022-01142-z