Abstract
We provide a perturbative expansion for the empirical spectral distribution of a Hermitian matrix with large size perturbed by a random matrix with small operator norm whose entries in the eigenvector basis of the first matrix are independent with a variance profile. We prove that, depending on the order of magnitude of the perturbation, several regimes can appear, called perturbative and semi-perturbative regimes. Depending on the regime, the leading terms of the expansion are related either to the one-dimensional Gaussian free field or to free probability theory.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
It is a natural and central question, in mathematics and physics, to understand how the spectral properties of an operator are altered when the operator is subject to a small perturbation. This question is at the center of perturbation theory and has been studied in many different contexts. We refer the reader to Kato’s book [17] for a thorough account on this subject. In this text, we provide a perturbative expansion for the empirical spectral distribution of a Hermitian matrix with large size perturbed by a random matrix with small operator norm whose entries in the eigenvector basis of the first one are independent with a variance profile. More explicitly, let \(D_n\) be an \(n\times n\) Hermitian matrix, that, up to a change in basis, we suppose diagonal.Footnote 1 We denote by \(\mu _n\) the empirical spectral distribution of \(D_n\). This matrix is additively perturbed by a random Hermitian matrix \(\varepsilon _n X_n\) whose entries are chosen at random independently and scaled so that the operator norm of \(X_n\) has order one. We are interested in the empirical spectral distribution \(\mu _n^\varepsilon \) of
in the regime where the matrix size n tends to infinity and \(\varepsilon _n\) tends to 0. We shall prove that, depending on the order of magnitude of the perturbation, several regimes can appear. We suppose that \(\mu _n\) converges to a limiting measure \(\rho (\lambda )\mathrm {d}\lambda \) and that the variance profile of the entries of \( X_n\) has a macroscopic limit \(\sigma _{{\text {d}}}\) on the diagonal and \(\sigma \) elsewhere. We then prove that there is a deterministic function F and a Gaussian random linear form \(\mathrm {d}Z\) on the space of \(\mathcal {C}^6\) functions on \(\mathbb {R}\), both depending only on the limit parameters of the model \(\rho ,\sigma \) and \(\sigma _{{\text {d}}}\) such that if one defines the distribution \(\mathrm {d}F: \phi \longmapsto -\int \phi '(s)F(s)\mathrm {d}s\), then, for large n:
and if, moreover, \( n^{-1}\ll \varepsilon _n\ll n^{-1/3},\) then convergence (2) can be refined as follows:
In Sect. 3, several figures show a very good matching of random simulations with these theoretical results. The definitions of the function F and of the process Z are given below in (6) and (7). In many cases, the linear form \(\mathrm {d}F\) can be interpreted as the integration with respect to the signed measure \(F'(x)\mathrm {d}x\). The function F is related to free probability theory, as explained in Sect. 4 below, whereas the linear form \(\mathrm {d}Z\) is related to the so-called one-dimensional Gaussian free field defined, for instance, at [14, Sect. 4.2]. If the variance profile of \(X_n\) is constant, then it is precisely the Laplacian of the Gaussian free field, defined in the sense of distributions.
The transition at \(\varepsilon _n\sim n^{-1}\) is the well-known transition, in quantum mechanics, where the perturbative regime ends. Indeed, one can distinguish the two following regimes:
-
The regime \(\varepsilon _n\ll n^{-1}\), called the perturbative regime (see [15]): the size of the perturbation (i.e. its operator norm) is much smaller than the typical spacing between two consecutive eigenvalues (level spacing), which is of order \(n^{-1}\) in our setting.
-
The regime \(n^{-1}\ll \varepsilon _n\ll 1\), sometimes called the semi-perturbative regime, where the size of the perturbation is not small compared to the level spacing. This regime concerns many applications [1, 19] in the context of covariance matrices and applications to finance.
A surprising fact discovered during this study is that the semi-perturbative regime \(n^{-1}\ll \varepsilon _n\ll 1\) decomposes into infinitely many sub-regimes. In the case \(n^{-1}\ll \varepsilon _n\ll n^{-1/3}\), the expansion of \(\mu _n^\varepsilon -\mu _n\) contains a single deterministic term before the random term \(\frac{\varepsilon _n}{n}\mathrm {d}Z\). In the case \(n^{-1/3}\ll \varepsilon _n\ll n^{-1/5}\), the expansion of \(\mu _n^\varepsilon -\mu _n\) contains two of them. More generally, for all positive integer p, when \(n^{-1/(2p-1)}\ll \varepsilon _n\ll n^{-1/(2p+1)}\), the expansion contains p of them. For computational complexity reasons, the only case we state explicitly is the first one. We refer the reader to Sect. 6.5 for a discussion around this point.
In the papers [1,2,3,4, 23], Wilkinson, Walker, Allez, Bouchaud et al. have investigated some problems related to this one. Some of these works were motivated by the estimation of a matrix out of the observation of its noisy version. Our paper differs from these ones mainly by the facts that firstly, we are interested in the perturbations of the global empirical distribution of the eigenvalues and not of a single one, and secondly, we push our expansion up to the random term, which does not appear in these papers. Besides, the noises they consider have constant variance profiles (either a Wigner-Dyson noise in the four first cited papers or a rotationally invariant noise in the fifth one). The transition at \(\varepsilon _n\sim n^{-1}\) between the perturbative and the semi-perturbative regimes is already present in these texts. They also consider the transition between the perturbative regime \(\varepsilon _n\ll 1\) and the non-perturbative regime \(\varepsilon _n\asymp 1\). As explained above, we exhibit the existence of an infinity of sub-regimes in this transition and focus on \(\varepsilon _n\ll 1\) for the first order of the expansion and to \(\varepsilon _n\ll n^{-1/3}\) for the second (and last) order. The study of other sub-regimes is postponed to forthcoming papers.
The paper is organized as follows. Results, examples and comments are given in Sects. 2–4, while the rest of the paper, including an appendix, is devoted to the proofs, except for Sect. 6.5, where we discuss the sub-regimes mentioned above.
Notations For \(a_n,b_n\) some real sequences, \(a_n\ll b_n\) (resp. \(a_n\sim b_n\)) means that \(a_n/b_n\) tends to 0 (resp. to 1). Also, \({\mathop {\longrightarrow }\limits ^{P}}\) and \({\mathop {\longrightarrow }\limits ^{\mathrm {dist.}}}\) stand, respectively, for convergence in probability and convergence in distribution for all finite marginals.
2 Main Result
2.1 Definition of the Model and Assumptions
For all positive integer n, we consider a real diagonal matrix \(D_n ={\text {diag}}(\lambda _n(1), \ldots , \lambda _n(n))\), as well as a Hermitian random matrix
and a positive number \(\varepsilon _n\). The normalizing factor \( n^{-1/2}\) and our hypotheses below ensure that the operator norm of \(X_n\) is of order one. We then define, for all n,
We now introduce the probability measures \(\mu _n\) and \(\mu _n^\varepsilon \) as the respective uniform distributions on the eigenvalues (with multiplicity) of \(D_n\) and \(D_n^\varepsilon \). Our aim is to give a perturbative expansion of \(\mu _n^\varepsilon \) around \(\mu _n\).
We make the following hypotheses:
-
(a)
the entries \(x^n_{i,j}\) of \(\sqrt{n}X_n\) are independent (up to symmetry) random variables, centered, with variance denoted by \(\sigma _n^2(i,j)\), such that \(\mathbb {E}|x^n_{i,j}|^{8} \) is bounded uniformly on n, i, j,
-
(b)
there are \(f,\sigma _{{\text {d}}},\sigma \) real functions defined, respectively, on [0, 1], [0, 1] and \([0,1]^2\) such that, for each \(x\in [0,1]\),
$$\begin{aligned} \lambda _n(\lfloor nx\rfloor )\underset{n\rightarrow \infty }{\longrightarrow }f(x)\quad \text { and }\quad \sigma _n^2(\lfloor nx\rfloor ,\lfloor nx\rfloor )\underset{n\rightarrow \infty }{\longrightarrow }\sigma _{{\text {d}}}(x)^2 \end{aligned}$$and for each \(x\ne y\in [0,1]\),
$$\begin{aligned} \sigma _n^2(\lfloor nx\rfloor ,\lfloor ny\rfloor )\underset{n\rightarrow \infty }{\longrightarrow }\sigma ^2(x,y). \end{aligned}$$We make the following hypothesis about the rate of convergence:
$$\begin{aligned} \eta _n := & {} \max \{n\varepsilon _n,1\}\times \sup _{1\le i\ne j\le n}( |\sigma _n^2(i,j)\\&-\sigma ^2(i/n,j/n)|+|\lambda _n(i)-f(i/n)|)\quad \underset{n\rightarrow \infty }{\longrightarrow }\quad 0. \end{aligned}$$
Let us now make some assumptions on the limiting functions \(\sigma \) and f:
-
(c)
the function f is bounded and the push-forward of the uniform measure on [0, 1] by the function f has a density \(\rho \) with respect to the Lebesgue measure on \(\mathbb {R}\) and a compact support denoted by \(\mathcal {S}\),
-
(d)
the variance of the entries of \(X_n\) essentially depends on the eigenspaces of \(D_n\), namely there exists a symmetric function \(\tau (\,\cdot \,,\,\cdot \,)\) on \(\mathbb {R}^2\) such that for all \(x\ne y\), \(\sigma ^2(x,y)=\tau (f(x),f(y))\),
-
(e)
the following regularity property holds: there exist \(\eta _0>0, \alpha >0\) and \(C<\infty \) such that for almost all \(s\in \mathbb {R}\), for all \(t\in [s-\eta _0, s+\eta _0]\), \(|\tau (s,t)\rho (t)-\tau (s,s)\rho (s)|\le C|t-s|^\alpha \).
We add a last assumption which strengthens assumption (c) and makes it possible to include the case where the set of eigenvalues of \(D_n\) contains some outliers:
-
(f)
there is a real compact set \(\widetilde{\mathcal {S}}\) such that
$$\begin{aligned}\max _{1\le i\le n}{\text {dist}}(\lambda _n(i), \widetilde{\mathcal {S}})\underset{n\rightarrow \infty }{\longrightarrow }0.\end{aligned}$$
Remark 1
(About the hypothesis that \(D_n\) is diagonal)
-
(i)
If the perturbing matrix \(X_n\) belongs to the GOE (resp. to the GUE), then its law is invariant under conjugation by any orthogonal (resp. unitary) matrix. It follows that in this case, our results apply to any real symmetric (resp. Hermitian) matrix \(D_n\) with eigenvalues \(\lambda _n(i)\) satisfying the above hypotheses.
-
(ii)
As explained after Proposition 2 below, we conjecture that when the variance profile of \(X_n\) is constant, for \(\varepsilon _n\gg n^{-1}\), we do not need the hypothesis that \(D_n\) is diagonal neither. However, if the perturbing matrix does not have a constant variance profile, then for a non-diagonal \(D_n\) and \(\varepsilon _n\gg n^{-1}\), the spectrum of \(D_n^\varepsilon \) should depend heavily on the relation between the eigenvectors of \(D_n\) and the variance profile, which implies that our results should not remain true.
-
(iii)
At last, it is easy to see that the random process \((Z_\phi )\) introduced at (7) satisfies, for any test function \(\phi \),
$$\begin{aligned} \frac{1}{\varepsilon _n}\sum _{i=1}^n\left( \phi (\lambda _n(i)+\frac{\varepsilon _n}{\sqrt{n}}x_{ii})-\phi (\lambda _n(i))\right) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{\mathrm {dist.}}}\quad Z_\phi . \end{aligned}$$
Thus, regardless of the variance profile, the convergence of (8) rewrites, informally,
A so simple expression, up to a \(o(\varepsilon _n/n)\) error, of the empirical spectral distribution of \(D_n^\varepsilon \), with some independent translations \(\frac{\varepsilon _n}{\sqrt{n}}x_{ii}\), should not remain true without the hypothesis that \(D_n\) is diagonal or that the distribution of \(X_n\) is invariant under conjugation.
2.2 Main Result
Recall that the Hilbert transform, denoted by H[u], of a function u, is the function
and define the function
Note that, by assumptions (c) and (e), F is well defined and supported by \(\mathcal {S}\). Besides, for any \(\phi \) supported on an interval where F is \(\mathcal {C}^1\),
where \(\mathrm {d}F(s)\) denotes the measure \(F'(s)\mathrm {d}s\).
We also introduce the centered Gaussian field, \((Z_\phi )_{\phi \in \mathcal {C}^6}\), indexed by the set of \(\mathcal {C}^6\) complex functions on \(\mathbb {R}\), with covariance defined by
Note that the process \((Z_\phi )_{\phi \in {\mathcal {C}}^6}\) can be represented, for \((B_t)\) is the standard one-dimensional Brownian motion, as
Theorem 1
For all compactly supported \(\mathcal {C}^6\) function \(\phi \) on \(\mathbb {R}\), the following convergences hold:
-
Perturbative regime if \(\varepsilon _n\ll n^{-1}\), then,
$$\begin{aligned} n\varepsilon _n^{-1}(\mu _n^\varepsilon -\mu _n)(\phi ) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{\mathrm {dist.}}}\quad Z_\phi .\end{aligned}$$(8) -
Critical regime if \(\varepsilon _n\sim c/n\), with c constant, then,
$$\begin{aligned} n\varepsilon _n^{-1}(\mu _n^\varepsilon -\mu _n)(\phi ) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{\mathrm {dist.}}}\quad -c\int \phi '(s)F(s)\mathrm {d}s+Z_\phi .\end{aligned}$$(9) -
Semi-perturbative regime if \( n^{-1}\ll \varepsilon _n\ll 1\), then,
and if, moreover, \( n^{-1}\ll \varepsilon _n\ll n^{-1/3}\), then,
Remark 2
(Sub-regimes for \(n^{-1/3}\ll \varepsilon _n\ll 1\)) In the semi-perturbative regime, the reason why we provide an expansion up to a random term, only for \(\varepsilon _n\ll n^{-1/3}\), is that the study of the regime \(n^{-1/3}\ll \varepsilon _n\ll 1\) up to such a precision requires further terms in the expansion of the resolvent of \(D_n^\varepsilon \) that make appear, beside \(\mathrm {d}F\), additional deterministic terms of smaller order, which are much larger than the probabilistic term containing \(Z_\phi \). The computation becomes rather intricate without any clear recursive formula. As we will see in Sect. 6.5, there are infinitely many regimes. Precisely, for any positive integer p, when \(n^{-1/(2p-1)}\ll \varepsilon _n\ll n^{-1/(2p+1)}\), there are p deterministic terms in the expansion before the term in \(Z_\phi \).
Remark 3
(Local law) The approximation
of (10) should stay true even for intervals I with size tending to 0 as the dimension n grows, as long as the size of I stays much larger than the right-hand side term of (30), as can be seen from Proposition 5.
Remark 4
The second part of Hypothesis (b), concerning the speed of convergence of the profile of the spectrum of \(D_n\) as well as of the variance of its perturbation, is needed in order to express the expansion of \(\mu _n^\varepsilon -\mu _n\) in terms of limit parameters of the model \( \sigma \) and \(\rho \). We can remove this hypothesis and get analogous expansions where the terms \(\mathrm {d}F\) and \(\mathrm {d}Z\) are replaced by their discrete counterparts \(\mathrm {d}F_n\) and \(\mathrm {d}Z_n\), defined thanks to the “finite n” empirical versions of the limit parameters \(\sigma \) and \(\rho \).
3 Examples
3.1 Uniform Measure Perturbation by a Band Matrix
Here, we consider the case where \(f(x)=x\), \(\sigma _{{\text {d}}}(x)\equiv m\) and \(\sigma (x,y)=\mathbb {1}_{|y-x|\le \ell }\), for some constants \(m\ge 0\) and \(\ell \in [0,1]\) (the relative width of the band). In this case, \(\tau (\,\cdot \,,\,\cdot \,)=\sigma (\,\cdot \,,\,\cdot \,)^2\), hence
and \((Z_\phi )_{\phi \in \mathcal {C}^6}\) is the centered complex Gaussian process with covariance defined by
Theorem 1 is then illustrated in Fig. 1, where we plotted the cumulative distribution functions.
3.2 Triangular Pulse Perturbation by a Wigner Matrix
Here, we consider the case where \(\rho (x)=(1-|x|)\mathbb {1}_{[-1,1]}(x)\), \(\sigma _d \equiv m\), for some real constant m, and \(\sigma \equiv 1\) (what follows can be adapted to the case \(\sigma (x,y)=\mathbb {1}_{|y-x|\le \ell }\), with a bit longer formulas). In this case, thanks to the formula (9.6) of \(H[\rho (\,\cdot \,)]\) given p. 509 of [18], we get
and the covariance of \((Z_\phi )_{\phi \in \mathcal {C}^6}\) is given by
Theorem 1 is then illustrated in Fig. 2 in the case where \(\varepsilon _n\gg n^{-1/2}\). In Fig. 2, we implicitly use some test functions of the type \(\phi (x)=\mathbb {1}_{x\in I}\) for some intervals I. These functions are not \(\mathcal {C}^6\), and one can easily see that for \(\varepsilon _n\ll n^{-1/2}\), Theorem 1 cannot work for such functions. However, considering imaginary parts of Stieltjes transforms, i.e. test functions
give a perfect matching between the predictions from Theorem 1 and numerical simulations, also for \(\varepsilon _n\ll n^{-1/2}\) (see Fig. 3, where we use Proposition 4 and (17) to compute the theoretical limit).
3.3 Parabolic Pulse Perturbation by a Wigner Matrix
Here, we consider the case where \(\rho (x)=\frac{3}{4}(1-x^2)\mathbb {1}_{[-1,1]}(x)\), \(\sigma _d \equiv m\), for some real constant m, and \(\sigma \equiv 1\) (again, this can be adapted to the case \(\sigma (x,y)=\mathbb {1}_{|y-x|\le \ell }\)). Theorem 1 is then illustrated in Fig. 4. In this case, thanks to the formula (9.10) of \(H[\rho (\,\cdot \,)]\) given p. 509 of [18], we get
and the covariance of \((Z_\phi )_{\phi \in \mathcal {C}^6}\) is given by
4 Relation to Free Probability Theory
Let us now explain how this work is related to free probability theory. If, instead of letting \(\varepsilon _n\) tend to zero, one considers the model
for a fixed \(t>0\), then, by [5, 12, 13, 22], the empirical eigenvalue distribution of \(D_n^t\) has a limit as \(n\rightarrow \infty \), that we shall denote here by \(\mu _t\). The law \(\mu _t\) can be interpreted as the law of the sum of two elements in a non-commutative probability space which are free with an amalgamation over a certain sub-algebra (see [22] for more details). The following proposition relates the function F from (6) to the first order expansion of \(\mu _t\) around \(t=0\).
Proposition 2
For any \(z\in \mathbb {C}\backslash \mathbb {R}\), we have
This is related to the fact that in Eqs. (1)–(4), for \(\varepsilon _n\) large enough, the term \(\varepsilon _n^2\mathrm {d}F\) is the leading term.
In the particular case where \(X_n\) is a Wigner matrix, \(\mu _t\) is the free convolution of the measure \(\rho (\lambda )\mathrm {d}\lambda \) with a semicircle distribution and admits a density \(\rho _t\), by [8, Cor. 2]. Then, Theorem 1 makes it possible to formally recover the free Fokker–Planck equation with null potential:
where \(H[\rho _t ]\) denotes the Hilbert transform of \(\rho _t\). This equation is also called McKean-Vlasov (or Fokker–Planck) equation with logarithmic interaction (see [9,10,11]).
Note also that when \(X_n\) is a Wigner matrix, the hypothesis that \(D_n\) is diagonal is not required to have the convergence of the empirical eigenvalue distribution of \(D_n^t\) to \(\mu _t\) as \(n\rightarrow \infty \). This suggests that, even for non-diagonal \(D_n\), the convergence of (10) still holds when \(X_n\) is a Wigner matrix.
Proof of Proposition 2
By [22, Th. 4.3], we have
where \(C_t(x,z)\) is bounded by \(|\mathfrak {Im}z|^{-1}\) and satisfies the fixed-point equation
Hence as \(t\rightarrow 0\), \(C_t(x, z)\longrightarrow \frac{1}{z-f(x)}\) uniformly in x. Thus
where each o(t) is uniform in \(x\in [0,1]\). Then, by (15), we deduce that
The right-hand side term of the previous equation is precisely the number B(z) introduced at (17) below. Then, one concludes using Proposition 4 from Sect. 6.1. \(\square \)
5 Strategy of the Proof
We shall first prove the convergence results of Theorem 1 for test functions \(\phi \) of the form \(\varphi _z(x):=\frac{1}{z-x}\). This is done in Sect. 6 by writing an expansion of the resolvent of \(D_n^\varepsilon \).
Once we have proved that the convergences hold for the resolvent of \(D_n^\varepsilon \), we can extend them to the larger class of compactly supported \({\mathcal {C}}^6\) functions on \(\mathbb {R}\).
In Sect. 7, we use the Helffer–Sjöstrand formula to extend the convergence in probability in the semi-perturbative regime (10) to the case of compactly supported \({\mathcal {C}}^6\) functions on \(\mathbb {R}\).
In Sect. 8, the convergences in distribution (8), (9) and (11) are proved in two steps. The overall strategy is to apply an extension lemma of Shcherbina and Tirozzi which states that a CLT that applies to a sequence of centered random linear forms on some space can be extended, by density, to a larger space, as long as the variance of the image of these random linear forms by a function \(\phi \) of the larger space is uniformly bounded by the norm of \(\phi \). Therefore, our task is twofold. We need first to prove that the sequences of variables involved in the convergences (8), (9) and (11) can be replaced by their centered counterparts \(n\varepsilon _n^{-1}(\mu _n^\varepsilon (\phi )-\mathbb {E}[\mu _n^\varepsilon (\phi )])\) (i.e. they differ by o(1)). In a second step, we dominate the variance of these latter variables, in order to apply the extension lemma which is precisely stated in Appendix as Lemma 10.
6 Stieltjes Transforms Convergence
As announced in the previous section, we start with the proof of Theorem 1 in the special case of test functions of the type \(\varphi _z:= \frac{1}{z-x}\). We decompose it into two propositions. Their statement and proof are the purpose of the three following subsections. The two last Sects. 6.4 and 6.5 are devoted, respectively, to a local type convergence result and to a discussion about the possibility of an extension of the expansion result to a wider range of rate of convergence of \(\varepsilon _n\), namely beyond \(n^ {-1/3}\).
6.1 Two Statements
Let denote, for \(z\in \mathbb {C}\backslash \mathbb {R}\),
where \((Z_\phi )_{\phi \in \mathcal {C}^6}\) is the Gaussian field with covariance defined by (7). We also introduce, for \(z\in \mathbb {C}\backslash \mathbb {R}\),
and
Proposition 3
Under Hypotheses (a), (b), (f),
-
if \(\varepsilon _n\ll n^{-1}\), then for all \(z\in \mathbb {C}\backslash \mathbb {R}\),
$$\begin{aligned} n\varepsilon _n^{-1}\Delta \mathrm {G}_n(z) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{\mathrm {dist.}}}\quad Z(z)\, \end{aligned}$$(19) -
if \(\varepsilon _n\sim c/n\), with c constant, then for all \(z\in \mathbb {C}\backslash \mathbb {R}\)
$$\begin{aligned} n\varepsilon _n^{-1}\Delta \mathrm {G}_n(z) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{\mathrm {dist.}}}\quad cB(z)+Z(z) \, , \end{aligned}$$(20) -
if \( n^{-1}\ll \varepsilon _n\ll n^{-1/3}\), then for all \(z\in \mathbb {C}\backslash \mathbb {R}\)
$$\begin{aligned} n\varepsilon _n^{-1} \left( \Delta \mathrm {G}_n(z)-\varepsilon _n^2B(z)\right) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{\mathrm {dist.}}}\quad Z(z)\,. \end{aligned}$$(21) -
if \( n^{-1}\ll \varepsilon _n\ll 1\), then for all \(z\in \mathbb {C}\backslash \mathbb {R}\),
$$\begin{aligned} \varepsilon _n^{-2}\Delta \mathrm {G}_n(z) - B(z) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{P}}\quad 0 \,. \end{aligned}$$(22)
Remark
Note that (20) is merely an extension of (21) in the critical regime.
The following statement expresses B(z) as the image of a \(\varphi _z\) by a linear form. So, in the expansion of the previous proposition, both quantities Z(z) and B(z) depend linearly on \(\varphi _z\). Note that as F vanishes at \(\pm \infty \), Proposition 4 does not contradicts the fact that as |z| gets large, \(B(z)=O(|z|^{-3})\).
Proposition 4
Under Hypotheses (c), (d), (e), for any \(z\in \mathbb {C}\backslash \mathcal {S}\), for F defined by (6),
6.2 Proof of Proposition 3
The proof is based on a perturbative expansion of the resolvent \(\frac{1}{n} {\text {Tr}}\frac{1}{z-D_n^\varepsilon }\). To make notations lighter, we shall sometimes suppress the subscripts and superscripts n, so that \(D_n^\varepsilon \), \(D_n\), \(X_n\) and \(x_{i,j}^n\) will be, respectively, denoted by \(D^\varepsilon \), D, X and \(x_{i,j}\). Let us fix \(z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}\). We can deduce from the expansion of the resolvent of \(D^\varepsilon \):
with
The purpose of the four following claims is to describe the asymptotic behavior of each of these four terms.
Claim 1
The finite dimension marginals of the centered process
converge in distribution to those of the centered Gaussian process \((Z(z))_{z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}}\). Besides, there is \(C>0\) such that for any \(z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}\),
Proof
Estimate (23) follows from
and from the existence of a uniform upper bound for \(\sigma _n^2(i,i)\) which comes from Hypothesis (a) which stipulates that the 8-th moments of the entries \(x_{i,j}\) are uniformly bounded.
We turn now to the proof of the convergence in distribution of \(n\varepsilon _n^{-1}A_n(z)\) which actually does not depend on the sequence \((\varepsilon _n)\). For all \(\alpha _1,\beta _1,\dots ,\alpha _p,\beta _p \in \mathbb {C}\) and for all \(z_1,\dots ,z_p \in \mathbb {C}\backslash \widetilde{\mathcal {S}}\),
for \(\displaystyle { \xi _n(i,j) = \frac{\alpha _i}{(z_i-\lambda _n(j))^2} + \frac{\beta _i}{(\overline{z_i}-\lambda _n(j))^2} }\).
On one the hand, by dominated convergence, the covariance matrix of the above two dimensional random vector converges.
On the other hand, \(\mathbb {E}|x_{i,j}|^4\) is uniformly bounded in i, j and n, by Hypothesis (a). Moreover, for n large enough, for all i, j,
Hence, the conditions of Lindeberg central limit theorem are satisfied and the finite dimension marginals of the process \((n\varepsilon _n^{-1} A_n(z))_{z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}}\) converge in distribution to those of the centered Gaussian process \((Z_z)_{z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}}\) defined by its covariance structure
and by the fact that \(\overline{Z(z)} = Z(\overline{z})\) which comes from \(\overline{A_n(z)} = A_n(\overline{z})\). \(\square \)
Claim 2
There is a constant C such that, for \(\eta _n\) as in Hypotheses (b),
-
if \( \varepsilon _n\ll n^{-1}\), then
$$\begin{aligned} \mathbb {E}[|n\varepsilon _n^{-1}B_n(z)|^2]\le \frac{C(n\varepsilon _n)^2}{{\text {dist}}(z, \widetilde{\mathcal {S}})^6}+ \frac{C\eta _n^2}{{\text {dist}}(z, \widetilde{\mathcal {S}})^8}, \end{aligned}$$ -
if \(\varepsilon _n\sim c/n\) or if \( n^{-1} \ll \varepsilon _n\ll 1\), then
$$\begin{aligned} \quad \mathbb {E}[| n\varepsilon _n^{-1}(B_n(z)-\varepsilon _n^2B(z))|^2] \le \frac{C \varepsilon _n^2}{{\text {dist}}(z, \widetilde{\mathcal {S}})^6}+\frac{C \eta _n^2 }{{\text {dist}}(z, \widetilde{\mathcal {S}})^8}. \end{aligned}$$
Proof
Remind that,
Introduce the variable \(b_n^{\circ }(z)\) obtained by centering the variable \(n\varepsilon _n^{-2} B_n(z)\):
and the defect variable
In the two regimes \(\varepsilon _n \ll n^{-1}\) and \(\varepsilon _n \ge c/n\), we want to dominate the \(L^2\) norms, respectively, of \( n\varepsilon _n^{-1}B_n(z)=\varepsilon _nb_n^{\circ }(z)+n\varepsilon _n(\delta _n(z)+B(z))\quad \text {and}\quad n\varepsilon _n^{-1}(B_n(z)-\varepsilon _n^2B(z))=\varepsilon _nb_n^{\circ }+ n\varepsilon _n\delta _n(z).\)
For this purpose, we successively dominate \(b_n^{\circ }\), \(\delta _n(z)\) and B(z).
Using the independence of the \(x_{i,j}\)’s, the fact that they are bounded in \(L^4\) and the fact that z stays at a macroscopic distance of the \(\lambda _n(i)\)’s, we can write for all \(z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}\)
Now, the term \(\delta _n(z)\) rewrites
Since, for \(M_\sigma :=\sup _{0\le x\ne y\le 1}\sigma (x,y)^2\) and for any fixed \(z\notin \widetilde{\mathcal {S}}\), the function
is \(C{\text {dist}}(z,\widetilde{\mathcal {S}})^{-4}\)-Lipschitz, for C a universal constant, by Hypothesis (b),
Finally, the expression of B(z) given in (17) implies,
Collecting estimations (24), (25) and (26), we conclude. \(\square \)
Claim 3
There is a constant C such that for any \(z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}\),
Proof
We start by writing for all \(z\in \mathbb {C}\backslash \widetilde{\mathcal {S}}\)
Generically, the set of “edges” \(\{(l,m), (m,p), (p,l)\}\) must be equal to the set \(\{(i,j), (j,k), (k,i)\}\) in order to get a nonzero term. Therefore, the complexity of the previous sum is \(O(n^3)\). Note that other nonzero terms involving third or fourth moments are much less numerous. Hence,
\(\square \)
Claim 4
There is a constant C such that for any \(z\in \mathbb {C}\backslash \mathbb {R}\),
Proof
Remind that,
Hence,
The inequality of the last line takes into account that
-
the \(L^8\) norm of the entries of \(\sqrt{n}X\) is uniformly bounded
-
the norm of the entries of X is of order \(n^{-1/2}\)
-
the norm of the coefficients of \((z-D)^{-1}\) is smaller than \({\text {dist}}(z,\widetilde{\mathcal {S}})^{-1}\)
-
the complexity of the sum defining the trace is of order \(O(n^5)\) since its non-null terms are encoded by four edges trees which have therefore five vertices.
\(\square \)
We gather now the results of the previous claims.
For any rate of convergence of \(\varepsilon _n\), Claim 1 proves that the process \(n\varepsilon _n^{-1}A_n(z)\) converges in distribution to the centered Gaussian variable Z(z). Moreover,
-
if \(\varepsilon _n \ll n^{-1}\), then as Claims 2, 3 and 4 imply that the processes \(n\varepsilon _n^{-1}B_n(z)\), \(n\varepsilon _n^{-1}C_n(z)\) and \(n\varepsilon _n^{-1}R_n^\varepsilon (z)\) converge to 0 in probability, we can conclude, by Slutsky’s theorem, that for any \(z\in \mathbb {C}\setminus \mathbb {R}\):
$$\begin{aligned} n\varepsilon _n^{-1}\Delta \mathrm {G}_n(z) \xrightarrow [n\rightarrow \infty ]{\text {dist}} Z(z) \end{aligned}$$ -
if \(\varepsilon _n \sim \frac{c}{n}\), then, as Claims 2, 3 and 4 imply that the processes \(n\varepsilon _n^{-1}B_n(z)\), \(n\varepsilon _n^{-1}C_n(z)\) and \(n\varepsilon _n^{-1}R_n^\varepsilon (z)\) converge, respectively, to cB(z), 0 and 0 in probability, we can conclude, by Slutsky’s theorem, that for any \(z\in \mathbb {C}\setminus \mathbb {R}\):
$$\begin{aligned} n\varepsilon _n^{-1}\Delta \mathrm {G}_n(z) \xrightarrow [n\rightarrow \infty ]{\text {dist}} Z(z) + cB(z) \end{aligned}$$ -
if \(n^{-1}\ll \varepsilon _n\ll n^{-1/3}\), then, as Claims 2, 3 and 4 imply that the three processes \(n\varepsilon _n^{-1}(B_n(z)-\varepsilon _n^2B(z))\), \(n\varepsilon _n^{-1}C_n(z)\) and \(n\varepsilon _n^{-1}R_n^\varepsilon (z)\) converge to 0 in probability, we can conclude, by Slutsky’s theorem, that for any \(z\in \mathbb {C}\setminus \mathbb {R}\):
$$\begin{aligned} n\varepsilon _n^{-1} \left( \Delta \mathrm {G}_n(z)-\varepsilon _n^2B(z)\right) \quad {\mathop {\underset{n\rightarrow \infty }{\longrightarrow }}\limits ^{\mathrm {dist.}}}\quad Z(z) \end{aligned}$$
Regarding the convergence in probability (22), in the case \(n^{-1} \ll \varepsilon _n\ll 1\), Claims 1, 2, 3 and 4 imply that the processes \(\varepsilon _n^{-2} A_n(z)\), \(\varepsilon _n^{-2} B_n(z) - B(z)\), \(\varepsilon _n^{-2} C_n(z)\) and \(\varepsilon _n^{-2} R_n^\varepsilon (z)\) converge to 0.
This finishes the proof of the convergences of Proposition 3. \(\square \)
6.3 Proof of Proposition 4
Recall that
Recall that \(\rho \) is the density of the push-forward of the uniform measure on [0, 1] by the map f.
Let \(\tau \) be as in Hypothesis (d). We have
By a partial fraction decomposition, we have for all \(a \ne b\)
Thus, as the Lebesgue measure of the set \(\left\{ (y_1,y_2) \in [0,1]^2 \ \,;\, \ y_1 = y_2 \right\} \) is null, we have
Moreover, for \(\varphi _z\) the function \(\varphi _z: x \longmapsto \frac{1}{z-x}\), we obtain
Now, we want to prove that \(\displaystyle B(z) = -\int _{\mathbb {R}^2} \frac{\tau (s,t) \ \rho (s) \ \rho (t)}{t-s} \ \varphi _z'(s) \ \mathrm {d}s \mathrm {d}t\).
To do this, we will use a symmetry argument: in fact both terms in \(\varphi _z(t)\) and \(\varphi _z(s)\) neutralize each other, and it remains only to prove that we did not remove \(\infty \) to \(\infty \) and that the remaining term has the desired form.
Let us define
By the Taylor–Lagrange inequality we obtain:
So that, since \(\rho \) is a density, by dominated convergence, we have
Moreover, by symmetry, for any \(\eta \),
So
where for \(\eta > 0\) and \(s \in \mathbb {R}\), we define
Note that that by definition of the function F given at (6), for any s, we have
Thus by (27) and (28), to conclude the proof of Proposition 4, by dominated convergence, one needs only to state that \(F_\eta \) is dominated, uniformly in \(\eta \), by an integrable function. This follows from the following computation.
Note first that by symmetry, we have
Let \(M>0\) such that the support of the function \(\rho \) is contained in \([-M,M]\). Then, for \(\eta _0,\alpha ,C\) as in Hypothesis (e), using the expression of \(F_\eta (s)\) given at (29), we have
\(\square \)
6.4 A Local Type Convergence Result
One can precise the convergence (22) by replacing the complex variable z by a complex sequence \((z_n)\) which converges slowly enough to the real axis. This convergence won’t be used in the sequel. As it is discussed in [7], this type of result is a first step toward a local result for the empirical distribution.
Proposition 5
Under Hypotheses (a), (b), (f), if \(n^{-1}\ll \varepsilon _n\ll 1\), then for any non-real complex sequence \((z_n)\), such that
the following convergence holds
Remark
In the classical case where \(\displaystyle {\dfrac{\eta _n}{n\varepsilon _n} = \sup _{i\ne j }( |\sigma _n^2(i,j)-\sigma ^2(i/n,j/n)|+|\lambda _n(i)}\)\(\displaystyle {-f(i/n)|)}\) is of order \(\dfrac{1}{n}\), the above assumption boils down to \(\mathfrak {Im}(z_n) \gg \max \left\{ (n\varepsilon _n)^{-1/2} \ , \ \varepsilon _n^{2/5} \right\} \).
Proof
Assume \(n^{-1} \ll \varepsilon _n\ll 1\). One can directly obtain, for all non-real complex sequences \((z_n)\), that
-
by Claim 1, if \({\text {dist}}(z_n,\widetilde{\mathcal {S}}) \gg (n\varepsilon _n)^{-1/2}\), then
$$\begin{aligned}\mathbb {E}\left[ \left| \varepsilon _n^{-2}A_n(z_n)\right| ^2\right] \le \frac{C}{ (n\varepsilon _n)^2 \ {\text {dist}}(z_n, \widetilde{\mathcal {S}})^4} \underset{n\rightarrow \infty }{\longrightarrow } 0,\end{aligned}$$ -
by Claim 2, if \({\text {dist}}(z_n,\widetilde{\mathcal {S}}) \gg \max \left\{ n^{-1/3} \ , \ (\eta _n/(n\varepsilon _n ))^{1/4}\right\} \), then
$$\begin{aligned}{\mathbb {E}}\left[ \left| \varepsilon _n^{-2}B_n(z_n) - B(z_n)\right| ^2\right] \le \frac{C}{n^{2}{\text {dist}}(z_n, \widetilde{\mathcal {S}})^6}+\frac{C \eta _n^2 }{(n\varepsilon _n)^{2} \ {\text {dist}}(z_n, \widetilde{\mathcal {S}})^8} \quad \underset{n\rightarrow \infty }{\longrightarrow }\quad 0,\end{aligned}$$ -
by Claim 3, if \({\text {dist}}(z_n,\widetilde{\mathcal {S}}) \gg \left( \varepsilon _n/n\right) ^{1/4}\), then
$$\begin{aligned} \mathbb {E}\left[ \left| \varepsilon _n^{-2}C_n(z_n)\right| ^2\right] \le \frac{C\varepsilon _n^2}{n^{2} \ {\text {dist}}(z_n,\widetilde{\mathcal {S}})^{8}} \quad \underset{n\rightarrow \infty }{\longrightarrow }\quad 0, \end{aligned}$$ -
by Claim 4, if \( |\mathfrak {Im}(z_n)| {\text {dist}}(z_n,\widetilde{\mathcal {S}})^4 \gg \varepsilon _n^{2}\), then
$$\begin{aligned} \mathbb {E}\left[ \left| \varepsilon _n^{-2}R_n^\varepsilon (z_n)\right| ^2\right] \le \frac{O(\varepsilon _n^4)}{|\mathfrak {Im}(z_n)|^2 {\text {dist}}(z_n,\widetilde{\mathcal {S}})^8} \quad \underset{n\rightarrow \infty }{\longrightarrow }\quad 0. \end{aligned}$$
Therefore, when
the four processes, \(\varepsilon _n^{-2}A_n(z_n)\), \(\varepsilon _n^{-2}B_n(z_n) - B(z_n)\), \(\varepsilon _n^{-2}C_n(z_n)\) and \(\varepsilon _n^{-2}R_n^\varepsilon (z_n)\) converge to 0 in probability. Since \({\text {dist}}(z_n,\widetilde{\mathcal {S}}) \ge \mathfrak {Im}(z_n)\), the above condition is implied by
Observing finally that the two terms \(n^{-1/3}\) and \(\left( \frac{\varepsilon _n}{n}\right) ^{1/4}\) are dominated by the maximum of the three other ones, we conclude the proof. \(\square \)
6.5 Possible Extensions to Larger \(\varepsilon _n\)
The convergence in distribution result of Theorem 1 is valid for \(\varepsilon _n\ll n^{-1/3}\) but fails above \(n^{-1/3}\). Let us consider, for example, the case where \(n^{-1/3}\ll \varepsilon _n\ll n^{-1/5}\). In this case, the contribution of the first term \(A_n(z)\) in the expansion of \(\Delta \mathrm {G}_n(z)\) which yields the random limiting quantity is dominated not only by the term \(B_n(z)\) as it used to be previously. It is also dominated by a further and smaller term \(D_n(z)\) of the expansion
with:
In this case, the random term Z(z) is still produced by \(A_n(z)\) and has an order of magnitude of \(\varepsilon _n/n\). Meanwhile, the term \(D_n(z)\) writes
All the indices satisfying \(j=l\) contribute to the previous sum, since they produce a term in \(|x_{i,l}|^2 |x_{k,l}|^2\). Their cardinality is of order \(n^3\). Therefore, the term \(D_n(z)\) is of order \(\varepsilon _n^4\) which prevails on the order \(\varepsilon _n/n\) of \(A_n(z)\), as soon as \(\varepsilon _n\gg n^{-1/3}\). One can also observe that the odd terms \(C_n(z)\) and \(E_n(z)\) in the expansion are negligible with respect to \(A_n(z)\) due to the fact that the entries \(x_{i,j}\) are centered. One can then state an analogous result to Proposition 3, but the deterministic limiting term D(z) arising from \(D_n(z)\) does not find a nice expression as the image of \(\varphi _z\) by a linear form as it was the case for B(z) in Proposition 4. Therefore, we did not state an extension of Theorem 1.
More generally, for all positive integer p, when \(n^{-1/(2p-1)}\ll \varepsilon _n\ll n^{-1/(2p+1)}\), the expansion will contain p deterministic terms, produced by the even variables, \(B_n(z)\), \(D_n(z)\), \(F_n(z)\), \(H_n(z)\)\(\dots \) All the other odd terms, \(C_n(z)\), \(E_n(z)\), \(G_n(z)\)\(\dots \) being negligible due to the centering of the entries. The limits of the even terms \(B_n(z)\), \(D_n(z)\), \(F_n(z)\), \(H_n(z)\)\(\dots \) can be expressed thanks to operator-valued free probability theory, using the results of [22] (namely Th. 4.1), but expressing these limits as the images of \(\varphi _z\) by linear forms is a quite involved combinatorial problem that we did not solve yet.
7 Convergence in Probability in the Semi-Perturbative Regime
Our goal now is to extend the convergence in probability result (22) of Proposition 3, proved for test functions \(\varphi _z(x) := \frac{1}{z-x}\), to any \({\mathcal {C}}^6\) and compactly supported function on \(\mathbb {R}\). We do it in the following lemma by using the Helffer–Sjöstrand formula which is stated in Proposition 9 of Appendix.
Lemma 6
If \(n^{-1} \ll \varepsilon _n \ll 1\), then, for any compactly supported \({\mathcal {C}}^6\) function \(\phi \) on \(\mathbb {R}\),
Proof
Let us introduce the Banach space \(\mathcal {C}^1_{{\text {b}},{\text {b}}}\) of bounded \(\mathcal {C}^1\) functions on \(\mathbb {R}\) with bounded derivative, endowed with the norm \(\Vert \phi \Vert _{\mathcal {C}^1_{{\text {b}},{\text {b}}}}:=\Vert \phi \Vert _\infty +\Vert \phi '\Vert _\infty \).
On this space, let us define the random continuous linear form
Convergence (22) of Proposition 3 can now be formulated as
Actually, we can be more precise by adding the upper bounds of Claims 1, 2, 3 and 4, and obtain, uniformly in z,
Now, let \(\phi \) be a compactly supported \(\mathcal {C}^{6}\) function on \(\mathbb {R}\) and let us introduce the almost analytic extension of degree 5 of \(\phi \) defined by
An elementary computation gives, by successive cancellations, that
Furthermore, by Helffer–Sjöstrand formula (Proposition 9), for \(\chi \in {\mathcal {C}}^\infty _c(\mathbb {C};[0,1])\) a smooth cutoff function with value one on the support of \(\phi \),
where \(\mathrm {d}^2 z\) denotes the Lebesgue measure on \(\mathbb {C}\).
Note that by (32), \(z \mapsto \mathbb {1}_{y\ne 0} \frac{\bar{\partial }(\widetilde{\phi }_5(z) \chi (z))}{y^5}\) is a continuous compactly supported function and that \(z\in \mathbb {C}\mapsto \mathbb {1}_{y\ne 0} y^5\varphi _z\in \mathcal {C}^1_{{\text {b}},{\text {b}}}\) is continuous, hence,
Therefore, using the Cauchy–Schwarz inequality and the fact that \(\chi \) has compact support at the second step, for a certain constant C, we have
Since the function \(\left| \frac{\bar{\partial }(\widetilde{\phi }_5(z) \chi (z))}{y^5}\right| ^2\) is continuous and compactly supported and that, by (31), for \(n^{-1} \ll \varepsilon _n \ll 1\), uniformly in z,
Thus, for any compactly supported \(\mathcal {C}^6\) function on \(\mathbb {R}\),
which implies that \(\Pi _n(\phi )\) converges to 0 in probability. \(\square \)
8 Convergence in Distribution Toward the Gaussian Variable \(Z_\phi \)
The purpose of this section is to extend the convergences in distribution of Proposition 3, from test functions of the type \(\varphi _z:= \frac{1}{z-x}\), to compactly supported \({\mathcal {C}}^6\) functions on \(\mathbb {R}\). To do so, we will use an extension lemma of Shcherbina and Tirozzi, stated in Lemma 10 of Appendix, which concerns the convergence of a sequence of centered random fields with uniformly bounded variance. Hence, we need to show first that our non-centered random sequence is not far from being centered, which is done in Sect. 8.1 by using again the Helffer–Sjöstrand formula (9). In Sect. 8.2, we dominate the variance of this centered random field thanks to another result of Shcherbina and Tirozzi stated in Proposition 11 of Appendix. Section 8.3 collects the preceding results to conclude the proof.
8.1 Coincidence of the Expectation of \(\mu _n^\varepsilon \) with Its Deterministic Approximation
The asymptotic coincidence of the expectation of \(\mu _n^\varepsilon \) with its deterministic approximation is the content of next lemma:
Lemma 7
Let us define, for \(\phi \) a \(\mathcal {C}^1\) function on \(\mathbb {R}\),
Then, as \(n\rightarrow \infty \), for any compactly supported \(\mathcal {C}^6\) function \(\phi \) or any \(\phi \) of the type \(\varphi _z(x)=\frac{1}{z-x}\), \(z\in \mathbb {C}\backslash \mathbb {R}\), we have
Proof
First note that, as the variables \(x_{i,j}\) are centered, \(\mathbb {E}[A_n(z)] = 0\). Moreover, by adding the renormalized upper bounds of Claims 2, 3 and 4 one can directly obtain the two following inequalities for any \(z\in \mathbb {C}\setminus \mathbb {R}\):
-
If \( \varepsilon _n\ll n^{-1}\), then
$$\begin{aligned} |\Lambda _n(\varphi _z)|= & {} n\varepsilon _n^{-1}| \mathbb {E}[\Delta \mathrm {G}_n(z)]| \\ \\\le & {} n\varepsilon _n^{-1}\left( |\mathbb {E}[A_n(z)]| + \mathbb {E}[|B_n(z)|] + \mathbb {E}[|C_n(z)|] + \mathbb {E}[|R_n^\varepsilon (z)|] \right) \\ \\\le & {} \frac{ C(n\varepsilon _n + \eta _n) }{ \min \left\{ {\text {dist}}(z, \widetilde{\mathcal {S}})^3, {\text {dist}}(z, \widetilde{\mathcal {S}})^4,|\mathfrak {Im}(z)| {\text {dist}}(z, \widetilde{\mathcal {S}})^4\right\} } \quad \underset{n\rightarrow \infty }{\longrightarrow }\quad 0\,. \end{aligned}$$ -
If \(\varepsilon _n\sim c/n\) or \(n^{-1}\ll \varepsilon _n\ll n^{-1/3}\), then
$$\begin{aligned} |\Lambda _n(\varphi _z)|= & {} n\varepsilon _n^{-1}|\mathbb {E}[\Delta \mathrm {G}_n(z)-\varepsilon _n^{2}B(z)]| \\ \\\le & {} n\varepsilon _n^{-1}\left( |\mathbb {E}[A_n(z)]|\! +\! \mathbb {E}[|B_n(z)-\varepsilon _n^{2}B(z)|]\! +\! \mathbb {E}[|C_n(z)|] + \mathbb {E}[|R_n^\varepsilon (z)|] \right) \\ \\\le & {} \frac{ C(\varepsilon _n + \eta _n + n\varepsilon _n^3) }{ \min \left\{ {\text {dist}}(z, \widetilde{\mathcal {S}})^3, {\text {dist}}(z, \widetilde{\mathcal {S}})^4,|\mathfrak {Im}(z)| {\text {dist}}(z, \widetilde{\mathcal {S}})^4\right\} } \quad \underset{n\rightarrow \infty }{\longrightarrow }\quad 0 \, . \end{aligned}$$
Hence, in all cases, \( \Lambda _n(\varphi _z)\underset{n\rightarrow \infty }{\longrightarrow }0 \).
The extension of this result to compactly supported \(\mathcal {C}^6\) test functions on \(\mathbb {R}\) goes the same way as for \(\Pi _n\) in the proof of Lemma 6. \(\square \)
8.2 Domination of the Variance of \(\mu _n^\varepsilon \)
The second ingredient goes through a domination of the variance of \(\mu _n^\varepsilon (\phi )\):
Lemma 8
Let \(s>5\). There is a constant C such that for each n and each \(\phi \in \mathcal {H}_s\),
Proof
By Proposition 11, it suffices to prove that
are bounded independently of n.
Note that for \(\Delta \mathrm {G}_n(z)\) defined in (18),
Moreover, the sum of the inequalities of Claims 1, 2, 3 and 4 yields
Let \(M>0\) such that \(\widetilde{\mathcal {S}}\subset [-M,M]\). Then
Thus \( {{\text {dist}}(z,\widetilde{\mathcal {S}})}\ge y\) if \(|x|\le M\) and, for \(|x|> M\),
and for any \(y>0\),
for a suitable constant k.
We deduce that, as soon as \(2s-10>0\), i.e. \(s>5\),
\(\square \)
8.3 Proof of the Convergences in Distribution of Theorem 1
Since we have proved in Lemma 7 that for all compactly supported \(\mathcal {C}^6\) function \(\phi \), the deterministic term \(\mu _n(\phi )\) could be replaced by \(\mathbb {E}[\mu _n^\varepsilon (\phi )]\), we only have to prove, that for all \(\phi \in \mathcal {C}^6\),
For the time being, we know this result to be valid for functions \(\phi \) belonging to the space \(\mathcal {L}_1\), defined as the linear span of the family of functions \(\varphi _z(x):=\frac{1}{z-x}\), \(z\in \mathbb {C}\backslash \mathbb {R}\).
By applying Lemma 10 to the centered field \(\mu _n^\varepsilon -\mathbb {E}[\mu _n^\varepsilon ]\), we are going to extend the result from the space \(\mathcal {L}_1\) to the Sobolev space \((\mathcal {H}_s, \Vert \cdot \Vert _{\mathcal {H}_s})\) with \(s\in (5,6)\). Note that, since \(s<6\), this latter space contains the space of \(\mathcal {C}^6\) compactly supported functions (see [16, Sec. 7.9]).
It remains to check the two hypotheses of Lemma 10. First, the subspace \(\mathcal {L}_1\) is dense in every space \((\mathcal {H}_s, \Vert \cdot \Vert _{\mathcal {H}_s})\). This is the content of Lemma 13 of Appendix. Second, by Lemma 8, since \(s>5\), \({\text {Var}}(n\varepsilon _n^{-1}\mu _n^\varepsilon ( \phi ))\le C\Vert \phi \Vert _{\mathcal {H}_s}^2\) for a certain constant C.
This concludes the proof.
Notes
If the perturbing matrix belongs to the GOE or GUE, then its law is invariant under this change in basis, hence our results in fact apply to any self-adjoint matrix \(D_n\).
References
Allez, R., Bouchaud, J.-P.: Eigenvector dynamics: general theory and some applications. Phys. Rev. E 86, 046202 (2012)
Allez, R., Bouchaud, J.-P.: Eigenvector dynamics under free addition random matrices. Rand. Matrices: Theory Appl. 3(3), 1450010 (2014)
Allez, R., Bun, J., Bouchaud, J.-P.: The eigenvectors of Gaussian matrices with an external source (2014). arXiv:1412.7108
Bun, J., Allez, R., Bouchaud, J.-P., Potters, M.: Rotational invariant estimator for general noisy matrices. IEEE Trans. Inf. Theory 62(12), 7475–7490 (2016)
Anderson, G., Zeitouni, O.: A CLT for a band matrix model. Probab. Theory Relat. Fields 134, 283–338 (2005)
Benaych-Georges, F., Guionnet, A., Male, C.: Central limit theorems for linear statistics of heavy tailed random matrices. Commun. Math. Phys. 329(2), 641–686 (2014)
Benaych-Georges, F., Knowles, A.: Lectures on the local semicircle law for Wigner matrices, arXiv. To appear in SMF series Panoramas et Synthèses
Biane, P.: On the free convolution by a semi-circular distribution. Indiana Univ. Math. J. 46, 705–718 (1997)
Biane, P.: Free Brownian motion, free stochastic calculus and random matrices. In: Free Probability Theory (Waterloo, ON, 1995), volume 12 of Fields Inst. Commun., pp. 1–19. American Mathematical Society, Providence (1997)
Biane, P., Speicher, R.: Stochastic calculus with respect to free Brownian motion and analysis on Wigner space. Probab. Theory Relat. Fields 112(3), 373–409 (1998)
Biane, P., Speicher, R.: Free diffusions, free entropy and free Fisher information. Ann. Inst. H. Poincaré Probab. Statist. 37(5), 581–606 (2001)
Casati, G., Girko, V.: Generalized Wigner law for band random matrices. Random Oper. Stoch. Equ. 1, 279–286 (1993)
Casati, G., Girko, V.: Wigner’s semicircle law for band random matrices. Random Oper. Stoch. Equ. 1, 15–21 (1993)
Dubédat, J.: SLE and the free field: partition functions and couplings. J. Am. Math. Soc. 22(4), 995–1054 (2009)
Fernandez, F.M.: Introduction to Perturbation Theory in Quantum Mechanics. CRC Press, Boca Raton (2000)
Hörmander, L.: The Analysis of Linear Partial Differential Operators. I. Distribution Theory and Fourier Analysis, 2nd edn. Springer, Berlin (1990). (Springer Study Edition)
Kato, T.: Perturbation Theory for Linear Operators, 2nd corr. print. of the 2nd ed. Grundlehren der Mathematischen Wissenschaften, 132. Springer, Berlin (1984)
King, F.W.: Hilbert Transforms. Encyclopedia of Mathematics and its Applications, 125, vol. 2. Cambridge University Press, Cambridge (2009)
Ledoit, O., Péché, S.: Eigenvectors of some large sample covariance matrix ensembles. Probab. Theory Relat. Fields 151(1–2), 233–264 (2011)
Shcherbina, M., Tirozzi, B.: Central limit theorem for fluctuations of linear eigenvalue statistics of large random graphs. J. Math. Phys. 51(2), 023523 (2010)
Shcherbina, M., Tirozzi, B.: Central limit theorem for fluctuations of linear eigenvalue statistics of large random graphs. Diluted regime. J. Math. Phys. 53(4), 043501 (2011)
Shlyakhtenko, D.: Random Gaussian band matrices and freeness with amalgamation. Int. Math. Res. Not. 1996(20), 1013–1025 (1996)
Wilkinson, M., Walker, P.: A Brownian motion model for the parameter dependence of matrix elements. J. Phys. A Math. Gen. 28, 6143 (1995)
Acknowledgements
We thank Jean-Philippe Bouchaud, Guy David and Vincent Vargas for some fruitful discussions. We are also glad to thank the GDR MEGA for partial support.
Author information
Authors and Affiliations
Corresponding author
Appendix
Appendix
The reader can find here the results we use along the paper, namely the Helffer–Sjöstrand formula, the CLT extension lemma of Shcherbina and Tirozzi and a functional density lemma with its proof.
1.1 Helffer–Sjöstrand Formula
The proof of the following formula can be found, e.g., in [7].
Proposition 9
(Helffer–Sjöstrand formula) Let \(n \in \mathbb {N}\) and \(\phi \in {\mathcal {C}}^{p+1}(\mathbb {R})\). We define the almost analytic extension of \(\phi \) of degree p through
Let \(\chi \in {\mathcal {C}}^\infty _c(\mathbb {C};[0,1])\) be a smooth cutoff function. Then for any \(\lambda \in \mathbb {R}\) satisfying \(\chi (\lambda ) = 1\) we have
where \(\mathrm {d}^2 z\) denotes the Lebesgue measure on \(\mathbb {C}\) and is the antiholomorphic derivative.
1.2 CLT Extension Lemma
The following CLT extension lemma is borrowed from the paper of Shcherbina and Tirozzi [20]. We state here the version that can be found in Appendix of [6].
Lemma 10
Let \((\mathcal {L}, \Vert \,\cdot \,\Vert )\) be a normed space with a dense subspace \(\mathcal {L}_1\) and, for each \(n\ge 1\), \((N_n(\phi ))_{\phi \in \mathcal {L}}\) a collection of real random variables such that:
-
for each n , \(\phi \longmapsto N_n(\phi )\) is linear,
-
for each n and each \(\phi \in \mathcal {L}\), \( \mathbb {E}[N_n(\phi )]=0 \),
-
there is a constant C such that for each n and each \(\phi \in \mathcal {L}\), \({\text {Var}}(N_n(\phi ))\le C\Vert \phi \Vert ^2\),
-
there is a quadratic form \(V:\mathcal {L}_1\rightarrow \mathbb {R}_+\) such that for any \(\phi \in \mathcal {L}_1\), we have the convergence in distribution \(N_n(\phi )\underset{n\rightarrow \infty }{\longrightarrow }\mathcal {N}(0, V(\phi ))\).
Then, V is continuous on \(\mathcal {L}_1\), can (uniquely) be continuously extended to \(\mathcal {L}\) and for any \(\phi \in \mathcal {L}\), we have the convergence in distribution \(N_n(\phi )\underset{n\rightarrow \infty }{\longrightarrow }\mathcal {N}(0, V(\phi ))\).
One of the assumptions of previous lemma concerns a variance domination. The next proposition provides a tool in order to check it. Let us first remind the definition of the Sobolev space \(\mathcal {H}_s\). For \(\phi \in L^1(\mathbb {R},\mathrm {d}x)\), we define
and, for \(s>0\),
We define the Sobolev space \(\mathcal {H}_s\) as the set of functions with finite \(\Vert \cdot \Vert _{\mathcal {H}_s}\) norm. Let us now state Proposition 2 of the paper [21] of Shcherbina and Tirozzi.
Proposition 11
For any \(s>0\), there is a constant \(C=C(s)\) such that for any n, any \(n\times n\) Hermitian random matrix M, and any \(\phi \in \mathcal {H}_s\), we have
1.3 A Density Lemma
We did not find Lemma 13 in the literature, so we provide its proof. Recall that for any \(z\in \mathbb {C}\backslash \mathbb {R}\),
Lemma 12
For any \(z\in \mathbb {C}\backslash \mathbb {R}\), we have, in the \(L^2\) sense,
and \(\varphi _z\) belongs to each \(\mathcal {H}_s\) for any \(s\in \mathbb {R}\).
Proof
It is well known that if \(\mathfrak {Re}z>0\), then \(\displaystyle \frac{1}{z}= \int _{t=0}^{+\infty } \mathrm {e}^{-tz}\mathrm {d}t.\)
Let \(z=E+\mathrm {i}\eta \), \(E\in \mathbb {R}, \eta >0\). For any \(\xi \in \mathbb {R}\), we have
We deduce (33) for \(\mathfrak {Im}z>0\). The general result can be deduced by complex conjugation. \(\square \)
Lemma 13
Let \(\mathcal {L}_1\) denote the linear span of the functions \(\varphi _z(x):=\frac{1}{z-x}\), for \(z\in \mathbb {C}\backslash \mathbb {R}\). Then the space \(\mathcal {L}_1\) is dense in \(\mathcal {H}_s\) for any \(s\in \mathbb {R}\).
Proof
We know, by Lemma 12, that \(\mathcal {L}_1\subset \mathcal {H}_s\). Recall first the definition of the Poisson kernel, for \(E\in \mathbb {R}\) and \(\eta >0\),
and that, by Lemma 12,
Hence for any \(f\in \mathcal {H}_s\), we have
so that, by dominated convergence, \(P_\eta *f\longrightarrow f\) in \(\mathcal {H}_s\) as \(\eta \rightarrow 0\).
To prove Lemma 13, it suffices to prove that any smooth compactly supported function can be approximated, in \(\mathcal {H}_s\), by functions of \(\mathcal {L}_1\). So let f be a smooth compactly supported function. By what precedes, it suffices to prove that for any fixed \(\eta >0\), \(P_\eta *f\) can be approximated, in \(\mathcal {H}_s\), by functions of \(\mathcal {L}_1\). For \(x\in \mathbb {R}\),
Without loss of generality, one can suppose that the support of f is contained in [0, 1]. Then, for any \(n\ge 1\),
where for \([t]_n:=\lceil nt\rceil /n\),
The error term \(R_n(x)\) rewrites
Now, note that for any \(t\in \mathbb {R}\) and \(\eta \in \mathbb {R}\backslash \{0\}\), we have by Lemma 12,
so that when, for example, \(\eta >0\), for any \(t\in \mathbb {R}\),
does not depend on t and for any \(t,t'\in \mathbb {R}\),
depends only on \(t'-t\) end tends to zero (by dominated convergence) when \(t'-t\rightarrow 0\).
We deduce that \(\Vert R_n\Vert _{\mathcal {H}_s}\longrightarrow 0\) as \(n\rightarrow \infty \), which closes the proof, by (34). \(\square \)
Rights and permissions
About this article
Cite this article
Benaych-Georges, F., Enriquez, N. & Michaïl, A. Empirical Spectral Distribution of a Matrix Under Perturbation. J Theor Probab 32, 1220–1251 (2019). https://doi.org/10.1007/s10959-017-0790-0
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10959-017-0790-0