Keywords

5.1 Introduction

We say that a random variable ξ has sub-Gaussian moment bounded above by K > 0 if

$$\displaystyle \begin{aligned} {\mathbb P} \{|\xi|\geq t\} \leq \exp\big(1-t^2/(2K^2)\big), \quad t\geq 0. \end{aligned}$$

Let A be an n × n random matrix with i.i.d. entries of zero mean, unit variance and sub-Gaussian moment bounded above by K, and denote by s i(A), 1 ≤ i ≤ n, its singular values arranged in non-increasing order. We will write \(s_{\max }(A)\) and \(s_{\min }(A)\) for s 1(A) and s n(A), respectively. Estimating the magnitude of the condition number,

$$\displaystyle \begin{aligned}\kappa(A)=s_{\max}(A)/s_{\min}(A),\end{aligned}$$

is a well studied problem, with connections to numerical analysis and computation of the limiting distribution of the matrix spectrum; we refer, in particular, to [20] for discussion. Since the largest singular value \(s_{\max }(A)\) is strongly concentrated (see the proof of Corollary 5.1.2 below), estimating κ(A) is essentially reduced to estimating \(s_{\min }(A)\) from above and below.

The main result of [12] provides small ball probability estimates for \(s_{\min }(A)\) of the form

$$\displaystyle \begin{aligned}{\mathbb P}\big\{s_{\min}(A)\leq t/\sqrt{n}\big\}\leq Ct+e^{-c n},\quad t\leq 1,\end{aligned}$$

for some C, c > 0 depending only on the sub-Gaussian moment. It seems natural to investigate the complementary regime—the large deviation estimates for \(s_{\min }(A)\). It was shown in [13] that

$$\displaystyle \begin{aligned} {\mathbb P}\big\{s_{\min}(A)\geq t/\sqrt{n}\big\}\leq \frac{C\ln t}{t}+e^{-c n},\quad t\geq 2 \end{aligned}$$

(see also [21] for an extension of this result to distributions with no assumptions on moments higher than 2). The probability estimate was improved in [10] to

$$\displaystyle \begin{aligned} {\mathbb P}\big\{s_{\min}(A)\geq t/\sqrt{n}\big\}\leq e^{-ct},\quad t\geq 2, \end{aligned}$$

for c > 0 depending only on the sub-Gaussian moment. The existing results on the distribution of the singular values of random Gaussian matrices [4, 18] suggest that the optimal dependence on t in the exponent on the right hand side is quadratic, i.e. the variable \(\sqrt {n}\,s_{\min }(A)\) is sub-Gaussian. Specifically, it is shown in [18] that \(s_{\min }(G)\) for the standard n × n Gaussian matrix G satisfies two-sided estimates

$$\displaystyle \begin{aligned} \exp(-C t^2)\leq {\mathbb P}\big\{s_{\min}(G)\geq t/\sqrt{n}\big\}\leq \exp(-c t^2),\quad t\geq C_1, \end{aligned}$$

where C, C 1, c > 0 are some universal constants. The main result of our note provides matching upper estimate for matrices with sub-Gaussian entries:

Theorem 5.1.1

Let A be an n × n random matrix with i.i.d. entries of zero mean, unit variance, and sub-Gaussian moment bounded above by K > 0. Then the smallest singular value \(s_{\min }(A)\) satisfies

$$\displaystyle \begin{aligned} {\mathbb P}\big\{s_{\min}(A)\geq t/\sqrt{n}\big\}\leq 2\exp(-c t^2),\quad t\geq 1, \end{aligned}$$

where c > 0 is a constant depending only on K.

As a simple corollary of the theorem, we obtain small ball probability estimates for the condition number:

Corollary 5.1.2

Let A be an n × n random matrix with i.i.d. entries of zero mean, unit variance, and sub-Gaussian moment bounded above by K > 0. Then the condition number κ(A) satisfies

$$\displaystyle \begin{aligned} {\mathbb P}\big\{\kappa(A)\leq n/t\big\}\leq 2\exp(-c t^2),\quad t\geq 1, \end{aligned}$$

where c > 0 is a constant depending only on K.

Theorem 5.1.1 is a consequence of the following theorem, which is of independent interest.

Theorem 5.1.3

Under conditions of Theorem 5.1.1 one has

$$\displaystyle \begin{aligned} {\mathbb P}\big\{\|A^{-1}\|{}_{HS} \leq \min(n/t, \, \sqrt{n/t})\big\}\leq 2\exp(-c t^2),\quad t\geq 0, \end{aligned}$$

where c > 0 is a constant depending only on K.

The proof of Theorem 5.1.3 uses, as a main step, the estimates

$$\displaystyle \begin{aligned}{\mathbb P}\big\{s_{n-k+1}(A)\leq ck/\sqrt{n}\big\}\leq 2 \exp(-c k^2),\quad \quad 1\leq k\leq n,\end{aligned}$$

for the singular values of the matrix A. These estimates, based on the restricted invertibility of matrices and certain averaging arguments, were recently obtained by Nguyen [9] under some additional assumptions (which will be discussed in the next section).

5.2 Preliminaries

Given a matrix A, it singular values s i = s i(A), i ≥ 1, are square roots of eigenvalues of AA . We always assume that s 1 ≥ s 2 ≥… By ∥A∥ and ∥AHS we denote the operator 2 →  2 norm of A (also called the spectral norm) and the Hilbert–Schmidt norm respectively. Note that

$$\displaystyle \begin{aligned} \|A\| = s_1 \quad \quad \mbox{ and } \quad \quad \|A\|{}_{HS}^2=\sum_{i\geq 1} s_i^2. \end{aligned}$$

The columns and rows of A are denoted by C i(A) and R i(A), i ≥ 1, respectively. Given J ⊂ [m], the coordinate projection in \({\mathbb R}^m\) onto \({\mathbb R}^J\) is denoted by P J. For convenience, we often write A J instead of AP J. Given m ≥ 1, the identity operator \({\mathbb R}^\ell \to {\mathbb R}^\ell \) we denote by I m. Given \(x, y\in {\mathbb {R}^n}\) by \(\left \langle x, \cdot \right \rangle y\) we denote the operator \(z\mapsto \left \langle x, z\right \rangle y\) (in the literature it is often denoted by x ⊗ y or yx ). The canonical Euclidean norm in \({\mathbb R}^m\) is denoted by ∥⋅∥2 and the unit Euclidean sphere by S m−1.

As the most important part of our argument, we will use the following result.

Theorem 5.2.1

Let A be an n × n random matrix with i.i.d. entries of zero mean, unit variance, and sub-Gaussian moment bounded above by K > 0. Then for any 1 ≤ k  n one has

$$\displaystyle \begin{aligned}{\mathbb P}\big\{s_{n-k+1}(A)\leq ck/\sqrt{n}\big\}\leq 2 \exp(-c k^2),\end{aligned}$$

where c > 0 is a constant depending only on K.

The above theorem, up to some minor modifications, was proved by Nguyen in [9]. Specifically, in the case \(k\geq C\log n\), the theorem follows from [9, Theorem 1.7] (or [9, Corollary 1.8]) if one additionally assumes either that the entries of A are uniformly bounded by a constant, or that the distribution density of the entries is bounded. Removing these conditions requires a minor change of the proof in [9]. Further, in the case \(k\leq C\log n\), the above result (in fact, in a stronger form) is stated as formula (4) in [9, Theorem 1.4]. However, [9, Theorem 3.6], which is used to derive [9, formula (4)], provides a non-trivial probability estimate only for the event \(\{s_{n-k+1}(A)\leq c_\gamma k^{1-\gamma }/\sqrt {n}\}\) (for any given γ ∈ (0, 1) and c γ depending on γ), see [9, formula (31)]. Again, a minor update of the argument of [9] provides the result needed for our purposes. In view of the above and for the reader’s convenience, we provide a proof of Theorem 5.2.1 in the last section.

The following result was proved in [17] as an extension of the classical Bourgain–Tzafriri restricted invertibility theorem [2]. With worse dependence on ε, the theorem was earlier proved in [22]. See also recent papers [1, 8] for further improvements and discussions.

Theorem 5.2.2 ([17])

Let T be n × n matrix. Then for any ε ∈ (0, 1) there is a set J ⊂ [n] such that

$$\displaystyle \begin{aligned} \ell:=|J|\geq \bigg\lfloor\frac{\varepsilon^2 \|T\|{}_{HS}^2}{\|T\|{}^2}\bigg\rfloor \quad \quad \mathit{\mbox{ and }} \quad \quad s_\ell (T_J)\geq \frac{(1-\varepsilon)\|T\|{}_{HS}}{\sqrt{n}}. \end{aligned}$$

We will use two following results by Rudelson–Verhsynin. The first one was one of the key ingredients in estimating the smallest singular value of rectangular matrices. The second one is an immediate consequence of the Hanson–Wright inequality [5, 23] generalized in [15].

Theorem 5.2.3 ([14, Theorem 4.1])

Let X be a vector in \({\mathbb R}^n\), whose coordinates are i.i.d. mean-zero, sub-Gaussian random variables with unit variance. Let F be a random subspace in \({\mathbb R}^n\) spanned by n  ℓ vectors, 1 ≤ ℓ  c n, whose coordinates are i.i.d. mean-zero, sub-Gaussian random variables with unit variance, jointly independent with X. Then, for every ε > 0, one has

$$\displaystyle \begin{aligned} {\mathbb P}\big\{{\mathrm{dist}} (X, F)\leq \varepsilon \sqrt{\ell} \big\}\leq (C\varepsilon)^\ell + \exp(-cn). \end{aligned}$$

where C > 0, c, c ∈ (0, 1) are constants depending only on the sub-Gaussian moments.

Theorem 5.2.4 ([15, Corollary 3.1])

Let X be a vector in \({\mathbb R}^n\), whose coordinates are i.i.d. mean-zero random variables with unit variance and with sub-Gaussian moment bounded by K. Let F be a fixed subspace in \({\mathbb R}^n\) of dimension n  ℓ. Then, for every t > 0, one has

$$\displaystyle \begin{aligned} {\mathbb P}\big\{| {\mathrm{dist}} (X, F)-\sqrt{\ell}|\geq t \big\}\leq 2 \exp(-c t^2/K^4). \end{aligned}$$

where c > 0 is an absolute constant.

We will also need the following standard claim, which can be proved by integrating the indicator functions (see e.g., [9, Claim 3.4], cf. [7, Claim 4.9]).

Claim 5.2.5

Let α, p ∈ (0, 1). Let \({\mathcal E}\) be an event. Let Z be a finite index set, and \( \{{\mathcal E}_z \} _{z\in Z} \) be a collection of |Z| events satisfying \({\mathbb P}({\mathcal E}_z)\leq p\) for every z ∈ Z. Assume that at least α|Z| of events \({\mathcal E}_z\) hold whenever the event \({\mathcal E}\) occurs. Then \({\mathbb P}({\mathcal E})\leq p/\alpha \).

5.3 Proofs of Main Results

Proof of Theorem 5.1.1

In the case t > n we have

$$\displaystyle \begin{aligned} {\mathbb P}\big\{ s_{\min}(A)\geq t/\sqrt{n} \big\} &= {\mathbb P}\Big\{s_1(A^{-1})\leq \sqrt{n}/t \Big\}\leq {\mathbb P}\Big\{\sum_{i=1}^n s_i(A^{-1})^2\leq n^2/t^2\Big\} \end{aligned} $$

and the result follows from Theorem 5.1.3.

Now we consider the case 1 ≤ t ≤ n. Let L ≥ 1 be a parameter which we will choose later. Then

$$\displaystyle \begin{aligned} {\mathbb P}\big\{s_{\min}(A)\geq t/\sqrt{n}\big\} &= {\mathbb P}\big\{s_{1}(A^{-1})\leq \sqrt{n}/t\big\} \\&\leq {\mathbb P}\Big\{s_1(A^{-1})^2\leq n/t^2\;\;\mbox{ and }\;\;\sum_{i\geq \lceil t\rceil} s_i(A^{-1})^2\geq L n/t\Big\}\\ & \quad +{\mathbb P}\Big\{s_1(A^{-1})^2\leq n/t^2\;\;\mbox{ and }\;\;\sum_{i\geq \lceil t\rceil} s_i(A^{-1})^2< L n/t\Big\}\\ &\leq {\mathbb P}\Big\{\sum_{i\geq \lceil t\rceil} s_i(A^{-1})^2\geq L n/t\Big\} \\ & \quad +{\mathbb P}\Big\{\sum_{i=1}^n s_i(A^{-1})^2\leq n/t+ L n/t\Big\}. \end{aligned} $$

For the first summand in the last expression, we apply Theorem 5.2.1. Since \(\sum _{i=\lfloor t\rfloor }^\infty \frac {1}{i^2}\leq \frac {2}{t}\), we obtain

$$\displaystyle \begin{aligned} {\mathbb P}\Big\{\sum_{i=\lceil t\rceil}^n s_i(A^{-1})^2\geq L n/t\Big\} &\leq \sum_{i=\lceil t\rceil}^n {\mathbb P}\big\{s_i(A^{-1})^2\geq L n/(2 i^2)\big\}\\ &= \sum_{i=\lceil t\rceil}^n {\mathbb P}\big\{s_{n-i+1}(A)\leq \sqrt{2} i/\sqrt{L n}\big\}. \end{aligned} $$

Choosing L so that \(\sqrt {2/L}\) is equal to the constant from Theorem 5.2.1, we get

$$\displaystyle \begin{aligned} \sum\limits_{i=\lfloor t\rfloor}^n {\mathbb P}\big\{s_{n-i+1}(A)\leq \sqrt{2} i/\sqrt{L n}\big\} \leq 2 \sum\limits_{i=\lfloor t\rfloor}^n \exp(-c i^2)\leq 3 \exp(-c^{\prime} t^2) \end{aligned}$$

for some c  > 0 depending only on K. The bound on the second summand follows from Theorem 5.1.3 applied with t∕(L + 1) instead of t. This completes the proof. □

Proof of Corollary 5.1.2

Theorem 5.2.4 implies that there exists an absolute constant c 1 > 0 depending only on K such that for every i ≤ n

$$\displaystyle \begin{aligned} {\mathbb P}(\|{\mathbf{C}}_i(A) \|{}_2\leq \sqrt{n}/2 )\leq \exp(- c_1 n) \end{aligned}$$

(this can be shown by direct calculations as well, see e.g. Fact 2.5 in [6]). Since the entries of A are independent, we obtain

$$\displaystyle \begin{aligned} {\mathbb P}(\|A \|\leq \sqrt{n}/2 )\leq \prod_{i=1}^n {\mathbb P}(\|{\mathbf{C}}_i(A) \|{}_2\leq \sqrt{n}/2 )\leq \exp(- c_1 n^2). \end{aligned}$$

Note that if \(\|A \|\geq \sqrt {n}/2\) and κ(A) ≤ n∕2t then \(s_n(A)= \|A\|/\kappa (A)\geq t/\sqrt {n}\). Therefore, by Theorem 5.1.1,

$$\displaystyle \begin{aligned} {\mathbb P}\{\kappa(A)\leq n/2t\} \leq 2\exp(-c t^2) + \exp(- c_1 n^2). \end{aligned}$$

By adjusting constants, this implies the conclusion for t ≤ n. Since κ(A) ≥ 1, the case t > n is trivial. □

Proof of Theorem 5.1.3

Adjusting the constant in the exponent if needed, without loss of generality, we assume that t ≥ C 0, where C 0 > 0 is a large enough constant depending only on K. Denote

$$\displaystyle \begin{aligned} {\mathcal E}_0:=\bigg\{\sum_{i=1}^n s_i(A^{-1})^2\leq n/t\bigg\}. \end{aligned}$$

We first consider the case t ≤ n. Applying the negative second moment identity (see e.g. Exercise 2.7.3 in [19]),

$$\displaystyle \begin{aligned} \sum_{i=1}^n s_i(A^{-1})^2=\sum_{i=1}^n {\mathrm{dist}}\big({\mathbf{C}}_i(A),{\mathrm{span}}\{{\mathbf{C}}_j(A),\; j\neq i\}\big)^{-2}, \end{aligned}$$

we observe that on the event \({\mathcal E}_0\),

$$\displaystyle \begin{aligned}\big|\big\{i\leq n:\;{\mathrm{dist}}\big({\mathbf{C}}_i(A),{\mathrm{span}}\{{\mathbf{C}}_j(A),\; j\neq i\}\big)\geq \sqrt{t/2}\big\}\big|\geq n/2.\end{aligned}$$

For each subset I ⊂ [n] of cardinality k ≤ n∕2 (the actual value of k will be defined later), let 1 I be the indicator of the event

$$\displaystyle \begin{aligned} \big\{{\mathrm{dist}}\big({\mathbf{C}}_i(A),{\mathrm{span}}\{{\mathbf{C}}_j(A),\; j\in [n]\setminus I\}\big)\geq \sqrt{t/2} \mbox{ for all }i\in I\big\}. \end{aligned}$$

Then, in view of the above, everywhere on the event \({\mathcal E}_0\) we have

$$\displaystyle \begin{aligned} \sum_{I\subset[n],\;|I|=k} {\mathbf{1}}_{I}\geq {\lceil n/2\rceil \choose k}\geq \bigg(\frac{n}{2k}\bigg)^k \geq (2e)^{-k} {n\choose k}. \end{aligned}$$

Hence, by Markov’s inequality and permutation invariance of the matrix distribution,

$$\displaystyle \begin{aligned} {\mathbb P}({\mathcal E}_0) \leq (2e)^{k} \, {\mathbb E}\, {\mathbf{1}}_{[k]}. \end{aligned}$$

As the last step of the proof, we estimate the expectation of 1 [k] (with a suitable choice of k). In view of independence and equidistribution of the matrix columns, we have

$$\displaystyle \begin{aligned} {\mathbb E}\, {\mathbf{1}}_{[k]}=\left( {\mathbb P}\big\{ {\mathrm{dist}}\big({\mathbf{C}}_1(A),{\mathrm{span}}\{{\mathbf{C}}_j(A),\; j\in [n]\setminus [k]\}\big)\geq \sqrt{t/2} \big\}\right)^k. \end{aligned}$$

Choose k := ⌊t∕4⌋≤ n∕2 and denote

$$\displaystyle \begin{aligned}D:={\mathrm{dist}}\big({\mathbf{C}}_1(A),{\mathrm{span}}\{{\mathbf{C}}_j(A),\; j\in [n]\setminus [k]\}\big).\end{aligned}$$

Using independence of columns of the matrix A and applying Theorem 5.2.4 with  = k and F = span{C j(A), j ∈ [n] ∖ [k]}, we obtain

$$\displaystyle \begin{aligned} {\mathbb P} \Big\{ D \geq \sqrt{t/2} \Big\} \leq {\mathbb P}\Big\{ D - \sqrt{k} \geq (\sqrt{2}-1)\, \sqrt{t/4} \Big\} \leq 2\exp(-\bar c\, t) \end{aligned}$$

for some \(\bar c>0\) depending only on K. Hence,

$$\displaystyle \begin{aligned} {\mathbb P}({\mathcal E}_0) \leq (2e)^{k} \,2^k\,\exp(-\bar c\, t\,k)\leq \exp(-\bar c t^2/16), \end{aligned}$$

provided that t is larger than a certain constant depending only on K. This implies the desired result for t ≤ n.

In the case t > n we essentially repeat the argument along the same lines. Define

$$\displaystyle \begin{aligned} {\mathcal E}_0^{\prime}:=\bigg\{\sum_{i=1}^n s_i(A^{-1})^2\leq n^2/t^2\bigg\}. \end{aligned}$$

Observe that on the event \({\mathcal E}_0^{\prime }\),

$$\displaystyle \begin{aligned} \big|\big\{i\leq n:\;{\mathrm{dist}}\big({\mathbf{C}}_i(A),{\mathrm{span}}\{{\mathbf{C}}_j(A),\; j\neq i\} \big)\geq t/\sqrt{2n}\big\}\big|\geq n/2. \end{aligned}$$

Repeating the above computations with the same notation and with k = ⌊n∕4⌋ we obtain

$$\displaystyle \begin{aligned} {\mathbb P} \Big\{ D \geq t/\sqrt{2n} \Big\} \leq {\mathbb P}\Big\{ D - \sqrt{k} \geq t/(5 \sqrt{n}) \Big\} \leq 2\exp(-\bar c\, t^2/n), \end{aligned}$$

which leads to

$$\displaystyle \begin{aligned} {\mathbb P}({\mathcal E}_0^{\prime}) \leq (2e)^{k} \,2^k\,\exp(-\bar c\, k t^2/n )\leq \exp(-\bar c t^2/16), \end{aligned}$$

provided that t > Cn for large enough C depending only on K. For n < t ≤ Cn the result follows by adjusting the absolute constants. □

5.4 Small Ball Estimates for Singular Values

The goal of this section is to prove Theorem 5.2.1. As we have noted, the argument essentially reproduces that of [9]. An important part of the proof is the use of restricted invertibility (see also [3] and [11] for some recent applications of restricted invertibility in the context of random matrices).

We will use a construction from [9]. Given an integer k and an n × n matrix A define a k × n matrix Z = Z(A, k) in the following way. Consider singular value decomposition \(A=\sum _{i=1}^n s_i \left \langle v_i, \cdot \right \rangle w_i\), where s i = s i(A) are singular values of A (arranged in non-increasing order) and {v i}i, {w i}i are two orthonormal systems in \({\mathbb {R}^n}\). For i ≤ k denote z i = v ni+1. Let Z be the matrix whose rows are R i(Z) = z i. Clearly, the rows of Z are orthonormal and for every i ≤ k,

$$\displaystyle \begin{aligned} \|Az_i\|{}_2 = s_{n-i+1}\leq s_{n-k+1}. \end{aligned} $$
(5.1)

Moreover,

$$\displaystyle \begin{aligned} \|Z\|=1\quad \quad \mbox{ and } \quad \quad \|Z\|{}_{HS}=\sqrt{k}. \end{aligned}$$

The matrix Z is not uniquely defined when some of the k smallest singular values of A have non-trivial multiplicity; we will however assume that for each realization of A, a single admissible Z is chosen in such a way that Z is a (measurable) random matrix.

5.4.1 Proof of Theorem 5.2.1, the Case \(k\geq \ln n\)

Let C, c, c be constants from Theorem 5.2.3. Let \(\gamma =\sqrt {c^{\prime }}\). Note that C, c, c , γ depend only on K. Let Z = Z(A, k) be the k × n matrix constructed above. Applying Theorem 5.2.2 to Z (one can add zero rows to make it an n × n matrix), there exists J ⊂ [n] such that

$$\displaystyle \begin{aligned} |J| = \ell:=\lfloor \gamma^2 k \big\rfloor\leq c^{\prime} k \quad \quad \mbox{ and } \quad \quad s_\ell (Z_J)\geq (1-\gamma ) \sqrt{k/n}. \end{aligned} $$

Fix a (small enough, depending on K) constant c 0 > 0. Define the event

$$\displaystyle \begin{aligned} {\mathcal E}_k := \big\{s_{n-k+1}(A)\leq c_0 k/\sqrt{n}\big\}. \end{aligned} $$

Consider the n × k matrix B = AZ . Using property (5.1), on the event \({\mathcal E}_k\), we have for every i ≤ k,

$$\displaystyle \begin{aligned} \|{\mathbf{C}} _i (B) \|{}_2= \|A z_i\|{}_2\leq c_0 k/\sqrt{n}, \end{aligned} $$

hence \(\|B\|{ }_{HS}\leq c_0 k^{3/2}/\sqrt {n}\). Now, since s (Z J) > 0, there exists a k ×  matrix M such that \(Z^\top _J M=I_\ell \). Then

$$\displaystyle \begin{aligned} \|M\| =1/s_\ell (Z)\leq (1-\gamma )^{-1} \sqrt{n/k}. \end{aligned} $$

Therefore,

$$\displaystyle \begin{aligned} \|B M\|{}_{HS}\leq \|B\|{}_{HS}\, \|M\|\leq c_0 (1-\gamma )^{-1} k. \end{aligned}$$

Writing \(B= A_J (Z_J)^\top + A_{J^c} (Z_{J^c})^\top \), we also have \(BM= A_J + A_{J^c} (Z_{J^c})^\top M\). Next denote

$$\displaystyle \begin{aligned} F=F(A, J):={\mathrm{span}}\{ {\mathbf{C}}_i(A_{J^c})\}_{i\in J^c}, \end{aligned}$$

and let P be the orthogonal projection on F . Then, on the event \({\mathcal E} _k\),

$$\displaystyle \begin{aligned} c_0^2 (1-\gamma )^{-2} k^2 &\geq \|P B M\|{}_{HS}^2 \geq \|P A_J\|{}_{HS}^2 = \sum _{i\in J} \|P \, {\mathbf{C}}_i(A_J)\|{}^2_2 \\ & = \sum _{i\in J} {\mathrm{dist}} ^2 ({\mathbf{C}}_i(A), F) . \end{aligned} $$

Therefore, for at least ∕2 indices i ∈ J, one has

$$\displaystyle \begin{aligned} {\mathrm{dist}} ({\mathbf{C}}_i(A), F) \leq \sqrt{2} c_0 (1-\gamma )^{-1} k/\sqrt{\ell} \leq 2 c_0 \sqrt{\ell}/((1-\gamma )\gamma^2) . \end{aligned}$$

Note that the subspace F is spanned by n −  random vectors, it is independent of columns C i(A), i ∈ J, and that columns of A are independent. Therefore, by Theorem 5.2.3 and the union bound we obtain

Choosing small enough c 0 and using \(k\geq \ln n\), we obtain \( {\mathbb P}({\mathcal E} _k) \leq \exp (-c_3 \ell ^2) , \) where c 3 > 0 depends only on K. By adjusting constants this proves the desired result for \(k\geq \ln n\). □

5.4.2 Proof of Theorem 5.2.1, the Case \(k\leq \ln n\)

Let A be as in Theorem 5.2.1. It is well known (see e.g. Fact 2.4 in [6]) that there is an absolute constant C 1 > 0 such that

$$\displaystyle \begin{aligned} {\mathbb P}\big\{\|A\|\leq C_1 K\sqrt{n} \big\}\geq 1- e^{-n}. \end{aligned} $$
(5.2)

Let \({\mathcal E}_{bd}\) denote the event from this equation. Further, from [16, Theorem 1.5] one infers that for any γ > 0 there are γ 1, γ 2, γ 3 > 0 depending only on γ and K such that, denoting

$$\displaystyle \begin{aligned} {\mathcal E}_{inc}(\gamma) := \big\{&\forall x \in S^{n-1} \, \mbox{ with } \, \|Ax \|{}_2\leq \gamma_1 \sqrt{n}, \, \, \forall I\subset [n]\\ &\mbox{ with } \, |I|\geq \gamma n \, \mbox{ one has } \, \|P_I x\|{}_2\geq \gamma_2 \big\}, \end{aligned} $$

the event satisfies

$$\displaystyle \begin{aligned} {\mathbb P}({\mathcal E}_{inc}(\gamma))\geq 1-2 e^{-\gamma_3 n}. \end{aligned} $$
(5.3)

The following statement was proved by Nguyen ([9, Corollary 3.8]).

Proposition 5.4.1

For any K > 0 there are C, c 1, c 2, γ > 0 depending only on K with the following property. Let A be an n × n random matrix with i.i.d. entries of zero mean, unit variance, and sub-Gaussian moment bounded above by K. Let \(2\leq k\leq n/(C\ln n)\), and let the random k × n matrix Z = Z(A, k) be defined as above. Then everywhere on the event \(\big \{s_{n-k+1}(A)\leq c_1 k/\sqrt {n}\big \}\cap {\mathcal E}_{inc}(\gamma )\cap {\mathcal E}_{bd}\) one has

$$\displaystyle \begin{aligned} \big|\big\{J\subset[n]:\;|J|=\lfloor k/2\rfloor,\;s_{\lfloor k/2\rfloor}(Z_J)\geq c_1\sqrt{k/n}\big\}\big| \geq c_2^{k\ln k}\, n^{\lfloor k/2\rfloor}. \end{aligned}$$

Now assume that \(k\leq \ln n\). Without loss of generality we may also assume that k is bounded below by a large constant. Let C, c, c be constants from Theorem 5.2.3 and c 1, c 2, γ from Proposition 5.4.1. Fix for a moment any realization of A from the event \(\big \{s_{n-k+1}(A)\leq c_0 k/\sqrt {n}\big \}\cap {\mathcal E}_{inc}(\gamma )\cap {\mathcal E}_{bd}\), where c 0 ∈ (0, c 1] will be chosen later. Let  := ⌊k∕2⌋ and

$$\displaystyle \begin{aligned} \mathcal{J} :=\big\{J\subset[n]:\;|J|=\lfloor k/2\rfloor,\;s_{\lfloor k/2\rfloor}(Z_J)\geq c_1\sqrt{k/n}\big\}. \end{aligned}$$

Fix \(J\in \mathcal {J}\) and repeat the procedure used in Sect. 5.4.1 with J and . We obtain that for at least ∕2 indices i ∈ J, one has

$$\displaystyle \begin{aligned} {\mathrm{dist}} ({\mathbf{C}}_i(A), F) \leq \sqrt{2} c_0 k/(c_1 \sqrt{\ell}) \leq 4 c_0\sqrt{\ell}/c_1, \end{aligned} $$
(5.4)

where \(F={\mathrm {span}}\{ {\mathbf {C}}_i(A_{J^c})\}_{i\in J^c}\). For any fixed subset J ⊂ [n] of cardinality consider the event

$$\displaystyle \begin{aligned} {\mathcal E}_J:=\big\{\mbox{for at least }\ell/2\mbox{ indices }i\in J\mbox{ inequality (5.4) holds}\big\}. \end{aligned}$$

Applying Theorem 5.2.3 and the union bound we observe

$$\displaystyle \begin{aligned}{\mathbb P}({\mathcal E} _J)&\leq 2^\ell \, \left((4 c_0 C /c_1)^{\ell} + \exp(-cn) \right)^{\ell/2}\\ &\leq \left(4\, \max\left\{\left(4 c_0 C/c_1 \right)^\ell,\, \exp(-cn)\right\}\right)^{\ell/2}. \end{aligned} $$

Choosing c 0 to be small enough we obtain that \({\mathbb P}({\mathcal E} _J)\leq \exp (-c_4 k^2)\), where c 4 > 0 depends only on K. Combining this with Claim 5.2.5 and Proposition 5.4.1 we obtain

$$\displaystyle \begin{aligned} {\mathbb P}\big(\big\{s_{n-k+1}(A)\leq c_0 k/\sqrt{n}\big\}\cap {\mathcal E}_{inc}(\gamma)\cap {\mathcal E}_{bd}\big)\leq c_2^{- k\ln k } \exp(-c_4 k^2) \leq \exp(-c_5 k^2) \end{aligned}$$

provided that k ≥ C 2, where C 2 ≥ 1 ≥ c 5 > 0 are constants depending on on K only. By Eqs. (5.2) and (5.3) this completes the proof in the case \(k\leq \ln n\). □