Abstract
New results on the asymptotic behavior of the probabilities of large deviations of combinatorial sums of independent random variables satisfying the Linnik condition are obtained. A zone, where these probabilities are equivalent to the tail of the standard normal law, is found. Such results were previously obtained by the author under Bernstein’s condition. The new results are proved by the truncation method.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 INTRODUCTION
We suppose {(Xnij), 1 \(\leqslant \) i, j \(\leqslant \) n, n = 2, 3, …} is a sequence of matrices of independent random variables and {\({{\vec {\pi }}_{n}}\) = (\({{\pi }_{n}}(1)\), \({{\pi }_{n}}(2)\), …, \({{\pi }_{n}}(n)\)), n = 2, 3, …} is a sequence of random permutations of the numbers 1, 2, …, n. Let \({{\vec {\pi }}_{n}}\) have a uniform distribution on the set of all permutations of 1, 2, …, n and be independent of (Xnij) for any n. We define the combinatorial sum Sn by the relation
We note that if the distributions of Xnij coincide for all \(1\;\leqslant \;j\;\leqslant \;n\) at all n, then Sn has the same distribution as the sum of independent random variables. Although this case is well studied, we should take it into account when estimating the optimality of the obtained results.
Under certain conditions, the sequence of distributions of normalized combinatorial sums weakly converges to the normal law. Any similar result is called a combinatorial central limit theorem (CLT). Research in this direction started long ago. The combinatorial CLT has been studied by Wald and Wolfowitz [1], Noether [2], Hoeffding [3], Motoo [4], Kolchin and Chistyakov [5]. Later on, nonasymptotic bounds of the type of Berry–Esseen and Esseen inequalities were found. Bolthausen [6], von Bahr [7], Ho and Chen [8], Goldstein [9], Neammanee and Suntornchost [10], Neammanee and Rattanawong [11], Chen, Goldstein and Shao [12], Chen and Fang [13], and Frolov [14, 15] obtained similar results. A.N. Frolov obtained the results for a random number of summands in [16].
Bounds in CLT allow the asymptotics of probabilities of large deviations in logarithmic zones to be found. Usually, in this case, they speak of moderate deviations. We obtained such results for combinatorial sums in [17].
In our work [18], we first obtained results on the asymptotic behavior of probabilities of large deviations of combinatorial sums in power zones. In that work, we assumed that random variables satisfy some analogue of Bernstein’s condition. But for separate particular cases, combinatorial sums do not have independent increments. This makes it difficult to use classical methods of summation theory for independent random variables. In [18], we obtained the bounds of the moment-generating function and its logarithmic derivatives of the normalized combinatorial sum rather than of its particular summands. This is what led to those results.
Bernstein’s condition is one of the forms of the existence condition for the exponential moment. A natural problem is to obtain new results on the asymptotics of probabilities of large deviations if it is violated. This is what we discuss in this work. We replace Bernstein’s condition with the weaker Linnik condition. To prove the results, we use the truncation method.
2 RESULTS
We suppose {(Xnij), 1 \(\leqslant \) i, j \(\leqslant \) n, n = 2, 3, …} is a sequence of matrices of independent random variables such that
for any n. We suppose {\({{\vec {\pi }}_{n}}\) = (\({{\pi }_{n}}(1)\), \({{\pi }_{n}}(2)\), …, \({{\pi }_{n}}(n)\)), n = 2, 3, …} is a sequence of random permutations of the numbers 1, 2, …, n. We assume that \({{\vec {\pi }}_{n}}\) has a uniform distribution on the set of permutations Pn and is independent of (Xnij) for any n.
We put
It is not difficult to check that
Thus, condition (1) ensures centrality of the combinatorial sums. Replacing \({\mathbf{D}}{{X}_{{nij}}}\) = \({\mathbf{E}}X_{{nij}}^{2}\) – (EXnij)2 in the latter formula, we have
If \({\mathbf{D}}{{S}_{n}} \to \infty \), the principal part of the variance is the normalized sum of the second moments
Therefore, in what follows, we use {Bn} as the norming sequence for Sn.
Further, we assume that the summands have all moments. We put
We note that γn \( \geqslant \) 1. This follows from the fact that \(n{{B}_{n}}\) = \(\sum\nolimits_{i,j = 1}^n {{\mathbf{E}}X_{{nij}}^{2}} \) \(\leqslant \) \(n\mathop {\max }\limits_i \sum\nolimits_{j = 1}^n {{\mathbf{E}}X_{{nij}}^{2}} \).
The result below was obtained in our work [18].
Theorem 1. We suppose {Mn} is a nondecreasing sequence of positive constants such that for s = 1, 2, 3 the inequalities
hold for all k \( \geqslant \) s, all \(1\;\leqslant \;i,j\;\leqslant \;n\), and all \(n\; \geqslant \;2\), where D is the absolute positive constant.
Then, for any sequence of real numbers {un} such that un → ∞, \(u_{n}^{3}\) = \(o(\sqrt n {\text{/}}{{\gamma }_{n}})\), and un = \(o(\sqrt {{{B}_{n}}} {\text{/}}{{M}_{n}})\) as n → ∞, the relation
holds, where Φ(x) is the standard normal distribution function.
The condition \(u_{n}^{3} = o(\sqrt n {\text{/}}{{\gamma }_{n}})\) is natural for relation (4), which is the exact (not logarithmic) asymptotic of large deviations. If we assume that all Xnij are identically distributed, this condition turns into the optimal condition un = o(n1/6).
Condition (3) is similar to Bernstein’s condition, which is a form of the exponential-moment condition. In [18], we give some variants of this conditions and the examples of random variables that satisfy the hypothesis of Theorem 1. In particular, these include bounded random variables. In the latter case, Spearman’s rank correlation coefficient is an important example of a combinatorial sum.
The theorem below is our principal result. In it, we replace Bernstein’s condition by the weaker Linnik condition, thus expanding the theorem’s application domain. For instance, the Weibull distribution with the parameter α that arises, in particular, as a limit distribution in the theory of extreme order statistics satisfies the Linnik condition and does not satisfy Bernstein’s condition when α < 1.
Theorem 2. We suppose for some β ∈ (0, 1) the inequalities
hold for s = 1, 2, 3, all \(1\;\leqslant \;i,j\;\leqslant \;n\) and all \(n\; \geqslant \;2\), where C is an absolute positive constant.
We suppose {un} is a sequence of real numbers such that \({{u}_{n}} \to \infty \), \(u_{n}^{3}\) = \(o(\sqrt n {\text{/}}{{\gamma }_{n}})\), un = \(o(B_{n}^{{\beta /(2(2 - \beta ))}})\), and \(\ln (n{{c}_{n}} + 1)\) = \(o(u_{n}^{\beta }B_{n}^{{\beta /2}})\) as \(n \to \infty \), where
Then, relation (4) holds.
Condition (5) holds if for any i, j, and n the random variable Xnij can have one of k given distributions. Moreover, if there exist positive constants A and B such that \({\mathbf{E}}{{e}^{{{{{\left| {{{X}_{{nij}}}} \right|}}^{\beta }}}}}\;\leqslant \;A\) and \({\mathbf{E}}{{\left| {{{X}_{{nij}}}} \right|}^{s}}\; \geqslant \;B\) for all i, j, n, and s, then condition (5) holds.
If Xnij have the same distributions for any i, j, and n, then by Theorem 2 relation (4) holds for un = \(o({{n}^{\alpha }})\), where α = min{1/6, β/(2(2 – β))}. From Linnik’s work [19], we know that in this case, for β ∈ (0, 1/2], the Linnik condition is optimal to fulfill for relation (4) in the zone un = \(o({{n}^{{\beta /(2(2 - \beta ))}}})\), while for β > 1/2 it requires additional conditions. Thus, the hypotheses of Theorem 2 cannot be improved in this case.
Remark 1. Theorem 2 still holds for β = 1. One should replace cn by zero.
3 PROOFS
We now prove our results.
Proof of Theorem 2. We put \({{y}_{n}} = {{u}_{n}}\sqrt {{{B}_{n}}} \), \({{\bar {X}}_{{nij}}}\) = \({{X}_{{nij}}}I\left\{ {{\kern 1pt} \left| {{{X}_{{nij}}}} \right| < {{y}_{n}}} \right\}\), \({{\tilde {X}}_{{nij}}}\) = \({{X}_{{nij}}}I\left\{ {{\kern 1pt} \left| {{{X}_{{nij}}}} \right|\; \geqslant \;{{y}_{n}}} \right\}\), and Tn = \(\sum\nolimits_{i = 1}^n {{{{\bar {X}}}_{{ni{{\pi }_{n}}(i)}}}} \). We suppose B0 = \(\bigcap\nolimits_{i = 1}^n {\left\{ {{{X}_{{ni{{\pi }_{n}}(i)}}} = {{{\bar {X}}}_{{ni{{\pi }_{n}}(i)}}}} \right\}} \) and B1 = \(\bigcup\nolimits_{i = 1}^n {\left\{ {{{X}_{{ni{{\pi }_{n}}(i)}}} \ne {{{\bar {X}}}_{{ni{{\pi }_{n}}(i)}}}} \right\}} \). We have
Similarly, we have
Hence, the inequality
holds.
For all \(1\;\leqslant \;i,j\;\leqslant \;n\), we put \({{\bar {a}}_{{nij}}} = {\mathbf{E}}{{\bar {X}}_{{nij}}}\), \(\bar {\sigma }_{{nij}}^{2} = {\mathbf{D}}{{\bar {X}}_{{nij}}}\), \({{\bar {a}}_{{ni.}}}\) = \(\frac{1}{n}\sum\nolimits_{j = 1}^n {{{{\bar {a}}}_{{nij}}}} \), \({{\bar {a}}_{{n.j}}} = \frac{1}{n}\sum\nolimits_{i = 1}^n {{{{\bar {a}}}_{{nij}}}} \), \({{\bar {a}}_{{n..}}} = \frac{1}{{{{n}^{2}}}}\sum\nolimits_{i,j = 1}^n {{{{\bar {a}}}_{{nij}}}} \), and \({{\mu }_{{nij}}} = {{\bar {a}}_{{ni.}}} + {{\bar {a}}_{{n.j}}} - {{\bar {a}}_{{n..}}}\). We designate \({{Y}_{{nij}}}\) = \({{\bar {X}}_{{nij}}} - {{\mu }_{{nij}}}\), Rn = \(\sum\nolimits_{i = 1}^n {{{Y}_{{ni{{\pi }_{n}}(i)}}}} \), \({{\bar {B}}_{n}}\) = \(\frac{1}{n}\sum\nolimits_{i,j = 1}^n {{\mathbf{E}}Y_{{nij}}^{2}} \).
First, we show that the relation
holds. To do this, we apply Theorem 1 and use the following result to check its hypotheses.
Lemma 1. We suppose y > 0, β ∈ (0, 1), μ ∈ (–1, 1), α > 0, and a ∈ (0, 1). Let X be a random variable and \(\bar {X}\) = \(XI\left\{ {\left| X \right| < y} \right\}\). We assume that \({\mathbf{E}}{{e}^{{{{{\left| X \right|}}^{\beta }}}}}\;\leqslant \;\alpha {\mathbf{E}}{{\left| {\bar {X} - \mu } \right|}^{s}}\) for s = 1, 2, 3 and \(\left| \mu \right|\;\leqslant \;M\ln 2\), where M = \({{y}^{{1 - \beta }}}{\text{/}}a\).
Then,
for all k > s for s = 1, 2, 3, where C(a, β) = \(8(2 + {{3}^{{1/\beta }}}{{((1 - a)\beta )}^{{ - 1/\beta }}})\).
Proof. For all z from the circle \(\left| z \right|\;\leqslant \;a{{y}^{{\beta - 1}}}\) = 1/M and s = 1, 2, 3, we have
By the Cauchy inequalities for the coefficients of expansion of the analytical function \({\mathbf{E}}{{(\bar {X} - \mu )}^{s}}{{e}^{{z(\bar {X} - \mu )}}}\), we obtain what is required.
\(\square \)
We show that condition (3) is fulfilled for Ynij with \({{M}_{n}} = y_{n}^{{1 - \beta }}{\text{/}}a\). Given condition (5), it is sufficient to show that \(\left| {{{\mu }_{{nij}}}} \right|\;\leqslant \;{{M}_{n}}\ln 2\) and \({\mathbf{E}}{{\left| {{{X}_{{nij}}}} \right|}^{s}}\;\leqslant \;\alpha {\mathbf{E}}{{\left| {{{{\bar {X}}}_{{nij}}} - {{\mu }_{{nij}}}} \right|}^{s}}\) for s = 1, 2, 3.
The functions \(x{{e}^{{ - {{x}^{\beta }}}}}\), \({{x}^{2}}{{e}^{{ - {{x}^{\beta }}}}}\), and \({{x}^{3}}{{e}^{{ - {{x}^{\beta }}}}}\) decrease for \(x\; \geqslant \;{{x}_{0}} > 0\). Further, we take that yn > x0.
Given condition (10), for all i, we have
Similarly, we have
Hence,
Further, for s = 1, 2, 3 and all i and j, the inequalities
hold for all sufficiently large n. Therefore,
Hence, \({\mathbf{E}}{{\left| {{{X}_{{nij}}}} \right|}^{s}}\;\leqslant \;90{\mathbf{E}}{{\left| {{{{\bar {X}}}_{{nij}}} - {{\mu }_{{nij}}}} \right|}^{s}}\). Moreover, \({\mathbf{E}}{{\left| {{{{\bar {X}}}_{{nij}}} - {{\mu }_{{nij}}}} \right|}^{s}}\) \(\leqslant \) \(4\left( {{\mathbf{E}}{{{\left| {{{{\bar {X}}}_{{nij}}}} \right|}}^{s}} + {{{\left| {{{\mu }_{{nij}}}} \right|}}^{s}}} \right)\) \(\leqslant \) \(5{\mathbf{E}}{{\left| {{{X}_{{nij}}}} \right|}^{s}}\). Thus, for s = 1, 2, 3 and all i and j, the inequalities
hold for all rather large n. Hence, by Lemma 1, condition (3) holds for Ynij with Mn = \(y_{n}^{{1 - \beta }}{\text{/}}a\).
Now, we estimate the difference of the variances Bn and \({{\bar {B}}_{n}}\). We have
Further,
By relations (8)–(11), we have
Hence,
Moreover,
In particular, \({{B}_{n}}\sim {{\bar {B}}_{n}}\). Given inequalities (12), we can conclude that the variable \(\bar {\gamma }\) specified by formula (2) with \({{X}_{{nij}}}\) replaced by \({{Y}_{{nij}}}\) satisfies the relation \({{\gamma }_{n}} = O({{\bar {\gamma }}_{n}})\) as n → ∞.
By Theorem 1, relation (7) holds for any sequence of real numbers {un} such that un → ∞, \(u_{n}^{3}\) = \(o(\sqrt n {\text{/}}{{\bar {\gamma }}_{n}})\) = \(o(\sqrt n {\text{/}}{{\gamma }_{n}})\), and un = \(o(\sqrt {{{{\bar {B}}}_{n}}} {\text{/}}{{M}_{n}})\) = \(o(B_{n}^{{\beta /(2(2 - \beta ))}})\) as n → ∞.
Further, we have
where
Since
we can conclude that 1 – \(\Phi ({{{v}}_{n}})\) ~ 1 – \(\Phi ({{u}_{n}})\) as n → ∞. Given (7), we obtain
It follows from inequalities (6) and (13) and relation (14) that
The theorem is proved completely.
\(\square \)
Proof of Remark 1. If β = 1, then by Lemma 1 with \(\bar {X} = X\) and μ = 0 (truncation and centering are not required in this case) condition (3) holds with \({{M}_{n}}\) = 1/a. Remark 1 follows from Theorem 1. The conditions on cn are superfluous in this case.
\(\square \)
REFERENCES
A. Wald and J. Wolfowitz, “Statistical tests based on permutations of observations,” Ann. Math. Stat. 15, 358–372 (1944).
G. E. Noether, “On a theorem by Wald and Wolfowitz,” Ann. Math. Stat. 20, 455–458 (1949).
W. Hoeffding, “A combinatorial central limit theorem,” Ann. Math. Stat. 22, 558–566 (1951).
M. Motoo, “On Hoeffding’s combinatorial central limit theorem,” Ann. Inst. Stat. Math. 8, 145–154 (1957).
V. F. Kolchin and V. P. Chistyakov, “On a combinatorial limit theorem,” Theory Probab. Its Appl. 18, 728–739 (1973). https://doi.org/10.1137/1118093
E. Bolthausen, “An estimate of the remainder in a combinatorial central limit theorem,” Z. Wahrscheinlichkeitstheor. Verw. Geb. 66, 379–386 (1984).
B. von Bahr, “Remainder term estimate in a combinatorial central limit theorem,” Z. Wahrscheinlichkeitstheor. Verw. Geb. 35, 131–139 (1976).
S. T. Ho and L. H. Y. Chen, “An Lp bounds for the remainder in a combinatorial central limit theorem,” Ann. Probab. 6, 231–249 (1978).
L. Goldstein, “Berry–Esseen bounds for combinatorial central limit theorems and pattern occur- rences, using zero and size biasing,” J. Appl. Probab. 42, 661–683 (2005).
K. Neammanee and J. Suntornchost, “A uniform bound on a combinatorial central limit theorem,” Stoch. Anal. Appl. 3, 559– 578 (2005).
K. Neammanee and P. Rattanawong, “A constant on a uniform bound of a combinatorial central limit theorem,” J. Math. Res. 1, 91–103 (2009).
L. H. Y. Chen, L. Goldstein, and Q.-M. Shao, Normal Approximation by Stein’s Method (Springer-Verlag, Berlin, 2011).
L. H. Y. Chen and X. Fang, “On the error bound in a combinatorial central limit theorem,” Bernoulli 21, 335–359 (2015).
A. N. Frolov, “Esseen type bounds of the remainder in a combinatorial CLT,” J. Stat. Plann. Inference 149, 90–97 (2014).
A. N. Frolov, “Bounds of the remainder in a combinatorial central limit theorem,” Stat. Probab. Lett. 105, 37–46 (2015).
A. N. Frolov, “On the probabilities of moderate deviations for combinatorial sums,” Vestn. St. Petersburg Univ.: Math. 48, 23–28 (2015). https://doi.org/10.3103/S1063454115010045
A. N. Frolov, “On Esseen type inequalities for combinatorial random sums,” Commun. Stat. - Theory Meth. 46, 5932–5940 (2017).
A. N. Frolov, “On large deviations for combinatorial sums,” J. Stat. Plann. Inference 217, 24–32 (2022).
Yu. V. Linnik, “Limit theorems for sums of independent random variables. I,” Teor. Veroyatn. Ee Primen. 6, 145–163 (1961);
Yu. V. Linnik, “Limit theorems for sums of independent random variables. II,” Teor. Veroyatn. Ee Primen. 6, 377–391 (1961);
Yu. V. Linnik, “Limit theorems for sums of independent random variables. III,” Teor. Veroyatn. Ee Primen. 7, 121–134 (1962).
ACKNOWLEDGMENTS
I thank the reviewers for a number of useful remarks that helped improve the text of the work.
Funding
This work is supported by the Russian Science Foundation (grant no. 23-21-00078). https://rscf.ru/project/23-21-00078/
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
The author declares that he has no conflicts of interest.
Additional information
Translated by M. Talacheva
About this article
Cite this article
Frolov, A.N. On the Probabilities of Large Deviations of Combinatorial Sums of Independent Random Variables That Satisfy the Linnik Condition. Vestnik St.Petersb. Univ.Math. 56, 385–391 (2023). https://doi.org/10.1134/S1063454123030044
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1134/S1063454123030044