Abstract
Two new symmetry tests, of integral and Kolmogorov type, based on the characterization by squares of linear statistics are proposed. The test statistics are related to the family of degenerate U-statistics. Their asymptotic properties are explored. The maximal eigenvalue, needed for the derivation of their logarithmic tail behavior, was calculated or approximated using techniques from the theory of linear operators and the perturbation theory. The quality of the tests is assessed using the approximate Bahadur efficiency as well as the simulated powers. The tests are shown to be comparable with some recent and classical tests of symmetry.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Consider the classical problem of testing the univariate symmetry with respect to zero. Let F be the distribution function (d.f.) of an i.i.d. sample \(X_1,\ldots ,X_n\), and suppose it is continuous. We are interested in testing the hypothesis
against the alternative \(H_1\) under which the equality in (1) is violated at least in one point.
Well-known and simple test statistics for this problem are the sign statistic and the Wilcoxon signed-rank statistic. Their properties are thoroughly explored and described in classical literature, as are some more sophisticated signed-rank statistics (see, e.g., [5, 11, 16, 27]).
Another class contains symmetry tests based on the empirical d.f.’s. Many examples, including the Kolmogorov–Smirnov- and \(\omega ^2\)-type tests, are described in [20]. This monograph offers an extensive review of various symmetry tests, together with the calculation of their efficiencies.
In recent times, introducing tests based on characterizations became a popular direction in goodness-of-fit testing. Such tests are attractive because they employ some intrinsic properties of the probability laws related to the characterization, and therefore they can exhibit high efficiency and power.
The first to introduce such symmetry tests were Baringhaus and Henze in [4]. They proposed suitable U-empirical Kolmogorov–Smirnov- and \(\omega ^2\)-type tests of symmetry based on their characterization. The calculation of Bahadur efficiencies, for the Kolmogorov-type test, was then performed in [21], see also [22]. An integral-type symmetry test, based on the same characterization, was proposed and analyzed by Litvinova [18].
Recently, Nikitin and Ahsanullah [23] built new tests of symmetry with respect to zero, based on the characterization by Ahsanullah [1]. This characterization was generalized and used for construction of similar symmetry tests by Milošević and Obradović [19]. The quality of all these tests was examined using the Bahadur efficiency, which is applicable to the case of non-normal limiting distributions.
Here we consider the characterization obtained independently by Wesolowski [28, Corollary 1], and Donati-Martin, Song and Yor [6, Lemma 1]. They proved the following proposition:
Let X and Y be i.i.d. random variables such that \((X-Y)^2\) and \((X+Y)^2\) are equidistributed. Then X and Y are symmetric with respect to zero.
Our aim is to build the integral- and the Kolmogorov-type U-empirical tests of symmetry based on this characterization; to explore their asymptotic properties; and to assess their quality via the approximate Bahadur efficiency and the simulated powers.
Our test statistics are based on U-statistics. This large class of statistics, which was first defined in the middle of last century in the problems of the unbiased estimation [8], and its theory established in the seminal paper of Hoeffding [10], are very important since numerous well-known statistics belong to this class. The most complete treatment of the theory can be found in [13, 15]. In our paper, unlike in many others in this domain of research, the emerging U-statistics and the families of U-statistics turn out to be degenerate. This feature highly complicates the problem and makes it new and attractive.
The limiting distribution of the underlying U-statistics is the second-order Gaussian chaos (see [14, Chapter 3]). Their tail behavior depends on the maximal eigenvalue of the corresponding integral operator. In the case of the uniform distribution, we obtain it theoretically. When studying the U-empirical Kolmogorov-type test, we are forced to work with the family of degenerate U-statistics depending on a real parameter t. Here a challenging problem lies in finding the supremum of the first eigenvalues (also depending on t) of the corresponding family of integral operators. This is mathematically the most interesting and original point. In the case of the uniform distribution, we solve it using a suitable decomposition of the family of operators to some simpler “triangular” operators (see “Appendix”). In the case of other distributions, we approximate the corresponding maximal eigenvalue using the appropriate sequences of discrete linear operators. The application of these mathematical means, for the first time in this field, is the most innovative and important feature of the paper.
The rest of the paper is organized as follows. In Sect. 2, we propose the test statistics and study their limiting distributions. Section 3 is devoted to the calculation of the approximate Bahadur efficiencies of our tests. In Sect. 4, we assess the powers of our tests through a simulation study. For convenience, some proofs are given in “Appendix”.
2 Test Statistics
Let \(F_n^{*}\) be the empirical d.f. of \(|X_1|,\ldots ,|X_n|\), and let \(F^{*}(t) = P\{ |X_1| < t \}\) be its theoretical counterpart. In view of the characterization, we consider the following two test statistics:
where
are U-empirical d.f.’s. After the integration, we obtain
The statistic \({\bar{J}}_n\) is a hybrid of a U- and a V-statistics. Instead of it, we propose the corresponding U-statistic
with symmetrized kernel
where \(\Pi (m)\) is the set of all permutations of the set \(\{1,2,\ldots ,m\}\).
This statistic is more natural, and, moreover, an unbiased estimator of
Magnified by the factor n, \(J_n\) and \({\bar{J}}_n\) have somewhat different limiting distributions. However, in terms of the logarithmic tail behavior they are equivalent, and both statistics lead to consistent tests for our hypothesis.
We consider large values of \(|J_n|\) and \(K_n\) to be significant.
2.1 Statistic \(J_n\)
The statistic \(J_n\) is a U-statistic with symmetric kernel \(\Phi \) given in (4). Its first projection on \(X_1\) under \(H_0\) is equal to zero, while the second projection on \((X_1,X_2)\), at the point (s, t), is
where F is the d.f. of a null symmetric distribution.
Theorem 2.1
Under \(H_0\), the following convergence in distribution holds
where \(\{W_i\}\) is the sequence of i.i.d. standard normal random variables, and \(\{\nu _i(F)\}\) is the non-increasing sequence of eigenvalues of the integral operator \({\mathcal {J}}_F\) with kernel \(\varphi _F\).
Proof
Since the kernel \(\Phi \) is bounded and degenerates, the result follows from the theorem for the asymptotic distribution of U-statistics with degenerate kernels [13, Corollary 4.4.2]. \(\square \)
Since our test statistic is not distribution free, the eigenvalues need to be derived for each null distribution. In the following theorem we consider the case of the uniform distribution.
Theorem 2.2
Let F be the d.f. of the uniform \(U[-\vartheta ,\vartheta ]\) distribution. Then the sequence of eigenvalues \({\nu _i}\) from (6) is the solution of the following equation
Proof
It is easy to see that our statistic is scale free. Therefore, we may suppose that F is the d.f. of the uniform \(U[-1,1]\) distribution.
From (5), it is obvious that the function \(\varphi \) is odd, as a function of s as well as a function of t. Hence, it suffices to present the kernel for \(s>0\) and \(t>0\).
By definition, the eigenvalues and their eigenfunctions e satisfy
Since the kernel is an odd function, the eigenfunctions must be odd, too. Therefore, they can be represented using their Fourier expansion
where \(u_k(s)=\sin k\pi s\).
Applying the operator \({\mathcal {J}}\) to the function e given in (9), we have that (8) is equivalent to
The left-hand side is a function from \(L^2[-1,1]\), and its Fourier expansion is
where \(b_l=\sum _{k=1}^{\infty }a_k\langle {\mathcal {J}}[u_k(t)],u_{l}(t)\rangle /||u_{l}(t)||=2\sum _{k=1}^{\infty }\langle {\mathcal {J}}[u_k(t)],u_{l}(t)\rangle \), and \(\langle \cdot ,\cdot \rangle \) is the scalar product. After some calculations, we get
From (10) and (11), we obtain the system
and substituting (12) in (13) we get
where
Transforming (14), we obtain
Summing both sides of (15) for \(l=1,2\ldots \), we obtain
from where follows (7). \(\square \)
2.2 Statistic \(K_n\)
For a fixed \(t>0\), the expression \(K_n^*(t)=G_n(t)-H_n(t)\) is a U-statistic with the symmetric kernel
It is easy to see that the kernel is degenerated with the second projection
For studying the asymptotics of the statistic \(K^{*}_n\), it is of interest to consider the integral operator with the same kernel.
For any function \(v\in L^2({\mathbb {R}})\), we define it as
Let \(\{\nu _i(t;F)\}\) be the sequence of the eigenvalues of the integral operator \({\mathcal {Q}}_F(t)\). In the following theorem, we give the limiting process of \(nK^{*}(t)\). This process is called the second-order Gaussian chaos process (see, e.g., [14, Chapter 3]).
Theorem 2.3
Under \(H_0\), the limiting process of \(nK_n^{*}(t)\), \(n\rightarrow \infty \), is
where \(\{e_i\}\) is an orthonormal basis of \(L^2({\mathbb {R}})\), and \(\{W_i\}\) are i.i.d. standard normal random variables.
Proof
Our class of kernels \(\xi (x,y;t)\) is Euclidean in the sense of [25], so the conditions of [26, Theorem 7] are satisfied, and (17) follows. \(\square \)
Hence, \(nK_n\) converges to the random variable \(\sup _{t\in [0,\infty ]}|\zeta _F(t)|\).
3 Approximate Bahadur Efficiency
Let \({\mathcal {G}}=\{G(x;\theta )\}\) be the family of d.f.’s with densities \(g(x;\theta )\), such that \(G(x,\theta )\) is symmetric only for \(\theta =0\). We assume that the d.f.’s from the class \({\mathcal {G}}\) satisfy the regularity conditions from [24, Assumptions WD]. Denote \(h(x)=g'_{\theta }(x;0)\).
Suppose that \(T_n=T_n(X_1,\ldots ,X_n)\) is a sequence of test statistics whose large values are significant, i.e., the null hypothesis \(H_0:\theta \in \Theta _0\) is rejected in favor of \(H_1:\theta \in \Theta _1\), whenever \(T_n>t_n\). Let the sequence of d.f.’s of the test statistic \(T_n\) converge in distribution to a non-degenerate d.f. F. Additionally, suppose that
and the limit in probability under the alternative
exists for \(\theta \in \Theta _1\).
The approximate relative Bahadur efficiency with respect to another test statistic \(V_n=V_n(X_1,\ldots ,X_n)\) is defined as
where
is called the Bahadur approximate slope of \(T_n\). This is a popular measure of the test efficiency suggested by Bahadur in [3].
3.1 Integral-Type Test
In the case of our integral-type test statistic, the role of \(T_n\) is played by the statistic \({\widetilde{J}}_n=\sqrt{n|J_n|}\). Its Bahadur approximate slope is obtained in the following lemma.
Lemma 3.1
For the statistic \({\widetilde{J}}_n\) and a given alternative density \(g(x;\theta )\) from \({\mathcal {G}}\), the Bahadur approximate slope satisfies the relation
where \(b_J(\theta )\) is the limit in \({P_\theta }\) probability of \(J_n\), and \(\nu _1\) is the largest eigenvalue of the sequence \(\{\nu _i(F)\}\) in (6).
Proof
Using the result of Zolotarev [29], we have that the logarithmic tail behavior of \({\widetilde{J}}\) is
and hence, \({\widetilde{a}}_{{\widetilde{J}}}=\frac{1}{3\nu _1}\). The limit in probability of \({\widetilde{J}}_n/\sqrt{n}\) is
Inserting the expressions for \({\widetilde{a}}_{{\widetilde{J}}}\) and \({\widetilde{b}}_{{\widetilde{J}}}(\theta )\) into (18), we obtain the statement of the lemma. \(\square \)
The largest eigenvalue in the case of the uniform distribution is calculated from (7) and is equal to 0.1898. The equations for the eigenvalues of the operator
for other distributions F are too complicated to derive. Thus, we calculate the largest eigenvalues using the following approximation. First, notice that the “symmetrized” operator
has the same spectrum as the operator \({\mathcal {J}}_F\).
Consider the case where \(\inf _{x}\{x: F(x)=1\}=A<\infty \). The sequence of symmetric linear operators defined by \((2m+1)\times (2m+1)\) matrices \(M^{(m)}_F=||m_{i,j}^{(m)}||,\; |i|\le m,|j|\le m\), where
converges in norm to \({\mathcal {S}}_F\).
Indeed, for a function \(v\in L^2[-A,A]\), the operator \(M^{(m)}_F\), for \(x\in [Ai/m,A(i+1)/m)\), can be written as
This sum converges to the Lebesgue integral from (20). Since this is true for every function v, then
The operators \({\mathcal {S}}\) and \(M^{(m)}_F\) are symmetric and self-adjoint, and the norm of their difference tends to zero as m tends to infinity. Using the perturbation theory, see [12, Theorem 4.10, page 291], we have that the spectra of these two operators are at the distance that tends to zero. Hence, \(\nu _1^{(m)}(F)\)—the sequence of the largest eigenvalues of \(M^{(m)}_F\)—must converge to \(\nu _1(F)\), the largest eigenvalue of \({\mathcal {S}}\).
In the case where \(\inf _{x}\{x: F(x)=1\}=\infty \), we consider its truncation to the interval \([-A,A]\), such that \(F(A)>1-\varepsilon \) for a desired value of \(\varepsilon >0\). The values for some common symmetric distributions are given in Table 1.
Concerning the limit in probability \(b_J(\theta )\), we give its formula in the following lemma. Its proof follows from [24].
Lemma 3.2
Under a close alternative \(g(x;\theta )\) from \({\mathcal {G}}\), such that \(G(x;0)=F(x)\), the limit in probability of \(J_n\) is
as \(\theta \rightarrow 0\), where \(h(x)=g_\theta '(x;0)\).
We consider null d.f.’s to be uniform, normal, logistic and Laplace, and the following alternative distributions are close to a null d.f. F:
a Lehmann alternative with d.f.
$$\begin{aligned} G_1(x;\theta )=F^{1+\theta }(x), \;\; \theta >0; \end{aligned}$$(21)a first Ley–Paindaveine [17] alternative with d.f.
$$\begin{aligned} G_2(x;\theta )=F(x)e^{-\theta (1-F(x))}, \;\; \theta >0; \end{aligned}$$(22)a second Ley–Paindaveine alternative [17] with d.f.
$$\begin{aligned} G_3(x;\theta )=F(x)-\theta \sin \big (\pi F(x)\big ), \;\; \theta \in [0,\pi ^{-1}]; \end{aligned}$$(23)a contamination (with \(G_1\)) alternative with d.f.
$$\begin{aligned} G_4(x;\theta ,\beta )=(1-\theta )F(x)+\theta F^{\beta }(x),\;\;\beta >1,\;\theta \in [0,1]; \end{aligned}$$(24)a location alternative with d.f.
$$\begin{aligned} G_5(x;\theta )=F(x-\theta ); \end{aligned}$$(25)a contamination (with shift) alternative with d.f.
$$\begin{aligned} G_6(x;\theta ,\beta )=(1-\theta )F(x)+\theta F(x-\beta ),\;\;\;\beta >0,\;\theta \in [0,1]; \end{aligned}$$(26)a skew alternative in the sense of Azzalini [2] with density
$$\begin{aligned} g_7(x;\theta )=2F(\theta x)f(x). \end{aligned}$$(27)
Example 3.3
Let the alternative distribution be \(G_1(x;\theta )\) when F is the d.f. of the uniform distribution. In this case, we have
By Lemma 3.2, after calculating the corresponding integral, we have that \(b_J(\theta )\sim 0.348\theta ^2\), \(\theta \rightarrow 0\). The largest solution of (7) is \(\nu _1\approx 0.1898\). Therefore, the Bahadur approximate slope is
The calculations are similar for the other alternatives. The Bahadur approximate indices (the leading coefficient in the Maclaurin expansion of the Bahadur approximate slope) are presented in Table 3 at the end of this section, together with the results for the Kolmogorov-type test.
3.2 Kolmogorov-Type Test
Similarly to the integral-type statistic, we study tail behavior of the limiting distribution of the statistic \({\widetilde{K}}_n=\sqrt{nK_n}\). Notice that \({\widetilde{K}}_n\) converges to \(\sup _{t}|\zeta _F(t)|^{1/2}\).
Theorem 3.4
For the limiting distribution of the test statistic \({\widetilde{K}}_n\), it holds true that
where \(\varkappa _0(F)=\sup _{t}|\nu _1(t;F)|\), and \(\{\nu _1(t;F)\}\) is the set of eigenvalues of the family of operators \({\mathcal {Q}}_F(t),\;t>0\), with the largest absolute value.
Proof
The limiting process of \(nK_n^{*}\) is the second-order Gaussian chaos process \(|\zeta _F(t)|\). The tail behavior of its supremum \(nK_n\) is obtained in [14, Corollary 3.9]. The constant \(\sigma \) appearing there is equal to the supremum of the maximal eigenvalues of the corresponding linear operator. This is because
Writing it in our notation, we obtain
and transforming the variable we get (28). \(\square \)
The following lemma gives us the limit in probability of the statistic \(K_n\).
Lemma 3.5
Under a close alternative \(g(x;\theta )\) from \({\mathcal {G}}\), such that \(G(x;0)=F(x)\), the limit in probability of \(K_n\) is
Proof
Denote by \(a(t;\theta )\) the limit in probability of the statistic \(K_n^*(t)\) under the alternative \(g(x;\theta )\). Using the Glivenko–Cantelli theorem for U-statistics [9], we have that \(b(\theta )=\sup _{t>0}a(t;\theta )\), where
It is easy to show that \(a(t;0)=a'(t;0)=0\). The second derivative of \(a(t;\theta )\) along \(\theta \) at \(\theta =0\) is
Expanding \(a(t;\theta )\) in the Maclaurin series completes the proof. \(\square \)
Finally, the Bahadur approximate slope is given in the following theorem.
Theorem 3.6
For the statistic \({\widetilde{K}}_n\) and a given alternative density \(g(x;\theta )\) from \({\mathcal {G}}\), the Bahadur approximate slope satisfies the relation
where \(\varkappa _0\) is given in (30).
Proof
Using Theorem 3.4 and Lemma 3.5, and the same arguments as in the case of the statistic \({\widetilde{J}}_n\), we get the statement of the theorem. \(\square \)
We apply the same approximation procedure used in the case of the operator \({\mathcal {J}}_F\). For the eigenvalues of the operator \({\mathcal {Q}}_F(t)\), for \(n=1000\), we get the functions \(\nu _1^{(1000)}(F;t)\) in the case of uniform, normal, logistic and Laplace distributions. Since in all the cases the functions have the same shape, we show only the function \(\nu _1^{(1000)}(F;t)\) for the uniform distribution case (Fig. 1).
From Fig. 1, we can see that the function has a unique maximum \(\varkappa _0\approx 0.766\) for \(t\approx 2/3\). In this case, however, we are able to derive \(\varkappa _0\) theoretically in the following lemma.
Lemma 3.7
Let \(F(x)=(x+1)/2\), \(x\in [-1,1]\) and let \({\mathcal {Q}}(t)\) be the integral operator (16) corresponding to this F. Let \(\varkappa _0=\sup _{t\in [0,2]}|\nu _1(t)|\), where \(\{\nu _1(t), 0 \le t \le 2\} \) is the set of eigenvalues of the family of operators \({\mathcal {Q}}(t),\;t\in (0,2)\) with the largest absolute value. Then
The proof is given in “Appendix”.
For the other distributions, we rely on the values obtained using the approximation. We present the obtained values in Table 2.
Example 3.8
Let the alternative distribution be \(G_3(x;\theta )\). In this case, we have
By Lemma 3.5, after calculating the corresponding integral, we have that
The supremum is attained at \(t=0.708\) and is approximately equal to 3.639. The Bahadur approximate slope is then
The calculations are similar for the other alternatives. In Table 3, we give the values of the local approximate indices. Naturally, we confine ourselves to the cases where the alternatives (21)–(27) belong to the class \({\mathcal {G}}\), i.e., satisfy the regularity conditions.
For comparison purpose, we also include Bahadur indices of some competitor tests. We choose some recent characterization-based symmetry tests from [19, 23] (labeled, respectively, as \(c_{\mathrm {MOI}_k}\) and \(c_{\mathrm {NAI}_k}\) for integral-type tests, and \(c_{\mathrm {MOK}_k}\) and \(c_{\mathrm {NAK}_k}\) for Kolmogorov-type tests), as well as the classical sign test (\(c_S\)). The relative efficiency of any two tests can be calculated as the ratio of their Bahadur indices.
We find that our tests, in all cases, are more efficient than the sign test. In comparison with the recent characterization-based tests, in some cases (e.g., \(g_3\) alternative for the uniform distribution) they outperform all the others, while in some other cases (e.g., \(g_1\) for normal distribution) new tests are the least efficient. Moreover, as it is often the case, no test is uniformly the most efficient.
When comparing two new tests to each other, we can notice that in the case of the uniform null distribution, the integral-type test is more efficient than the Kolmogorov-type test. This is a widespread situation in the comparison of tests, see [20]. However, for the other null distributions, it is mostly the other way around.
4 Power Study
In Tables 4 and 5, we present the empirical sizes and powers of our tests (\(J_n\) and \(K_n\)) against the alternatives \(g_5\), \(g_6(1)\), and \(g_7\), for some values of parameter \(\theta \).
The null distributions are normal, Laplace, logistic, and Cauchy. The simulated powers for \(J_n\) and \(K_n\), at the level of significance of 0.05, are obtained using a warp speed Monte Carlo bootstrap procedure [7] given below.
Warp speed Monte Carlo bootstrap algorithm
- (i)
Generate the sequence \(x_1,\ldots ,x_n\) from an alternative distribution and compute the value of the test statistic \(T_n(x_1,\ldots ,x_n)\);
- (ii)
Generate \(y^*_k=x_ku_k, \ k=1,\ldots ,n\), using i.i.d sequence of Rademacher random variables \(u_k\) taking values -1 and 1 with equal probabilities, to obtain the symmetrized sampling distribution;
- (iii)
Calculate the value of the test statistic \(T^*_n=T_n(y_1^*,\ldots ,y^*_n)\);
- (iv)
Repeat the steps (i)–(iii) N times, and obtain the empirical sampling distributions of \(T_n\) and \(T^*_n\) that correspond to the alternative and the null distribution of the test statistic, respectively;
- (v)
Calculate the empirical power as the percentage of values of \(T^*_n\) greater than the 95th percentile of the empirical sampling distribution of \(T_n\).
The procedure is done for \(N=10{,}000\) replications, for the sample sizes of 20 and 50. For comparison purpose, we also include the same characterization-based tests as in Table 3, as well as the classical Kolmogorov–Smirnov symmetry test (KS) and the sign test (S), whose powers are calculated using the standard Monte Carlo procedure with 10,000 replications.
From Tables 4 and 5, one can see that the empirical sizes of our tests are satisfactory. Besides, \(J_n\) and \(K_n\) have almost equal empirical powers for all the alternatives.
We can observe that our tests have the highest powers in the case of the contamination alternative \(g_6(1)\), for all nulls, for smaller sample sizes (\(n=20\)), and for the logistic null for \(n=50\). Similar conclusion can be made for the location alternative \(g_5\) of the Cauchy null, for both sample sizes.
In other cases, while not being uniformly the best, the powers of our tests are comparable to all the competitors’.
5 Conclusion
In this paper, we presented two new tests of symmetry based on a characterization and examined their asymptotic properties. We calculated the local approximate Bahadur efficiencies of our tests, and performed a small-scale power study. We found out that our tests are comparable to some commonly used classical tests of symmetry.
When exploring the asymptotics of our tests, the most challenging problem was to obtain the maximal eigenvalue of some integral operators. In some cases, we were able to do it theoretically, using Fourier analysis and a decomposition of linear operators. For the rest of the cases, we suggested an approximation method based on a discretization of the corresponding integral operators. The described procedure could be useful in general for approximating the asymptotic distribution of degenerate U-statistics, which emerge often in the problems of goodness-of-fit testing.
References
Ahsanullah, M.: On some characteristic property of symmetric distributions. Pak. J. Stat. 8(3), 19–22 (1992)
Azzalini, A., with the collaboration of Capitanio, A.: The Skew-Normal and Related Families. Cambridge University Press, New York (2014)
Bahadur, R.R.: Stochastic comparison of tests. Ann. Math. Stat. 31, 276–295 (1960)
Baringhaus, L., Henze, N.: A characterization of and new consistent tests for symmetry. Commun. Stat. Theory Methods 21(6), 1555–1566 (1992)
Behboodian, J.: A note of the skewness and symmetry. The Statistician 38(1), 21–23 (1980)
Donati-Martin, C., Song, S., Yor, M.: On symmetric stable random variables and matrix transposition. Ann. de l’IHP (Probab. et Statist.) 30(3), 397–413 (1994)
Giacomini, R., Politis, D., White, H.: A warp-speed method for conducting Monte Carlo experiments involving bootstrap. Econ. Theory 29(3), 567–589 (2013)
Halmos, P.R.: The theory of unbiased estimation. Ann. Math. Stat. 17(1), 34–43 (1946)
Helmers, R., Janssen, P., Serfling, R.: Glivenko–Cantelli properties of some generalized empirical df’s and strong convergence of generalized L-statistics. Probab. Theory Relat. Fields 79(1), 75–93 (1988)
Hoeffding, W.: A class of statistics with asymptotically normal distribution. Ann. Math. Stat. 19(3), 293–395 (1948)
Huškova, M.: Hypothesis of symmetry. In: Handbook of Statistics: Nonparametric Methods, vol. 4, pp. 63–78. North Holland, Amsterdam (1984)
Kato, T.: Perturbation Theory of Linear Operators, 2nd edn. Springer, Berlin (1980)
Korolyuk, V.S., Borovskikh, YuV: Theory of U-Statistics. Kluwer, Dordrecht (1994)
Ledoux, M., Talagrand, M.: Probability in Banach Spaces: Isoperimetry and Processes. Springer, Berlin (2013)
Lee, J.: U-statistics, Theory and Practice. CRC Press, New York (1990)
Lehmann, E.L., D’Abrera, H.J.M.: Nonparametrics: Statistical Methods Based on Ranks. Springer, Heidelberg (2006)
Ley, C., Paindaveine, D.: Le Cam optimal tests for symmetry against Ferreira and Steels general skewed distribution. J. Nonparametr. Stat. 21(8), 943–967 (2008)
Litvinova, V.V.: New nonparametric test for symmetry and its asymptotic efficiency. Vestn. St. Petersb. Univ. Math. 34(4), 12–14 (2001)
Milošević, B., Obradović, M.: Characterization based symmetry tests and their asymptotic efficiencies. Stat. Probab. Lett. 119, 155–162 (2016)
Nikitin, Ya. Yu.: Asymptotic Efficiency of Nonparametric Tests. Cambridge University Press, Cambridge (1995)
Nikitin, Ya. Yu.: On Baringhaus–Henze test for symmetry: Bahadur efficiency and local optimality for shift alternatives. Math. Methods Stat. 5(2), 214–226 (1996)
Nikitin, Ya. Yu.: Large deviation of U-empirical Kolmogorov–Smirnov tests and their efficiency. J. Nonparametr. Stat. 22(5), 649–668 (2010)
Nikitin, Ya. Yu., Ahsanullah, M.: New U-empirical tests of symmetry based on extremal order statistics, and their efficiencies. In: Hallin, M., Mason, D.M., Pfeifer, D., Steinebach, J.G. (eds.) Mathematical Statistics and Limit Theorems, Festschrift in Honour of Paul Deheuvels, pp. 231–248. Springer, Berlin (2015)
Nikitin, Ya. Yu., Peaucelle, I.: Efficiency and local optimality of nonparametric tests based on U- and V-statistics. Metron LXII(2), 185–200 (2004)
Nolan, D., Pollard, D.: U-processes: rates of convergence. Ann. Stat. 15(2), 780–799 (1987)
Nolan, D., Pollard, D.: Functional limit theorems for U-processes. Ann. Probab. 16(3), 1291–1298 (1988)
Šidák, Z., Sen, P.K., Hájek, J.: Theory of Rank Tests. Academic Press, New York (1999)
Wesolowski, J.: Distributional properties of squares of linear statistics. J. Appl. Stat. Sci. 1(1), 89–94 (1993)
Zolotarev, V.M.: Concerning a certain probability problem. Theory Probab. Appl. 6(2), 201–204 (1961)
Acknowledgements
We would like to thank the referees for their useful remarks that improved our paper. The research was supported by MNTRS, Serbia, Grant No. 174032 (first author), MNTRS, Serbia, Grant No. 174012 (second author), and the SPbGU-DFG Grant 6.65.37.2017, and RFBR Grant 16-01-00258 (third author).
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Anton Abdulbasah Kamil.
Appendix
Appendix
1.1 Proof of Lemma 3.7
To prove the lemma, we show that the local maximum, which coincides with the global one (see Fig. 1), is attained at \(t=2/3\) and find its value.
The idea is to demonstrate that both the right and the left derivative of \(\nu _1(t)\) at \(t=2/3\) are equal to zero. For this, we need the functional equations that \(\nu _1(t)\) satisfies in the neighborhood of \(t=2/3\). We need both derivatives since these functional equations happen to be different on different sides of \(t=2/3\).
We start from the eigenfunction equation
Then, for t close to 2 / 3 let us decompose the operator to some simpler operators whose spectra are obtainable in closed form.
First, since the kernels of the family of operators \({\mathcal {Q}}(t),\;t\in [0,2]\), are odd functions, the corresponding eigenfunctions must be odd, too. Therefore, instead of \({\mathcal {Q}}(t)\), we may consider its restriction \({\mathcal {Q}}^{\star }(t)\), for functions defined on [0, 1], which has the same spectrum. The kernels of the operators \({\mathcal {Q}}^{\star }(t)\) for \(t \in (2/3,1)\) and \(t \in (1/2,2/3)\) are shown in Fig. 2a, b, respectively. The kernel is equal to one inside the shaded region, and equal to zero outside.
From Fig. 2a, b, one can notice (dashed lines) that the eigenfunctions can be decomposed in such a way that the only operators applied to these “subfunctions” are “triangular” or “constant”.
We now introduce some notation to formalize this argument. Let \({\mathcal {D}}_1\) be the “upper right triangular” operator acting on integrable functions f defined on [0, 1], i.e., \({\mathcal {D}}_1[f](x)=\int _{1-x}^1f(y)\text {d}y.\) Analogously, we define the “upper left triangular” operator \({\mathcal {L}}_1\), and the “lower right triangular” operator \({\mathcal {R}}_1\). Let \({\mathcal {M}}_1\) be the mean value operator \({\mathcal {M}}_1[f]=\int _{0}^1f(y)\text {d}y: = {\bar{f}}\).
We say that two functions f and g defined on [0, 1] are reverse if \(f(x)=g(1-x)\), for all \(x\in [0,1]\). It is easy to show the following:
- (i)
The image \({\mathcal {M}}_1[f]\) is a constant function \({\bar{f}}\);
- (ii)
The images \({\mathcal {D}}_1[f]\) and \({\mathcal {L}}_1[f]\) are reverse functions;
- (iii)
If the functions f and g are reverse, then \({\mathcal {D}}_1[f]={\mathcal {R}}_1[g]\).
Let f be the function defined on [0, 1]. Denote with \({\hat{f}}_{a,\delta }\) its contraction to the interval of length \(\delta \), i.e., for any subinterval \([a,a+\delta ]\subset [0,1]\),
Let \({\mathcal {D}}_{\delta },{\mathcal {R}}_{\delta },{\mathcal {L}}_{\delta }\) and \({\mathcal {M}}_{\delta }\) be the corresponding natural \(\delta \)-contraction operators
The operator \({\mathcal {M}}\) is a bit different because it can map any \({\hat{f}}_{b,\delta }\) to a function defined on an interval of a different length, say \({\hat{f}}_{a,\delta _1}\). However, it is a constant operator, so its restriction is just, regardless of a and \(\delta _1\),
Change of variable in (33) gives us a useful relation
The same holds for the other three operators.
Let \(t \in [2/3,1]\) (Fig. 2a). We can decompose our eigenfunction e to four subfunctions \({\hat{f}},{\hat{u}},{\hat{g}}\) and \({\hat{h}}\), as follows
Applying \({\mathcal {Q}}^{\star }(t)\) to e(x), for \(x \in [0,1-t]\) we get
This is exactly what we can see in Fig. 2a for \(x\in (0,1-t)\): the first two operators are zero; then comes an upper right operator \({\mathcal {D}}\); and finally the lower right operator \({\mathcal {R}}\). On the other hand, e(x) restricted to this interval is simply \({\hat{f}}(x)\). Hence from (31) we get
Dilating all the functions to [0, 1] using (33) and (34), we get
where \(x_1=\frac{x}{1-t}\).
Putting x through all four intervals, we transform Eq. (31) into the system
where for simplicity, we write the equations in terms of the functions only, omitting their arguments.
From the last two equations, using (ii), and the fact that the reverse functions have the same mean value, we get that g and h are reverse functions. Then, using (i) and (iii), we transform the system to
Expressing f from the first equation and rearranging the remaining equations, we get
where \({\mathcal {E}}\) is the identity operator acting on functions defined on [0, 1]. The constant function \({\bar{g}}\) (and \({\bar{u}}\)) can be expressed as \({\bar{g}}=\langle g,v\rangle v\), where \(v(x)=1\), \(x\in [0,1]\), and \(\langle g,v\rangle =\int _{0}^{1}g(x)v(x)dx\) is the scalar product.
Denote, for brevity, \(c_1(t)=(3t-2)/\nu (t)\) and \(c_2(t)=2(1-t)^2/\nu ^2(t)\).
Define the functions \(\Psi _1(c_1(t))=\langle v,({\mathcal {E}}-c_1(t){\mathcal {D}}_1)^{-1}[v]\rangle \) and \(\Psi _2(c_2(t))=\langle v,({\mathcal {E}}-c_2(t){\mathcal {D}}^2_1)^{-1}[v]\rangle \). Applying the appropriate inverse operators to the left-hand side of both equations in the system, and multiplying scalarly with v, the system becomes:
Then, solving the system we obtain the following equation
To find the functions \(\Psi _1\) and \(\Psi _2\), we need the spectrum of \({\mathcal {D}}_1\). We get it from the following proposition.
Proposition
Let \(\{\mu _n\}\) and \(\{e_n\}\), \(n \in {\mathbb {Z}}\) be the sequences of eigenvalues and normalized eigenfunctions of \({\mathcal {D}}_1\). Let \(v=\sum _{n}a_ne_n\) be the representation of the function \(v(x)=1,\;x\in [0,1]\) in the basis \(\{e_n\}\). Then \(\mu _n=\frac{2}{(4n+1)\pi }\) and \(a_n=\frac{2\sqrt{2}}{(4n+1)\pi }.\)
The proof can be done either mimicking the proof of Theorem 2.2, or by reducing the eigenfunction equation to the appropriate Sturm–Liouville boundary problem.
Using the decomposition of linear operators in the basis of its eigenfunctions, and Proposition, we obtain
For \(t=2/3\), Eq. (36) reduces to
The solution with the largest absolute value is \(\nu _1\big (\frac{2}{3}\big )=\frac{\sqrt{2}}{3}(\arctan \frac{1}{\sqrt{2}})^{-1}\).
Using the implicit function theorem, we get that \(\nu _1(t)\) is differentiable along t in the right neighborhood of \(t=2/3\). Furthermore, the right first derivative of right-hand side of equation (36) at \(t=2/3\) is equal to
Since this must be equal to zero, we conclude that the right derivative \(\nu '_1(2/3)=0\). The right second derivative of the right-hand side of (36) gives us that \(\nu _1''(2/3)<0\). Hence, \(t=2/3\) is a “right maximum”.
Using a completely analogous procedure, one can show that it is a “left maximum”, too, and therefore a local maximum of \(\nu _1(t).\)\(\square \)
Rights and permissions
About this article
Cite this article
Božin, V., Milošević, B., Nikitin, Y.Y. et al. New Characterization-Based Symmetry Tests. Bull. Malays. Math. Sci. Soc. 43, 297–320 (2020). https://doi.org/10.1007/s40840-018-0680-3
Received:
Revised:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40840-018-0680-3
Keywords
- Symmetry tests
- Characterization
- Degenerate U-statistics
- Second-order Gaussian chaos process
- Approximation of maximal eigenvalue
- Asymptotic efficiency