Keywords

These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

The Hotelling’s \(T^2\) test is used to test \(H_0: {\varvec{\mu }}= {\varvec{\mu }}_0\) when there is one sample, and \(H_0: {\varvec{\mu }}_1 = {\varvec{\mu }}_2\) when there are two samples. Other applications include the multivariate matched pairs test and a test in the repeated measurements setting. These tests are robust to nonnormality.

The one-sample Hotelling’s \(T^2\) test, multivariate matched pairs test, and two-sample Hotelling’s \(T^2\) test are analogs of the univariate one-sample t test, matched pairs t test, and two-sample t test, respectively. For the multivariate Hotelling’s \(T^2\) tests, there are \(p > 1\) variables and their correlations are important.

9.1 One Sample

The one-sample Hotelling’s \(T^2\) test is used to test \(H_0: {\varvec{\mu }}= {\varvec{\mu }}_0\) versus \(H_A: {\varvec{\mu }}\ne {\varvec{\mu }}_0\). The test rejects \(H_0\) if

$$T_H^2 = n(\overline{\varvec{x}} - {\varvec{\mu }}_0)^T {\varvec{S}}^{-1} (\overline{\varvec{x}} - {\varvec{\mu }}_0) > \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha }$$

where \(P(Y \le F_{p, d,\alpha }) = \alpha \) if \(Y \sim F_{p, d}\).

If a multivariate location estimator T satisfies

$$\sqrt{n}(T - {\varvec{\mu }}) {\mathop {\rightarrow }\limits ^{D}} N_p({\varvec{0}}, {\varvec{D}}),$$

then a competing test rejects \(H_0\) if

$$T_C^2 = n(T - {\varvec{\mu }}_0)^T \hat{{\varvec{D}}}^{-1} (T - {\varvec{\mu }}_0) > \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha }$$

where \(\hat{{\varvec{D}}}\) is a consistent estimator of \({\varvec{D}}\). The scaled F cutoff can be used since \(T^2_C {\mathop {\rightarrow }\limits ^{D}} \chi ^2_p\) if \(H_0\) holds, and

$$\frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha } \rightarrow \chi ^2_{p, 1-\alpha }$$

as \(n \rightarrow \infty .\) This idea is used for small p by Srivastava and Mudholkar (2001) where T is the coordinatewise trimmed mean. The one-sample Hotelling’s \(T^2\) test uses \(T = \overline{\varvec{x}}\), \({\varvec{D}}= {\varvec{\varSigma }}_{\varvec{x}}\), and \(\hat{{\varvec{D}}} = {\varvec{S}}\).

The Hotelling’s \(T^2\) test is a large sample level \(\alpha \) test in that if \(\varvec{x}_1, ..., \varvec{x}_n\) are iid from a distribution with mean \({\varvec{\mu }}_0\) and nonsingular covariance matrix \({\varvec{\varSigma }}_{\varvec{x}}\), then the type I error \(=\) P(reject \(H_0\) when \(H_0\) is true) \(\rightarrow \alpha \) as \(n \rightarrow \infty \). We want \(n \ge 10 p\) if the DD plot is linear through the origin and subplots in the scatterplot matrix all look ellipsoidal. For any n, there are distributions with nonsingular covariance matrix where the \(\chi ^2_p\) approximation to \(T^2_H\) is poor.

Let pval be an estimate of the pvalue. We typically use \(T_C^2 = T_H^2\) in the following four-step test. i) State the hypotheses \(H_0: {\varvec{\mu }}= {\varvec{\mu }}_0\) \(H_1: {\varvec{\mu }}\ne {\varvec{\mu }}_0\).

ii) Find the test statistic \(T_C^2 = n(T - {\varvec{\mu }}_0)^T \hat{{\varvec{D}}}^{-1} (T - {\varvec{\mu }}_0)\).

iii) Find pval \(=\)

$$P\left( T_C^2< \frac{(n-1)p}{n-p} F_{p, n-p}\right) = P\left( \frac{n-p}{(n-1)p} \ \ T_C^2 < F_{p, n-p}\right) .$$

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that \({\varvec{\mu }}\ne {\varvec{\mu }}_0\) while if you fail to reject \(H_0\) conclude that the population mean \({\varvec{\mu }}= {\varvec{\mu }}_0\) or that there is not enough evidence to conclude that \({\varvec{\mu }}\ne {\varvec{\mu }}_0\). Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \). As a benchmark for this text, use \(\alpha = 0.05\) if \(\alpha \) is not given.

If \({\varvec{W}}\) is the data matrix, then \(R({\varvec{W}})\) is a large sample \(100(1 - \alpha )\%\) confidence region for \({\varvec{\mu }}\) if \(P[{\varvec{\mu }}\in R({\varvec{W}})] \rightarrow 1 - \alpha \) as \(n \rightarrow \infty .\) If \(\varvec{x}_1, ..., \varvec{x}_n\) are iid from a distribution with mean \({\varvec{\mu }}\) and nonsingular covariance matrix \({\varvec{\varSigma }}_{\varvec{x}}\), then

$$ R({\varvec{W}}) = \{ {\varvec{w}}| n(\overline{\varvec{x}} - {\varvec{w}})^T {\varvec{S}}^{-1} (\overline{\varvec{x}} - {\varvec{w}}) \le \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha } \} $$

is a large sample \(100(1 - \alpha )\%\) confidence region for \({\varvec{\mu }}\). This region is a hyperellipsoid centered at \(\overline{\varvec{x}}.\) Note that the estimated covariance matrix for \(\overline{\varvec{x}}\) is \({\varvec{S}}/n\) and \(n(\overline{\varvec{x}} - {\varvec{\mu }})^T {\varvec{S}}^{-1} (\overline{\varvec{x}} - {\varvec{\mu }}) = D^2_{{\varvec{\mu }}}(\overline{\varvec{x}},{\varvec{S}}/n).\) If \({\varvec{\mu }}\) is close to \(\overline{\varvec{x}}\) with respect to the Mahalanobis distance based on dispersion matrix \({\varvec{S}}/n\), then \({\varvec{\mu }}\) will be in the confidence region.

Recall from Theorem 1.1e that \(\displaystyle \max _{{\varvec{a}}\ne {\varvec{0}}} \frac{{\varvec{a}}^T (\overline{\varvec{x}} - {\varvec{\mu }})(\overline{\varvec{x}} - {\varvec{\mu }})^T {\varvec{a}}}{{\varvec{a}}^T {\varvec{S}}{\varvec{a}}} = n (\overline{\varvec{x}} - {\varvec{\mu }})^T {\varvec{S}}^{-1} (\overline{\varvec{x}} - {\varvec{\mu }}) = T^2\). This fact can be used to derive large sample simultaneous confidence intervals for \({\varvec{a}}^T {\varvec{\mu }}\) in that separate confidence statements using different choices of \({\varvec{a}}\) all hold simultaneously with probability tending to \(1 - \alpha \). Let \(\varvec{x}_1, ..., \varvec{x}_n\) be iid with mean \({\varvec{\mu }}\) and covariance matrix \({\varvec{\varSigma }}_{\varvec{x}} > 0\). Then simultaneously for all \({\varvec{a}}\ne {\varvec{0}}\), \(P(L_{{\varvec{a}}} \le {\varvec{a}}^T {\varvec{\mu }}\le U_{{\varvec{a}}})\) \(\rightarrow 1 - \alpha \) as \(n \rightarrow \infty \) where

$$[L_{{\varvec{a}}}, U_{{\varvec{a}}}] = {\varvec{a}}^T \overline{\varvec{x}} \pm \sqrt{\frac{p(n-1)}{n(n-p)} F_{p,n-p, 1-\alpha } {\varvec{a}}^T {\varvec{S}}{\varvec{a}}}.$$

Simultaneous confidence intervals (CIs) can be made after collecting data and hence are useful for “data snooping.” Following Johnson and Wichern (1988, pp. 184–5), the p confidence intervals (CIs) for \(\mu _i\) and the \(p(p-1)/2\) CIs for \(\mu _i - \mu _k\) can be made such that for each of the two types of CI, they all hold simultaneously with confidence \(\rightarrow 1 - \alpha \). Hence if \(\alpha = 0.05\), then in 100 samples, we expect all p CIs to contain \(\mu _i\) about 95 times, and we expect all \(p(p-1)/2\) CIs to contain \(\mu _i - \mu _k\) about 95 times. For each of the two types of CI, about 5 times at least one of the CIs will fail to contain its parameter (\(\mu _i\) or \(\mu _i - \mu _k\)). The simultaneous CIs for \(\mu _i\) are

$$[L, U] = \overline{x}_i \pm \sqrt{\frac{p(n-1)}{(n-p)} F_{p,n-p, 1-\alpha }} \sqrt{\frac{S_{ii}}{n}}$$

while the simultaneous CIs for \(\mu _i - \mu _k\) are

$$[L, U] = \overline{x}_i - \overline{x}_k \pm \sqrt{\frac{p(n-1)}{(n-p)} F_{p,n-p, 1-\alpha }} \sqrt{\frac{S_{ii} - 2S_{ik} + S_{kk}}{n}}.$$

Example 9.1.

Following Mardia et al. (1979, p. 126), data for first and second adult sons had \(n = 25\) and variables \(X_1\) = head length of first son and \(X_2\) = head length of second son. Suppose \({\varvec{\mu }}_0 = (182, 182)^T\) and \(T_C^2 = 1.28.\) Perform the one-sample Hotelling’s \(T^2\) test.

Solution: i) \(H_0: {\varvec{\mu }}= {\varvec{\mu }}_0\) \(H_1: {\varvec{\mu }}\ne {\varvec{\mu }}_0\)

ii) \(T_C^2 = 1.28\)

iii) \(\displaystyle \frac{n-p}{(n-1)p} T_C^2 = \frac{25-2}{(24(2)} 1.28 = 0.6133\), and pval = \(P(0.613 < F_{2,23}) > 0.05\)

iv) Fail to reject \(H_0\), so \({\varvec{\mu }}= (182, 182)^T.\)

9.1.1 A Diagnostic for the Hotelling’s \(T^2\) Test

Now the RMVN estimator is asymptotically equivalent to a scaled DGK estimator that uses \(k=5\) concentration steps and two “reweight for efficiency” steps. Lopuhaä (1999, pp. 1651–1652) showed that if (E1) holds, then the classical estimator applied to cases with \(D_i(\overline{\varvec{x}}, S) \le h\) is asymptotically normal with

$$\sqrt{n}(T_{0,D} - {\varvec{\mu }}) {\mathop {\rightarrow }\limits ^{D}} N_p({\varvec{0}}, \kappa _p {\varvec{\varSigma }}).$$

Here h is some fixed positive number, such as \(h = \chi ^2_{p, 0.975}\), so this estimator is not quite the DGK estimator after one concentration step.

We conjecture that a similar result holds after concentration:

$$\sqrt{n}(T_{RMVN} - {\varvec{\mu }}) {\mathop {\rightarrow }\limits ^{D}} N_p({\varvec{0}}, \tau _p {\varvec{\varSigma }})$$

for a wide variety of elliptically contoured distributions where \(\tau _p\) depends on both p and the underlying distribution. Since the “test” is based on a conjecture, it is ad hoc and should be used as an outlier diagnostic rather than for inference.

For MVN data, simulations suggest that \(\tau _p\) is close to 1. The ad hoc test that rejects \(H_0\) if

$$\frac{T_R^2}{f_{n, p}} = n(T_{RMVN} - {\varvec{\mu }}_0)^T \hat{{\varvec{C}}}_{RMVN}^{-1} (T_{RMVN} - {\varvec{\mu }}_0)/f_{n, p} > \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha }$$

where \(f_{n, p} = 1.04 + 0.12/p + (40+p)/n\) gave fair results in the simulations described later in this subsection for \(n \ge 15p\) and \(2 \le p \le 100\).

Table 9.1 Hotelling simulation

The correction factor \(f_{n, p}\) was found by simulating the “robust” and classical test statistics for 100 runs, plotting the test statistics, then finding a correction factor so that the identity line passed through the data. The following R commands were used to make Figure 9.1, which shows that the plotted points of the scaled “robust” test statistic versus the classical test statistic scatter about the identity line.

Table 9.2 Hotelling power simulation
figure a
Fig. 9.1
figure 1

Scaled “Robust” Statistic Versus \(T_H^2\) Statistic

For the Hotelling’s \(T_H^2\) simulation, the data is \(N_p(\delta {\varvec{1}}, diag(1,2,..., p))\) where \(H_0: {\varvec{\mu }}= {\varvec{0}}\) is being tested with 5000 runs at a nominal level of 0.05. In Table 9.1, \(\delta = 0\) so \(H_0\) is true, while hcv and rhcv are the proportion of rejections by the \(T_H^2\) test and by the ad hoc robust test. Sample sizes are \(n = 15p, 20p,\) and 30p. The robust test is not recommended for \(n < 15 p\) and appears to be conservative (the proportion of rejections is less than the nominal 0.05) except when \(n = 15 p\) and \(75 \le p \le 100.\) See Zhang (2011).

If \(\delta > 0\), then \(H_0\) is false and the proportion of rejections estimates the power of the test. Table 9.2 shows that \(T_H^2\) has more power than the robust test, but suggests that the power of both tests rapidly increases to one as \(\delta \) increases.

9.1.2 Bootstrapping Hotelling’s \(T^2\) Type Tests

The prediction region method of Section 5.3 is useful for bootstrapping the test \(H_0: {\varvec{\mu }}_T = {\varvec{\mu }}_0\) versus \(H_A: {\varvec{\mu }}_T \ne {\varvec{\mu }}_0\) where the test statistic T estimates the parameter \({\varvec{\mu }}_T\). Take a sample of size n with replacement from the cases \(\varvec{x}_1, ..., \varvec{x}_n\) to make the bootstrap statistic \(T^*_1\). Repeat to get the bootstrap sample \(T^*_1, ..., T^*_B\). Apply the nonparametric prediction region to the bootstrap sample and see if \({\varvec{\mu }}_0\) is in the region. Equivalently, apply the nonparametric prediction region to \({\varvec{w}}_i = T^*_i - {\varvec{\mu }}_0\), \(i = 1, ..., B\), and fail to reject \(H_0\) if \({\varvec{0}}\) is in the region, otherwise reject \(H_0\).

The mpack function rhotboot bootstraps T where T is the coordinatewise median or T is the RMVN location estimator. The function medhotsim simulates the test with \({\varvec{\mu }}_0 = {\varvec{0}}\) when T is the coordinatewise median. The simulated data are as in Section 6.3, with \(\varvec{x}= {\varvec{A}}{\varvec{z}}\), except that \({\varvec{z}}= {\varvec{u}}- {\varvec{1}}\) was used for the multivariate lognormal distribution with \(u_i = \exp (w_i)\) and \(w_i \sim N(0,1)\), so that the population coordinatewise median of \(\varvec{x}\) and \({\varvec{z}}\) was \({\varvec{0}}\) when \(H_0\) is true. When \(H_0\) was false, \({\varvec{\mu }}_0 = \delta {\varvec{1}}\) with \(\delta > 0\).

The term hotcov was the proportion of times the bootstrap test rejected \(H_0\) with a nominal level of 0.05. With \(n = 100\) and \(p = 2\), hotcov was near 0.05 when \(H_0\) was true. The test usually had good power if \({\varvec{\mu }}= (0.5,0.5)^T\). See output below where 1000 runs were used.

figure b

9.2 Matched Pairs

Assume that there are \(k = 2\) treatments, and both treatments are given to the same n cases or units. Then p measurements are taken for both treatments. For example, systolic and diastolic blood pressure could be compared before and after the patient (case) receives blood pressure medication. Then \(p = 2.\) Alternatively use n correlated pairs, for example, pairs of animals from the same litter or neighboring farm fields. Then use randomization to decide whether the first member of the pair gets treatment 1 or treatment 2. Let \(n_1 = n_2 = n\) and assume \(n - p\) is large.

Let \({\varvec{y}}_i = (Y_{i1}, Y_{i2}, ..., Y_{ip})^T\) denotes the p measurements from the 1st treatment, and \({\varvec{z}}_i = (Z_{i1}, Z_{i2}, ..., Z_{ip})^T\) denotes the p measurements from the 2nd treatment. Let \({\varvec{d}}_i \equiv \varvec{x}_i = {\varvec{y}}_i - {\varvec{z}}_i\) for \(i = 1, ..., n\). Assume that the \(\varvec{x}_i\) are iid with mean \({\varvec{\mu }}\) and covariance matrix \({\varvec{\varSigma }}_{\varvec{x}}\). Let \(T^2 = n (\overline{\varvec{x}} - {\varvec{\mu }})^T {\varvec{S}}^{-1} (\overline{\varvec{x}} - {\varvec{\mu }}).\) Then \(T^2 {\mathop {\rightarrow }\limits ^{P}} \chi ^2_p\) and \(p F_{p, n-p} {\mathop {\rightarrow }\limits ^{P}} \chi ^2_p.\) Let \(P(F_{p,n} \le F_{p, n,\delta }) = \delta .\) Then the one-sample Hotelling’s \(T^2\) inference is done on the differences \(\varvec{x}_i\) using \({\varvec{\mu }}_0 = {\varvec{0}}\). If the p random variables are continuous, make three DD plots: one for the \(\varvec{x}_i\), one for the \({\varvec{y}}_i\), and one for the \({\varvec{z}}_i\) to detect outliers.

Let pval be an estimate of the pvalue. The large sample multivariate matched pairs test has four steps.

i) State the hypotheses \(H_0: {\varvec{\mu }}= {\varvec{0}}\) \(H_1: {\varvec{\mu }}\ne {\varvec{0}}\).

ii) Find the test statistic \(T^2_M = n \overline{\varvec{x}}^T {\varvec{S}}^{-1} \overline{\varvec{x}}.\)

iii) Find pval =

$$P\left( T^2_M< \frac{(n-1)p}{n-p} F_{p, n-p}\right) = P\left( \frac{n-p}{(n-1)p} \ T^2_M < F_{p, n-p}\right) .$$

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that \({\varvec{\mu }}\ne {\varvec{0}}\) while if you fail to reject \(H_0\) conclude that the population mean \({\varvec{\mu }}= {\varvec{0}}\) or that there is not enough evidence to conclude that \({\varvec{\mu }}\ne {\varvec{0}}\). Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \). As a benchmark for this text, use \(\alpha = 0.05\) if \(\alpha \) is not given.

A large sample \(100(1 - \alpha )\%\) confidence region for \({\varvec{\mu }}\) is

$$\{ {\varvec{w}}| \ n(\overline{\varvec{x}} - {\varvec{w}})^T {\varvec{S}}^{-1} (\overline{\varvec{x}} - {\varvec{w}}) \le \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha } \},$$

and the p large sample simultaneous confidence intervals (CIs) for \(\mu _i\) are

$$[L, U] = \overline{x}_i \pm \sqrt{\frac{p(n-1)}{(n-p)} F_{p,n-p, 1-\alpha }} \sqrt{\frac{S_{ii}}{n}}$$

where \(S_{ii} = S_i^2\) is the ith diagonal element of \({\varvec{S}}\).

Example 9.2.

Following Johnson and Wichern (1988, pp. 213–214), wastewater from a sewage treatment plant was sent to two labs for measurements of biochemical demand (BOD) and suspended solids (SS). Suppose \(n = 11\), \(p = 2\), and \(T^2_M = 13.6.\) Perform the appropriate test.

Solution: i) \(H_0: {\varvec{\mu }}= {\varvec{0}}\) \(H_1: {\varvec{\mu }}\ne {\varvec{0}}\)

ii) \(T^2_M = 13.6\)

iii) \(\displaystyle \frac{n - p}{(n-1)p} T^2_M = \frac{11-2}{(11-1)2} 13.6 = 6.12\), and pval = \( P(6.12< F_{2,9}) < 0.05\)

iv) Reject \(H_0\). Hence \({\varvec{\mu }}\ne (0,0)^T\), and the two labs are giving different mean measurements for \((\mu _{BOD}, \mu _{SS})^T\).

To get a bootstrap analog of this test, bootstrap the \({\varvec{d}}_i = \varvec{x}_i\) as in Section 9.1.2 where usually \(H_0: {\varvec{\mu }}\equiv {\varvec{\mu }}_T = {\varvec{0}}.\) Again robust location estimators, such as the coordinatewise median or RMVN location estimator \(T_{RMVN}\), could be used on the \(\varvec{x}_i\).

9.3 Repeated Measurements

Repeated measurements \(=\) longitudinal data analysis. Take p measurements on the same unit, often the same measurement, e.g., blood pressure, at several time periods. Hence each unit or individual is measured repeatedly over time. The variables are \(X_1\), ..., \(X_p\) where often \(X_k\) is the measurement at the kth time period. Then \(E(\varvec{x}) = (\mu _1, ..., \mu _p)^T = (\mu + \tau _1, ..., \mu + \tau _p)^T.\) Let the \((p-1) \times 1\) vector \({\varvec{y}}_j = (x_{1j}-x_{2j}, x_{2j}-x_{3j}, ..., x_{p-1,j} - x_{pj})^T\) for \(j = 1, ..., n\). Hence \(y_{ij} = x_{ij} - x_{i+1,j}\) for \(j = 1, ..., n\) and \(i = 1, ..., p-1\). Then \(\overline{{\varvec{y}}} = (\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2, \overline{\varvec{x}}_2 - \overline{\varvec{x}}_3, ..., \overline{\varvec{x}}_{p-1} - \overline{\varvec{x}}_p)^T.\) If \({\varvec{\mu }}_{{\varvec{y}}} = E({\varvec{y}}_i)\), then \({\varvec{\mu }}_{{\varvec{y}}} = {\varvec{0}}\) is equivalent to \(\mu _1 = \cdots = \mu _p\) where \(E(X_k) = \mu _k\). Let \({\varvec{S}}_{{\varvec{y}}}\) be the sample covariance matrix of the \({\varvec{y}}_i\).

The large sample repeated measurements test has four steps.

i) State the hypotheses \(H_0: {\varvec{\mu }}_{{\varvec{y}}} = {\varvec{0}}\) \(H_1: {\varvec{\mu }}_{{\varvec{y}}} \ne {\varvec{0}}\).

ii) Find the test statistic \(T_R^2 = n \overline{{\varvec{y}}}^T {\varvec{S}}_{{\varvec{y}}}^{-1} \overline{{\varvec{y}}}.\)

iii) Find pval =

$$P\left( \frac{n-p+1}{(n-1)(p-1)} \ T_R^2 < F_{p-1,n-p+1}\right) .$$

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that \({\varvec{\mu }}_{{\varvec{y}}} \ne {\varvec{0}}\) so not all p of the \(\mu _i\) are equal, while if you fail to reject \(H_0\) conclude that the population mean \({\varvec{\mu }}_{{\varvec{y}}} = {\varvec{0}}\) or that there is not enough evidence to conclude that \({\varvec{\mu }}_{{\varvec{y}}} \ne {\varvec{0}}\). Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \). Give a nontechnical sentence, if possible.

Example 9.3.

Following Morrison (1967, pp. 139–141), reaction times to visual stimuli were obtained for \(n = 20\) normal young men under conditions A, B, and C of stimulus display. Let \(\overline{x}_A = 21.05, \overline{x}_B = 21.65,\) and \(\overline{x}_C = 28.95.\) Test whether \(\mu _A = \mu _B = \mu _C\) if \(T_R^2 = 882.8\).

Solution: i) \(H_0: {\varvec{\mu }}_{{\varvec{y}}} = {\varvec{0}}\) \(H_1: {\varvec{\mu }}_{{\varvec{y}}} \ne {\varvec{0}}\)

ii) \(T_R^2 = 882.8\)

iii) \(\displaystyle \frac{n-p+1}{(n-1)(p-1)} T_R^2 = \frac{20-3+1}{(20-1)(3-1)} 882.8 = 418.168\), and

pval = \(P(418.168 < F_{2,18}) \approx 0\)

iv) Reject \(H_0\). The three mean reaction times are different.

An alternative test would use a statistic T, such as the coordinatewise median or RMVN location estimator, on the \({\varvec{y}}_j\), and the bootstrap method of Section 9.1.2 can be applied with \({\varvec{\mu }}_{{\varvec{y}}} = {\varvec{0}}\). This test is equivalent to \(H_0: \mu _1 = \cdots = \mu _p\) where \(\mu _k\) is a population location parameter for the kth measurement. Hence if the coordiatewise median is being used, then \(\mu _k\) is the population median of the kth measurement.

9.4 Two Samples

Suppose there are two independent random samples \(\varvec{x}_{1,1}, ..., \varvec{x}_{n_1,1}\) and \(\varvec{x}_{1,2}, ..., \varvec{x}_{n_2,2}\) from populations with mean and covariance matrices \(({\varvec{\mu }}_i, {\varvec{\varSigma }}_{\varvec{x}_i})\) for \(i = 1,2\). Assume the \({\varvec{\varSigma }}_{\varvec{x}_i}\) are positive definite and that it is desired to test \(H_0: {\varvec{\mu }}_1 = {\varvec{\mu }}_2\) versus \(H_1: {\varvec{\mu }}_1 \ne {\varvec{\mu }}_2\) where the \({\varvec{\mu }}_i\) are \(p \times 1\) vectors. To simplify large sample theory, assume \(n_1 = k n_2\) for some positive real number k.

By the multivariate central limit theorem,

$$ \left( \begin{array}{c} \sqrt{n_1} \ (\overline{\varvec{x}}_1 - {\varvec{\mu }}_1) \\ \sqrt{n_2} \ (\overline{\varvec{x}}_2 - {\varvec{\mu }}_2) \end{array} \right) {\mathop {\rightarrow }\limits ^{D}} N_{2p}\left[ \left( \begin{array}{c} {\varvec{0}}\\ {\varvec{0}}\\ \end{array} \right) , \left( \begin{array}{cc} {\varvec{\varSigma }}_{\varvec{x}_1} &{} {\varvec{0}}\\ {\varvec{0}}&{} {\varvec{\varSigma }}_{\varvec{x}_2} \end{array} \right) \right] ,$$

or

$$\left( \begin{array}{c} \sqrt{n_2} \ (\overline{\varvec{x}}_1 - {\varvec{\mu }}_1) \\ \sqrt{n_2} \ (\overline{\varvec{x}}_2 - {\varvec{\mu }}_2) \end{array} \right) {\mathop {\rightarrow }\limits ^{D}} N_{2p}\left[ \left( \begin{array}{c} {\varvec{0}}\\ {\varvec{0}}\\ \end{array} \right) , \left( \begin{array}{cc} \frac{{\varvec{\varSigma }}_{\varvec{x}_1}}{k} &{} {\varvec{0}}\\ {\varvec{0}}&{} {\varvec{\varSigma }}_{\varvec{x}_2} \end{array} \right) \right] .$$

Hence

$$\sqrt{n_2} \ [(\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2) - ({\varvec{\mu }}_1 - {\varvec{\mu }}_2)] {\mathop {\rightarrow }\limits ^{D}} N_p({\varvec{0}}, \frac{{\varvec{\varSigma }}_{\varvec{x}_1}}{k} + {\varvec{\varSigma }}_{\varvec{x}_2}).$$

Using \(\displaystyle n {\varvec{B}}^{-1} = \left( \frac{{\varvec{B}}}{n}\right) ^{-1}\) and \(n_2 k = n_1\), if \({\varvec{\mu }}_1 = {\varvec{\mu }}_2,\) then

$$n_2 (\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2)^T \left( \frac{{\varvec{\varSigma }}_{\varvec{x}_1}}{k} + {\varvec{\varSigma }}_{\varvec{x}_2} \right) ^{-1} (\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2) =$$
$$(\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2)^T \left( \frac{{\varvec{\varSigma }}_{\varvec{x}_1}}{n_1} + \frac{{\varvec{\varSigma }}_{\varvec{x}_2}}{n_2} \right) ^{-1} (\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2) {\mathop {\rightarrow }\limits ^{D}} \chi ^2_p.$$

Hence

$$T^2_0 = (\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2)^T \left( \frac{{\varvec{S}}_1}{n_1} + \frac{{\varvec{S}}_2}{n_2} \right) ^{-1} (\overline{\varvec{x}}_1 - \overline{\varvec{x}}_2) {\mathop {\rightarrow }\limits ^{D}} \chi ^2_p.$$

The above result is easily generalized to other statistics. See Rupasinghe Arachchige Don and Pelawa Watagoda (2017). If the sequence of positive integers \(d_n \rightarrow \infty \) and \(Y_n \sim F_{p, d_n}\), then \(\displaystyle Y_n {\mathop {\rightarrow }\limits ^{D}} \chi ^2_p/p\). Using an \(F_{p, d_n}\) distribution instead of a \(\chi ^2_p\) distribution is similar to using a \(t_{d_n}\) distribution instead of a standard normal N(0, 1) distribution for inference. Instead of rejecting \(H_0\) when \(T^2_0 > \chi ^2_{p, 1 - \alpha },\) reject \(H_0\) when

$$T^2_0 > p F_{p, d_n, 1 - \alpha } = \frac{p F_{p, d_n, 1 - \alpha }}{\chi ^2_{p, 1 - \alpha }} \chi ^2_{p, 1 - \alpha }.$$

The term \(\displaystyle \frac{p F_{p, d_n, 1 - \alpha }}{\chi ^2_{p, 1 - \alpha }}\) can be regarded as a small sample correction factor that improves the test’s performance for small samples. We will use \(d_n = \min (n_1-p, n_2-p)\). Here \(P(Y_n \le \chi ^2_{p,\alpha }) = \alpha \) if \(Y_n\) has a \(\chi ^2_p\) distribution, and \(P(Y_n \le F_{p, d_n,\alpha }) = \alpha \) if \(Y_n\) has an \(F_{p, d_n}\) distribution.

Let pval denote the estimated pvalue. The four-step test is

i) State the hypotheses \(H_0: {\varvec{\mu }}_1 = {\varvec{\mu }}_2\) \(H_1: {\varvec{\mu }}_1 \ne {\varvec{\mu }}_2\).

ii) Find the test statistic \(t_0 = T^2_0/p\).

iii) Find pval = \(P(t_0 < F_{p, d_n})\).

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that the population means are not equal while if you fail to reject \(H_0\) conclude that the population means are equal or that there is not enough evidence to conclude that the population means differ. Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \). Give a nontechnical sentence if possible. As a benchmark for this text, use \(\alpha = 0.05\) if \(\alpha \) is not given.

Example 9.4.

Following Mardia et al. (1979, p. 153), cranial length and breadth (\(X_1\) and \(X_2\)) were measured on \(n_1 = 35\) female frogs and \(n_2 = 14\) male frogs with \(\overline{\varvec{x}}_1 = (22.86,24.397)^T\) and \(\overline{\varvec{x}}_2 =(21.821, 22.442)^T\). Test \({\varvec{\mu }}_1 = {\varvec{\mu }}_2\) if \(T_0^2 = 2.550\).

Solution: i) \(H_0: {\varvec{\mu }}_1 = {\varvec{\mu }}_2\) \(H_1: {\varvec{\mu }}_1 \ne {\varvec{\mu }}_2\)

ii) \(t_0 = T^2_0/p = 2.550/2 = 1.275\)

iii) pval = \(P(1.275 < F_{2,14-2}) > 0.05\)

iv) Fail to reject \(H_0\). There is not enough evidence to conclude that the mean lengths and breadths differ for the male and female frogs.

The plots for the one way MANOVA model in Section 10.2 are also useful for the two-sample Hotelling’s \(T^2\) test. An alternative to the above test is to used the pooled covariance matrix. This Hotelling’s \(T^2\) test is a special case of the one way MANOVA model with two groups covered in Section 10.3.

9.4.1 Bootstrapping Two-Sample Tests

Bootstrapping the two-sample test is similar to bootstrapping discriminant analysis and one way MANOVA models. Take a sample of size \(n_i\) with replacement from random sample i for \(i =1, 2\), and compute \(T^*_{11} - T^*_{21}\). Repeat B times to get the bootstrap sample \({\varvec{w}}_1 = T^*_{11} - T^*_{21}, ..., {\varvec{w}}_B = T^*_{1B} - T^*_{2B}.\) Apply the nonparametric prediction region on the \({\varvec{w}}_i\), and fail to reject \(H_0: {\varvec{\mu }}_1 = {\varvec{\mu }}_2\) if \({\varvec{0}}\) is in the prediction region, and reject \(H_0\), otherwise. See Rupasinghe Arachchige Don and Pelawa Watagoda (2017).

Some R output is below for the Gladstone (1905) data where several infants are outliers. We first tested the first 133 cases versus the last 134 cases. It turned out that the first group was younger and had all of the infants, so \(H_0\) was rejected. Then a random sample of 133 was used as the first group and the remaining 134 as the second group. Then the test failed to reject \(H_0\). Using the nominal level \(\alpha = 0.05\) of the large sample bootstrap test, reject \(H_0\) if the test statistic is larger than the cutoff, where 4.102 was the cutoff for the first test which used RMVN.

figure c

9.5 Summary

1) The one-sample Hotelling’s \(T^2\) test is used to test \(H_0: {\varvec{\mu }}= {\varvec{\mu }}_0\) versus \(H_A: {\varvec{\mu }}\ne {\varvec{\mu }}_0\). The test rejects \(H_0\) if \(\displaystyle T_H^2 = n(\overline{\varvec{x}} - {\varvec{\mu }}_0)^T {\varvec{S}}^{-1} (\overline{\varvec{x}} - {\varvec{\mu }}_0) > \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha }\) where \(P(Y \le F_{p, d,\alpha }) = \alpha \) if \(Y \sim F_{p, d}\).

If a multivariate location estimator T satisfies \(\displaystyle \sqrt{n}(T - {\varvec{\mu }}) {\mathop {\rightarrow }\limits ^{D}} N_p({\varvec{0}}, {\varvec{D}}),\) then a competing test rejects \(H_0\) if \(\displaystyle T_C^2 = n(T - {\varvec{\mu }}_0)^T \hat{{\varvec{D}}}^{-1} (T - {\varvec{\mu }}_0) > \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha }\) where \(\hat{{\varvec{D}}}\) is a consistent estimator of \({\varvec{D}}\). The scaled F cutoff can be used since \(T^2_C {\mathop {\rightarrow }\limits ^{D}} \chi ^2_p\) if \(H_0\) holds, and \(\displaystyle \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha } \rightarrow \chi ^2_{p, 1-\alpha }\) as \(n \rightarrow \infty .\)

2) Let pval be an estimate of the pvalue. As a benchmark for hypothesis testing, use \(\alpha = 0.05\) if \(\alpha \) is not given.

3) Typically, use \(T_C^2 = T_H^2\) in the following four-step one-sample Hotelling’s \(T_C^2\) test. i) State the hypotheses \(H_0: {\varvec{\mu }}= {\varvec{\mu }}_0\) \(H_1: {\varvec{\mu }}\ne {\varvec{\mu }}_0\).

ii) Find the test statistic \(T_C^2 = n(T - {\varvec{\mu }}_0)^T \hat{{\varvec{D}}}^{-1} (T - {\varvec{\mu }}_0).\)

iii) Find pval =

$$P\left( \frac{n-p}{(n-1)p} \ \ T_C^2 < F_{p, n-p}\right) .$$

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that \({\varvec{\mu }}\ne {\varvec{\mu }}_0\) while if you fail to reject \(H_0\) conclude that the population mean \({\varvec{\mu }}= {\varvec{\mu }}_0\) or that there is not enough evidence to conclude that \({\varvec{\mu }}\ne {\varvec{\mu }}_0\). Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \).

4) The multivariate matched pairs test is used when there are \(k = 2\) treatments applied to the same n cases with the same p variables used for each treatment. Let \({\varvec{y}}_i\) be the p variables measured for treatment 1 and \({\varvec{z}}_i\) be the p variables measured for treatment 2. Let \(\varvec{x}_i = {\varvec{y}}_i - {\varvec{z}}_i\). Let \({\varvec{\mu }}= E(\varvec{x}) = E({\varvec{y}}) - E({\varvec{z}})\). We want to test if \({\varvec{\mu }}= {\varvec{0}},\) so \(E({\varvec{y}}) = E({\varvec{z}}).\) The test can also be used if \(({\varvec{y}}_i, {\varvec{z}}_i)\) are matched (highly dependent) in some way. For example, if identical twins are in the study, \({\varvec{y}}_i\) and \({\varvec{z}}_i\) could be the measurements on each twin. Let \((\overline{\varvec{x}},{\varvec{S}}_{\varvec{x}})\) be the sample mean and covariance matrix of the \(\varvec{x}_i\).

5) The large sample multivariate matched pairs test has four steps.

i) State the hypotheses \(H_0: {\varvec{\mu }}= {\varvec{0}}\) \(H_1: {\varvec{\mu }}\ne {\varvec{0}}\).

ii) Find the test statistic \(T_M^2 = n \overline{\varvec{x}}^T {\varvec{S}}_{\varvec{x}}^{-1} \overline{\varvec{x}}.\)

iii) Find pval =

$$P\left( \frac{n-p}{(n-1)p} \ T_M^2 < F_{p, n-p}\right) .$$

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that \({\varvec{\mu }}\ne {\varvec{0}}\) while if you fail to reject \(H_0\) conclude that the population mean \({\varvec{\mu }}= {\varvec{0}}\) or that there is not enough evidence to conclude that \({\varvec{\mu }}\ne {\varvec{0}}\). Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \). Give a nontechnical sentence if possible.

6) Repeated measurements = longitudinal data analysis. Take p measurements on the same unit, often the same measurement, e.g., blood pressure, at several time periods. The variables are \(X_1\), ..., \(X_p\) where often \(X_k\) is the measurement at the kth time period. Then \(E(\varvec{x}) = (\mu _1, ..., \mu _p)^T = (\mu + \tau _1, ..., \mu + \tau _p)^T.\) Let \({\varvec{y}}_j = (x_{1j}-x_{2j}, x_{2j}-x_{3j}, ..., x_{p-1,j} - x_{pj})^T\) for \(j = 1, ..., n\). Then \(\overline{{\varvec{y}}} = (\overline{x}_1 - \overline{x}_2, \overline{x}_2 - \overline{x}_3, ..., \overline{x}_{p-1} - \overline{x}_p)^T.\) If \({\varvec{\mu }}_{{\varvec{y}}} = E({\varvec{y}}_i)\), then \({\varvec{\mu }}_Y = {\varvec{0}}\) is equivalent to \(\mu _1 = \cdots = \mu _p\) where \(E(X_k) = \mu _k\). Let \({\varvec{S}}_{{\varvec{y}}}\) be the sample covariance matrix of the \({\varvec{y}}_i\).

7) The large sample repeated measurements test has four steps.

i) State the hypotheses \(H_0: {\varvec{\mu }}_y = {\varvec{0}}\) \(H_1: {\varvec{\mu }}_y \ne {\varvec{0}}\).

ii) Find the test statistic \(T_R^2 = n \overline{{\varvec{y}}}^T {\varvec{S}}_y^{-1} \overline{{\varvec{y}}}.\)

iii) Find pval =

$$P\left( \frac{n-p+1}{(n-1)(p-1)} \ T_R^2 < F_{p-1,n-p+1}\right) .$$

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that \({\varvec{\mu }}_y \ne {\varvec{0}}\) while if you fail to reject \(H_0\) conclude that the population mean \({\varvec{\mu }}_y = {\varvec{0}}\) or that there is not enough evidence to conclude that \({\varvec{\mu }}_y \ne {\varvec{0}}\). Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \). Give a nontechnical sentence, if possible.

8) The F tables give left tail area and the pval is a right tail area. The Section 15.5 table gives \(F_{k,d, 0.95}.\) If \(\alpha = 0.05\) and \(\displaystyle \frac{n-p}{(n-1)p} \ \ T_C^2 < F_{k,d, 0.95}\), then fail to reject \(H_0\). If \(\displaystyle \frac{n-p}{(n-1)p} \ \ T_C^2 \ge F_{k,d, 0.95}\), then reject \(H_0\).

a) For the one-sample Hotelling’s \(T_C^2\) test and the matched pairs \(T_M^2\) test, \(k = p\) and \(d = n-p\).

b) For the repeated measures \(T_R^2\) test, \(k = p-1\) and \(d = n-p+1\).

9) If \(n \ge 10 p\), the tests in 3), 5), and 7) are robust to nonnormality. For the one-sample Hotelling’s \(T_C^2\) test and the repeated measurements test, make a DD plot. For the multivariate matched pairs test, make a DD plot of the \(\varvec{x}_i\), of the \({\varvec{y}}_i\), and of the \({\varvec{z}}_i.\)

10) Suppose there are two independent random samples \(\varvec{x}_{1,1}, ..., \varvec{x}_{n_1,1}\) and \(\varvec{x}_{1,2}, ..., \varvec{x}_{n_2,2}\) from populations with mean and covariance matrices \(({\varvec{\mu }}_i, {\varvec{\varSigma }}_{\varvec{x}_i})\) for \(i = 1,2\) where the \({\varvec{\mu }}_i\) are \(p \times 1\) vectors. Let \(d_n = \min (n_1 - p, n_2 - p).\) The large sample two-sample Hotelling’s \(T_0^2\) test is a four-step test:

i) State the hypotheses \(H_0: {\varvec{\mu }}_1 = {\varvec{\mu }}_2\) \(H_1: {\varvec{\mu }}_1 \ne {\varvec{\mu }}_2\).

ii) Find the test statistic \(t_0 = T^2_0/p\).

iii) Find pval = \(P(t_0 < F_{p, d_n})\).

iv) State whether you fail to reject \(H_0\) or reject \(H_0\). If you reject \(H_0\) then conclude that the population means are not equal while if you fail to reject \(H_0\) conclude that the population means are equal or that there is not enough evidence to conclude that the population means differ. Reject \(H_0\) if pval \(\le \alpha \) and fail to reject \(H_0\) if pval \(> \alpha \). Give a nontechnical sentence if possible.

9.6 Complements

The mpack function rhotsim is useful for simulating the robust diagnostic for the one-sample Hotelling’s \(T^2\) test. See Zhang (2011) for more simulations. Willems et al. (2002) used similar reasoning to present a diagnostic based on the FMCD estimator.

Yao (1965) suggested a more complicated denominator degrees of freedom than \(d_n = \min (n_1 - p, n_2 - p)\) for the two-sample Hotelling’s \(T^2\) test. Good (2012, pp. 55–57), which provides randomization tests as competitors for the two-sample Hotelling’s \(T^2\) test. Bootstrapping the tests with robust estimators seems to be effective. For bootstrapping the two-sample Hotelling’s \(T^2\) test, see Rupasinghe Arachchige Don and Pelawa Watagoda (2017). Gregory et al. (2015) and Feng and Sun (2015) considered the two-sample test when \(p \ge n\).

9.7 Problems

PROBLEMS WITH AN ASTERISK * ARE ESPECIALLY USEFUL.

9.1. Following Morrison (1967, pp. 122–123), the Wechsler Adult Intelligence Scale scores of \(n = 101\) subjects aged 60 to 64 were recorded, giving a verbal score (\(X_1\)) and performance score \((X_2)\) for each subject. Suppose \({\varvec{\mu }}_0 = (60, 50)^T\) and \(T_C^2 = 357.43.\) Perform the one-sample Hotelling’s \(T^2\) test.

9.2. Following Morrison (1967, pp. 137–138), the levels of free fatty acid (FFA) in the blood were measured in \(n = 15\) hypnotized normal volunteers who had been asked to experience fear, depression, and anger effects while in the hypnotic state. The mean FFA changes were \(\overline{x}_1 = 2.669,\) \(\overline{x}_2 = 2.178\), and \(\overline{x}_3 = 2.558\). Let \(\mu _F = \mu + \tau _1,\) \(\mu _D = \mu + \tau _2\), and \(\mu _A = \mu + \tau _3\). We want to know if the mean stress FFA changes were equal. So test whether \(\mu _F = \mu _D = \mu _F\) if \(T_R^2 = 2.68.\)

9.3. Data is taken or modified from Johnson and Wichern (1988, pp. 185, 224).

a) Suppose \(S_2^2 = S_{22} = 126.05,\) \(\overline{x}_2 = 54.69,\) \(n = 87,\) and \(p = 3\). Find a large sample simultaneous 95% CI for \(\mu _2\).

b) Suppose a random sample of 50 bars of soap from method 1 and a random sample of 50 bars of soap from method 2 are obtained. Let \(X_1 =\) lather and \(X_2 =\) mildness with \(\overline{\varvec{x}}_1 = (8.4,4.1)^T\) and \(\overline{\varvec{x}}_2 =(10.2, 3.9)^T\). Test \({\varvec{\mu }}_1 = {\varvec{\mu }}_2\) if \(T_0^2 = 52.4722\).

R Problems

Warning: Use the command source(“G:/mpack.txt”) to download the programs. See Preface or Section 15.2 Typing the name of the mpack function, e.g., rhotsim, will display the code for the function. Use the args command, e.g., args(rhotsim), to display the needed arguments for the function. For some of the following problems, the R commands can be copied and pasted from (http://lagrange.math.siu.edu/Olive/mrsashw.txt) into R.

9.4 \(^*\). Use the R commands in Subsection 9.1.1 to make a plot similar to Figure 9.1. The program may take a minute to run.

9.5. Conjecture:

$$\sqrt{n}(T_{RMVN} - {\varvec{\mu }}) {\mathop {\rightarrow }\limits ^{D}} N_p({\varvec{0}}, \tau _p {\varvec{\varSigma }})$$

for a wide variety of elliptically contoured distributions where \(\tau _p\) depends on both p and the underlying distribution. The following “test” is based on a conjecture and should be used as an outlier diagnostic rather than for inference. The ad hoc “test” that rejects \(H_0\) if

$$\frac{T_R^2}{f_{n, p}} = n(T_{RMVN} - {\varvec{\mu }}_0)^T \hat{{\varvec{C}}}_{RMVN}^{-1} (T_{RMVN} - {\varvec{\mu }}_0)/f_{n, p} > \frac{(n-1)p}{n-p} F_{p,n-p, 1-\alpha }$$

where \(f_{n, p} = 1.04 + 0.12/p + (40+p)/n\). The simulations use \(n = 150\) and \(p = 10\).

a) The R commands for this part use simulated data is

$$\varvec{x}_i \sim N_p({\varvec{0}}, diag(1,2,..., p))$$

where \(H_0: {\varvec{\mu }}= {\varvec{0}}\) is being tested with 5000 runs at a nominal level of 0.05. So \(H_0\) is true, and hcv and rhcv are the proportion of rejections by the \(T_H^2\) test and by the ad hoc robust test. We want hcv and rhcv near 0.05. THIS SIMULATION MAY TAKE A FEW MINUTES. Record hcv and rhcv. Were hcv and rhcv near 0.05?

b) The R commands for this part use simulated data

$$\varvec{x}_i \sim N_p(\delta {\varvec{1}}, diag(1,2,..., p))$$

where \(H_0: {\varvec{\mu }}= {\varvec{0}}\) is being tested with 5000 runs at a nominal level of 0.05. In the simulation, \(\delta = 0.2\), so \(H_0\) is false, and hcv and rhcv are the proportion of rejections by the \(T_H^2\) test and by the ad hoc robust test. We want hcv and rhcv near 1 so that the power is high. Paste the output into Word. THIS SIMULATION MAY TAKE A FEW MINUTES. Record hcv and rhcv. Were hcv and rhcv near 1?