Keywords

1 Introduction

The inverse Gaussian distribution is used to describe and analyze positive and right-skewed data and has been applied in useful applications in a variety of fields, such as cardiology, pharmacokinetics, economics, medicine, and finance. The inverse Gaussian distribution was first derived by Schordinger [1] for the first passage of time of a Wiener process to an absorbing barrier and has been used to describe the cycle time distribution of particles in the blood [2]. Liu et al. [3] used data sets from an inverse Gaussian distribution applied to a lifetime model for reliability analysis. Banerjee and Bhattacharyya [4] applied this distribution in a study of market incidence models, while Lancaster [5] used it as a model for the duration of strikes, and Sheppard [6] proposed applying it for the time duration of injected labeled substances called tracers in a biological system. For more informations and applications, Chhikara and Folks [7], Krishnamoorthy and Lu [8], Tian and Wu [9], Lin et al. [10], and Ye et al. [11].

The coefficient of variation, which is the ratio of the standard deviation to the mean, can be used to measure data dispersion with non-homogeneous units. It has been used in many fields, such as biology, economics, medicine, agriculture, and finance. Many researchers have proposed confidence intervals for the coefficient of variation. Wongkhao et al. [12] presented confidence intervals for the ratio of two independent coefficients of variation of normal distributions based on the concept of the general confidence interval (GCI) and the method of variance estimates recovery (MOVER). Banik and Kibria [13] estimated the population coefficient of variation and compared it with bootstrap interval estimators. Mahmoudvand and Hassani [14] proposed an unbiased estimator to construct confidence intervals for the population coefficient of variation of normal distribution. Hasan and Krishnamoorthy [15] proposed two new approximate confidence intervals for the ratio of the coefficients of variation of lognormal populations using MOVER and the fiducial confidence intervals method. Sangnawakij et al. [16] proposed two confidence intervals for the ratio of coefficients of variation of gamma distributions based on MOVER with the methods of Score and Wald. Thangjai and Niwitpong [17] constructed confidence intervals for the weighted coefficient of constructed of two-parameter exponential distributions based on the adjusted method of variance estimates recovery method (adjusted MOVER) and compared it with the general confidence interval (GCI) and the large sample method. Yosboonruang et al. [18] proposed confidence intervals for the coefficient of variation for a delta-lognormal distribution based on GCI and the modified Fletcher method.

As mentioned previously, there have been many comprehensive studies on confidence intervals for the difference between the coefficients of variation of inverse Gaussian distributions. However, few researchers have investigated confidence intervals for the parameters of inverse Gaussian distributions. For instance, Ye et al. [11] proposed confidence intervals for the common mean of several inverse Gaussian populations when the scalar parameters are unknown and unequal. Tian and Wilding [19] presented confidence intervals for the ratio of the means of two independent inverse Gaussian distribution. Krishnamoorthy and Tian [20] developed confidence intervals for the difference between and ratio of the means of two inverse Gaussian distributions based on GCI. The purpose of the current study is to establish new confidence intervals for the difference between the coefficients of variation of inverse Gaussian distributions based on GCI, the adjusted generalized confidence interval (AGCI), the bootstrap percentile confidence interval (BPCI), and MOVER.

The organization of this paper are as follows. Section 2 provides preliminaries for the difference between the coefficients of variation of inverse Gaussian distributions. Simulation studies are presented in Sect. 3. Section 4 presents empirical studies. Finally, concluding remarks are summarized in Sect. 5.

2 Confidence Intervals for the Difference Between the Coefficients of Variation of Inverse Gaussian Distributions

Let X = (\(X_{1}\), \(X_{2}\),..., \(X_{n}\)) be an independent random sample of size n from the two-parameter inverse Gaussian distribution, IG(\(\mu \), \(\lambda \)), is defined as

$$\begin{aligned} f(x,\mu ,\lambda )=\left( \frac{\lambda }{2\pi x^{3}}\right) ^{\frac{1}{2}}exp\left\{ -\frac{\lambda \left( x-\mu \right) ^{2}}{2\mu ^{^{2}}x}\right\} ,x>0,\mu>0,\lambda >0, \end{aligned}$$
(1)

where \(\mu \) and \(\lambda \) are the mean parameter and the scale parameter. The population mean and variance of X are define as

$$\begin{aligned} E(X)=\mu \end{aligned}$$
(2)

and

$$\begin{aligned} Var(X)=\mu ^{3}/\lambda . \end{aligned}$$
(3)

Then, the coefficient of variation of X is expressed by

$$\begin{aligned} \omega = CV(X) = \sqrt{\frac{\mu }{\lambda }} \end{aligned}$$
(4)

and the difference between of coefficients of variation is defined as

(5)

2.1 The Generalized Confidence Interval (GCI) Method

Weerahandi [21] introduced the generalized confidence interval (GCI) method. The concept of this method is to the definition of generalized pivotal quantity (GPQ). Suppose that X is a random sample from a distribution having the parameter \((\theta ,\delta )\), where \(\theta \) is the parameter of interest and \(\delta \) is the nuisance parameter.

Definition 1

Let \(X=(X_{1},X_{2},...,X_{n})\) be the observed value of X and the probability density function of \((X;x,\theta ,\delta )\) is \(R(X;x,\theta ,\delta )\). The generalized pivotal quantity \(R(X;x,\theta ,\delta )\) satisfies the following two properties:

  1. (a)

    The probability distribution of the function \(R(X;x,\theta ,\delta )\) is independent of unknown parameters.

  2. (b)

    The observed value of \(R(X;x,\theta ,\delta )\), \(X=x\), does not depend on nuisance parameters.

Therefore, the \(100\left( 1-\alpha \right) \%\) two-sided GCI for the parameter of interest is defined as \([R(\alpha /2),R(1-\alpha /2)]\), where \(R(\alpha /2)\) and \(R(1-\alpha /2)\) are the \(100(\alpha /2)-th\) and \(100(1-\alpha /2)-th\) percentile of \(R(X;x,\theta , \delta )\).

Ye et al. [11] developed the generalized confidence interval based on general pivotal quantities for the mean and scale parameter of the inverse Gaussian distribution. Suppose that k is independent populations of the inverse Gaussian distributions with mean parameters \(\mu _{i}\) and scale parameters \(\lambda _{i}\), \(i=1,2,..,k.\) Let \(X_{i1},X_{i2},...,X_{in_{i}}\) be the random sample from \(IG(\mu _{i}, \lambda _{i})\), \(i=1,...,k.\) Form the ith population, the maximum likelihood estimators (MLEs) of \(\mu _{i}\) and \(\lambda _{i}\) can be found as

$$\begin{aligned} \hat{\mu }=\bar{X}_{i}, \qquad \hat{\lambda }_{i}^{-1}=\frac{1}{n_{i}}\sum _{j=1}^{n_{i}}(X_{ij}^{-1}-\bar{X_{i}}^{-1}), \end{aligned}$$
(6)

where \(\bar{X_{_{i}}}=\sum _{j=1}^{n_{i}}X_{ij}/n_{i}\). Let \(V_{i} = \hat{\lambda }_{i}^{-1}\), it is well known that \(\bar{X}_{i}\) and \(V_{i}\) are mutually independent random variables. Note that

$$\begin{aligned} \bar{X}_{i}\sim IG(\mu ,n_{i}\lambda _{i}), \qquad n_{i}\lambda _{i}V_{i}\sim \chi _{n_{i-1}}^{2}, i=1,..,k, \end{aligned}$$
(7)

where \(\chi _{m}^{2}\) denotes as the Chi-square distribution with m degrees of freedom. It is easily proved that \((\bar{X_{i}},V_{i})\) forms a set complete sufficient statistics for \((\mu _{i},\lambda _{i})\). From Eq. (7), the generalized pivotal quantity \(R_{\lambda _{i}}\) for \(\lambda _{i}\) based on the ith sample is defined as

$$\begin{aligned} R_{\lambda _{i}}=\frac{n_{i}\lambda _{i}V_{i}}{n_{i}\upsilon _{i}} \sim \frac{\chi _{n_{i-1}}^{2}}{n_{i}\upsilon _{i}} , i=1,...,k, \end{aligned}$$
(8)

where \(v_{i}\) denotes as the observed value of \(V_{i}\). The distribution of \(R_{\lambda _{i}}\) is free of any unknown parameters and the observed values only relate to the parameter \(\lambda _{i}\). Therefore, \(R_{\lambda _{i}}\) is the generalized pivotal quantity for\(\lambda _{i}\).

According to Ye et al. [11], the generalized pivotal quantity \(R_{\mu _{i}}\) for \(\mu _{i}\) based on the ith sample is defined as

(9)

where \(\sim \) denotes as “approximately distributed” and \(Z_{i}\sim N(0,1)\). \(\bar{x_{i}}\) and \(v_{i}\) are the observed values of \(\bar{X_{i}}\) and \(V_{i}\). The approximations in Eq. (9) are derived by using Theorem 2.1 given by Chhikara and Folks [7]. Using the moment matching method, it can be shown that\(\sqrt{n_{i}\lambda _{i}}(\bar{X_{i}}-\mu )/\mu \sqrt{x_{i}}\) has a limiting distribution of \(Z_{i}\sim N(0,1)\). Note that the observed value of \(R_{\mu _{i}}\) is \(\mu _{i}\). Therefore, \(R_{\mu _{i}}\) satisfies conditions (a) and (b) in Definition 1. However, \(R_{\mu _{i}}\) is an approximate generalized pivotal quantity for \(\mu _{i}\) based on the ith sample.

Therefore, the generalized pivotal quantities for difference between the coefficients of variation are given by

(10)

Then, the \(100(1-\alpha )\%\) two-sided confident interval for the difference between the coefficients of variation based on generalized confidence interval method is given by

$$\begin{aligned} CI_{(GCI)}=(L_{(GCI)},U_{(GCI)})=(R_{\eta }(\alpha /2),R_{\eta }(1-\alpha /2)), \end{aligned}$$
(11)

where \(R_{\eta }(\alpha /2)\) and \(R_{\eta }(1-\alpha /2)\) are the \(100(\alpha /2)\%\) and \(100(1-\alpha /2)\%\) percentiles of the distribution of \(R=R(X;x,\theta ,\delta )\), respectively.

The following algorithm is used to construct the generalized confidence interval:

Algorithm 1.

Step 1. Generate \(x_{j}\) and \(y_{j}\) \(,j=1,2,...,n_{i}\) from the inverse Gaussian distribution.

Step 2. Compute \(\bar{x}_{j}\), \(\bar{y_{j}}\), \(\hat{v}_{x_{j}}\) and \(\hat{v}_{y_{j}}\).

Step 3. For t = 1 to T.

Step 4. Generate \(\chi _{n_{i}-1}^{2}\) from chi-square distribution and \(Z\sim N(0,1)\).

Step 5. Compute \(R_{\lambda _{i}}\) from Eq. (8).

Step 6. Compute \(R_{\mu _{i}}\) from Eq. (9).

Step 7. Compute \(R_{\eta }\) from Eq. (10).

Step 8. End t loop.

Step 9. Compute \(R_{\eta }(\alpha /2)\) and \(R_{\eta }(1-\alpha /2)\).

2.2 The Adjusted Generalized Confidence Interval (AGCI) Method

According to Ye et al. [11], We can use a similar method in GCI method for calculating the difference between the coefficients of variation \(\eta \). \(R_{\tilde{\lambda _{i}}}\) can be computed by the generalized pivotal quantity for \(\tilde{\lambda }\) based on ith sample shown in Eq. (8). Then, Krishnomoorthy and Tian [20] presented an approximate generalized pivotal quantity for \(R_{\tilde{\mu _{i}}}\) based on the ith sample, as follows:

$$\begin{aligned} R_{\tilde{\mu _{i}}}=\frac{\bar{x}_{i}}{ \text {max}\left\{ 0,t_{n_{i}-1}\sqrt{\frac{\bar{x_{i}}v_{i}}{n_{i}-1}} \right\} }, \end{aligned}$$
(12)

where \(t_{n_{i}-1}\) denotes the t distribution with \({n_{i}-1}\) degrees of freedom. Therefore, the denominator in Eq. (12) may be zero when \(t_{n_{i}-1}\) obtains the negative value. \( R_{\tilde{\mu _{i}}}\) is an approximate generalized pivotal quantity.

Therefore, the generalized pivotal quantities for the difference between the coefficients of variation are given by

$$\begin{aligned} R_{\tilde{\eta }} =R_{{\tilde{\omega }}_{X}}-R_{{\tilde{\omega }}_{Y}}=\sqrt{\frac{R_{\tilde{\mu }_{X}}}{R_{\tilde{\lambda }_{X}}}}-\sqrt{\frac{R_{\tilde{\mu }_{Y}}}{R_{\tilde{\lambda }_{Y}}}}. \end{aligned}$$
(13)

Then, the \(100(1-\alpha )\%\) two-sided confident interval for the difference between the coefficients of variation based on generalized confidence interval method is given by

$$\begin{aligned} CI_{(AGCI)}=(L_{(AGCI)},U_{(AGCI)})=(R_{\tilde{\eta }}(\alpha /2),{R_{\tilde{\eta }} }(1-\alpha /2)), \end{aligned}$$
(14)

where \(R_{\tilde{\eta }}(\alpha /2)\) and \(R_{\tilde{\eta } }(1-\alpha /2)\) which are the \(100(\alpha /2 )\%\) and \(100(1-\alpha /2 )\%\) percentiles of the distribution of \(R=R(X;x,\theta ,\delta )\) can be obtained from the concept of Algorithm 1.

2.3 The Bootstrap Percentile Confidence Interval (BPCI) Method

Efron and Tibshirani [22] introduced the bootstrap percentile method. The bootstrap is a re-sampling method for assigning measures of accuracy to statistical estimate a random selection of resamples from the original sample with replacement. Let x be a random sample of size n from the inverse Gaussian distribution. Suppose that \(x=x_{1},x_{2},...,x_{n}\) is a random sample of size n from the inverse Gaussian distribution. Sampling is replaced by \(x^{*}=x_{1}^{*},x_{2}^{*},...,x_{n}^{*}\), which can be obtained by the bootstrap sample with B times. When the re-sampling bootstrap sample is operated, the difference between coefficients of variation is then calculated.

The \(100(1-\alpha )\%\) two-sided confidence interval for the difference between the coefficients of variation based on the bootstrap percentile confidence interval is defined by

$$\begin{aligned} CI_{(BPCI)}=(L_{(BPCI)},U_{(BPCI)})=(\eta ^{*}(\alpha /2),\eta ^{*}(1-\alpha /2)), \end{aligned}$$
(15)

where \(\eta ^{*}(\alpha /2)\) and \(\eta ^{*}(1-\alpha /2)\) are the \(100(\alpha /2 )\%\) and \(100(1-\alpha /2 )\%\) percentiles of the distribution.

Algorithm 2.

Step 1. Generate\( X_{1},X_{2},...X_{n}\) from the inverse Gaussian distribution

Step 2. Obtain a bootstrap sample \(X^{*}=X_{1}^{*},X_{2}^{*},...,X_{n}^{*}\) from Step 1.

Step 3. Compute \(\eta ^{*}\)

Step 4. Repeat Steps 2 and 3, B times.

Step 5. Compute \(\eta ^{*}(\alpha /2)\) and \(\eta ^{*}(1-\alpha /2)\)

2.4 The Method of Variance Estimates Recovery (MOVER)

Gulhar et al. [23] proposed the confidence interval for a coefficient of variation of X which is

$$\begin{aligned} (l_{x},u_{x})=\left( \frac{\sqrt{n-1}(\hat{\omega _{x}})}{\sqrt{\chi _{1-\alpha /2 ,n-1}^{2}}},\frac{\sqrt{n-1}(\hat{\omega _{x}})}{\sqrt{\chi _{\alpha /2 ,n-1}^{2}}} \right) \end{aligned}$$
(16)

and the confidence interval for a coefficient of variation of Y is defined as

$$\begin{aligned} (l_{y},u_{y})=\left( \frac{\sqrt{n-1}(\hat{\omega _{y}})}{\sqrt{\chi _{1-\alpha /2 ,n-1}^{2}}},\frac{\sqrt{n-1}(\hat{\omega _{y}})}{\sqrt{\chi _{\alpha /2 ,n-1}^{2}}} \right) , \end{aligned}$$
(17)

where \(\chi _{1-\alpha /2,n-1}^{2}\) and \(\chi _{\alpha /2,n-1}^{2}\) are respectively the \(100(\alpha )\%\) -th and \(100(1-\alpha )\%\) -th percentile of the chi-square distribution with \(n-1\) degrees of freedom.

Donner and Zou [24] introduced the confidence interval estimation for the difference of parameters of interest using by MOVER. The lower limit and upper limit are given by

$$\begin{aligned} L_{\eta }=\hat{\omega }_{x}-\hat{\omega }_{y}-\sqrt{(\hat{\omega }_{x}-l_{x})^{2}+(u_{y}-\hat{\omega }_{y})^{2}} \end{aligned}$$
(18)

and

$$\begin{aligned} U_{\eta }=\hat{\omega }_{x}-\hat{\omega }_{y}-\sqrt{(u_{x}-\hat{\omega }_{x})^{2}+(\hat{\omega }_{y}-l_{y})^{2}} , \end{aligned}$$
(19)

where \(\hat{\omega }_{x}\) and \(\hat{\omega }_{y}\) are denoted in Eq. (5), \(l_{x}\) and \(u_{x}\) are denoted in Eq. (16), and \(l_{y}\) and \(u_{y}\) are denoted in Eq. (17).

Then, the \(100(1-\alpha )\%\) two-sides confidence interval for the difference between the coefficients of variation of inverse Gaussian distribution based on the MOVER is given by

$$\begin{aligned} CI_{MOVER}=\left( L_{MOVER},U_{MOVER} \right) =\left( L_{\eta },U_{\eta } \right) , \end{aligned}$$
(20)

where \(L_{\eta }\) and \(U_{\eta }\) are defined in Eqs. (18) and (19), respectively.

3 Simulation Studies

A Monte Carlo simulation studies out to evaluate the coverage probabilities and average lengths of the confidence intervals for the difference between the coefficients of variation of inverse Gaussian distributions based on GCI, AGCI, BPCI, and MOVER. The simulations were run by using R statistics programming language. In the simulation, The sample sizes were \((n_{x},n_{y})\) = (5, 5), (5, 10), (10, 10), (10, 30), (30, 30), (30, 50), (50, 50), (50, 100), and (100, 100); \(\mu _{x}=0.5\), \(\mu _{y}=0.5,1\), \(\lambda _{x}=10\) and \(\lambda _{y}=1,2,5,10\). The nominal confidence level is at 0.95. The number of simulation replications for each situation was 10,000 replications, 1,000 bootstrap samples and 5,000 pivotal quantities for GCI and AGCI. The confidence interval which has the coverage probability was greater than or close to the nominal confidence level and the shortest expected lengths are chosen.

The estimated coverage probability and estimated average length for this simulation study are respectively given as:

$$\begin{aligned} CP=\frac{c(L_{(g)}\le \eta \le U_{(g)})}{M}, \qquad AL =\frac{\sum _{i=1}^{m}(U_{(g)}-L_{(g)})}{M}, \end{aligned}$$
(21)

where \(\textit{M}\) is the number of simulation replications, and \(c(L_{(g)}\le \eta \le U_{(g)})\) is the numbers of simulation replications for \(\eta \) which lies within the confidence interval.

The following algorithm is used to construct the coverage probability for the difference between the coefficients of variation:

Algorithm 3.

Step 1. For a given \(M,m,n_{1},n_{2},\mu _{x},\mu _{y},\lambda _{x},\) and \(\lambda _{y}\).

Step 2. For g = 1 to M.

Step 3. Generate \( X_{1},X_{2},...,X_{n}\) and \(Y_{1},Y_{2},...,Y_{n}\) from inverse Gaussian distribution.

Step 4. Use Algorithm 1, Eq. (14), Algorithm 2, and Eq. (20) to construct lower and upper limits for GCI, AGCI, BPCI, and MOVER, respectively.

Step 5. If \((L_{(g)}\le \eta \le U_{(g)})\), set \(P_{(g)}=1\); else set \(P_{(g)}=0\).

Step 6. Calculate \(U_{(g)}-L_{(g)}\).

Step 7. End g loop.

Step 8: Compute the coverage probability and the average length.

The results in Table 1 showed that the difference between the coefficients of variation, the coverage probabilities of GCI method and AGCI method were greater than or equal to the nominal level in all cases. However, the average lengths of the AGCI method were shorter than the GCI method. For small sample sizes, the coverage probabilities of the MOVER were close to the nominal confidence level of 0.95 and the shortest average length.

Table 1. The coverage probability (CP) and average length (AL) of 95% two-sided confidence intervals for the difference between the coefficients of variation of inverse Gaussian distribution: \((\mu _{X}, \lambda _{x})=(0.5, 10)\).

4 An Empirical Study

Herein, we illustrate the methods used to computation of confidence intervals proposed.

Example 1:

The data were given by Mudholkar and Hutson [25] for the consecutive annual flood discharge rates of the Floyd river at James, Iowa. The data are as follows:

In 1935−1944 : 1460, 4050, 3570, 2060, 1300, 1390, 1720, 6280, 1360, 7440.

In 1945−1954 : 5320, 1400, 3240, 2710, 4520, 4840, 8320, 13900, 71500, 6250.

The summary statistics of data are \(n_{1}=10\), \(n_{2}=10\), \(\hat{\mu }_{1}=3063\), \(\hat{\mu }_{2}=12200\), \(\hat{\lambda }_{1}=6529.078\), \(\hat{\lambda }_{2}=6434.176\), \(\hat{\omega }_{1} = 0.6849\), \(\hat{\omega }_{2}= 1.3770\), and the difference between the coefficients of variation \(\hat{\eta } = -0.6921\). Based on the 95% two sided confidence interval for the difference between the coefficients of variation using GCI method was (−0.0051, 0.0148) with interval length of 0.0199; AGCI method was (−0.0045, 0.0144) with interval length of 0.0189; BPCI method was (−1.4848, 0.3342) with interval length of 1.8189, and MOVER was (−1.8489, 0.0182) with interval length of 1.8667. Therefore, the results confirm that the simulation results for difference between the coefficients of variation are not different from the results of the previous study.

Example 2:

The real data were provided by Eilam et al. [26]. The plasma bradykininogen levels were measured in healthy subjects, in patients with active Hodgkin’s disease and in patients with inactive Hodgkin’s disease. The outcome variable is measured in micrograms of bradykininogen per milliliter of plasma. The data are as follows:

Active Hodgkin’s disease: 3.96, 3.04, 5.28, 3.40, 4.10, 3.61, 6.16, 3.22, 7.48, 3.87, 4.27, 4.05, 2.40, 5.81, 4.29, 2.77, 4.40.

Inactive Hodgkin’s disease: 5.37, 10.60, 5.02, 14.30, 9.90, 4.27, 5.75, 5.03, 5.74, 7.85, 6.82, 7.90, 8.36, 5.72, 6.00, 4.75, 5.83, 7.30, 7.52, 5.32, 6.05, 5.68, 7.57, 5.68, 8.91, 5.39, 4.40, 7.13.

The summary statistics of data are \(n_{1}=17\), \(n_{2}=28\), \(\hat{\mu }_{1}=4.2418\), \(\hat{\mu }_{2}=6.7914\), \(\hat{\lambda }_{1}=50.8394\), \(\hat{\lambda }_{2}=86.0289\), \(\hat{\omega }_{1}=0.2889\), \(\hat{\omega }_{2}=0.2810\), and the difference between the coefficients of variation \(\hat{\eta }=0.0079\). Based on the 95% two sided confidence interval for the difference between the coefficients of variation, the results showed that GCI method \(CI_{GCI}=(0.0188,0.1497)\) with interval length of 0.1309, AGCI method \(CI_{AGCI}=(0.0213,0.1447)\) with interval length of 0.1234, BPCI method \(CI_{BPCI}=(-0.1245,0.1274)\) with interval length of 0.2519 and MOVER method \(CI_{MOVER}=(0.1697,0.2873)\) with interval length of 0.2873. Therefore, the results from above examples support our simulation results.

5 Conclusions

The new confidence intervals for the difference between the coefficients of variation of inverse Gaussian distributions based on GCI, AGCI, BPCI, and MOVER were presented. The performances of these confidence intervals were assessed in terms of their coverage probabilities and the average lengths. The results obtained from the GCI and the AGCI methods were satisfactory in all cases. However, the AGCI method was better than the GCI method in terms of the average length. For small sample sizes, MOVER provided the shortest average length. Meanwhile, based on the findings of this study, BPCI is not recommended because the coverage probabilities were under the nominal confidence level in all cases.