1 Introduction

One of the important right skewed distributions with a long tail is lognormal distribution. It is widely used in many fields, such as environmental study, survival analysis, biostatistics and other statistical fields. The lognormal distribution has closely resembled a normal distribution. Simple to implement and easy to understand, by taking the natural logarithm of a random variable, the random variable will have a normal distribution.

Interval estimation of lognormal means for one, two and several populations have received widespread attention in papers of science and statistical literature. Statistical methods for interval estimation involving common mean for several lognormal distributions have also appeared frequently in many journals, such as biometrical journal by Tian and Wu [1] defined the concept of generalized variable and the large sample criteria to provide approach for the confidence interval estimation and hypothesis testing of the common mean of several log-normal populations. Lin and Wang [2] in a journal of applied statistics focused on making inferences on several log-normal means based on the modification of the quadratic method. There are also many other journals such as journal of statistical research by Ahmed et al. [3], journal of probability and statistical science by Baklizi and Ebrahem [4] and measurement science review by Cimermanová [5] interested in construction of the confidence intervals for common mean of several lognormal distributions.

In this paper, the interest is to construct confidence intervals for common mean of lognormal distributions. A simple approach to construct the confidence intervals for lognormal mean by transform the lognormal data would be to log-transfer data prior to analyzing statistical. There have been several researchers used the log-transformed data to construct confidence intervals for mean of lognormal distribution. For example, the paper by Krishnamoorthy and Mathew [6] and Olsson [7]. It seems evident that the results for interval estimation of the common mean of several lognormal populations by taking the natural logarithm have not been studied. Furthermore, there has not been much discussion on methods for interval estimation of common lognormal means by taking the natural logarithm of data. Therefore, researchers have proposed new simple approaches to construct the confidence intervals for the common lognormal mean. The first approach is generalized confidence intervals (GCI) which is based on the concepts of generalized confidence interval and was introduced by Weerahandi [8]. The second and the third approaches are adjusted method of variance estimates recovery approach (adjusted MOVER) based on cox’s method (AM1) and Angus’s conservative method (AM2) which are based on the concepts of the method of variance of estimates recovery (MOVER) introduced by Zou and Donner [9]. The GCI approach, the MOVER approach and the adjusted MOVER approach have been successfully used to construct the confidence interval for many common parameters. As reviewed in Tian [10], Tian and Wu [1], Krishnamoorthy and Lu [11], Ye et al. [12], Donner and Zou [13], Suwan and Niwitpong [14], Li et al. [15], Wongkhao [16] and Thangjai and Niwitpong [17]. Therefore, the focus is to develop interval estimation procedures with three approaches for the common mean of lognormal distributions and then compare them to each of the situation in terms of coverage probability and average length.

This paper is organized as follows. The properties of lognormal distribution and the parameter of interest will be briefly introduced in Sect. 2. The three approaches developed and descriptions of computational procedures are presented in Sect. 3. Section 4 presents simulation results to evaluate performances of the three approaches on coverage probabilities and average lengths. Section 5 illustrates the proposed approaches with real examples. Finally, conclusions are given in Sect. 6.

2 Lognormal Distribution and the Parameter of Interest

Let \(Y_{1}, Y_{2},..., Y_{n}\) be a random variable having lognormal distribution with two parameters. This means that the log-transformed variables \(X_{1}= \log Y_{1}, X_{2}= \log Y_{2},..., X_{n}= \log Y_{n}\) are normally distributed, that has mean value \( E\left( X\right) =\mu \) and variance \( var\left( X\right) =\sigma ^2 \). The mean of Y is \( E\left( Y\right) = \exp \left( \mu +\frac{\sigma ^2}{2}\right) \), by taking the natural logarithm of a random variable we get \( \log \left( E\left( Y\right) \right) = \mu +\frac{\sigma ^2}{2}\).

According to Olsson [7], An estimator of \( \log \left( E\left( Y\right) \right) \) can be calculated from sample data as \( \log \left( E(\widehat{Y})\right) = \overline{X} +\frac{S^2}{2} \) and an estimator of the variance of \( \log \left( E(\widehat{Y})\right) \) is given by \( var\left[ \log \left( E(\widehat{Y})\right) \right] = \frac{S^2}{2} + \frac{S^4}{2(n-1)}\).

Consider k independent lognormal populations with a common mean \( \alpha \). Let \(Y_{i1}, Y_{i2},..., Y_{in_{i}}\) be a random sample from the i-th lognormal population as follows:

               \( X_{ij} = \log Y_{ij} \sim \left( \mu _{i},\sigma ^2_{i}\right) \), for \(i = 1,2,...,k , j= 1,2,...,n_{i}\).

Thus, the common mean is \( \alpha = \exp \left( \mu _{i} +\frac{\sigma _{i}^2}{2}\right) , \log \alpha = \left( \mu _{i} +\frac{\sigma _{i}^2}{2}\right) \).

3 The Approaches of Confidence Interval Estimation

3.1 The Generalized Confidence Interval Approach

Weerahandi [8] introduced the concept of generalized confidence intervals (GCI) which is based on the generalized pivotal quantity (GPQ) for a parameter of interest \( \theta \) and \(\nu \) is a vector of nuisance parameters. A generalized pivot \( R\left( X, x,\theta ,\nu \right) \) for interval estimation, where x is an observed value of X, as a random variable having the following two properties:

  1. 1.

    \( R\left( X, x,\theta ,\nu \right) \) has a distribution free of the vector of nuisance parameters \( \nu \).

  2. 2.

    The value of \( R\left( X, x,\theta ,\nu \right) \) is \( \theta \).

Let \( R_{\alpha } \) be the \( 100\alpha \)-th percentile of R. Then \( R_{\alpha } \) becomes the \( 100\left( 1-\alpha \right) \)% lower bound for \( \theta \) and \(\left( R_{\alpha /2}, R_{1-\alpha /2}\right) \) becomes a \( 100\left( 1-\alpha \right) \)% two-side generalized confidence interval for \( \theta \).

Consider k independent lognormal populations with a common mean \(\alpha \).

Thus, we have             \( \alpha = \exp \left( \mu _{i} +\frac{\sigma _{i}^2}{2}\right) \).

The common log-mean,   \( \theta = \log \alpha = \left( \mu _{i} +\frac{\sigma _{i}^2}{2}\right) \).

Let \( \overline{X}_{i} \) and \( S_{i}^{2}\) denote the sample mean and variance for data \( X_{ij} \) for the i-th sample and let \( \overline{x}_{i} \) and \( s_{i}^{2}\) denote the observed sample mean and variance respectively.

Thus \(\sigma _{i}^{2} = \frac{\left( n_{i}-1\right) S^{2}_{i}}{V_{i}} where\,\,V_{i} \sim \chi ^{2}_{n_{i}-1}\).

where \( V_{i} \) is \(\chi ^{2}\) variates with degrees of freedom and \(n_{i}-1\), we have the generalized pivot

$$\begin{aligned} R_{\sigma ^{2}_{i}}= \frac{\left( n_{i}-1\right) s^{2}_{i}}{V_{i}} \sim \frac{\left( n_{i}-1\right) s^{2}_{i}}{\chi ^{2}_{n_{i}-1}}. \end{aligned}$$
(1)

The generalized pivotal quantity to estimate \(\mu _{i}\) based on the i-th sample can be defined as

$$\begin{aligned} R_{\mu _{i}}= \overline{x_{i}} - \frac{Z_{i}}{\sqrt{U_{i}}}\sqrt{\frac{(n_{i}-1)s_{i}^2}{n_{i}}}, \end{aligned}$$
(2)

where \( Z_{i} \) and \( U_{i} \) denote standard normal variate and \(\chi ^2\) variate with degree of freedom \( n_{i}-1 \) respectively.

The generalized pivotal quantity for estimating \( \theta \) based on the i-th sample is

$$\begin{aligned} R_{\theta }^{\left( i\right) }= R_{\mu _{i}} + \frac{R_{\sigma ^{2}_{i}}}{2}. \end{aligned}$$
(3)

From the i-th sample, the maximum likelihood estimator of \( \theta \) is

$$\begin{aligned} \widehat{\theta }^{\left( i\right) } = \widehat{\mu _{i}} + \frac{\widehat{\sigma }^{2}_{i}}{2}, where \,\, \widehat{\mu _{i}} = \overline{X}_{i}, \widehat{\sigma }^{2}_{i} = S^{2}_{i}. \end{aligned}$$
(4)

The variance for \(\widehat{\theta }^{\left( i\right) }\) is

$$\begin{aligned} var\left( \widehat{\theta }^{\left( i\right) }\right) = \frac{\sigma _{i}^{2}}{n_{i}} + \frac{\sigma _{i}^{4}}{2(n_{i}-1)},see~Olsson~[7]. \end{aligned}$$
(5)

According to Ye et al. [12], the generalized pivotal quantity proposed for the common log-mean \( \theta = \log \alpha \) is a weighted average of the generalized pivot \( R_{\theta }^{\left( i\right) } \) based on k individual samples defined as

$$\begin{aligned} R_{\theta } = \frac{\sum \limits ^{k}_{i=1}R_{w}R_{\theta }^{\left( i\right) }}{\sum \limits ^{k}_{i=1}R_{w_{i}}}, \end{aligned}$$
(6)

where

$$\begin{aligned} R_{w_{i}} = \frac{1}{R_{var\left( \widehat{\theta }^{\left( i\right) }\right) }}, \end{aligned}$$
(7)
$$\begin{aligned} R_{var\left( \hat{\theta }^{\left( i\right) }\right) }=\frac{R_{\sigma _{i}^{2}}}{n_{i}} + \frac{R_{\sigma _{i}^{4}}}{2(n_{i}-1)}. \end{aligned}$$
(8)

That is, \( R_{var\left( \widehat{\theta }^{\left( i\right) }\right) } \) is \( var\left( \widehat{\theta }^{\left( i\right) }\right) \) with \( \sigma ^{2}_{i}\) replaced by \( R_{\sigma ^{2}_{i}} \).

\( \big [L_{Gci},U_{Gci}\big ] = \left( R_{\alpha /2}, R_{1-\alpha /2}\right) \) is the \( 100\left( 1-\alpha \right) \)% two-side generalized confidence interval of the common log-mean \(\theta = \log \alpha \).

\( \big [ \exp \left( L_{Gci}\right) , \exp \left( U_{Gci}\right) \big ] = \left( \exp \left( R_{\alpha /2}\right) , \exp \left( R_{1-\alpha /2}\right) \right) \) is the \( 100\left( 1-\alpha \right) \)% two-side generalized confidence interval of the common mean \(\theta \).

Computing algorithms

For a given data set \( X_{ij} \) for \( i = 1,2,...,k, j= 1,2,...,n_{i}\), the generalized confidence intervals for \( \theta \) can be computed by the following steps.

  1. 1.

    Compute \( \overline{x}_{i}\) and \( s_{i}^{2} \) for \( i = 1,2,...,k \).

  2. 2.

    Generate \( V_{i} \sim \chi ^{2}_{n_{i}-1}\) and then calculate \( R_{\sigma ^{2}_{i}} \) from (1) for \( i = 1,2,...,k \).

  3. 4.

    Generate \( Z_{i} \sim N(0,1)\) and \( U_{i} \sim \chi ^{2}_{n_{i}-1}\) then calculate \( R_{\mu _{i}} \) from (2) for \( i = 1,2,...,k \).

  4. 4.

    Calculate \( R_{\theta }^{\left( i\right) } \) from (3) for \( i = 1,2,...,k \).

  5. 5.

    Repeat steps 2–3, calculate \( R_{w_{i}} \) from (7) and (8) for \( i = 1,2,...,k \) .

  6. 6.

    Compute \( R_{\theta } \) following (6).

  7. 7.

    Repeat step 2–6 a total m times and obtain an array of \( R_{\theta }\)’s.

  8. 8.

    Rank this array of \( R_{\theta }\)’s from small to large. The \(100\alpha \)-th percentile of \(R_{\theta }\)’s, \(R_{\theta }\left( \alpha \right) \), is an estimate of the lower bound of the one-sided \( 100\left( 1-\alpha \right) \)% confidence interval and \( \left( R_{\theta }\left( \alpha /2\right) ,R_{\theta }\left( 1-\alpha /2\right) \right) \) is a two-sided \( 100\left( 1-\alpha \right) \)% confidence interval.

  9. 9.

    Calculate the interval length.

  10. 10.

    Count the number of successes in 5,000 independent generated datasets.

  11. 11.

    Calculate coverage probability and average length.

3.2 The Adjusted Method of Variance Estimates Recovery Approach

The concepts of the method of variance estimates recovery (the Mover approach) and the large sample method are used to create the adjusted method of variance estimates recovery (The adjusted MOVER approach).

The Mover approach was introduced by Zou and Donner [9] which considers two parameters \( \theta _{1}+\theta _{2}\) which have \( 100\left( 1-\alpha \right) \)% confidence limits \(\left( l_{1},u_{1}\right) \) and \(\left( l_{2},u_{2}\right) \), respectively. Under the assumption of the point estimates \(\widehat{\theta _{1}}\) and \(\widehat{\theta _{2}}\) are independence, the lower limit L and the upper limit U are given by

$$\begin{aligned} \big [L_{1},U_{1}\big ]=\left( \widehat{\theta _{1}}+\widehat{\theta _{2}}\right) \pm z_{\alpha /2}\sqrt{\widehat{var}\left( \widehat{\theta _{1}}\right) +\widehat{var}\left( \widehat{\theta _{2}}\right) }, \end{aligned}$$
(9)

where \(\widehat{var}\left( \widehat{\theta _{i}}\right) =\frac{\left( \widehat{\theta _{i}}-l_{i}\right) ^{2}}{z^{2}_{\alpha /2}}, \widehat{var}\left( \widehat{\theta _{i}}\right) =\frac{\left( u_{i}-\widehat{\theta _{i}}\right) ^{2}}{z^{2}_{\alpha /2}}\).

For \(i = 1,2\). Using these estimates with from (9), two-side \( 100\left( 1-\alpha \right) \)% confidence limits for \( \theta _{1}+\theta _{2}\) given as

$$ L=\left( \widehat{\theta _{1}}+\widehat{\theta _{2}}\right) -\sqrt{\left( \widehat{\theta _{1}}-l_{1}\right) ^{2}+\left( \widehat{\theta _{2}}-l_{2}\right) ^{2}} $$
$$ U=\left( \widehat{\theta _{1}}+\widehat{\theta _{2}}\right) +\sqrt{\left( U_{1}+\widehat{\theta _{1}}\right) ^{2}+\left( U_{2}+\widehat{\theta _{2}}\right) ^{2}}.$$

Let \(\theta ^{\left( 1\right) }, \theta ^{\left( 2\right) },..., \theta ^{\left( k\right) }\) be k parameters of interest, where the estimates \(\widehat{\theta }^{\left( 1\right) }, \widehat{\theta }^{\left( 2\right) },..., \widehat{\theta }^{\left( k\right) }\) are independent. Use concept of Donner and Zou [13] to construct of a \( 100\left( 1-\alpha \right) \)% two-sided confidence interval (L, U) for \(\theta ^{\left( 1\right) }, \theta ^{\left( 2\right) },..., \theta ^{\left( k\right) }\).

Thus \(\big [L,U\big ]=\left( \widehat{\theta }^{\left( 1\right) }+\widehat{\theta }^{\left( 2\right) }+...+\widehat{\theta }^{\left( k\right) }\right) \pm z_{\alpha /2}\sqrt{\widehat{var}\left( \widehat{\theta }^{\left( 1\right) }\right) +...+\widehat{var}\left( \widehat{\theta }^{\left( k\right) }\right) }\).

Where the variance estimate for \( \widehat{\theta }^{\left( i\right) }\) at \( \theta ^{\left( i\right) }=l_{i}\) is \(\widehat{var}\left( \widehat{\theta }^{\left( i\right) }\right) =\frac{\left( \widehat{\theta }^{\left( i\right) }-l_{i}\right) ^{2}}{z^{2}_{\alpha /2}}\).

And the variance estimate at \( \theta ^{\left( i\right) }=u_{i}\) is \(\widehat{var}\left( \widehat{\theta }^{\left( i\right) }\right) =\frac{\left( u_{i}-\widehat{\theta }^{\left( i\right) }\right) ^{2}}{z^{2}_{\alpha /2}}\).

Therefore, the lower limit L and upper limit U for \(\theta ^{\left( 1\right) }, \theta ^{\left( 2\right) },..., \theta ^{\left( k\right) }\) is given by

$$ L=\left( \widehat{\theta }^{\left( 1\right) }+\widehat{\theta }^{\left( 2\right) }+...+\widehat{\theta }^{\left( k\right) }\right) -\sqrt{\left( \widehat{\theta }^{\left( 1\right) }-l_{1}\right) ^{2}+...+\left( \widehat{\theta }^{\left( k\right) }-l_{k}\right) ^{2}},$$
$$ U=\left( \widehat{\theta }^{\left( 1\right) }+\widehat{\theta }^{\left( 2\right) }+...+\widehat{\theta }^{\left( k\right) }\right) +\sqrt{\left( u_{1}-\widehat{\theta }^{\left( 1\right) }\right) ^{2}+...+\left( u_{k}-\widehat{\theta }^{\left( k\right) }\right) ^{2}}.$$

The next step uses concept of the method of the large sample by Tain and Wu [1] for parameter with pooled estimate of the common parameter from k populations. It is defined as

$$\begin{aligned} \widehat{\theta } =\frac{\sum \limits ^{k}_{i=1}\frac{\widehat{\theta }^{\left( i\right) }}{var\left( \widehat{\theta }^{\left( i\right) }\right) }}{\sum \limits ^{k}_{i=1}\frac{1}{var\left( \widehat{\theta }^{\left( i\right) }\right) }}, \end{aligned}$$
(10)

which gives a variance estimate for \(\widehat{\theta }^{\left( i\right) }\) at \(\theta ^{\left( i\right) }=l_{i}\) and \( \theta ^{\left( i\right) }=u_{i}\) of

$$\begin{aligned} var\left( \widehat{\theta }^{\left( i\right) }\right) =\frac{1}{2}\left( \frac{\left( \widehat{\theta }^{\left( i\right) }-l_{i}\right) ^{2}}{z^{2}_{\alpha /2}}+\frac{\left( u_{i}-\widehat{\theta }^{\left( i\right) }\right) ^{2}}{z^{2}_{\alpha /2}}\right) . \end{aligned}$$
(11)

Therefore, the lower limit L for the common parameter \( \theta \) is given by

$$\begin{aligned} L= \widehat{\theta }-z_{1-\alpha /2}\sqrt{\frac{1}{\sum \limits ^{k}_{i=1}\displaystyle \frac{1}{\frac{\left( \widehat{\theta }^{\left( i\right) }-l_{i}\right) ^{2}}{z^{2}_{\alpha /2}}}}}. \end{aligned}$$
(12)

Similarly, the upper limit U for the common parameter \( \theta \) is given by

$$\begin{aligned} U= \widehat{\theta }+z_{1-\alpha /2}\sqrt{\frac{1}{\sum \limits ^{k}_{i=1}\displaystyle \frac{1}{\frac{\left( u_{i}-\widehat{\theta }^{\left( i\right) }\right) ^{2}}{z^{2}_{\alpha /2}}}}}. \end{aligned}$$
(13)

Hence, the adjusted MOVER solution for confidence interval estimation is

$$\left( \widehat{\theta }-z_{1-\alpha /2}\sqrt{1/\sum \limits ^{k}_{i=1}\displaystyle \frac{z^{2}_{\alpha /2}}{\left( \widehat{\theta }^{\left( i\right) }-l_{i}\right) ^{2}}},\widehat{\theta }+z_{1-\alpha /2}\sqrt{1/\sum \limits ^{k}_{i=1}\displaystyle \frac{z^{2}_{\alpha /2}}{\left( u_{i}-\widehat{\theta }^{\left( i\right) }\right) ^{2}}}\right) .$$

From the i-th sample, where \( i = 1,2,...,k \). The common log-mean \( \theta \) is

$$ \theta = \log \alpha = \left( \mu _{i} +\frac{\sigma _{i}^2}{2}\right) .$$

The maximum likelihood estimator of common log-mean \( \theta \) is

$$\widehat{\theta }^{\left( i\right) } = \widehat{\mu _{i}} + \frac{\widehat{\sigma }^{2}_{i}}{2}, where \,\, \widehat{\mu _{i}} = \overline{X}_{i}, \widehat{\sigma }^{2}_{i} = S^{2}_{i}.$$

Chami et al. [18] have presented Cox’s method and Angus’s conservative method for constructing confidence interval for log-mean of lognormal.

According to Cox’s method, the confidence interval \(\left( CI_C\right) \) for \( \widehat{\theta }^{(i)}\) is

$$\begin{aligned} l_{i1}=\overline{X}_{i} + \frac{S_{i}^2}{2}-Z_{1-\alpha /2}\sqrt{\frac{S_{i}^2}{n_{i}} + \frac{S_{i}^4}{2(n_{i}-1)}}, \end{aligned}$$
(14)
$$\begin{aligned} u_{i1}=\overline{X}_{i} + \frac{S_{i}^2}{2}+Z_{1-\alpha /2}\sqrt{\frac{S_{i}^2}{n_{i}} + \frac{S_{i}^4}{2(n_{i}-1)}}, \end{aligned}$$
(15)

According to Angus’s conservative method, the confidence interval \(\left( CI_A\right) \) for \( \widehat{\theta }^{(i)}\) is

$$\begin{aligned} l_{i2}=\overline{X}_{i} + \frac{S_{i}^2}{2}-\frac{t_{1-\alpha /2}}{\sqrt{n_{i}}}\sqrt{S_{i}^2\left( 1+\frac{S_{i}^2}{2}\right) }, \end{aligned}$$
(16)
$$\begin{aligned} u_{i2}=\overline{X}_{i} + \frac{S_{i}^2}{2}+\frac{q_{\alpha /2}}{\sqrt{n_{i}}}\sqrt{S_{i}^2\left( 1+\frac{S_{i}^2}{2}\right) }, \end{aligned}$$
(17)

Which \( t_{1-\alpha /2} \) be \( 1-\alpha \) percentile of a t-distribution with \( n_{i}-1\) degrees of freedom, and let \( q_{\alpha /2} = \sqrt{\frac{n}{2}\left( \frac{n-1}{\chi _{\alpha }^2}-1\right) } \) where \(\chi _{\alpha }^2 \) is the \( \alpha \)- percentile of the chi-square distributions with \( n_{i}-1\) degrees of freedom.

Researchers use the method of the large sample for log-mean with pooled estimate. It is defined in Eq. (10) and variance estimate for \( \widehat{\theta }^{(i)} \) in Eq. (11). Consequently, L and U are defined in Eqs. (12) and (13). One gains these two groups of confidence intervals \(\left( CI_C\right) \) for \( \widehat{\theta }^{(i)} \) in Eqs. (14), (15) and \(\left( CI_A\right) \) for \( \widehat{\theta }^{(i)} \) in Eqs. (16) and (17) defines AM1 and AM2.

Hence, the adjusted MOVER approach based on cox’s method (AM1) for confidence interval estimation of common log-mean \(\theta = \log \alpha \) is \( \big [L_{AM1},U_{AM1}\big ]\)

$$\begin{aligned} = \left( \widehat{\theta }-z_{1-\alpha /2}\sqrt{1/\sum \limits ^{k}_{i=1}\displaystyle \frac{z^{2}_{\alpha /2}}{\left( \widehat{\theta }^{\left( i\right) }-l_{i1}\right) ^{2}}}, \widehat{\theta }+z_{1-\alpha /2}\sqrt{1/\sum \limits ^{k}_{i=1}\displaystyle \frac{z^{2}_{\alpha /2}}{\left( u_{i1}-\widehat{\theta }^{\left( i\right) }\right) ^{2}}}\right) . \end{aligned}$$
(18)

or the confidence interval for common mean \( \alpha \) is \( \big [\exp \left( L_{AM1}\right) ,\exp \left( U_{AM1}\right) \big ] \).

The adjusted MOVER approach based on Angus’s conservative method (AM2) for confidence interval estimation of common log-mean \( \theta = \log \alpha \) is \( \big [L_{AM2},U_{AM2}\big ]\)

$$\begin{aligned} = \left( \widehat{\theta }-z_{1-\alpha /2}\sqrt{1/\sum \limits ^{k}_{i=1}\displaystyle \frac{z^{2}_{\alpha /2}}{\left( \widehat{\theta }^{\left( i\right) }-l_{i1}\right) ^{2}}}, \widehat{\theta }+z_{1-\alpha /2}\sqrt{1/\sum \limits ^{k}_{i=1}\displaystyle \frac{z^{2}_{\alpha /2}}{\left( u_{i1}-\widehat{\theta }^{\left( i\right) }\right) ^{2}}}\right) . \end{aligned}$$
(19)

or the confidence interval for common mean \( \alpha \) is \( \big [\exp \left( L_{AM2}\right) ,\exp \left( U_{AM2}\right) \big ] \).

Computing algorithms

For a given data set \( X_{ij} \) for \( i = 1,2,...,k, j= 1,2,...,n_{i}\), the adjusted method of variance estimates recovery for \( \theta \) can be computed by the following steps.

  1. 1.

    Compute \( \overline{x}_{i}\) and \( s_{i}^{2} \) for \( i = 1,2,...,k \).

  2. 2.

    Compute \(l_{i1},u_{i1},l_{i2},u_{i2} \) from (14), (15), (16), (17) for \( i = 1,2,...,k \).

  3. 3.

    Calculate \( var((\widehat{\theta }^{\left( i\right) }) \) from (11) for \( i = 1,2,...,k \).

  4. 4.

    Compute \( \widehat{\theta } \) following (10).

  5. 5.

    Calculate confidence interval estimation from (18), (19) for \( i = 1,2,...,k \).

  6. 6.

    Calculate the interval length.

  7. 7.

    Count the number of successes in 5,000 independent generated datasets.

  8. 8.

    Calculate coverage probability and average length.

4 Simulation Studies

A simulation study was performed with the coverage probabilities and average lengths of the common mean of the lognormal distributions for various combinations of the number of samples k = 2, 3 and 10, the sample sizes \( n =\left( n_{1},,..., n_{k}\right) \) and the population variance \( \sigma ^2 =\left( \sigma _1^2,,..., \sigma _k^2\right) \), the values were different and the common \(\theta = \log \alpha \) take 0 and 10. In this simulation study, 95\(\%\) confidence intervals from three approaches were compared, comprising of the proposed procedure generalized confidence interval approach (GCI), the adjusted MOVER approach based on cox’s method (AM1) and based on Angus’s conservative method (AM2). For each parameter setting, 5000 random samples were generated, 2500 \( R_{\theta } \)’s were obtained for each of the random samples.

In Tables 1, 2 and 3, the following notation applies \( n =\left( n_{1},,..., n_{k}\right) \) and \( \sigma ^2 =\) \(\left( \sigma _1^2,,..., \sigma _k^2\right) \). For \( k=2 \), we have \( n_1^{\left( 2\right) }=(15,15), n_2^{\left( 2\right) }=(10,20), n_3^{\left( 2\right) }=(20,10), n_4^{\left( 2\right) }=(50,50), n_5^{\left( 2\right) }=(50,100)\) and \( \sigma _1^{2\left( 2\right) }=(1,1), \sigma _2^{2\left( 2\right) }=(1,9), \sigma _3^{2\left( 2\right) }=(1,25), \sigma _4^{2\left( 2\right) }=(1,100)\). For \( k=3 \), we have \( n_1^{\left( 3\right) }=(15,15,15), n_2^{\left( 3\right) }=(10,15,20), n_3^{\left( 3\right) }=(20,15,10),n_4^{\left( 3\right) }=(30,50,100)\) and \( \sigma _1^{2\left( 3\right) } =(0.02,0.2,2), \sigma _2^{2\left( 3\right) }=(1,1,1), \sigma _3^{2\left( 3\right) }=(1,4,9), \sigma _4^{2\left( 3\right) }=(1,9,100)\). For \( k=10 \), we have \( n_1^{\left( 10\right) }=(15,15,15,15,15,15,15,15,15,15), n_2^{\left( 10\right) }=(15,15,1515,15,10,10,10,10,10),\) \(n_3^{\left( 10\right) }=(30,30,30,30,30,30,30,30,30,30), n_4^{\left( 10\right) }=(50,50,80,100,100,50,50,80,100,100)\) and \(\sigma _1^{2\left( 10\right) }=(1,1,1,1,1,1,1,1,1,1),\) \(\sigma _2^{2\left( 10\right) }=(1,1,9,9,25,25,49,49,81,81), \sigma _3^{2\left( 10\right) }=(0.01,0.01,0.1,0.1,1,1,10,10,100,100)\).

Tables 1, 2, and 3 presents the coverage probabilities and average lengths for 2, 3 and 10 sample cases respectively. In sample case 2, the GCI approach overestimated the coverage probabilities for all of the scenarios. In other sample cases, the GCI approach tended to overestimate the coverage probabilities for all of the scenarios but this depends on \( n_{i} \) and \( \sigma _{i}^2 \), especially when the \( n_{i} \) is small and \( \sigma _{i}^2 \) are similar values. It is also shown that the GCI approach tends to be drop from the nominal level 0.95. In all sample cases, The AM1 approach provides the underestimate coverage probabilities and has the coverage probabilities close to the nominal level 0.95 when the number of samples went up. The AM2 approach performs very well for all of the scenarios, although it tends to produce a wider average lengths more than the GCI approach, but the average length is slightly higher than the GCI approach.

In this paper, the average lengths of all intervals are considered since the approaches provide the coverage probability above the nominal level of 0.95 for all cases. Finally, it was discovered that the AM2 approach provided much better results over the another approaches in terms of average lengths for all cases. However, the average lengths of the GCI approach were the shortest when the coverage probabilities were close to the nominal level of 0.95.

Table 1. Coverage probabilities (CP) and average length (AL) of approximate 95\(\%\) two – side confidence bounds for common mean \( \alpha \) of lognormal distributions, \( \mu _{i} = \log \alpha -\left( \sigma _{i}^2/2\right) \) (based on 5000 simulations, m of gci = 2500): 2 sample cases.
Table 2. Coverage probabilities (CP) and average length (AL) of approximate 95\(\%\) two – side confidence bounds for common mean \( \alpha \) of lognormal distributions, \( \mu _{i} = \log \alpha -\left( \sigma _{i}^2/2\right) \) (based on 5000 simulations, m of gci = 2500): 3 sample cases.
Table 3. Coverage probabilities (CP) and average length (AL) of approximate 95\(\%\) two – side confidence bounds for common mean \( \alpha \) of lognormal distributions, \( \mu _{i} = \log \alpha -\left( \sigma _{i}^2/2\right) \) (based on 5000 simulations, m of gci = 2500): 10 sample cases.

5 An Empirical Application

In this section, two real data examples are exhibited to illustrate the generalized confidence interval approach (GCI), the adjusted MOVER approach based on cox’s method (AM1) and Angus’s conservative method (AM2). All examples have been studied and reported by Lin and Wang [2]. The first example (A) was the medical charge data divided into two groups, 119 of them were an American group and 106 of them were the white group. The second example (B) was the pharmacokinetics data equally divided into three groups, 22 of them were group 1, group 2 and group 3. The data sets are presented in Table 4 and the results of confidence interval for two data sets are presented in Table 5. It can be seen that the interval of the adjusted MOVER approach based on Angus’s conservative method (AM2) has the confidence interval close to the sample mean and wider lengths more than other approaches.

Table 4. The summary statistics of the log-transformed data sets.
Table 5. The results of confidence interval for four data sets.

6 Discussion and Conclusions

This paper has presented three simple approaches to construct confidence intervals for common mean of lognormal distributions. The proposed confidence intervals were constructed by the generalized confidence interval approach (GCI), the adjusted MOVER approach based on cox’s method (AM1) and based on Angus’s conservative method (AM2). By the simulation studies, coverage probabilities form the generalized confidence interval approach (GCI) was always close to the nominal confidence level at 0.95. But there were a few cases, it seems that the generalized confidence interval approach (GCI) performed well only when \(\sigma _i^2\) were more various. The adjusted MOVER approach based on cox’s method (AM1) provided underestimated coverage probabilities for all cases. The adjusted MOVER approach based on Angus’s conservative method (AM2) yielded coverage probabilities which tended to be high compared with the nominal level of 0.95 for all almost cases and the average lengths were wide to a little bit as compared with the GCI approach. Overall, the generalized confidence interval approach (GCI) provided coverage probabilities close to nominal level 0.95 and average lengths is shorter than other approaches for \( k=2 \). The adjusted MOVER approach based on Angus’s conservative method (AM2) provides stable coverage probabilities for all k. In conclusion, the adjusted MOVER approach based on Angus’s conservative method (AM2) can be successfully used to estimate the common mean of lognormal distributions.

The results of the generalized confidence interval approach (GCI) for common mean (\( k\ge 2\)) are similar to the simulation of the generalized confidence interval approach (GCI) for single mean of lognormal distribution which is studied by Olsson [7]. However, the coverage probabilities for \( k\ge 2\) is decrease when k increased. In addition, this paper is constructing the confidence intervals for common mean of lognormal distributions by transform the lognormal data prior to use the generalized confidence interval approach (GCI). It is simple to construct the confidence intervals for the common mean. The results show that, it provided coverage probabilities close to nominal level 0.95. However, the results are not good equal to results of Lin and Wang [2], but average lengths are shorter.